Scheduled System Maintenance
On Wednesday, July 26, IEEE Xplore will undergo scheduled maintenance from 1:00-3:00 PM ET.
During this time, there may be intermittent impact on performance. We apologize for any inconvenience.
By Topic

Proceedings. Compression and Complexity of SEQUENCES 1997 (Cat. No.97TB100171)

13-13 June 1997

Filter Results

Displaying Results 1 - 25 of 36
  • Proceedings. Compression and Complexity of SEQUENCES 1997 (Cat. No.97TB100171)

    Publication Year: 1997
    Request permission for commercial reuse | PDF file iconPDF (170 KB)
    Freely Available from IEEE
  • Hashing on strings, cryptography, and protection of privacy

    Publication Year: 1997
    Cited by:  Papers (10)
    Request permission for commercial reuse | Click to expandAbstract | PDF file iconPDF (80 KB)

    Summary form only given. The issues of privacy and reliability of personal data are of paramount importance. If L is a list of people carrying some harmful defective gene, we want questions as to whether a person is in L to be reliably answered without compromising the data concerning anybody else. Reliability means that once the list is formed, nobody can play with the answer. Thus the answer sho... View full abstract»

    Full text access may be available. Click article title to sign in or learn about subscription options.
  • On the role of data compression in new products

    Publication Year: 1997
    Request permission for commercial reuse | Click to expandAbstract | PDF file iconPDF (59 KB)

    Summary form only given. Discusses the role of data compression in storage subsystems (including caches and controllers) and operating systems (code compression). View full abstract»

    Full text access may be available. Click article title to sign in or learn about subscription options.
  • A parallel decoder for LZ2 compression using the ID update heuristic

    Publication Year: 1997, Page(s):368 - 373
    Cited by:  Papers (3)
    Request permission for commercial reuse | Click to expandAbstract | PDF file iconPDF (372 KB)

    The LZ2 compression method seems hardly parallelizable since some related heuristics are known to be P-complete. In spite of such negative result, the decoding process can be parallelized efficiently for the next character heuristic. We show an other parallel decoding algorithm for LZ2 compression using the ID update heuristic. The algorithm works in O(log/sup 2/n) time with O(n/log(n)) processors... View full abstract»

    Full text access may be available. Click article title to sign in or learn about subscription options.
  • Index of authors

    Publication Year: 1997, Page(s): 399
    Request permission for commercial reuse | PDF file iconPDF (52 KB)
    Freely Available from IEEE
  • Optimization of the SW algorithm for high-dimensional compression

    Publication Year: 1997, Page(s):194 - 203
    Request permission for commercial reuse | Click to expandAbstract | PDF file iconPDF (640 KB)

    This paper describes an algorithm and a software package SW (Spherical Wavelets) that implements a method for compression of scalar functions defined on 3D objects. This method combines discrete second generation wavelet transforms with an extension of the embedded zerotree coding method. We present some results on optimizing the performance of the SW algorithm via the use of arithmetic coding, di... View full abstract»

    Full text access may be available. Click article title to sign in or learn about subscription options.
  • Compression of low entropy strings with Lempel-Ziv algorithms

    Publication Year: 1997, Page(s):107 - 121
    Cited by:  Papers (1)
    Request permission for commercial reuse | Click to expandAbstract | PDF file iconPDF (668 KB)

    We compare the compression ratio of the Lempel-Ziv algorithms with the empirical entropy of the input string, We show that although these algorithms are optimal according to the generally accepted definition, we can find families of low entropy strings which are not compressed optimally. More precisely, we show that the compression ratio achieved by LZ78 (resp. LZ77) can be much higher than the ze... View full abstract»

    Full text access may be available. Click article title to sign in or learn about subscription options.
  • A signature technique for similarity-based queries

    Publication Year: 1997, Page(s):2 - 20
    Cited by:  Papers (9)  |  Patents (1)
    Request permission for commercial reuse | Click to expandAbstract | PDF file iconPDF (952 KB)

    Jagadish et al. (see Proc. ACM SIGACT-SIGMOD-SIGART PODS, p.36-45, 1995) developed a general framework for posing queries based on similarity. The framework enables a formal definition of the notion of similarity for an application domain of choice, and then its use in queries to perform similarity-based search. We adapt this framework to the specialized domain of real-valued sequences. (Although ... View full abstract»

    Full text access may be available. Click article title to sign in or learn about subscription options.
  • Dense coding-a fast alternative to arithmetic coding

    Publication Year: 1997, Page(s):295 - 304
    Cited by:  Papers (2)
    Request permission for commercial reuse | Click to expandAbstract | PDF file iconPDF (524 KB)

    With dense coding a new method for minimum redundancy coding is introduced. An analysis of arithmetic coding shows, that it is essentially identical to an encoding of discrete intervals. Interval coding is introduced, which encodes symbols directly by encoding the corresponding discrete intervals. Dense coding is an enhanced variant of interval coding, where redundancies are mostly removed with a ... View full abstract»

    Full text access may be available. Click article title to sign in or learn about subscription options.
  • A universal upper bound on the performance of the Lempel-Ziv algorithm on maliciously-constructed data

    Publication Year: 1997, Page(s):123 - 135
    Request permission for commercial reuse | Click to expandAbstract | PDF file iconPDF (536 KB)

    We consider the performance of the Lempel-Ziv (1978) algorithm on finite strings and infinite sequences having unbalanced statistics. We show that such strings and sequences are compressed by the Lempel-Ziv algorithm. We show that the converse does not hold, i.e., that there are sequences with perfectly balanced asymptotic statistics that the Lempel-Ziv algorithm compresses optimally View full abstract»

    Full text access may be available. Click article title to sign in or learn about subscription options.
  • Thresholding wavelets for image compression

    Publication Year: 1997, Page(s):374 - 389
    Cited by:  Patents (7)
    Request permission for commercial reuse | Click to expandAbstract | PDF file iconPDF (920 KB)

    The paper addresses the problem of thresholding wavelet coefficients in a transform-based algorithm for still image compression. Processing data before the quantization phase is a crucial step in a compression algorithm, especially in applications which require high compression ratios. In the paper, after a review on the applications of wavelets to image compression, a new solution to the problem ... View full abstract»

    Full text access may be available. Click article title to sign in or learn about subscription options.
  • Vector quantization and density estimation

    Publication Year: 1997, Page(s):172 - 193
    Cited by:  Papers (11)  |  Patents (2)
    Request permission for commercial reuse | Click to expandAbstract | PDF file iconPDF (1268 KB)

    The connection between compression and the estimation of probability distributions has long been known for the case of discrete alphabet sources and lossless coding. A universal lossless code which does a good job of compressing must implicitly also do a good job of modeling. In particular, with a collection of codebooks, one for each possible class or model, if codewords are chosen from among the... View full abstract»

    Full text access may be available. Click article title to sign in or learn about subscription options.
  • A probabilistic approach to some asymptotics in source coding

    Publication Year: 1997, Page(s):97 - 106
    Request permission for commercial reuse | Click to expandAbstract | PDF file iconPDF (544 KB)

    Renewal theory is a powerful tool in the analysis of source codes. In this paper, we use renewal theory to obtain some asymptotic properties of finite-state noiseless channels. We discuss the relationship between these results and earlier uses of renewal theory to analyze the Lempel-Ziv codes and the Tunstall code. As a new application of our results, we provide a simple derivation of the asymptot... View full abstract»

    Full text access may be available. Click article title to sign in or learn about subscription options.
  • Multi-string search in BSP

    Publication Year: 1997, Page(s):240 - 252
    Request permission for commercial reuse | Click to expandAbstract | PDF file iconPDF (724 KB)

    We have studied the worst-case complexity of the multi-string search problem in the bulk synchronous parallel (BSP) model (Valiant 1990). For this purpose, we have devised a very simple way to distribute the blind trie data structure among the p processors so that the communication cost is balanced. In the light of the very efficient algorithms and data structures known for external memory and the... View full abstract»

    Full text access may be available. Click article title to sign in or learn about subscription options.
  • Multialphabet coding with separate alphabet description

    Publication Year: 1997, Page(s):56 - 65
    Cited by:  Papers (15)
    Request permission for commercial reuse | Click to expandAbstract | PDF file iconPDF (564 KB)

    For lossless universal source coding of memoryless sequences with an a priori unknown alphabet size (multialphabet coding), the alphabet of the sequence must be described as well as the sequence itself. Usually an efficient description of the alphabet can be made only by taking into account some additional information. We show that these descriptions can be separated in such a way that the encodin... View full abstract»

    Full text access may be available. Click article title to sign in or learn about subscription options.
  • Error resilient data compression with adaptive deletion

    Publication Year: 1997, Page(s):285 - 294
    Request permission for commercial reuse | Click to expandAbstract | PDF file iconPDF (604 KB)

    In earlier work we presented the k-error protocol, a technique for protecting a dynamic dictionary method from error propagation as the result of any k errors on the communication channel or compressed file. Here we further develop this approach and provide experimental evidence that this approach is highly effective in practice against a noisy channel or faulty storage medium. That is, for LZ2-ba... View full abstract»

    Full text access may be available. Click article title to sign in or learn about subscription options.
  • Group testing problems with sequences in experimental molecular biology

    Publication Year: 1997, Page(s):357 - 367
    Cited by:  Papers (6)
    Request permission for commercial reuse | Click to expandAbstract | PDF file iconPDF (580 KB)

    Group testing is a basic paradigm for experimental design. In computational biology, group testing problems come up in designing experiments with sequences for mapping, screening libraries, etc. While a great deal of classical research has been done on group testing over the last fifty years, the current biological applications bring up many new issues in group testing which had not been previousl... View full abstract»

    Full text access may be available. Click article title to sign in or learn about subscription options.
  • A criterion for model selection using minimum description length

    Publication Year: 1997, Page(s):204 - 214
    Request permission for commercial reuse | Click to expandAbstract | PDF file iconPDF (500 KB)

    Rissanen (1978) proposed the idea that the goodness of fit of a parametric model of the probability density of a random variable could be thought of as an information coding problem. He argued that the best model was that which was able to describe the training data together with the model parameters using the fewest number of bits of information (Occam's razor). This paper builds upon that basic ... View full abstract»

    Full text access may be available. Click article title to sign in or learn about subscription options.
  • On the resemblance and containment of documents

    Publication Year: 1997, Page(s):21 - 29
    Cited by:  Papers (102)  |  Patents (78)
    Request permission for commercial reuse | Click to expandAbstract | PDF file iconPDF (472 KB)

    Given two documents A and B we define two mathematical notions: their resemblance r(A, B) and their containment c(A, B) that seem to capture well the informal notions of “roughly the same” and “roughly contained.” The basic idea is to reduce these issues to set intersection problems that can be easily evaluated by a process of random sampling that can be done independently ... View full abstract»

    Full text access may be available. Click article title to sign in or learn about subscription options.
  • Text compression using recency rank with context and relation to context sorting, block sorting and PPM*

    Publication Year: 1997, Page(s):305 - 319
    Cited by:  Papers (3)  |  Patents (1)
    Request permission for commercial reuse | Click to expandAbstract | PDF file iconPDF (748 KB)

    A block sorting compression scheme was developed and its relation to a statistical scheme was studied, but a theoretical analysis of its performance has not been studied fully. Context sorting is a compression scheme based on context similarity and it is regarded as an on-line version of block sorting and it is asymptotically optimal. However, the compression speed is slower and the real performan... View full abstract»

    Full text access may be available. Click article title to sign in or learn about subscription options.
  • A progressive Ziv-Lempel algorithm for image compression

    Publication Year: 1997, Page(s):136 - 144
    Cited by:  Papers (2)  |  Patents (1)
    Request permission for commercial reuse | Click to expandAbstract | PDF file iconPDF (496 KB)

    We describe an algorithm that gives a progression of compressed versions of a single image. Each stage of the progression is a lossy compression of the image, with the distortion decreasing in each stage, until the last image is losslessly compressed. Progressive encodings are useful in applications such as Web browsing and multicast, where the best rate/distortion tradeoff often is not known in a... View full abstract»

    Full text access may be available. Click article title to sign in or learn about subscription options.
  • Practical implementation of the lossless compression algorithm

    Publication Year: 1997, Page(s):390 - 397
    Request permission for commercial reuse | Click to expandAbstract | PDF file iconPDF (224 KB)

    A combination of the LZ78 method with a new scheme of model contexting is introduced. In the proposed scheme the hashing function is also used. This approach speeds up the searching process and has an improvement over model contexting View full abstract»

    Full text access may be available. Click article title to sign in or learn about subscription options.
  • On the approximate pattern occurrences in a text

    Publication Year: 1997, Page(s):253 - 264
    Cited by:  Papers (1)
    Request permission for commercial reuse | Click to expandAbstract | PDF file iconPDF (588 KB)

    Consider a given pattern H and a random text T generated randomly according to the Bernoulli model. We study the frequency of approximate occurrences of the pattern H in a random text when overlapping copies of the approximate pattern are counted separately. We provide exact and asymptotic formulae for mean, variance and probability of occurrence as well as asymptotic results including the central... View full abstract»

    Full text access may be available. Click article title to sign in or learn about subscription options.
  • Near-lossless image compression schemes based on weighted finite automata encoding and adaptive context modelling

    Publication Year: 1997, Page(s):66 - 77
    Request permission for commercial reuse | Click to expandAbstract | PDF file iconPDF (552 KB)

    We study high-fidelity image compression with a given tight bound on the maximum error magnitude. We propose a weighted finite automata (WFA) recursive encoding scheme on the adaptive context modelling based quantizing prediction residue images. By incorporating the proposed recursive WFA encoding techniques into the context modelling based nearly-lossless CALIC (context based adaptive lossless im... View full abstract»

    Full text access may be available. Click article title to sign in or learn about subscription options.
  • Annotated statistical indices for sequence analysis

    Publication Year: 1997, Page(s):215 - 229
    Cited by:  Papers (2)
    Request permission for commercial reuse | Click to expandAbstract | PDF file iconPDF (640 KB)

    A statistical index for string x is a digital-search tree or trie that returns, for any query string ω and in a number of comparisons bounded by the length of ω, the number of occurrences of ω in x. Clever algorithms are available that support the construction and weighting of such indices in time and space linear in the length of x. This paper addresses the problem of annotating... View full abstract»

    Full text access may be available. Click article title to sign in or learn about subscription options.