scispace - formally typeset
Search or ask a question
Author

Maja Bystrom

Bio: Maja Bystrom is an academic researcher from Boston University. The author has contributed to research in topics: Motion compensation & Multiview Video Coding. The author has an hindex of 13, co-authored 60 publications receiving 813 citations. Previous affiliations of Maja Bystrom include Drexel University & Rensselaer Polytechnic Institute.


Papers
More filters
Journal ArticleDOI
TL;DR: A complexity reduction algorithm for an H.264 encoder is proposed which incorporates an adaptive model for the Lagrange multiplier parameter based on local sequence statistics and can achieve computational savings of 19%-67% (depending on the source sequence).
Abstract: A complexity reduction algorithm for an H.264 encoder is proposed. Computational savings are achieved by identifying, prior to motion estimation, macroblocks (MBs) that are likely to be skipped and hence saving further computational processing of these MBs. This early prediction is made by estimating a Lagrangian rate-distortion cost function which incorporates an adaptive model for the Lagrange multiplier parameter based on local sequence statistics. Simulation results demonstrate that the algorithm can achieve computational savings of 19%-67% (depending on the source sequence) with no significant loss of rate-distortion performance.

128 citations

Journal ArticleDOI
TL;DR: A set of universal operational distortion-rate characteristics is developed which balances the tradeoff between source coding accuracy and channel error protection for a fixed overall transmission rate and provides the basis for the optimal bit allocation approach.
Abstract: There has been an increased interest in the transmission of digital video over real-world transmission media, such as the direct broadcast satellite (DBS) channel. Video transmitted over such a channel is subject to degradation due, in part, to additive white Gaussian noise (AWGN). Some form of forward error-control (FEC) coding may be applied in order to reduce the effect of the noise on the transmitted bitstream; however, determination of the appropriate level of FEC coding is generally an unwieldy and computationally intensive problem, as it may depend upon a variety of parameters such as the type of video, the available bandwidth, and the channel SNR. More specifically, a combined source-channel coding approach is necessary in optimally allocating rate between source and channel coding subject to a fixed constraint on overall transmission bandwidth. In this paper we develop a method of optimal bit allocation under the assumption that the distortion is additive and independent on a frame-by-frame basis. A set of universal operational distortion-rate characteristics is developed which balances the tradeoff between source coding accuracy and channel error protection for a fixed overall transmission rate and provides the basis for the optimal bit allocation approach. The results for specific source and channel coding schemes show marked improvement over suboptimum choices of channel error protection. In addition, we show that our results approach information-theoretic performance bounds which are developed in this work.

109 citations

Proceedings ArticleDOI
24 Oct 2004
TL;DR: From the experimental results it is observed that based on the average PSNR the nonscalable system outperforms the data partitioning system, however, with the data partitions system the percentage of entirely lost frames can be lowered, and the probability of poor quality decoded video can be reduced.
Abstract: In this work we compare nonscalable video coding with data partitioning using H.264/AVC under similar application and channel constraints for conversational applications over mobile channels. For both systems optimized rate allocation and network feedback has been applied. From the experimental results it is observed that based on the average PSNR the nonscalable system outperforms the data partitioning system. However, with the data partitioning system the percentage of entirely lost frames can be lowered, and the probability of poor quality decoded video can be reduced.

83 citations

Journal ArticleDOI
TL;DR: The objective behind this work is to provide motivation for decoding of data compressed by standard source coding schemes, that is, to view the compressed bitstreams as being the output of variable-length coders and to make use of the redundancy in the bit Streams to assist in decoding.
Abstract: Motivated by previous results in joint source-channel coding and decoding, we consider the problem of decoding of variable-length codes using soft channel values. We present results of decoding of selected codes using the maximum a posteriori (MAP) decoder and the sequential decoder, and show the performance gains over decoding using hard decisions alone. The objective behind this work is to provide motivation for decoding of data compressed by standard source coding schemes, that is, to view the compressed bitstreams as being the output of variable-length coders and to make use of the redundancy in the bitstreams to assist in decoding. In order to illustrate the performance achievable by soft decoding, we provide results for decoding of MPEG-4 reversible variable-length codes as well as for decoding of MPEG-4 overhead information, under the assumption that this information is transmitted without channel coding over an additive white Gaussian noise channel. Finally, we present a method of unequal error protection for an MPEG-4 bitstream using the MAP and sequential source decoders, and show results comparable to those achievable by serial application of source and channel coding.

65 citations

Proceedings ArticleDOI
18 Jun 2000
TL;DR: It is shown that maximum likelihood (ML) sequential decoding and maximum a posteriori (MAP) sequence estimation gives significant decoding improvements over hard decisions alone and that it is possible to make use of the inherent meaning of the codewords without additional transmission of side information which results in a further gain.
Abstract: We present the results of two methods for soft decoding of variable-length codes. We first show that maximum likelihood (ML) sequential decoding and maximum a posteriori (MAP) sequence estimation gives significant decoding improvements over hard decisions alone, then we show that further improvements can be gained by additional transmission of the symbol length. Finally, we show that it is possible to make use of the inherent meaning of the codewords without additional transmission of side information which results in a further gain.

37 citations


Cited by
More filters
Proceedings Article
01 Jan 1991
TL;DR: It is concluded that properly augmented and power-controlled multiple-cell CDMA (code division multiple access) promises a quantum increase in current cellular capacity.
Abstract: It is shown that, particularly for terrestrial cellular telephony, the interference-suppression feature of CDMA (code division multiple access) can result in a many-fold increase in capacity over analog and even over competing digital techniques. A single-cell system, such as a hubbed satellite network, is addressed, and the basic expression for capacity is developed. The corresponding expressions for a multiple-cell system are derived. and the distribution on the number of users supportable per cell is determined. It is concluded that properly augmented and power-controlled multiple-cell CDMA promises a quantum increase in current cellular capacity. >

2,951 citations

Book ChapterDOI
01 Jan 1993
TL;DR: For a wide class of distortion measures and discrete sources of information there exists a functionR(d) (depending on the particular distortion measure and source) which measures the equivalent rateR of the source (in bits per letter produced) whendis the allowed distortion level.
Abstract: Consider a discrete source producing a sequence of message letters from a finite alphabet. A single-letter distortion measure is given by a non-negative matrix (d ij ). The entryd ij measures the ?cost? or ?distortion? if letteriis reproduced at the receiver as letterj. The average distortion of a communications system (source-coder-noisy channel-decoder) is taken to bed= ? i.j P ij d ij whereP ij is the probability ofibeing reproduced asj. It is shown that there is a functionR(d) that measures the ?equivalent rate? of the source for a given level of distortion. For coding purposes where a leveldof distortion can be tolerated, the source acts like one with information rateR(d). Methods are given for calculatingR(d), and various properties discussed. Finally, generalizations to ergodic sources, to continuous sources, and to distortion measures involving blocks of letters are developed. In this paper a study is made of the problem of coding a discrete source of information, given afidelity criterionor ameasure of the distortionof the final recovered message at the receiving point relative to the actual transmitted message. In a particular case there might be a certain tolerable level of distortion as determined by this measure. It is desired to so encode the information that the maximum possible signaling rate is obtained without exceeding the tolerable distortion level. This work is an expansion and detailed elaboration of ideas presented earlier [1], with particular reference to the discrete case. We shall show that for a wide class of distortion measures and discrete sources of information there exists a functionR(d) (depending on the particular distortion measure and source) which measures, in a sense, the equivalent rateRof the source (in bits per letter produced) whendis the allowed distortion level. Methods will be given for evaluatingR(d) explicitly in certain simple cases and for evaluatingR(d) by a limiting process in more complex cases. The basic results are roughly that it is impossible to signal at a rate faster thanC/R(d) (source letters per second) over a memoryless channel of capacityC(bits per second) with a distortion measure less than or equal tod. On the other hand, by sufficiently long block codes it is possible to approach as closely as desired the rateC/R(d) with distortion leveld. Finally, some particular examples, using error probability per letter of message and other simple distortion measures, are worked out in detail.

658 citations

Journal ArticleDOI
01 Sep 1986
TL;DR: This chapter discusses algorithmics and modular computations, Theory of Codes and Cryptography (3), and the theory and practice of error control codes (3).
Abstract: algorithmics and modular computations, Theory of Codes and Cryptography (3).From an analytical 1. RE Blahut. Theory and practice of error control codes. eecs.uottawa.ca/∼yongacog/courses/coding/ (3) R.E. Blahut,Theory and Practice of Error Control Codes, Addison Wesley, 1983. QA 268. Cached. Download as a PDF 457, Theory and Practice of Error Control CodesBlahut 1984 (Show Context). Citation Context..ontinued fractions.

597 citations

Journal ArticleDOI
TL;DR: The majority of the article is devoted to the techniques developed for block-based hybrid coders using motion-compensated prediction and transform coding, and a separate section covers error resilience techniques for shape coding in MPEG-4.
Abstract: We review error resilience techniques for real-time video transport over unreliable networks. Topics covered include an introduction to today's protocol and network environments and their characteristics, encoder error resilience tools, decoder error concealment techniques, as well as techniques that require cooperation between encoder, decoder, and the network. We provide a review of general principles of these techniques as well as specific implementations adopted by the H.263 and MPEG-4 video coding standards. The majority of the article is devoted to the techniques developed for block-based hybrid coders using motion-compensated prediction and transform coding. A separate section covers error resilience techniques for shape coding in MPEG-4.

578 citations

Journal ArticleDOI
Liquan Shen1, Zhi Liu1, Xinpeng Zhang1, Wenqiang Zhao1, Zhaoyang Zhang1 
TL;DR: A fast CU size decision algorithm for HM that can significantly reduce computational complexity while maintaining almost the same RD performance as the original HEVC encoder is proposed.
Abstract: The emerging high efficiency video coding standard (HEVC) adopts the quadtree-structured coding unit (CU). Each CU allows recursive splitting into four equal sub-CUs. At each depth level (CU size), the test model of HEVC (HM) performs motion estimation (ME) with different sizes including 2N × 2N, 2N × N, N × 2N and N × N. ME process in HM is performed using all the possible depth levels and prediction modes to find the one with the least rate distortion (RD) cost using Lagrange multiplier. This achieves the highest coding efficiency but requires a very high computational complexity. In this paper, we propose a fast CU size decision algorithm for HM. Since the optimal depth level is highly content-dependent, it is not efficient to use all levels. We can determine CU depth range (including the minimum depth level and the maximum depth level) and skip some specific depth levels rarely used in the previous frame and neighboring CUs. Besides, the proposed algorithm also introduces early termination methods based on motion homogeneity checking, RD cost checking and SKIP mode checking to skip ME on unnecessary CU sizes. Experimental results demonstrate that the proposed algorithm can significantly reduce computational complexity while maintaining almost the same RD performance as the original HEVC encoder.

406 citations