Topic
Inter frame
About: Inter frame is a research topic. Over the lifetime, 4154 publications have been published within this topic receiving 63549 citations.
Papers published on a yearly basis
Papers
More filters
•
25 Jun 2009TL;DR: In this paper, a virtual viewpoint is used to determine expected contributions of individual portions of the frames to a synthesized image of the scene from the viewpoint position using the frames, and the frames are transmitted in compressed form via a network to a remote device, which is configured to render the scene using the compressed frames.
Abstract: Multi-view video that is being streamed to a remote device in real time may be encoded. Frames of a real-world scene captured by respective video cameras are received for compression. A virtual viewpoint, positioned relative to the video cameras, is used to determine expected contributions of individual portions of the frames to a synthesized image of the scene from the viewpoint position using the frames. For each frame, compression rates for individual blocks of a frame are computed based on the determined contributions of the individual portions of the frame. The frames are compressed by compressing the blocks of the frames according to their respective determined compression rates. The frames are transmitted in compressed form via a network to a remote device, which is configured to render the scene using the compressed frames.
54 citations
•
13 Apr 2007TL;DR: In this paper, the authors proposed an apparatus and method for encoding and decoding using alternative transform units according to the correlation of residual signals, which includes a first transforming unit for performing discrete cosine transform (DCT), first quantization, first inverse quantization and inverse DCT on a block basis onto residual coefficients generated after intra frame prediction or inter frame prediction.
Abstract: Provided is an apparatus and method for encoding and decoding using alternative transform units according to the correlation of residual signals. The video encoding apparatus includes a first transforming unit for performing discrete cosine transform (DCT), first quantization, first inverse quantization, and inverse DCT on a block basis onto residual coefficients generated after intra frame prediction or inter frame prediction; a second transforming unit for performing discrete sine transform (DST), second quantization, second inverse quantization, and inverse DST on a block basis onto the residual coefficients; a selecting unit for selecting one having a high compression rate between the first and second transforming units for each block through performing rate-distortion optimization; and a flag marking unit for recording information about the selected transforming unit at a flag bit provided on a macroblock basis.
54 citations
•
17 Feb 2004TL;DR: In this article, a receiver uses the map and frame count information to find data in the fields of received frames, where the odd fields contain the current map and part of the frame count.
Abstract: Frames comprise odd fields and even fields. The frame sync segments of the odd fields contains a current map specifying the location of data in frames, a next map specifying the location of data in a future frame, and a frame count designating the future frame. The frame sync segments of the even field may contain the same information. Alternatively, the frame sync segments of the odd fields contain the current map and part of the frame count, and the frame sync segments of the corresponding even fields contain the next map and the rest of the frame count. A receiver uses the map and frame count information to find data in the fields of received frames.
53 citations
•
AT&T1
TL;DR: In this article, the transmission end of a content distribution system selectively employs a redundancy mechanism to encoded video data based on the frame type of each frame being transmitted, where the loss of some or all of the data of the frame may not appreciably affect the presentation of the video content at the receiving end.
Abstract: The transmitting end of a content distribution system selectively employs a redundancy mechanism to encoded video data based on the frame type of each frame being transmitted. In the event that a particular frame contains information upon which the decoding of multiple frames may depend, the transmitting end can apply a redundancy mechanism to redundantly distribute the data of the frame throughout a set of data segments, each of which is separately transmitted via the network to the receiving end. Otherwise, in the event that a particular frame to be transmitted does not contain substantial information upon which the decoding of multiple frames may depend, the loss of some or all of the data of the frame may not appreciably affect the presentation of the video content at the receiving end and thus the transmitting end can forgo application of the redundancy mechanism to such frames so as to avoid unnecessarily processing and reduce the overall network bandwidth used to transmit the encoded video data.
53 citations
••
TL;DR: Evaluation on the subjective quality video database demonstrates that the proposed method outperforms the representative RR video quality metric and the full-reference VQAs, such as peak signal-to-noise ratio and structure similarity index in matching subjective ratings.
Abstract: In this paper, a novel reduced-reference (RR) video quality assessment (VQA) is proposed by exploiting the spatial information loss and the temporal statistical characteristics of the interframe histogram. From the spatial perspective, an energy variation descriptor (EVD) is proposed to measure the energy change of each individual encoded frame, which results from the quantization process. Besides depicting the energy change, EVD can further simulate the texture masking property of the human visual system (HVS). From the temporal perspective, the generalized Gaussian density (GGD) function is employed to capture the natural statistics of the interframe histogram distribution. The city-block distance (CBD) is used to calculate the histogram distance between the original video sequence and the encoded one. For simplicity, the difference image between adjacent frames is employed to characterize the temporal interframe relationship. By combining the spatial EVD together with the temporal CBD, an efficient RR VQA is developed. Evaluation on the subjective quality video database demonstrates that the proposed method outperforms the representative RR video quality metric and the full-reference VQAs, such as peak signal-to-noise ratio and structure similarity index in matching subjective ratings. This means that the proposed metric is more consistent with the HVS perception. Furthermore, as only a small number of RR features are extracted for representing the original video sequence (each frame requires only one parameter for describing EVD and three parameters for recording GGD), the RR features can be embedded into the video sequences or transmitted through the ancillary data channel, which can be used in the video quality monitoring system.
53 citations