scispace - formally typeset
Search or ask a question

Showing papers on "Residual frame published in 2016"


Proceedings ArticleDOI
03 Mar 2016
TL;DR: A novel and efficient approach for moving object detection under a static background by selecting the maximum pixel intensity value between both the difference frames, which is considered as an improvement over the previous approaches.
Abstract: Moving-object detection is one of the basic and most active research domains in the field of computer vision. This paper proposes a novel and efficient approach for moving object detection under a static background. Proposed approach first performs pre-processing tasks to remove noise from video frames. Secondly, it finds the difference between the current frame and previous consecutive frame as well as current frame and next consecutive frame separately. The algorithm then selects the maximum pixel intensity value between both the difference frames, which we consider as an improvement over the previous approaches. Next we divide the resultant difference frame into non-overlapping blocks and calculate the intensity sum and mean of each block. Subsequently, it finds the foreground and background pixels of each block using threshold and intensity mean. In the next step morphology operation along with connected component analysis are applied to correctly detect the target objects. The proposed approach is accurate for detecting the moving object with varying object size and numbers. This work has been formulated, implemented and tested on real video data sets and the results are found to be satisfactory as it evident from the performance analysis.

37 citations


Patent
Ho-Sang Sung1, Nam-Suk Lee1
04 Mar 2016
TL;DR: In this article, a frame error concealment (FEC) method was proposed for decoding time domain signals generated after time-frequency inverse transform processing, where the current frame is an error frame or a normal frame when the previous frame is a error frame.
Abstract: Disclosed are a frame error concealment method and apparatus and an audio decoding method and apparatus The frame error concealment (FEC) method includes: selecting an FEC mode based on at least one of a state of at least one frame and a phase matching flag, with regard to a time domain signal generated after time-frequency inverse transform processing; and performing corresponding time domain error concealment processing on the current frame based on the selected FEC mode, wherein the current frame is an error frame or the current frame is a normal frame when the previous frame is an error frame

34 citations


Journal ArticleDOI
TL;DR: In this paper, a modulation scheme in the time domain based on On-Off-Keying was proposed for different types of image sensors, which can support both typical frame rate cameras (in the oversampling mode) as well as very low frame rate camera(in the error detection mode) whose frame rate is lower than the transmission packet rate.
Abstract: This paper presents a modulation scheme in the time domain based on On-Off-Keying and proposes various compatible supports for different types of image sensors. The content of this article is a sub-proposal to the IEEE 802.15.7r1 Task Group (TG7r1) aimed at Optical Wireless Communication (OWC) using an image sensor as the receiver. The compatibility support is indispensable for Image Sensor Communications (ISC) because the rolling shutter image sensors currently available have different frame rates, shutter speeds, sampling rates, and resolutions. However, focusing on unidirectional communications (i.e., data broadcasting, beacons), an asynchronous communication prototype is also discussed in the paper. Due to the physical limitations associated with typical image sensors (including low and varying frame rates, long exposures, and low shutter speeds), the link speed performance is critically considered. Based on the practical measurement of camera response to modulated light, an operating frequency range is suggested along with the similar system architecture, decoding procedure, and algorithms. A significant feature of our novel data frame structure is that it can support both typical frame rate cameras (in the oversampling mode) as well as very low frame rate cameras (in the error detection mode for a camera whose frame rate is lower than the transmission packet rate). A high frame rate camera, i.e., no less than 20 fps, is supported in an oversampling mode in which a majority voting scheme for decoding data is applied. A low frame rate camera, i.e., when the frame rate drops to less than 20 fps at some certain time, is supported by an error detection mode in which any missing data sub-packet is detected in decoding and later corrected by external code. Numerical results and valuable analysis are also included to indicate the capability of the proposed schemes.

23 citations


Proceedings ArticleDOI
09 Jun 2016
TL;DR: To extract key frame from video, haar wavelet transform with various levels and Thepade's sorted pentnary block truncation coding is used and the increase in accuracy is observed till Haar wavelets of level 5, then higher levels have shown drop in accuracy.
Abstract: Due to advance growth in videos available across the internet, it is required to navigate and handle them properly. It is essential to select only valuable and accurate information from video. Video summarization helps in acquiring essential information. Video summary produces concise and exact data of the video. With help of key frame extraction video summary can be generated. Key frames from video represent main content of video. In the proposed methodology, to extract key frame from video, haar wavelet transform with various levels and Thepade's sorted pentnary block truncation coding is used. For experimentation purpose test bed of 30 videos is used here. To measure the diversity among successive frames various similarity measures are used. Alias Canberra distance, Sorencen distance, Wavehedge distance, Euclidean distance and mean square error similarity measures are used. The Euclidean distance has given better performance. The increase in accuracy is observed till Haar wavelet of level 5, then higher levels have shown drop in accuracy.

21 citations


Patent
17 Nov 2016
TL;DR: In this article, a reduced-bandwidth wireless 3D video transmission method is proposed, which includes receiving initial first-eye frame data, reprojecting the next second-eyeframe data to the first eye, and performing infilling on the next firsteye frame with the perspective-warped initial firsteye frames.
Abstract: A method for reduced-bandwidth wireless 3D video transmission includes receiving initial first-eye frame data, reprojecting the initial first-eye frame data to the second eye (which creates initial second-eye frame data), receiving sensor data; time-warping the initial first-eye and second-eye frame data, and receiving next second-eye frame data S 140 . The method can additionally or alternatively include perspective warping the initial first-eye frame data; reprojecting the next second-eye frame data to the first eye (which creates next first-eye frame data); performing infilling on the next first-eye frame data with the perspective-warped initial first-eye frame data; time-warping the next first-eye and second-eye frame data; and/or encoding transmitted frame data with sensor data.

18 citations


Journal ArticleDOI
09 Mar 2016-Entropy
TL;DR: A novel approach handling both common and wavelet video sequences, in which the extreme Studentized deviate test is exploited to identify shot boundaries for segmenting a video sequence into shots, which obtains very encouraging results in video key frame extraction.
Abstract: This paper studies the relative entropy and its square root as distance measures of neighboring video frames for video key frame extraction. We develop a novel approach handling both common and wavelet video sequences, in which the extreme Studentized deviate test is exploited to identify shot boundaries for segmenting a video sequence into shots. Then, video shots can be divided into different sub-shots, according to whether the video content change is large or not, and key frames are extracted from sub-shots. The proposed technique is general, effective and efficient to deal with video sequences of any kind. Our new approach can offer optional additional multiscale summarizations of video data, achieving a balance between having more details and maintaining less redundancy. Extensive experimental results show that the new scheme obtains very encouraging results in video key frame extraction, in terms of both objective evaluation metrics and subjective visual perception.

14 citations


Journal ArticleDOI
TL;DR: Experimental results confirm that the proposed 3DME-McFIS technique outperforms the HEVC-3D coding standard by improving 0.90dB PSNR on average, by reducing computational time by 50%, and by reducing RAFD problem compared to the existing HE VC-3d coding standard.

13 citations


Proceedings ArticleDOI
01 Jan 2016
TL;DR: This paper proposes to discover motion models and their associated masks and then use these models and masks to form a prediction of the current frame and shows that a savings in bit rate of 2.3% is achievable over standalone HEVC if this predicted frame is used as an additional reference frame.
Abstract: Traditional video coding uses the motion model to approximate geometric boundaries of moving objects where motion discontinuities occur. Motion hints based inter-frame prediction paradigm moves away from this redundant approach and employs an innovative framework consisting of motion hint fields that are continuous and invertible, at least, over their respective domains. However, estimation of motion hint is computationally demanding, in particular for high resolution video sequences. In this paper, we propose to discover motion models and their associated masks over the current frame and then use these models and masks to form a prediction of the current frame. The prediction process is computationally simpler and experimental results show that a savings in bit rate of 2.3% is achievable over standalone HEVC if this predicted frame is used as an additional reference frame.

11 citations


Proceedings ArticleDOI
01 Jul 2016
TL;DR: A binary tree based lossless depth coding scheme that arranges the residual frame into integer or binary residual bitmap that enables avoiding rendering artifacts in synthesized views due to depth compression artifacts.
Abstract: Depth maps are becoming increasingly important in the context of emerging video coding and processing applications. Depth images represent the scene surface and are characterized by areas of smoothly varying grey levels separated by sharp edges at the position of object boundaries. To enable high quality view rendering at the receiver side, preservation of these characteristics is important. Lossless coding enables avoiding rendering artifacts in synthesized views due to depth compression artifacts. In this paper, we propose a binary tree based lossless depth coding scheme that arranges the residual frame into integer or binary residual bitmap. High spatial correlation in depth residual frame is exploited by creating large homogeneous blocks of adaptive size, which are then coded as a unit using context based arithmetic coding. On the standard 3D video sequences, the proposed lossless depth coding has achieved compression ratio in the range of 20 to 80.

10 citations


Patent
24 Aug 2016
TL;DR: In this article, a change in brightness in ambient lighting is detected when an electronic apparatus operates in a dual-camera mode utilizing images captured by a first image sensor and a second image sensor.
Abstract: Techniques and examples pertaining to frame synchronization for dynamic frame rate in dual-camera applications are described. A change in brightness in ambient lighting is detected when an electronic apparatus operates in a dual-camera mode utilizing images captured by a first image sensor and a second image sensor. In response to the detected change in brightness in the ambient lighting, exposure times and frame rates of the first image sensor and the second image sensor are adjusted and the frame rates are synchronized.

9 citations


Journal ArticleDOI
TL;DR: A new method to obtain the objects according to temporal and depth information is provided and gives better visual quality and PSNR in most case with lower complexity.
Abstract: In real-time video transmission, the loss of packet results in visual quality degraded for the succeeding frames This paper proposes an efficient full frame algorithm for depth-based 3-D videos Each frame can be regarded as combination of objects True motion estimation (TME) and depth map are exploited to calculate the motion vector (MV) for each object In this paper, the object is defined as the pixels with the same MV and similar depth value The object-based MV can extrapolate each object in reference frame to reconstruct the damage frame In the consideration of computational complexity, in this method, only the high frequency regions need to execute TME In this paper, we provide a new method to obtain the objects according to temporal and depth information From the simulation results, our algorithm gives better visual quality and PSNR in most case with lower complexity

Journal ArticleDOI
TL;DR: The proposed divide-and-conquer based hierarchical video compressive sensing coding framework is proposed, in which the whole video is independently divided into non-overlapped blocks of the hierarchical frames, which achieves better performance against many state-of-the-art still-image CS and video CS recovery algorithms.

Patent
26 Jan 2016
TL;DR: In this paper, the same authors proposed techniques for performing near visually lossless video recompression, where they generate video frames having relatively small bitrates and relatively small file sizes while retaining approximately the same level of visually perceivable video quality as the originally recorded video frames.
Abstract: Techniques are described for performing near visually lossless video recompression. The disclosed techniques generate video frames having relatively small bitrates and relatively small file sizes while retaining approximately a same level of visually perceivable video quality as the originally recorded video frames. In general, recompression of a video frame takes an input video frame and produces a second copy of the video frame that has the same or lower bitrate. The proposed techniques address the problem of recompressing a video frame with no perceivable loss in visual quality (i.e., visually lossless recompression) compared to the original recording of the video frame. In addition, the disclosed techniques provide one-step recompression of video frames that includes a single decoding and encoding of each video frame.

Patent
06 Jun 2016
TL;DR: In this article, a method for determining an error condition during a bandwidth transition period of an encoded audio signal is proposed. The error condition corresponds to a second frame of the encoded audio signals, where the second frame sequentially follows a first frame in the encoded signal.
Abstract: A method includes determining an error condition during a bandwidth transition period of an encoded audio signal. The error condition corresponds to a second frame of the encoded audio signal, where the second frame sequentially follows a first frame in the encoded audio signal. The method also includes generating audio data corresponding to a first frequency band of the second frame based on audio data corresponding to the first frequency band of the first frame. The method further includes re-using a signal corresponding to a second frequency band of the first frame to synthesize audio data corresponding to the second frequency band of the second frame.

Journal ArticleDOI
TL;DR: The preliminary conclusion of the study is that the compression behaviors of CBRs in different coding sources are adjusted in a specific proportion in order to cope with the change in frame complexity.
Abstract: The study applied a charge-coupled device (CCD) camera to send video signals to 4 DaVinciTM development boards (TMS320DM6446) of Texas Instruments (TI) to carry out H.264 Baseline Profile video coding. One of the development boards coded in the Variable Bit Rate (VBR) mode, and the other three development boards coded in the Constant Bit Rate (CBR) mode. In addition, the constant rates are 2 Mbps, 1.5 Mbps and 1 Mbps respectively. The H.264 video compression files produced by the boards were analyzed via video analysis software (CodecVisa) in the study. This software can analyze and present the compression data characteristics of the video files under each video frame, i.e., bits/MB, QP, and PSNR. In this research, the characteristics of data of each frame under four different compression conditions were compared. Their differences were calculated and averaged, and the standard deviation was evaluated. It was further connected with the values of quality characteristics and the peak signal to noise ratio (PSNR) of each frame to analyze the relation among the frame quality, the compression rate of CBR, as well as the quantitative granularity. The preliminary conclusion of the study is that the compression behaviors of CBRs in different coding sources are adjusted in a specific proportion in order to cope with the change in frame complexity. The frame will be severely damaged by a critical value during the process of network transmission while the source rate is less than the value of the characteristic.

Patent
Jun Sung Ho1, Jang Hyuk Jae1
26 May 2016
TL;DR: In this paper, a codec processor receives a current frame, determines a type of a received current frame and sets rate control parameters of the current frame; a bit-rate estimator allocates total target bits to a first group of picture (GOP) including current frame.
Abstract: A codec according to an exemplary embodiment includes a codec processor which receives a current frame, determines a type of a received current frame, and sets rate control parameters of the current frame, and a bit-rate estimator which allocates total target bits to a first group of picture (GOP) including the current frame, and allocates a target bit to each of frames included in the first GOP based on a determined type of the current frame and set rate control parameters.

Journal ArticleDOI
TL;DR: Experimental results show that the proposed algorithm does not affect the visual quality of video frames and the scheme is robust against a variety of attacks.
Abstract: This paper presents a novel technique for embedding a digital watermark into video frames based on motion vectors and discrete wavelet transform (DWT). In the proposed scheme, the binary image watermark is divided into blocks and each watermark block is embedded several times in each selected video frame at different locations. The block-based motion estimation algorithm is used to select the video frame blocks having the greatest motion vectors magnitude. The DWT is applied to the selected frame blocks, and then, the watermark block is hidden into these blocks by modifying the coefficients of the Horizontal sub-bands (HL). Adding the watermark at different locations in the same video frame makes the scheme more robust against different types of attacks. The method was tested on different types of videos. The average peak signal to noise ratio (PSNR) and the normalized correlation (NC) are used to measure the performance of the proposed method. Experimental results show that the proposed algorithm does not affect the visual quality of video frames and the scheme is robust against a variety of attacks.

Proceedings ArticleDOI
01 Dec 2016
TL;DR: This paper considers the case when the dynamic processes of the sample occur in a small proportion of the entire scanning area of AFM while the background is relatively static and only slowly changing, and proposes a greedy algorithms to select measured pixels in the reference frame that are likely to be from the common static regions and share them to the target frame.
Abstract: Undersampling is a simple but efficient way to increase the imaging rate of atomic force microscopy (AFM). The undersampled AFM images typically can be faithfully reconstructed with signal recovery techniques such as inpainting or algorithms from compressive sensing. In this paper, we consider the case when the dynamic processes of the sample occur in a small proportion of the entire scanning area of AFM while the background is relatively static and only slowly changing. In this setting, two consecutive video frames, termed the reference frame and the target frame, share a significant amount of static regions in common. Based on the measurements, we use greedy algorithms to select measured pixels in the reference frame that are likely to be from the common static regions and share them to the target frame. The target frame can then be reconstructed from both the original and shared pixels, yielding a more accurate reconstruction. This algorithm is then extended to the more realistic problem of multiple frames. Through simulation, we demonstrate that the proposed algorithm can achieve better overall video reconstruction quality compared to the frame-to-frame based single image reconstruction.

Patent
01 Jul 2016
TL;DR: In this article, an image frame processing method for processing a plurality of input image frames with an image processing device is presented. But the method is limited to the case where the resolution of the first output image frames is higher than that of the second image frames.
Abstract: The present disclosure discloses an image frame processing method for processing a plurality of input image frames with an image processing device. An embodiment of the method comprises: receiving a plurality of input image frames; and processing the plurality of input image frames to produce a first number of first output image frames and a second number of second output image frames, in which the resolution of the first output image frames is higher than the resolution of the second output image frames and the first number is less than the second number, wherein a first frame of the first output image frames and a second frame of the second output image frames are derived from the same one of the plurality of input image frames.

Patent
09 Feb 2016
TL;DR: In this article, an application executed at a central processing unit of a head mounted display (HMD) system generates sets of frame drawing commands for a graphics processing unit (GPU), and for each set of frame drawings, the GPU renders a corresponding frame into one of a plurality of frame buffers.
Abstract: An application executed at a central processing unit (CPU) of a head mounted display (HMD) system generates sets of frame drawing commands for a graphics processing unit (GPU), and for each set of frame drawing commands the GPU renders a corresponding frame into one of a plurality of frame buffers. Each frame is generated to include or be associated with a frame number that indicates the location of the frame in the sequence of frames generated over time. In addition, each frame is generated to include or be associated with pose information indicating the pose of the HMD system when the frame is generated. At periodic preemption points, the GPU selects the frame stored at the plurality of frame buffers having the most recent frame number and applies an electronic display stabilization warp to the frame based on the difference between the current pose and the pose information stored with the selected frame.

Patent
07 Jan 2016
TL;DR: In this article, a method for tracking positions of object(s) in video frames is proposed, which includes processing an initial frame of a set of frames of the video frames using feature extraction to identify locations of features of the objects, obtaining a next frame of the set and applying a motion estimation algorithm as between the next frame and a prior frame to identify updated locations of the features.
Abstract: A method includes tracking positions of object(s) in video frames, including: processing an initial frame of a set of frames of the video frames using feature extraction to identify locations of features of the object(s), obtaining a next frame of the set and applying a motion estimation algorithm as between the next frame and a prior frame to identify updated locations of the features in the next frame, where locations of the features as identified based on the prior frame are used as input to the motion estimation algorithm to identify the updated locations of the features in the next frame based one searching less than an entirety of the next frame. The tracking further includes recognizing occurrence of an event, halting the iteratively performing, and repeating, for at least one subsequent set of frames, the processing an initial frame and the using motion estimation.

01 Jan 2016
TL;DR: This paper proposed a new technique for hiding in video files, Secret message has been hidden as number of bits after converts every frame of image colored bit to its value as 24 bit and there is no suspicion of the existence of secret text.
Abstract: This paper proposed a new technique for hiding in video files, Secret message has been hidden as number of bits after converts every frame of image colored bit to its value as 24 bit. The result video show high capacity and there is no suspicion of the existence of secret text. The result video show no change when played after embedding process, the receiver need the length of text that will send in secret locations within video file to know the number of bits embedded in every frame and width number of image in frame. Mean Square Error (MSE) and Peak Signal to Noise Ratio (PSNR) are used to measure the performance of the work and the result is that, the MSE , is decreased ,the PSNR is increased when increase the number of bits embedded and when the size of video file is increased.

Patent
Gustaf Pettersson1
04 Apr 2016
TL;DR: In this article, a method and an apparatus for video content stabilization is presented, which includes causing capture of a series of image frames of a video by a camera, by detecting a movement of the camera during capture of each image frame and effecting optical image stabilization to compensate for detected movement.
Abstract: A method and an apparatus for video content stabilization are presented. The method includes causing capture of a series of image frames of a video by a camera. The video is captured by detecting a movement of the camera during capture of each image frame and effecting optical image stabilization to compensate for the detected movement. At least one of a lens assembly and an image sensor is reset to a pre-capture position subsequent to the capture of the each image frame. The method further includes performing digital image stabilization (DIS) for an image frame based on the movement of the camera detected for the image frame and the movement of the camera detected for at least one other image frame. The DIS is performed for one or more image frames to substantially remove an effect of the movement of the camera from the one or more image frames.

Proceedings ArticleDOI
01 Dec 2016
TL;DR: The use of pel-wise motion enables the extrapolated frame to be generated under the assumption of linear uniform motions within a short time period and the prediction performance of the proposed method is higher than that of the traditional method.
Abstract: We propose an efficient motion compensation method based on a temporally extrapolated frame by using a pel-wise motion (optical flow) estimation. In traditional motion compensation methods, motion vectors are generally detected on a block-by-block basis and sent to the decoder as side information. However, such block-wise motions are not always suitable for motions such as local scaling, rotation, and deformation. On the other hand, pel-wise motion can be estimated on both the side of the encoder and decoder from two successive frames that were previously encoded without side information. The use of pel-wise motion enables the extrapolated frame to be generated under the assumption of linear uniform motions within a short time period. This frame is an approximation of the frame to be encoded. The proposed bi-prediction method uses the extrapolated frame as one of the reference frames. The experimental results indicate that the prediction performance of the proposed method is higher than that of the traditional method.

Book ChapterDOI
01 Jan 2016
TL;DR: The features to train the SVM classifier and accordingly classify frames of given video as tampered or non-tampered frames, i.e. detects the tampering of frame drop, show significant improvement in classification accuracy.
Abstract: For last many years video authentication and detection of tampering in a video are major challenges in the domain of digital video forensics. This paper presents detection of one of the temporal tampering (frame drop) under no reference mode of tampering detection. Inspirit of the scheme presented by Upadhyay and Singh, this paper extends the features to train the SVM classifier and accordingly classify frames of given video as tampered or non-tampered frames, i.e. detects the tampering of frame drop. Subsequently given video is classified as tampered or non-tampered video. The obtained results with enhanced features show significant improvement in classification accuracy.

Patent
18 Mar 2016
TL;DR: In this article, motion vectors are predicted for a current block of a current frame using motion vectors from previous frames in the video stream, and temporal distances between the previous frame and its reference frames are determined.
Abstract: Motion vectors are predicted for a current block of a current frame using motion vectors from previous frames in the video stream. Temporal distances between a current frame and the one or more reference frames used to predict a current block and temporal distances between the previous frame and its reference frames are determined. Temporal distances for current frames and previous frames can be combined to weight the motion vectors and improve motion vector prediction.

Patent
17 Mar 2016
TL;DR: In this paper, a reference frame selection technique for an error resilience of video codec using multiple reference frames is proposed, which consists of: performing a motion estimation individually in each of the multiple reference frame so as to search a reference macroblock of the macroblock to be coded, checking the degree of match between the reference macro block region found by the motion estimation and the intra-coded macro block of the reference frame, and selecting the motion vector (MV) of the frame which is relatively best matched to the intra coded macroblock.
Abstract: The present invention relates to a reference frame selection technique for an error resilience of video codec using multiple reference frames. The reference frame selection technique for an error resilience of video codec using multiple reference frames according to the present invention comprises the steps of: performing a motion estimation individually in each of the multiple reference frames so as to search a reference macroblock of the macroblock to be coded; checking the degree of match between the reference macroblock region found by the motion estimation and the intra-coded macroblock of the reference frame; and selecting the motion vector (MV) of the reference macroblock which is relatively best matched to the intra-coded macroblock of the reference frame from among the reference macroblocks found in each of the multiple reference frames.

Proceedings ArticleDOI
01 Nov 2016
TL;DR: A frame rate up-conversion (FRUC) method for 3D video (3DV) using depth guided extended block matching (DGE-BM) to maintain the completeness of the foreground object achieves significantly improvement comparing with existing algorithms.
Abstract: A frame rate up-conversion (FRUC) method for 3D video (3DV) is presented in this paper. Inspired by the fact that moving foreground objects draw more attention of the viewers, in our method depth guided extended block matching (DGE-BM) is adopted to maintain the completeness of the foreground object. We first obtain the motion vector field (MVF) of the interpolated frame via block-based bi-directional motion estimation (ME). And the blocks of the interpolated frame are classified according to the depth information. Then, the boundary blocks are divided into sub-blocks, whose motion vectors (MVs) are estimated using DGE-BM. Finally, the refined MVF is applied to do motion compensation. Experimental results show that the frame interpolation quality of the proposed method achieves significantly improvement comparing with existing algorithms.

Patent
08 Jun 2016
TL;DR: In this paper, a frame processing method and device relating to the technical field of images is described. And the specific scheme comprises the following steps of: separating a frame sequence into a main frame sequence and an affiliated frame sequence; obtaining an insertion frame sequence of the main frame sequences; subtracting the affiliated frame sequences and the insertion frame sequences so as to obtain a first residual frame sequence.
Abstract: The embodiment of the invention discloses a frame processing method and device, relating to the technical field of images. By means of the frame processing method and device, the quality of an image output by a terminal can be improved. The specific scheme comprises the following steps of: separating a frame sequence into a main frame sequence and an affiliated frame sequence; obtaining an insertion frame sequence of the main frame sequence; subtracting the affiliated frame sequence and the insertion frame sequence so as to obtain a first residual frame sequence; combining a main code stream, which is formed by coding the main frame sequence, and an affiliated code stream, which is formed by coding the first residual frame sequence, so as to form a transmission code stream; and sending the transmission code stream to the terminal, such that the terminal recovers the frame sequence according to the transmission code stream. The frame processing method and device disclosed by the invention are used for frame processing.

Patent
09 Mar 2016
TL;DR: In this article, a method of stabilizing a video is proposed, which includes detecting a feature point from a first frame, predicting a location of the feature point in a second frame, and determining a distance between the predicted location and a location detected from the second frame.
Abstract: A method of stabilizing a video is provided. The method includes detecting a feature point from a first frame; predicting a location of the feature point in a second frame based on a location of the feature point in the first frame and a predetermined parameter; detecting the feature point from the second frame; determining a distance between the predicted location and a location of the feature point detected from the second frame; and updating the parameter based on a location difference between the feature point detected in the first frame and the feature point detected in the second frame, in response to determining that the distance is within a predetermined threshold.