scispace - formally typeset
Search or ask a question

Showing papers on "Smacker video published in 2013"


Journal ArticleDOI
TL;DR: The design for these extensions represents the latest state of the art for video coding and its applications, including work on range extensions for color format and bit depth enhancement, embedded-bitstream scalability, and 3D video.
Abstract: This paper describes extensions to the High Efficiency Video Coding (HEVC) standard that are active areas of current development in the relevant international standardization committees. While the first version of HEVC is sufficient to cover a wide range of applications, needs for enhancing the standard in several ways have been identified, including work on range extensions for color format and bit depth enhancement, embedded-bitstream scalability, and 3D video. The standardization of extensions in each of these areas will be completed in 2014, and further work is also planned. The design for these extensions represents the latest state of the art for video coding and its applications.

420 citations


Journal ArticleDOI
TL;DR: This paper describes an extension of the high efficiency video coding (HEVC) standard for coding of multi-view video and depth data, and develops and integrated a novel encoder control that guarantees that high quality intermediate views can be generated based on the decoded data.
Abstract: This paper describes an extension of the high efficiency video coding (HEVC) standard for coding of multi-view video and depth data. In addition to the known concept of disparity-compensated prediction, inter-view motion parameter, and inter-view residual prediction for coding of the dependent video views are developed and integrated. Furthermore, for depth coding, new intra coding modes, a modified motion compensation and motion vector coding as well as the concept of motion parameter inheritance are part of the HEVC extension. A novel encoder control uses view synthesis optimization, which guarantees that high quality intermediate views can be generated based on the decoded data. The bitstream format supports the extraction of partial bitstreams, so that conventional 2D video, stereo video, and the full multi-view video plus depth format can be decoded from a single bitstream. Objective and subjective results are presented, demonstrating that the proposed approach provides 50% bit rate savings in comparison with HEVC simulcast and 20% in comparison with a straightforward multi-view extension of HEVC without the newly developed coding tools.

365 citations


Patent
18 Dec 2013
TL;DR: In this paper, a computer-implemented method and system for incorporating advertisement into video-based digital media, comprising: providing a video in digital format wherein the video comprises a sequence of video frames; providing meta-data wherein the meta data defines at least one surface in at least 1 of the video frames.
Abstract: A computer-implemented method and system for incorporating advertisement into video-based digital media, comprising: providing a video in digital format wherein the video comprises a sequence of video frames; providing meta-data wherein the meta-data defines at least one surface in at least one of the video frames; providing an image, wherein the image is external to the video; incorporating the image into the at least one surface; and displaying the sequence of video frames wherein the sequence of video frames includes the image incorporated in the at least one surface. Attributes of the surface, such as lighting, shading, texture, curvature, etc. may be factored into a transformation function projecting the image onto the surface. To a viewer of the video, the incorporated advertisement may appear to be an integral part of the video.

270 citations


Proceedings ArticleDOI
01 Dec 2013
TL;DR: A brief technical overview of VP9 is provided along with comparisons with other state-of-the-art video codecs H.264/AVC and HEVC on standard test sets and results show VP9 to be quite competitive with mainstream state of theart codecs.
Abstract: Google has recently finalized a next generation open-source video codec called VP9, as part of the libvpx repository of the WebM project (http://www.webmproject.org/). Starting from the VP8 video codec released by Google in 2010 as the baseline, various enhancements and new tools were added, resulting in the next-generation VP9 bit-stream. This paper provides a brief technical overview of VP9 along with comparisons with other state-of-the-art video codecs H.264/AVC and HEVC on standard test sets. Results show VP9 to be quite competitive with mainstream state-of-the-art codecs.

215 citations


Proceedings ArticleDOI
Li Song1, Xun Tang, Wei Zhang1, Xiaokang Yang1, Pingjian Xia 
03 Jul 2013
TL;DR: This short paper presents a set of 15 new 4K resolution ultra-high definition (UHD) video sequences for catering the requirement of active UHD video quality assessment algorithms in coming years, as well as help to fully evaluate coding efficiency of latest HEVC (High Efficient Video Coding).
Abstract: This short paper presents a set of 15 new 4K resolution ultra-high definition (UHD) video sequences for catering the requirement of active UHD video quality assessment algorithms in coming years, as well as help to fully evaluate coding efficiency of latest HEVC (High Efficient Video Coding). The details of processing procedures and sequences characteristics are illustrated.

176 citations


Proceedings ArticleDOI
11 Nov 2013
TL;DR: The analysis of the footprints left when tampering with a video sequence is presented, and a detection algorithm is proposed that allows a forensic analyst to reveal video forgeries and localize them in the spatio-temporal domain.
Abstract: Video sequences are often believed to provide stronger forensic evidence than still images, e.g., when used in lawsuits. However, a wide set of powerful and easy-to-use video authoring tools is today available to anyone. Therefore, it is possible for an attacker to maliciously forge a video sequence, e.g., by removing or inserting an object in a scene. These forms of manipulation can be performed with different techniques. For example, a portion of the original video may be replaced by either a still image repeated in time or, in more complex cases, by a video sequence. Moreover, the attacker might use as source data either a spatio-temporal region of the same video, or a region taken from an external sequence. In this paper we present the analysis of the footprints left when tampering with a video sequence, and propose a detection algorithm that allows a forensic analyst to reveal video forgeries and localize them in the spatio-temporal domain. With respect to the state-of-the-art, the proposed method is completely unsupervised and proves to be robust to compression. The algorithm is validated against a dataset of forged videos available online.

129 citations


Journal ArticleDOI
21 Jul 2013
TL;DR: This work proposes a novel differential-geometry-based scheme that interpolates these transformations in a manner that minimizes their curvature, similarly to curvature flows, and automatically determines a set of keyframes that best represent this interpolated transformation curve.
Abstract: In most professional cinema productions, the color palette of the movie is painstakingly adjusted by a team of skilled colorists -- through a process referred to as color grading -- to achieve a certain visual look. The time and expertise required to grade a video makes it difficult for amateurs to manipulate the colors of their own video clips. In this work, we present a method that allows a user to transfer the color palette of a model video clip to their own video sequence. We estimate a per-frame color transform that maps the color distributions in the input video sequence to that of the model video clip. Applying this transformation naively leads to artifacts such as bleeding and flickering. Instead, we propose a novel differential-geometry-based scheme that interpolates these transformations in a manner that minimizes their curvature, similarly to curvature flows. In addition, we automatically determine a set of keyframes that best represent this interpolated transformation curve, and can be used subsequently, to manually refine the color grade. We show how our method can successfully transfer color palettes between videos for a range of visual styles and a number of input video clips.

116 citations


Patent
14 Mar 2013
TL;DR: In this article, a stitched manifest identifying a combination of video assets is created in response to a client request for one or more of the video assets, which is used to playback as one continuous video stream.
Abstract: Methods and apparatus are provided for combining assets for a streaming video signal. Video assets are provided in an adaptive bit rate format. The video assets may comprise fragments of various types of video content. A stitched manifest identifying a combination of the video assets is created in response to a client request for one or more of the video assets. The stitched manifest is provided to a client and used by the client to request the combination of video assets for playback as one continuous video stream in an order specified by the stitched manifest.

102 citations


Patent
Ying Chen1, Wang Ye-Kui1
11 Jan 2013
TL;DR: In this article, a video coder is configured to code a video parameter set (VPS) for one or more layers of video data, wherein each of the one or multiple layers refer to the VPS, and code the video data based at least in part on VPS.
Abstract: In one example, a video coder, such as a video encoder or video decoder, is configured to code a video parameter set (VPS) for one or more layers of video data, wherein each of the one or more layers of video data refer to the VPS, and code the one or more layers of video data based at least in part on the VPS. The video coder may code the VPS for video data conforming to High-Efficiency Video Coding, Multiview Video Coding, Scalable Video Coding, or other video coding standards or extensions of video coding standards. The VPS may include data specifying parameters for corresponding sequences of video data within various different layers (e.g., views, quality layers, or the like). The parameters of the VPS may provide indications of how the corresponding video data is coded.

91 citations


Journal ArticleDOI
TL;DR: The proposed metric performs significantly better than state-of-the-art stereoscopic image and video quality metrics in predicting the subjective scores and will have important uses in advanced 3D media delivery systems.
Abstract: The quality assessment of impaired stereoscopic video is a key element in designing and deploying advanced immersive media distribution platforms. A widely accepted quality metric to measure impairments of stereoscopic video is, however, still to be developed. As a step toward finding a solution to this problem, this paper proposes a full reference stereoscopic video quality metric to measure the perceptual quality of compressed stereoscopic video. A comprehensive set of subjective experiments is performed with 14 different stereoscopic video sequences, which are encoded using both the H.264 and high efficiency video coding compliant video codecs, to develop a subjective test results database of 116 test stimuli. The subjective results are analyzed using statistical techniques to uncover different patterns of subjective scoring for symmetrically and asymmetrically encoded stereoscopic video. The subjective result database is subsequently used for training and validating a simple but effective stereoscopic video quality metric considering heuristics of binocular vision. The proposed metric performs significantly better than state-of-the-art stereoscopic image and video quality metrics in predicting the subjective scores. The proposed metric and the subjective result database will be made publicly available, and it is expected that the proposed metric and the subjective assessments will have important uses in advanced 3D media delivery systems.

83 citations


Patent
03 Dec 2013
TL;DR: In this article, the authors present systems and methods for reducing video communication bandwidth requirements of a video surveillance network camera system that includes network communication paths between network video cameras and content-aware computer networking devices analyzing by video analytics video visual content of the video streams to provide managed video representing, at specified quality levels, samples of the scenes observed.
Abstract: Disclosed are systems and methods for reducing video communication bandwidth requirements of a video surveillance network camera system that includes network communication paths between network video cameras producing video streams of scenes observed by the network video cameras and content-aware computer networking devices analyzing by video analytics video visual content of the video streams to provide managed video representing, at specified quality levels, samples of the scenes observed. Distribution of the managed video consumes substantially less network bandwidth than would be consumed by delivery through network communication paths a video stream at the specified quality level in the absence of analysis by the video analytics.

Patent
19 Sep 2013
TL;DR: In this article, a method for playing an interactive video on a user device is presented, which comprises receiving an interactive content file associated with a video to be played by the user device.
Abstract: Disclosed herein is a method for playing an interactive video on a user device. The method comprises receiving an interactive content file associated with a video to be played by the user device, the interactive content file comprising: one or more interactive tags arranged to be overlaid on the video when the video is played by the user device, wherein the one or more interactive tags have associated information which is accessible by a user when a respective tag is selected via a user interface of the user device; and information defining a location and a time for the tag to be overlaid on the video. The method also comprises receiving the video and combining, at a processor, the video and the one or more interactive tags in accordance with the information defining the location and the time for the tag to be overlaid on the video to produce an interactive video for display, and playing the interactive video for display. A method for creating an interactive content file associated with a video is also disclosed. Corresponding systems and computer program products are also disclosed.

Journal ArticleDOI
Sung-Ho Bae1, Jaeil Kim1, Munchurl Kim1, Cho Suk Hee, Jin Soo Choi 
TL;DR: It is shown that the HEVC-encoded YUV420 4K-UHD video at the target bit rate of 18 Mb/s is perceptually good enough in the usage of legacy DTV broadcasting systems with single channel bandwidths of 6 MHz and higher perceptual quality in the 420 YUV color format than in the YUV444 color format at the same target bitrates.
Abstract: In this paper, we present the experimental results with analysis for subjective quality assessment on 4K-ultra-high definition (UHD) video encoded by high efficiency video coding (HEVC) to investigate its feasibility for 4K-UHD TV broadcasting services. The subjective quality assessment on the HEVC-encoded (impaired) 4K-UHD video is conducted for the three target bitrates of 18, 23, and 36 Mb/s, the two color formats of YUV420 and YUV444, and the two viewing distances of 0.75 times the height of a display screen (H) and 1.5 H. The subjective quality assessment experiments on 4K-UHD test sequences show that the HEVC-encoded (impaired) YUV420 4K-UHD video even at the target bit of 18 Mb/s yields perceptual similar visual quality: 1) with MOS of 4.7 and 4.8 in 0.75 H and 1.5 H viewing distances, respectively, compared to the original YUV420 4K-UHD test video; and 2) with MOS of 4.8 in 1.5 H viewing distance, compared to the original YUV444 4K-UHD test video. This implies that the HEVC-encoded YUV420 4K-UHD video at the target bit rate of 18 Mb/s is perceptually good enough in the usage of legacy DTV broadcasting systems with single channel bandwidths of 6 MHz. It is also shown that the HEVC-encoded 4K-UHD test sequences mostly yield higher perceptual quality in the 420 YUV color format than in the YUV444 color format at the same target bitrates.

Patent
09 Sep 2013
TL;DR: In this article, a software application for mobile devices enables users to easily create a fully-edited short video by combining video clips of various lengths to form a final video that resembles a Hollywood-style, professionally edited video clip.
Abstract: A software application for mobile devices enables users to easily create a fully-edited short video by combining video clips of various lengths to form a final video that resembles a Hollywood-style, professionally edited video clip. The videos are automatically edited to the music cuts using pre-programmed storyboards and transitions that align with the user's thematic selection. There are few steps involved in the process making for a user-friendly experience. The professional style video clip is produced on a user's phone in only 45 seconds and can then be shared with friends via email, YouTube, Facebook and other forms of social media.

Journal ArticleDOI
TL;DR: A novel approach for video summarization that works in the compressed domain and allows the progressive generation of a video summary is presented.

Patent
25 Oct 2013
TL;DR: In this paper, a system and method for navigating through image frames of video content to display particular frames of the video content while the video contents continues to play is presented, which provides a viewer the ability to interact with and operate a device playing video content using touch input or input from a peripheral device.
Abstract: A system and method are presented for navigating through image frames of video content to display particular frames of the video content while the video content continues to play. The system provides a viewer the ability to interact with and operate a device playing video content using touch input or input from a peripheral device to permit locating and displaying particular frames of video content concurrently with ongoing display of the video content.

Patent
24 Dec 2013
TL;DR: In this article, a video player presents a video stream to a user during presentation of the video stream, the video player displays at least a portion of a library of selectable pre-recorded video segments, and a selection of one or more of the pre recorded video segments from the library is received from the user.
Abstract: Method and accompanying system for providing an in-video library A video player presents a video stream to a user During presentation of the video stream, the video player displays at least a portion of a library of selectable prerecorded video segments, and a selection of one or more of the prerecorded video segments from the library is received from the user The video player further displays a video segment watch-list that includes the selected prerecorded video segments The video player then presents the segments in the video segment watch-list as part of the video stream

Patent
27 Dec 2013
TL;DR: In this paper, a digital video camera system that provides a video summary using a method that includes: designating a reference image containing a particular person; capturing a video sequence of the scene using the image sensor, the video sequence including a time sequence of image frames; processing the captured video sequence using a video processing path to form a video file; and storing a representation of the video summary in the storage memory.
Abstract: A digital video camera system that provides a video summary using a method that includes: designating a reference image containing a particular person; capturing a video sequence of the scene using the image sensor, the video sequence including a time sequence of image frames; processing the captured video sequence using a video processing path to form a digital video file; during the capturing of the video sequence, analyzing the captured image frames using a person recognition algorithm to identify a subset of the image frames that contain the particular person; forming the video summary including fewer than all of the image frames in the captured video sequence, wherein the video summary includes at least part of the identified subset of image frames containing the particular person; storing the digital video file in the storage memory; and storing a representation of the video summary in the storage memory.

Patent
25 Apr 2013
TL;DR: In this article, a source encoder configured to encode source video as a number of alternative video streams includes a parallel processing system configured to collect statistics on source video data and write the statistics to shared memory in a first pass through received multimedia content.
Abstract: Systems and methods for encoding multiple video streams in accordance with embodiments of the invention are disclosed. In one embodiment, a source encoder configured to encode source video as a number of alternative video streams includes a parallel processing system configured to collect statistics on source video data and write the statistics to shared memory in a first pass through received multimedia content, determine initial encoding information for source video data and write the initial encoding information to shared memory during the first pass, encode the source video data in parallel using collected statistics and initial encoding information to produce a plurality of alternative video streams during a second pass, and parallel encoding processes are configured to reuse additional encoding information that has already been determined for a portion of video and to generate additional encoding information that has not already been determined for a portion of video.

Proceedings ArticleDOI
02 Dec 2013
TL;DR: This demo presents an extensible video processing framework in Apache Hadoop to parallelize video processing tasks in a cloud environment, using FFmpeg for a video coder and OpenCV for a image processing engine.
Abstract: Digital video is prominent big data spread all over the Internet. It is large not only in size but also in required processing power to extract useful information. Fast processing of excessive video reels is essential on criminal investigations, such as terrorism. This demo presents an extensible video processing framework in Apache Hadoop to parallelize video processing tasks in a cloud environment. Except for video transcending systems, there have been few systems that can perform various video processing in cloud computing environments. The framework employs FFmpeg for a video coder, and OpenCV for a image processing engine. To optimize the performance, it exploits MapReduce implementation details to minimize video image copy. Moreover, FFmpeg source code was modified and extended, to access and exchange essential data and information with Hadoop, effectively. A face tracking system was implemented on top of the framework for the demo, which traces the continuous face movements in a sequence of video frames. Since the system provides a web-based interface, people can try the system on site. In an 8-core environment with two quad-core systems, the system shows 75% of scalability.

Patent
15 Mar 2013
TL;DR: In this paper, a real-time video feed is received into a buffer, where user configurations are set for a time lapse video recording, based on a user configuration for frequency, certain frames in the buffer can be discarded prior to generating the video file.
Abstract: A time lapse video is created in real-time, without post-production. User configurations are set for a time lapse video recording. A real-time video feed is received into a buffer. Based on a user configuration for frequency, certain frames in the buffer can be discarded prior to generating the video file. The time lapse video is output as a video file according to user configurations.

Journal ArticleDOI
TL;DR: This article presents a novel approach, called Near-Lossless Semantic Summarization (NLSS), to summarize a video stream with the least high-level semantic information loss by using an extremely small piece of metadata.
Abstract: The ever increasing volume of video content on the Web has created profound challenges for developing efficient indexing and search techniques to manage video data. Conventional techniques such as video compression and summarization strive for the two commonly conflicting goals of low storage and high visual and semantic fidelity. With the goal of balancing both video compression and summarization, this article presents a novel approach, called Near-Lossless Semantic Summarization (NLSS), to summarize a video stream with the least high-level semantic information loss by using an extremely small piece of metadata. The summary consists of compressed image and audio streams, as well as the metadata for temporal structure and motion information. Although at a very low compression rate (around 1/40 of H.264 baseline, where traditional compression techniques can hardly preserve an acceptable visual fidelity), the proposed NLSS still can be applied to many video-oriented tasks, such as visualization, indexing and browsing, duplicate detection, concept detection, and so on. We evaluate the NLSS on TRECVID and other video collections, and demonstrate that it is a powerful tool for significantly reducing storage consumption, while keeping high-level semantic fidelity.

Journal ArticleDOI
TL;DR: The security and statistical analysis performed verify the effectiveness of the proposed security system for H.264/SVC and present a solution for the bit-rate and format compliance problems by careful selection of entropy coder syntax elements (bin-strings) for selective encryption (SE).
Abstract: The distribution of copyrighted scalable video content to differing digital devices requires protection during rendering and transmission. In this paper, we propose a complete security system for H.264/scalable video coding (SVC) video codec and present a solution for the bit-rate and format compliance problems by careful selection of entropy coder syntax elements (bin-strings) for selective encryption (SE), and the problem of managing multiple layer encryption keys for scalable video distribution. A standard key management protocol, multimedia Internet keying protocol, is implemented for the hierarchical key generation mechanism, in which a subscriber has only one encryption key to unlock all scalable layers that have been subscribed to. The evaluation demonstrates the resulting video quality degradation arising from SE for many CIF and 4CIF test video sequences, without there being any impact upon the bit-rate or format compliancy, and with small computational delay. The security and statistical analysis performed further verify the effectiveness of the proposed security system for H.264/SVC. The proposed system is highly suitable for video distribution to users who have subscribed to a varying degree of video quality on devices with medium to high computational resources.

Patent
06 Sep 2013
TL;DR: In this article, a memory and a processor are configured to determine a value of a current video unit of enhancement layer video information based at least on a weighted interlayer predictor and a weighted intra-layer predictor of at least one color component of the current unit.
Abstract: An apparatus for coding video data according to certain aspects includes a memory and a processor in communication with the memory. The memory is configured to store video information, such as base layer video information and enhancement layer video information. The processor is configured to determine a value of a current video unit of enhancement layer video information based at least on a weighted inter-layer predictor and a weighted intra-layer predictor of at least one color component of the current video unit.

Patent
Do Tae Won1, Gong Ki Choon1, Dong-Won Lee1, Kang Sang Ki1, Jung Kwang Soo1 
25 Feb 2013
TL;DR: In this paper, a transmission communication device provides information on a rotation angle of a video call image transmitted during a call to a reception communication device corresponding to a counterpart of the video call.
Abstract: A method and an apparatus perform a video call service. A transmission communication device provides information on a rotation angle of a video call image transmitted during a video call to a reception communication device corresponding to a counterpart of the video call. Then, the reception communication device renders the received video call image by using the information on the rotation angle of the received video call image, and displays a rendered video call image. Accordingly, an image of a counterpart of the video call can be normally provided during the video call.

Patent
01 Nov 2013
TL;DR: In this article, a system for loading videos includes an interactive video player with a loader, and external assets are configured for a design of an interactive layer of the video, which includes a plurality of video segments.
Abstract: A system for loading videos includes an interactive video player is with a loader. A product configuration file in operation configures files for a user in creation of a custom video. External assets are configured for a design of an interactive layer of the video. The interactive video player in operation creates in real-time a custom video that includes a plurality of video segments.

Patent
Ye-Kui Wang1, Ying Chen1
30 Jan 2013
TL;DR: In this article, a video file creation module encapsulates the plurality of slices of coded video content within one or more access units of a video stream, and a second type of parameter set may be encapsulated within a sample description.
Abstract: A device comprising a video file creation module is configured to obtain a plurality of slices of coded video content. Parameter sets are associated with the coded video content. The video creation module encapsulates the plurality of slices of coded video content within one or more access units of a video stream. A first type of parameter set may be encapsulated within one or more access units of the video stream. A second type of parameter set may be encapsulated within a sample description. The sample description may include stream properties associated with the video stream.

Patent
20 Aug 2013
TL;DR: In this paper, the authors present a system for delivering overhead video to a computing device by generating multiple versions of the overhead video having respective resolutions, bitrates, or frame rates.
Abstract: Examples of systems and methods for delivering overhead video to a computing device are provided. Delivering the overhead video can include generating multiple versions of the overhead video having respective resolutions, bitrates, or frame rates. The system can then generate georeferenced video data for each version by incorporating georeferencing coordinates and time synchronization information. The georeferenced video data can be stored in transcoded video files. In response to a request from a computing device for video data of a geographical region, the system can transmit a primary video stream comprising georeferenced video data of a relatively high quality and a secondary video stream comprising georeferenced video data of a relatively low quality, wherein the primary video stream includes the geographical region and the secondary video stream includes a proximal geographical region.

Patent
04 Oct 2013
TL;DR: In this paper, a set of video clips with related content can be ordered according to an ordering parameter and the ordered set can be stitched into an aggregate video, and overlapping and/or missing content can also be identified.
Abstract: Systems and methods for identifying sources associated with video clips uploaded by users and stitching those video clips into a single aggregate video according to a desired parameter and/or order. In particular, video clips uploaded by users can be matched to a source. Based upon processing of the video clip and/or source, a set of video clips with related content can be identified. That set of video clips can be ordered according to an ordering parameter. Overlapping and/or missing content can be identified, and the ordered set can be stitched into an aggregate video.

Patent
26 Sep 2013
TL;DR: In this paper, the video frames of the lower-resolution chroma sampling format can be unpacked to reconstruct the video frame of the higher-resolution sampling format, while still retaining higher resolution chroma information.
Abstract: Video frames of a higher-resolution chroma sampling format such as YUV 4:4:4 are packed into video frames of a lower-resolution chroma sampling format such as YUV 4:2:0 for purposes of video encoding. For example, sample values for a frame in YUV 4:4:4 format are packed into two frames in YUV 4:2:0 format. After decoding, the video frames of the lower-resolution chroma sampling format can be unpacked to reconstruct the video frames of the higher-resolution chroma sampling format. In this way, available encoders and decoders operating at the lower-resolution chroma sampling format can be used, while still retaining higher resolution chroma information. In example implementations, frames in YUV 4:4:4 format are packed into frames in YUV 4:2:0 format such that geometric correspondence is maintained between Y, U and V components for the frames in YUV 4:2:0 format.