scispace - formally typeset
Search or ask a question

Showing papers on "Smacker video published in 2007"


Journal ArticleDOI
TL;DR: The purpose of this article is to provide a systematic classification of various ideas and techniques proposed towards the effective abstraction of video contents, and identify and detail, for each approach, the underlying components and how they are addressed in specific works.
Abstract: The demand for various multimedia applications is rapidly increasing due to the recent advance in the computing and network infrastructure, together with the widespread use of digital video technology. Among the key elements for the success of these applications is how to effectively and efficiently manage and store a huge amount of audio visual information, while at the same time providing user-friendly access to the stored data. This has fueled a quickly evolving research area known as video abstraction. As the name implies, video abstraction is a mechanism for generating a short summary of a video, which can either be a sequence of stationary images (keyframes) or moving images (video skims). In terms of browsing and navigation, a good video abstract will enable the user to gain maximum information about the target video sequence in a specified time constraint or sufficient information in the minimum time. Over past years, various ideas and techniques have been proposed towards the effective abstraction of video contents. The purpose of this article is to provide a systematic classification of these works. We identify and detail, for each approach, the underlying components and how they are addressed in specific works.

879 citations


Proceedings ArticleDOI
12 Nov 2007
TL;DR: The impact on image quality of rendered arbitrary intermediate views is investigated and analyzed in a second part, comparing compressed multi-view video plus depth data at different bit rates with the uncompressed original.
Abstract: A study on the video plus depth representation for multi-view video sequences is presented. Such a 3D representation enables functionalities like 3D television and free viewpoint video. Compression is based on algorithms for multi-view video coding, which exploit statistical dependencies from both temporal and inter-view reference pictures for prediction of both color and depth data. Coding efficiency of prediction structures with and without inter-view reference pictures is analyzed for multi-view video plus depth data, reporting gains in luma PSNR of up to 0.5 dB for depth and 0.3 dB for color. The main benefit from using a multi-view video plus depth representation is that intermediate views can be easily rendered. Therefore the impact on image quality of rendered arbitrary intermediate views is investigated and analyzed in a second part, comparing compressed multi-view video plus depth data at different bit rates with the uncompressed original.

485 citations


Patent
28 Jun 2007
TL;DR: A bar code reading terminal can, in one embodiment, include a two-dimensional image sensor and an imaging assembly as discussed by the authors, which can be configured, in response to receipt of a trigger signal, to buffer a set of video frames and subject each frame of the set of frames to decode attempt.
Abstract: A bar code reading terminal can, in one embodiment, include a two-dimensional image sensor and an imaging assembly. The terminal can be configured, in response to receipt of a trigger signal, to buffer a set of frames and subject each frame of the set of frames to a decode attempt. The terminal can also be configured to convert the convert the captured set of video frames into a compressed video format, and transfer the resulting video file to a file server or to one or more playback devices, in response to a user interface action, or in response to difficult read condition having been detected. The terminal can also be configured to establish two-way video- and audio-communication with a remote video- and audio-streaming device. The video capturing and streaming operation of the terminal can be controlled remotely from a playback device via web interface.

356 citations


Patent
23 Jan 2007
TL;DR: In this paper, the authors describe a system and methods for automatically inserting advertisements into source video content playback streams, based on which a contextually relevant and/or targeted video advertisement clip is inserted into the playback stream for presentation to the user.
Abstract: Systems and methods for automatically inserting advertisements into source video content playback streams are described. In one aspect, the systems and methods communicate a source video content playback stream to a video player to present source video to a user. During playback of the source video, and in response to receipt of a request from the user to navigate portions of the source video (e.g., a user command to fast forward the source video, rewind the source video, or other action), the systems and methods dynamically define a video advertisement clip insertion point (e.g., and insertion point based on a current playback position). The systems and methods then insert a contextually relevant and/or targeted video advertisement clip into the playback stream for presentation to the user.

214 citations


Patent
27 Jun 2007
TL;DR: In this paper, a multimedia device integration system for car stereo or video systems is presented. And a docking station is provided for docking a portable audio or video device for integration with the car stereo and video system.
Abstract: An multimedia device integration system is provided. One or more after-market audio or video devices, such as a CD player, CD changer, digital media device (e.g., MP3 player, MP4 player, WMV player, Apple iPod, portable media center, or other device) satellite receiver (e.g., XM or Sirius receiver), DAB receiver, video device (e.g., DVD player), cellular telephone, or any other device or combinations thereof, is integrated for use with an existing OEM or after-market car stereo or video system, wherein control commands can be issued at the car stereo or video system and data from the after-market device can be displayed on the car stereo or video system. Control commands generated at the car stereo or video system are received, processed, converted into a format recognizable by the after-market device, and dispatched to the after-market device for execution. Information from the after-market device is converted into a format recognizable by the car stereo or video system, and dispatched to the car stereo or video system for display thereon. One or more auxiliary input sources can be integrated with the car stereo or video system, and selected using the controls of the car stereo or video system. A docking station is provided for docking a portable audio or video device for integration with the car stereo or video system. Wireless integration between the portable audio or video device and a car stereo or video system is provided, and voice recognition and speech synthesis capabilities are provided in the portable audio or video device or the car stereo or video system.

211 citations


Patent
29 Oct 2007
TL;DR: In this paper, a user provides an indication that one or more particular video segments are to be shared over the network, and the video segment is/are then automatically assessed and determined to be compatible with streaming video, or not.
Abstract: A user can create a video segment or employ an existing video segment, and then share it over a computer network. The user provides an indication that one or more particular video segments are to be shared over the network. The video segment(s) is/are then automatically assessed and determined to be compatible with streaming video, or not. If the video segment(s) is/are not compatible with streaming video, it/they are converted to a compatible format automatically. An identifier for the video segment is automatically created and the segment and the identifier are automatically uploaded to a host computer over the network such as the Internet. The video segment and the identifier (optionally with other identifying material such as an identity of the sender, an access authorization for the video, a number of accesses permitted, and a duration for the availability of the video) can be stored at the direction of the host computer. A viewer can be sent an identifier of the video, and can request that the video be served as a streaming video to the viewer's computer. Alternatively, the viewer can be sent a location of the video such as a URL, can be served the video as an embedded portion of a Web page, or can be served the video as a consequence of being sent a link in an e-mail or as an e-mail greeting card.

206 citations


Journal ArticleDOI
Peter Amon1, T. Rathgen, D. Singer
TL;DR: This paper describes the file format defined for scalable video coding, which enables rapid extraction of scalable data, corresponding to the desired operating point, in a variety of usages and application scenarios.
Abstract: This paper describes the file format defined for scalable video coding. Techniques in the file format enable rapid extraction of scalable data, corresponding to the desired operating point. Significant assistance to file readers can be provided, and there is also great flexibility in the ways that the techniques can be used and combined, corresponding to different usages and application scenarios.

173 citations



Patent
28 May 2007
TL;DR: In this paper, a method of storing video data from a video surveillance system having one or more cameras is presented, where the video data is stored in a hierarchical video storage area, the meta-data, indexed by date and time stamp to the video, is stored on a meta data storage area; and the attribute data and associated weights are stored in an attribute storage area.
Abstract: One embodiment is a method of storing video data from a video surveillance system having one or more cameras. Video data is captured from one or more surveillance cameras. Meta-data is automatically generated by performing video analysis on the captured video data from the surveillance cameras. A human operator may manually enter additional meta-data. Attribute data and associated weights, representing information about the relevance of the meta-data, is received. The video data is stored in a hierarchical video storage area; the meta-data, indexed by date and time stamp to the video data, is stored in a meta-data storage area; and the attribute data is stored in an attribute storage area. One or more alerts may be issued based on the past and present meta-data. The video data is secured by encrypting and storing the video data remotely, and audit trails are generated about who and when viewed the video data.

151 citations


Journal ArticleDOI
TL;DR: This work states that the proposed encoder architecture, which combines a block- based transform and interframe predictive coding approach, is well-suited for applications where the video is encoded once and decoded many times, i.e., one-to-many topologies, such as broadcasting or video-on-demand, where the cost of the decoder is more critical than thecost of the encoder.
Abstract: Growing percentage of the world population now uses image and video coding technologies on a regular basis. These technologies are behind the success and quick deployment of services and products such as digital pictures, digital television, DVDs, and Internet video communications. Today's digital video coding paradigm represented by the ITU-T and MPEG standards mainly relies on a hybrid of block- based transform and interframe predictive coding approaches. In this coding framework, the encoder architecture has the task to exploit both the temporal and spatial redundancies present in the video sequence, which is a rather complex exercise. As a consequence, all standard video encoders have a much higher computational complexity than the decoder (typically five to ten times more complex), mainly due to the temporal correlation exploitation tools, notably the motion estimation process. This type of architecture is well-suited for applications where the video is encoded once and decoded many times, i.e., one-to-many topologies, such as broadcasting or video-on-demand, where the cost of the decoder is more critical than the cost of the encoder.

142 citations


Patent
10 Oct 2007
TL;DR: In this paper, a method and system for accessing video file content is provided, where a user can select to view the video content and wait for the server to transcode the video file and to stream the transcoded video file to the user's client device.
Abstract: A method and system for accessing video file content is provided. When a user encounters a webpage with video content, the user can select to view the video content and wait for the server to transcode the video file and to stream the transcoded video file to the user's client device. Alternatively, the user may request that the server transcode the video file and to send the transcoded video file to the user's device, where the transcoded video file will be stored. While waiting for the video to be transcoded, the user may browse other websites, for example. The user may then view the video file at a later time that is convenient by accessing a video file inbox. The transcoded video file could alternatively be stored at the server, which can send a notification to the user's device to indicate that the video file has been transcoded.

Journal ArticleDOI
TL;DR: A new multiview video coding scheme that can improve the compression efficiency under such a limited inter-view prediction structure and adopt the following three modifications: object-based interpolation on 3-D warping; depth estimation with consideration of rate-distortion costs; and quarter-pel accuracy depth representation.
Abstract: Multiview video coding demands high compression rates as well as view scalability, which enables the video to be displayed on a multitude of different terminals. In order to achieve view scalability, it is necessary to limit the inter-view prediction structure. In this paper, we propose a new multiview video coding scheme that can improve the compression efficiency under such a limited inter-view prediction structure. All views are divided into two groups in the proposed scheme: base view and enhancement views. The proposed scheme first estimates a view-dependent geometry of the base view. It then uses a video encoder to encode the video of base view. The view-dependent geometry is also encoded by the video encoder. The scheme then generates prediction images of enhancement views from the decoded video and the view-dependent geometry by using image-based rendering techniques, and it makes residual signals for each enhancement view. Finally, it encodes residual signals by the conventional video encoder as if they were regular video signals. We implement one encoder that employs this scheme by using a depth map as the view-dependent geometry and 3-D warping as the view generation method. In order to increase the coding efficiency, we adopt the following three modifications: (1) object-based interpolation on 3-D warping; (2) depth estimation with consideration of rate-distortion costs; and (3) quarter-pel accuracy depth representation. Experiments show that the proposed scheme offers about 30% higher compression efficiency than the conventional scheme, even though one depth map video is added to the original multiview video.

Journal ArticleDOI
TL;DR: A brief review of the state of the art of video analysis, indexing and retrieval is given and research directions are pointed to which could make searching and browsing of video archives based on video content, as easy as search and browsing (text) web pages.

Patent
19 Jun 2007
TL;DR: In this paper, video fingerprints provide a compact representation of the temporal locations of discontinuities in the video that can be used to quickly and efficiently identify video content, such as shot boundaries in video frame sequence or silent points in audio stream.
Abstract: A method and system generates and compares fingerprints for videos in a video library. The video fingerprints provide a compact representation of the temporal locations of discontinuities in the video that can be used to quickly and efficiently identify video content. Discontinuities can be, for example, shot boundaries in the video frame sequence or silent points in the audio stream. Because the fingerprints are based on structural discontinuity characteristics rather than exact bit sequences, visual content of videos can be effectively compared even when there are small differences between the videos in compression factors, source resolutions, start and stop times, frame rates, and so on. Comparison of video fingerprints can be used, for example, to search for and remove copyright protected videos from a video library. Furthermore, duplicate videos can be detected and discarded in order to preserve storage space.

Patent
Phuong Nguyen1, Alan Messer1
01 Mar 2007
TL;DR: In this paper, a method and system that implement a process for managing video information on a video recording device is described, where the management process involves receiving a video input signal including a video stream, obtaining content information corresponding to the video stream and associating the content information with the video streams, and storing the video streaming and associated content information in a database as a searchable index.
Abstract: A method and system that implement a process for managing video information on a video recording device. The management process involves receiving a video input signal including a video stream, obtaining content information corresponding to the video stream, associating the content information with the video stream, and storing the video stream and associated content information in a database as a searchable index for the video stream. The searchable index includes location of the stored video stream and identifies the associated content information. Then, content information from an input signal corresponding to a video stream is captured for analysis and searching the index to identify stored video streams related to the video program for suggestion to the user.

Patent
24 Sep 2007
TL;DR: In this paper, the authors proposed a method for video summarization, and more specifically to a system for segmenting and classifying data from a video in order to create a summary video that preserves and summarizes relevant content.
Abstract: The subject invention relates to a system and method for video summarization, and more specifically to a system for segmenting and classifying data from a video in order to create a summary video that preserves and summarizes relevant content. In one embodiment, the system first extracts appearance, motion, and audio features from a video in order to create video segments corresponding to the extracted features. The video segments are then classified as dynamic or static depending on the appearance-based and motion-based features extracted from each video segment. The classified video segments are then grouped into clusters to eliminate redundant content. Select video segments from each cluster are selected as summary segments, and the summary segments are compiled to form a summary video. The parameters for any of the steps in the summarization of the video can be altered so that a user can adapt the system to any type of video, although the system is designed to summarize unstructured videos where the content is unknown. In another aspect, audio features can also be used to further summarize video with certain audio properties.

Proceedings ArticleDOI
27 Jun 2007
TL;DR: An overview of multi-view video coding (MVC) is provided and its applications, requirements, and the reference software model for MVC are described.
Abstract: With the advancement of computer graphics and computer vision technologies, the realistic visual system can come true in the near future. The multi-view video system can provide an augmented realism through selective viewing experience. The multi-view video is a collection of multiple videos capturing the same 3D scene at different viewpoints. Since the data size of the multi-view video increases proportionally to the number of cameras, it is necessary to compress multi-view video data for efficient storage and transmission. This paper provides an overview of multi-view video coding (MVC) and describes its applications, requirements, and the reference software model for MVC.

Patent
09 May 2007
TL;DR: In this paper, a method and system for generating and comparing fingerprints for videos in a video library is presented, which provides a compact representation of the spatial and sequential characteristics of the video that can be used to identify video content.
Abstract: A method and system generates and compares fingerprints for videos in a video library. The video fingerprints provide a compact representation of the spatial and sequential characteristics of the video that can be used to quickly and efficiently identify video content. Because the fingerprints are based on spatial and sequential characteristics rather than exact bit sequences, visual content of videos can be effectively compared even when there are small differences between the videos in compression factors, source resolutions, start and stop times, frame rates, and so on. Comparison of video fingerprints can be used, for example, to search for and remove copyright protected videos from a video library. Further, duplicate videos can be detected and discarded in order to preserve storage space.

Patent
05 Jan 2007
TL;DR: In this paper, a digital video editing and playback system and methods of editing and playing back digital video are provided, which includes a video processor adapted to receive video segments from multiple sources.
Abstract: A digital video editing and playback system and methods of editing and playing back digital video are provided. The system includes a video processor adapted to receive video segments from multiple sources. The video segments include synchronization information. The video processor includes software instructions adapted to be executed by the video processor. The software instructions are adapted to evaluate the synchronization information from various video segments and to form associations between video segments from different sources that correspond to a common event.

Proceedings ArticleDOI
02 Jul 2007
TL;DR: This paper presents an automatic procedure for constructing a compact synthesized collage from a video sequence and formulate the generation of video collage as an energy minimization problem in which each of above desirability is represented by an energy term.
Abstract: This paper presents an automatic procedure for constructing a compact synthesized collage from a video sequence. The synthesized image called video collage, is a kind of static video summary - to select most representative images from video, to extract salient regions of interest (ROI) from these images and resize them according to their saliency, and to seamlessly arrange ROI on a given canvas with the temporal structure of video content preserved. We formulate the generation of video collage as an energy minimization problem in which each of above desirability is represented by an energy term. Unlike most existing video presentation schemes, video collage is more compact and visually appealing. We have applied video collage on several home videos and report superior performance in a user study compared with key existing approaches to video presentation.

Patent
22 Mar 2007
TL;DR: In this paper, a method and system for transmitting video information from a sender to a receiver over wireless channels, by inputting a frame of video information at the sender, packetizing the video information and transmitting the video packet from the sender to the receiver over a wireless channel, is presented.
Abstract: A method and system for transmitting video information from a sender to a receiver over wireless channels, by inputting a frame of video information at the sender, packetizing the video information and transmitting the video packet from the sender to the receiver over a wireless channel. Packetizing the video information is performed by segmenting the frame into one or more segments of video information, constructing a data payload from one of the segments, constructing a video header including information describing said one segment, forming a video packet from the video header and the data payload. The video header in each video packet uniquely defines the video information in the data payload of the video packet for the receiver to reconstruct the video frame for proper display of the data payload in a video stream.

Patent
Vlad Margulis1
20 Sep 2007
TL;DR: In this paper, the authors present a system for presenting graphical data and time data for audio and video clips, which can include a video display element, a clip selection element, and a timeline element to present both chronological and durational information about the video clips on a defined time scale.
Abstract: Techniques and systems for presenting graphical data and time data for audio and video clips. In some embodiments, a computer program product, encoded on a computer-readable medium, is operable to cause data processing apparatus to perform operations that include obtaining video data including multiple video clips and presenting the video data in a user interface. The user interface can include a video display element to play video from the video data, a clip selection element to receive selections of the video clips, and a timeline element to present both chronological and durational information about the video clips on a defined time scale. The operations can include receiving a selection of one of the video clips through the clip selection element, and adjusting, in response to the selection, the defined time scale in the timeline element based on the one of the video clips selected.

Patent
James D. Bennett1
26 Jan 2007
TL;DR: In this paper, a plurality of sequences of sub-frames of video data are generated from the sequence of full frames of the video data, with a different center point within each sub-frame.
Abstract: Video player circuitry used with encoded source video and a display. Decoder circuitry receives encoded source video and decodes the encoded source video to produce a sequence of full frames of video data. Pre-processing circuitry, pursuant to sub-frame information, generates a plurality of sequences of sub-frames of video data from the sequence of full frames of video data, a first sequence of the plurality of sequences of sub-frames of video data having a different center point within the sequence of full frames of video data than that of a second sequence of the plurality of sequences of sub-frames of video data. Post-processing circuitry, pursuant to supplemental information, modifies the plurality of sequences of sub-frames of video data to produce an output. Interface circuitry that delivers the output for subsequent presentation on the display.

Patent
Raja Neogi1
31 Oct 2007
TL;DR: In this article, a collection of sub-fingerprints constitute a fingerprint block and the size and temporal position of the block is chosen to represent the video that is being served, based on the video server's prior information about the video semantics.
Abstract: Systems and methods generate video fingerprints for video data streams. A collection of sub-fingerprints constitute a fingerprint block. The size and temporal position of the block is chosen to represent the video that is being served. The video server has prior information about the video semantics. Based on this information, the server generates control codes that identify which blocks, based on time/byte offsets, in the sequence best represent the video object. The video receiver is equipped with facilities to generate these fingerprint blocks. The block(s) are sent back to the server. The server is equipped with facilities to verify and record the mismatches with respect to references. The mismatch statistics may be archived for use in a number of applications.

Patent
16 Oct 2007
TL;DR: In this article, the authors provide interactive access to video programming and video content information in a multimedia system, where a processing device of the multimedia system receives video content associated with video programming that is to be broadcast.
Abstract: Providing interactive access to video programming and video content information in a multimedia system. A processing device of the multimedia system receives video content information associated with video programming that is to be broadcast. The video content information includes at least video identification information and video content information. The display device of the multimedia system presents a first interface image that presents the video content information for one of the video programs of the video programming. In response to viewer input, a second interface image is displayed on the display device presenting video content information for another one of the video programs of the video programming or for more detailed information associated with the video program. In this manner, the viewer can conveniently navigate through the video content information and identify information that is useful in selecting a program to watch.

Patent
28 Aug 2007
TL;DR: In this paper, a method of advertising during sequences of video images includes alerting a viewer of available video advertising content associated with an object displayed in a sequence of videos, and a viewer selection of the object is received during the sequence of images responsive to alerting the viewer, and an end of the video images is detected.
Abstract: A method of advertising during sequences of video images includes alerting a viewer of available video advertising content associated with an object displayed in a sequence of video images. A viewer selection of the object is received during the sequence of video images responsive to alerting the viewer, and an end of the sequence of video images is detected. The video advertising content is displayed responsive to receiving the viewer selection and after the end of the sequence of video images such that viewing of the sequence of video images is not interrupted. Related devices and computer program products are also discussed.

Patent
13 Apr 2007
TL;DR: In this paper, a hardware independent virtual camera that can be seamlessly integrated with existing video camera and computer system equipment is presented, which supports the ability to track a defined set of three-dimensional coordinates within a video stream and dynamically insert rendered 3-D objects within the video stream on a real-time basis.
Abstract: A method and apparatus are described that provide a hardware independent virtual camera that may be seamlessly integrated with existing video camera and computer system equipment. The virtual camera supports the ability to track a defined set of three-dimensional coordinates within a video stream and to dynamically insert rendered 3-D objects within the video stream on a real-time basis. The described methods and apparatus may be used to manipulate any sort of incoming video signal regardless of the source of the video. Exemplary application may include real-time manipulation of a video stream associated, for example, with a real-time video conference generated by a video camera, or a video stream generated by a video player (e.g., a video tape player, DVD, or other device) reading a stored video recording.

Patent
Xian-Sheng Hua1, Lai Wei1, Shipeng Li1
19 Mar 2007
TL;DR: In this paper, a video presentation UI includes a video playing region, a video slider bar region, and a filmstrip region that presents multiple static thumbnails for a video that is playable in the playing region.
Abstract: Smart video presentation involves presenting one or more videos in a video presentation user interface (IU). In example implementation, a video presentation UI includes a listing of multiple video entries, with each video entry including multiple static thumbnailes to represent the corresponding video. In another example implementation, a video presentation UI includes a scalable number of static thumbnails to represent a video, with the scalable number adjustable by a user with a scaling interface tool. In yet another example implementation, a video presentation UI includes a video playing region, a video slider bar region, and a filmstrip region that presents multiple static thumbnails for a video that is playable in the video playing region.

Patent
17 Jul 2007
TL;DR: In this paper, the authors propose a scheme to detect when an incoming video format is different from a native format and make a local decision to convert incoming video formats to a format native to the personal multimedia device.
Abstract: Personal multimedia devices can detect when an incoming video format is different from a native format and make a local decision to convert incoming video formats to a format native to the personal multimedia device. The personal multimedia device may include a media processor that comprises an MPEG decoder/encoder, graphics processors, and a video decoder/encoder. These components increase a frame rate of a received video signal when the frame rate of the received video signal is less than a frame rate of the native video format of the personal multimedia device and decrease the frame rate of the received video signal when the frame rate of the received video signal is greater than the native frame rate of the set-top box. The graphics processor scales a frame resolution of the frames in the received video signal to correspond to the native video format.

Patent
21 Feb 2007
TL;DR: In this article, the authors describe a video encoding acceleration service to increase one or more of the speed and quality of video encoding, which acts as an intermediary between an arbitrary video encoder computer program application and arbitrary video acceleration hardware.
Abstract: A video encoding acceleration service to increase one or more of the speed and quality of video encoding is described. The service acts as an intermediary between an arbitrary video encoder computer program application and arbitrary video acceleration hardware. The service receives one or more queries from the video encoder to identify implementation specifics of the video acceleration hardware. The service interfaces with the video acceleration hardware to obtain the implementation specifics. The service communicates the implementation specifics to the video encoder. The implementation specifics enable the video encoder to: (a) determine whether one or more of speed and quality of software encoding operations associated with the video encoder can be increased with implementation of a pipeline of one or more supported encoding pipeline configurations and capabilities, and (b) implement the pipeline by interfacing with the service.