scispace - formally typeset
Search or ask a question

Showing papers on "Smacker video published in 2002"


Proceedings ArticleDOI
03 Nov 2002
TL;DR: This work reports the first results on a Wyner-Ziv coding scheme for motion video that uses intraframe encoding, but interframe decoding, and suggests that an asymmetric video codec could achieve similar efficiency.
Abstract: In current interframe video compression systems, the encoder performs predictive coding to exploit the similarities of successive frames. The Wyner-Ziv theorem on source coding with side information available only at the decoder suggests that an asymmetric video codec, where individual frames are encoded separately, but decoded conditionally (given temporally adjacent frames) could achieve similar efficiency. We report the first results on a Wyner-Ziv coding scheme for motion video that uses intraframe encoding, but interframe decoding.

591 citations


Patent
15 Apr 2002
TL;DR: In this paper, an integrated, fully automated video production system that provides a video director with total control over all of the video production devices used in producing a show is presented. But, the system is not suitable for the production of live shows.
Abstract: An integrated, fully automated video production system that provides a video director with total control over all of the video production devices used in producing a show. Such devices include, but are not limited to, cameras, robotic pan/tilt heads, video tape players and recorders (VTRs), video servers and virtual recorders, character generators, still stores, digital video disk players (DVDs), audio mixers, digital video effects (DVE), video switchers, and teleprompting systems. The video production system provides an automation capability that allows the video director to pre-produce a show, review the show in advance of “air time,” and then, with a touch of a button, produce the live show. In one embodiment, the invention provides a video production system having a processing unit in communication with one or more of the video production devices mentioned above. The processing unit displays on a monitor graphical controls for controlling the variety of video production devices that it is in communication with. A video director uses a keyboard and mouse that are interfaced with the processing unit to activate the graphical controls, and thereby remotely control the video production devices from one location. The processing unit also enables the video director to automate the production of a show. According to one embodiment, the video director pre-produces the show, defines a set of video production commands or instructions (hereafter “transition macro”) to be executed by the processing unit, and then, by activating a control button displayed by the processing unit, the video director instructs the processing unit to execute the transition macro. Each video production command in a transition macro directs the processing unit to transmit in series and/or parallel one or more control commands to one or more of the video production devices when required.

283 citations


Proceedings ArticleDOI
Yu-Fei Ma1, Hong-Jiang Zhang1
10 Dec 2002
TL;DR: A new computational model of motion attention and the approach to applying this model in video skimming is presented and the effectiveness of the architecture and model is demonstrated by user studies of visual skimming experiments.
Abstract: One of the key issues in video manipulation is video abstraction in the form of skimmed video. For this purpose, an important task is to determine the content significance of each chunk of frames in a video sequence. In this paper, we present a new computational model of motion attention and the approach to applying this model in video skimming. The effectiveness of our architecture and model is demonstrated by user studies of visual skimming experiments. The results indicate that the precision of motion attention detection is over 80%, and the user satisfaction of visual skimming is beyond 70%.

183 citations


Proceedings ArticleDOI
25 Jun 2002
TL;DR: A video editor, called Silver, that uses metadata to make digital video editing more accessible to novices and offers smart editing operations that help users resolve the inconsistencies that arise because of the different boundaries in audio and video.
Abstract: Digital video is becoming increasingly ubiquitous. However, editing video remains difficult for several reasons: it is a time-based medium, it has dual tracks of audio and video, and current tools force users to work at the smallest level of detail. Based on interviews with professional video editors, we developed a video editor, called Silver, that uses metadata to make digital video editing more accessible to novices. To help users visualize video, Silver provides multiple views with different semantic content and at different levels of abstraction, including storyboard, editable transcript, and timeline views. Silver offers smart editing operations that help users resolve the inconsistencies that arise because of the different boundaries in audio and video. We conducted a preliminary user study to investigate the effectiveness of the Silver smart editing. Participants successfully edited video after only a short tutorial, both with and without smart editing assistance. Our research suggests several ways in which video editing tools could use metadata to assist users in the reuse and composition of video.

161 citations


Patent
28 Feb 2002
TL;DR: In this paper detects transition points in the audio signal and the video signal, which are used to align in time the video and audio signals, and the resulting edited video signal is merged with the audio signals to form a music video.
Abstract: Music videos are automatically produced from source audio and video signals. The music video contains edited portions of the video signal synchronized with the audio signal. An embodiment detects transition points in the audio signal and the video signal. The transition points are used to align in time the video and audio signals. The video signal is edited according to its alignment with the audio signal. The resulting edited video signal is merged with the audio signal to form a music video.

158 citations


Patent
18 Jul 2002
TL;DR: In this article, the authors present a system for mapping elements from a first XML format to a second XML format using an interface that allows a user to associate elements from the first format to the second format.
Abstract: Methods, apparatuses and systems facilitating mapping of elements from a first XML format to a second XML format using an interface that allows a user to associate elements from the first format to the second format. In some embodiments, a mapping can cause a direct transfer of a value in an input document to an output document. Maps can also be augmented with textual strings and scripts, for example, that can save a value from a first file format to a variable that can be accessed by another script, or save a value in association with a result tag in the second file format. A single tag from the first format can be mapped to multiple tags from the second format, multiple tags from the first format can be mapped to a single tag in the second format, and a single tag in the first format can be mapped multiple times to a single tag in the second format.

131 citations


Journal ArticleDOI
TL;DR: A motion-compensated, transform-domain super-resolution procedure for creating high-quality video or still images that directly incorporates the transform- domain quantization information by working with the compressed bit stream is proposed.
Abstract: There are a number of useful methods for creating high-quality video or still images from a lower quality video source. The best of these involve motion compensating a number of video frames to produce the desired video or still. These methods are formulated in the space domain and they require that the input be expressed in that format. More and more frequently, however, video sources are presented in a compressed format, such as MPEG, H.263, or DV. Ironically, there is important information in the compressed domain representation that is lost if the video is first decompressed and then used with a spatial-domain method. In particular, quantization information is lost once the video has been decompressed. Here, we propose a motion-compensated, transform-domain super-resolution procedure for creating high-quality video or still images that directly incorporates the transform-domain quantization information by working with the compressed bit stream. We apply this new formulation to MPEG-compressed video and demonstrate its effectiveness.

130 citations


Patent
29 Mar 2002
TL;DR: In this paper, the authors present a video reproduction apparatus that reproduces externally supplied package media, which contains video content storing video data and playback control data controlling reproduction of the video data in a specified data format and extensible application software using the video content.
Abstract: A video reproduction apparatus according to the present invention reproduces externally supplied package media. The package media contains video content storing video data and playback control data controlling reproduction of the video data in a specified data format, and extensible application software using the video content. The video reproduction apparatus includes as software pre-stored and executed in internal memory an operating system chosen from operating systems of plural types, middleware for absorbing differences in function according to the type of operating system, and a player application that runs on the middleware level for reproducing the video content. The middleware has a class library including tools used by the player application to play back the package media or to run the extensible application software. The player application consistently reproduces the video content of the package media according to the specified format by way of the tools included in the middleware class libraries. The extensible application software is run through the tools included in the class libraries of the middleware using video content contained in the same package media.

129 citations


Patent
26 Jun 2002
TL;DR: In this paper, a format conversion method comprising decoding the bit stream of a first encoded video data format, converting decoded video data to the second encoded data data format and encoding the converted video data in a process for converting the bit streams of the first encoded data format to the bitstream of the second encoder-decoder pair, and controlling processing parameters of at least one of the decoding, the converting and the encoding is presented.
Abstract: A format conversion method comprising decoding the bit stream of a first encoded video data format, converting decoded video data to the second encoded video data format, encoding the converted video data in a process for converting the bit stream of the first encoded video data format to the bit stream of the second encoded video data format, and controlling processing parameters of at least one of the decoding, the converting and the encoding.

127 citations


Book ChapterDOI
28 May 2002
TL;DR: A novel method for mosaic comparison is used and a highly compact non-temporal representation of video is created that allows accurate comparison of scenes across different videos and serves as a basis for indexing video libraries.
Abstract: We present an approach for compact video summaries that allows fast and direct access to video data. The video is segmented into shots and, in appropriate video genres, into scenes, using previously proposed methods. A new concept that supports the hierarchical representation of video is presented, and is based on physical setting and camera locations. We use mosaics to represent and cluster shots, and detect appropriate mosaics to represent scenes. In contrast to approaches to video indexing which are based on key-frames, our efficient mosaic-based scene representation allows fast clustering of scenes into physical settings, as well as further comparison of physical settings across videos. This enables us to detect plots of different episodes in situation comedies and serves as a basis for indexing whole video sequences. In sports videos where settings are not as well defined, our approachallo ws classifying shots for characteristic event detection. We use a novel method for mosaic comparison and create a highly compact non-temporal representation of video. This representation allows accurate comparison of scenes across different videos and serves as a basis for indexing video libraries.

117 citations


Journal ArticleDOI
TL;DR: Experimental results on the proposed online processing scheme combined with efficient VOS show the proposed integrated scheme generates desirable summarizations of surveillance videos.
Abstract: Key frames are the subset of still images which best represent the content of a video sequence in an abstracted manner. In other words, video abstraction transforms an entire video clip to a small number of representative images. We present a scheme for object-based video abstraction facilitated by an efficient video-object segmentation (VOS) system. In such a framework, the concept of a "key frame" is replaced by that of a "key video-object plane (VOP)." In order to achieve an online object-based framework such as an object-based video surveillance system, it becomes essential that semantically meaningful video objects are directly accessed from video sequences. Moreover, the extraction of key VOPs needs to be automated and context dependent so that they maintain the important contents of the video while removing all redundancies. Once a VOP is extracted, the shape of the VOP needs to be well described. To this end, both region-based and contour-based shape descriptors are investigated, and the region-based descriptor is selected for the proposed system. The key VOPs are extracted in a sequential manner by successive comparison with the previously declared key VOP. Experimental results on the proposed online processing scheme combined with efficient VOS show the proposed integrated scheme generates desirable summarizations of surveillance videos.

Proceedings ArticleDOI
01 Dec 2002
TL;DR: Methods for automatic and semi-automatic creation of music videos, given an arbitrary audio soundtrack and source video, are presented, which enhance the impact of both the soundtrack music and the video.
Abstract: We present methods for automatic and semi-automatic creation of music videos, given an arbitrary audio soundtrack and source video Significant audio changes are automatically detected; similarly, the source video is automatically segmented and analyzed for suitability based on camera motion and exposure Video with excessive camera motion or poor contrast is penalized with a high unsuitability score, and is more likely to be discarded in the final edit High quality video clips are then automatically selected and aligned in time with significant audio changes Video clips are adjusted to match the audio segments by selecting the most suitable region of the desired length Besides a fully automated solution, our system can also start with clips manually selected and ordered using a graphical interface The video is then created by truncating the selected clips (preserving the high quality portions) to produce a video digest that is synchronized with the soundtrack music, thus enhancing the impact of both

Patent
04 Jan 2002
TL;DR: In this paper, the video information is stored as a video file consisting of packet-divided video streams compression-coded with average bit rates t(c) which cover the clients' expected channel bit rates s each packet and the video file are supplied with a header containing information for realizing a bandwidth-scalable video transmission over a suitable version of HTTP.
Abstract: In a method for transmission of video information between HTTP servers and clients in a shared network resource, particularly Internet, the video information is stored as a video file consisting of packet-divided video streams compression-coded with average bit rates t(c) which cover the clients' expected channel bit rates s Each packet and the video file are supplied with a header containing information for realizing a bandwidth-scalable video transmission over a suitable version of HTTP During transmission switching between the video streams takes place on the basis pf an estimation of the channel bit rate s and the information in the packet header, such that the bit rate t(c) is adapted to the client's actual channel bit rate s In a method for client-executed search and retrieval of video information in a shared network resource, particularly searching of a video frame Fx in a packet-divided video stream, the packets in a video stream are divipacket packet in each group On basis of given search criteria and search-specific information in the packet headers the packet with the video five Fx is found, such that a pseudo-random search and retrieval are realized with the use of a suitable version of HTTP as transport protocol

Patent
14 Nov 2002
TL;DR: In this article, multiple video game players access an encrypted video game library stored on a shared mass storage device, where a secure bus communicates video game instructions to each of the video game playing units.
Abstract: Multiple video game players access an encrypted video game library stored on a shared mass storage device. A multiplexer multiplexes data read from the mass storage device to provide output streams to multiple video game playing units consuming video game instructions. A secure bus communicates video game instructions from the shared mass storage device to each of the video game playing units. Video game software or other entertainment content is distributed to the shared mass storage device via electronic download in multi-level encrypted form. Before being transported, the content is encrypted and then further encrypted. Once the content has been successfully transported, it is decrypted to remove the further encryption layer—leaving the first encryption layer intact for protecting the video game during storage on mass storage at the remote distribution location.

Patent
14 May 2002
TL;DR: In this paper, an exemplary video encoder that quantizes pel blocks of a video signal based upon one or more of the following: (i) an actual bit distribution of a previously encoded video frame, (ii) a required decoding delay for a splice point, and (iii) an activity measure of a pel block being encoded.
Abstract: An exemplary video encoder is provided that quantizes pel blocks of a video signal based upon one or more of the following: (i) an actual bit distribution of a previously encoded video frame, (ii) a required decoding delay for a splice point, and (iii) an activity measure of a pel block being encoded. Moreover, an exemplary video encoder uses the above quantization techniques to obtain an encoded video stream having seamless in points and seamless out points. Video encoding methods that utilize the above quantization methods are also provided.

Patent
15 Apr 2002
TL;DR: In this article, an automated lip sync error corrector embeds a unique video source identifier ID into the video signal from each of a plurality of video sources, which is used to access a corresponding delay value for an adjustable audio delay device to re-time a common audio signal to the selected video signal.
Abstract: An automated lip sync error corrector embeds a unique video source identifier ID into the video signal from each of a plurality of video sources. The unique video source ID may be in the form of vertical interval time code user bits or in the form of a watermark in an active video portion of the video signal. When one of the video signals is selected, the embedded unique video source ID is extracted. The extracted source ID is used to access a corresponding delay value for an adjustable audio delay device to re-time a common audio signal to the selected video signal. A look-up table may be used to correlate the unique video source ID with the corresponding delay value.

Patent
19 Jul 2002
TL;DR: In this paper, a technique is described for providing a client on a packet based network with a stream of encoded video data, which is able to maximise the bit-rate of the video stream by adapting to fluctuations in network capacity.
Abstract: A technique is described for providing a client on a packet based network with a stream of encoded video data. The system is able to maximise the bit-rate of the video stream by adapting to fluctuations in network capacity. The technique is characterised in that adaptation of the bit-rate of the transmitted encoded video data is timed to occur upon a scene change in the video sequence. In this way the interruption to the viewer when the perceived quality of the video sequence increases or decreases is minimised as it is 'hidden' in the scene change. The technique is described as applied to hierarchically encoded video data but equally may be applied to other encoding techniques which adapt to network conditions.

Patent
08 Feb 2002
TL;DR: In this article, the authors provide interactive access to video programming and video content information in a multimedia system, where a processing device of the multimedia system receives video content associated with video programming that is to be broadcast.
Abstract: Providing interactive access to video programming and video content information in a multimedia system. A processing device of the multimedia system receives video content information associated with video programming that is to be broadcast. The video content information includes at least video identification information and video content information. The display device of the multimedia system presents a first interface image that presents the video content information for one of the video programs of the video programming. In response to viewer input, a second interface image is displayed on the display device presenting video content information for another one of the video programs of the video programming or for more detailed information associated with the video program. In this manner, the viewer can conveniently navigate through the video content information and identify information that is useful in selecting a program to watch.

Book ChapterDOI
28 May 2002
TL;DR: A key feature of the system is the analysis of video input as a single entity as opposed to a sequence of separate frames, allowing for the detection and recognition of video events.
Abstract: In this work we describe a novel statistical video representation and modeling scheme. Video representation schemes are needed to enable segmenting a video stream into meaningful video-objects, useful for later indexing and retrieval applications. In the proposed methodology, unsupervised clustering via Guassian mixture modeling extracts coherent space-time regions in feature space, and corresponding coherent segments (video-regions) in the video content. A key feature of the system is the analysis of video input as a single entity as opposed to a sequence of separate frames. Space and time are treated uniformly. The extracted space-time regions allow for the detection and recognition of video events. Results of segmenting video content into static vs. dynamic video regions and video content editing are presented.

Patent
11 Mar 2002
TL;DR: In this article, a system and method for purchasing broadcast audio and/or video (audio/video) works is presented, which includes a data stream processor that receives a broadcast audio/video data stream and produces from the broadcast audio or video data stream a first audio or visual output data stream.
Abstract: A system and method for purchasing broadcast audio and/or video (audio/video) works. According to one embodiment, the system includes a data stream processor that receives a broadcast audio/video data stream and produces from the broadcast audio/video data stream a first audio/video output data stream and a second audio/video output data stream, wherein the broadcast audio/video data stream contains information corresponding to an audio/video work and the first audio/video output stream is in a format that can be recognized and processed by conventional audio/video reproducing equipment; a storage device that stores data included in the second audio/video output data stream, wherein the data corresponds to the audio/video work; and a user interface that enables a user of the system to input to the system an indication that the user desires to purchase the audio/video work.

Patent
03 Sep 2002
TL;DR: A video distribution device as discussed by the authors is a device communicating with a video receiving device via a communication network, which includes a video acquisition unit 110 that acquires plural videos from various perspectives, a video analysis unit 120 that analyses a detail contained in the video on a video basis and generates its analysis result as content information, and a video matching unit 130 that verifies a conformity level of each content information with preference information notified by a viewer, decides a video to be distributed, and distributes the video.
Abstract: A video distribution device 10 is a device communicating with a video receiving device 20 via a communication network 30, which includes a video acquisition unit 110 that acquires plural videos taken from various perspectives, a video analysis unit 120 that analyzes a detail contained in the video on a video basis and generates its analysis result as content information, and a video matching unit 130 that verifies a conformity level of each content information with preference information notified by a viewer, decides a video to be distributed, and distributes the video.

Patent
01 Nov 2002
TL;DR: In this paper, the key-frames of the video data sequence are identified independent of shot boundary detection and a static summary of shots in the video sequence is then generated based on key-frame importance.
Abstract: Systems and methods to generate a video summary of a video data sequence are described. In one aspect, key-frames of the video data sequence are identified independent of shot boundary detection. A static summary of shots in the video data sequence is then generated based on key-frame importance. For each shot in the static summary of shots, dynamic video skims are calculated. The video summary consists of the calculated dynamic video skims.

Patent
Michael Horowitz1, Rick Flott1
23 Aug 2002
TL;DR: In this article, a system and method for concealing video errors is presented. Butts et al. proposed a method to encode, reorder, and packetize video information into video data packets for transmission over a communication network such that the system conceals errors caused by lost video packets when the system receives, depacketizes, orders, and decodes the data packets.
Abstract: The present invention provides, in one embodiment, a system and method for concealing video errors. The system encodes, reorders, and packetizes video information into video data packets for transmission over a communication network such that the system conceals errors caused by lost video data packets when the system receives, depacketizes, orders, (915) and decodes the data packets. In one embodiment, the system and method encodes and packetizes video information, such that adjacent macroblocks are not placed in the same video data packets (925) . Additionally, the system and method may provide information accompanying the video data packets to facilitate the decoding process. An advantage to such a scheme is that errors due to video data packet loss are spatially distributed over a video frame. Thus, if regions of data surrounding a lost macroblock are successfully decoded, the decoder may predict motion vectors and spatial content with a higher degree of accuracy, which leads to higher video quality.

Patent
30 Sep 2002
TL;DR: In this article, a method and system for presenting an interactive multi-window video display which provides particular advantages in a one-way television broadcast system in which user selections can not be transmitted to a head end or other central station.
Abstract: A method and system for presenting an interactive multi-window video display which provides particular advantages in a one-way television broadcast system in which user selections can not be transmitted to a head end or other central station. The system comprises a receiver for receiving a multi-window video signal representing a multi-window video display comprising a number of reduced size video displays. The reduced size video displays each contain video representing programming being currently offered on a different channel. The system presents the multi-window video display on a display device such as a television and allows viewers to select of one of the reduced size video displays. When the reduced size video is selected for highlighting, the audio associated with that video is presented. When the highlighted video is selected, the receiver tunes to the channel upon which the video is available in full screen mode.

Patent
23 Oct 2002
TL;DR: In this paper, a system for sending and receiving multimedia transmissions over a network includes two or more clients and a server, where each client is connected to the network and generates and receives audio and video data via the network.
Abstract: A system for sending and receiving multimedia transmissions over a network includes two or more clients and a server. Each client is connected to the network and generates and receives audio and video data via the network. The server receives the audio and video data from the clients and sends the audio and video data to the clients. During the transmission of the audio and video data, the client and server dynamically determine the rate at which to transmit the audio and video data.

Patent
20 Sep 2002
TL;DR: In this paper, the authors present a system for providing multiple video signals simultaneously to an end user using a grid pattern with each sector in the grid phoning a different video signal, provided with a suitable user interface.
Abstract: Systems and devices for providing multiple video signals simultaneously to an end user Multiple video tuners are each tuned to a different video signal at a content provider's location The output of each tuner is then received by a video multiplexer which multiplexes the disparate video signals into a single composite video signal The resulting composite video signal, which presents the different video signals in a grid pattern with each sector in the grid phoning a different video signal, is then provided with a suitable user interface The user interface provides such information such as the source of each of the video signals and allows an end user to navigate between the different video signals The resulting composite video signal with its user interface, is then encoded and transmitted to an end user device by way of a network An end user, using the end user device, can therefore simultaneously view multiple video signals without the need for switching channels

Patent
05 Feb 2002
TL;DR: In this article, a system and method for video cataloging is described, where the video is cataloged according to predefined or user definable metadata, and the metadata is used to index and then retrieve encoded video.
Abstract: One aspect of the invention is directed to a system and method for video cataloging. The video is cataloged according to predefined or user definable metadata. The metadata is used to index and then retrieve encoded video.

Patent
28 Jan 2002

Proceedings ArticleDOI
Yijin Wang1, Peng Zhao1, Dong Zhang1, Mingjing Li1, Hong-Jiang Zhang1 
01 Dec 2002
TL;DR: MyVideos is a prototype system for managing home digital videos that provides the basic functionalities of video segmentation, summarization, grouping, and editing.
Abstract: MyVideos is a prototype system for managing home digital videos. It provides the basic functionalities of video segmentation, summarization, grouping, and editing. It also supports the video playback by shot, group, or highlight.

Proceedings ArticleDOI
10 Dec 2002
TL;DR: An adaptive video transcoding method is obtained, which combines intra-coding and inter-coded with a fast motion vector reestimation method to strike a good balance between computational complexity and transcoded video quality.
Abstract: Fast forward and fast reverse playbacks are two common video browsing functions provided in many analog and digital video players They help users quickly find and access video segments of interest by scanning through the content of a video at a faster than normal playback speed We propose a video transcoding approach to realizing fast forward and reverse video playbacks by generating a new compressed video from a pre-coded video To reduce the computational requirements, we design and compare several fast algorithms for estimating the motion vectors required in transcoded video To accommodate changes due to frame skipping for fast video playback, we also alter the group-of-pictures structure of transcoded video In addition, subjective tests are conducted to assess the minimum video peak-signal-to-noise-ratio degradation that is perceptible to viewers at different fast playback speeds To this end, we obtain an adaptive video transcoding method, which combines intra-coding and inter-coding with a fast motion vector reestimation method to strike a good balance between computational complexity and transcoded video quality Experimental results are reported to show the efficacy of the proposed method