scispace - formally typeset
Search or ask a question

Showing papers on "Closed captioning published in 2004"


Proceedings ArticleDOI
22 Aug 2004
TL;DR: A novel, graph-based approach, "MMG", to discover cross-modal correlations across the media in a collection of multimedia objects, where it outperforms domain specific, fine-tuned methods by up to 10 percentage points in captioning accuracy.
Abstract: Given an image (or video clip, or audio song), how do we automatically assign keywords to it? The general problem is to find correlations across the media in a collection of multimedia objects like video clips, with colors, and/or motion, and/or audio, and/or text scripts. We propose a novel, graph-based approach, "MMG", to discover such cross-modal correlations.Our "MMG" method requires no tuning, no clustering, no user-determined constants; it can be applied to any multimedia collection, as long as we have a similarity function for each medium; and it scales linearly with the database size. We report auto-captioning experiments on the "standard" Corel image database of 680 MB, where it outperforms domain specific, fine-tuned methods by up to 10 percentage points in captioning accuracy (50% relative improvement).

501 citations


Journal ArticleDOI
Martine Danan1
TL;DR: Audiovisual material enhanced with captions or interlingual subtitles is a particularly powerful pedagogical tool which can help improve the listening comprehension skills of second-language learners.
Abstract: Audiovisual material enhanced with captions or interlingual subtitles is a particularly powerful pedagogical tool which can help improve the listening comprehension skills of second-language learners. Captioning facilitates language learning by helping students visualize what they hear, especially if the input is not too far beyond their linguistic ability. Subtitling can also increase language comprehension and leads to additional cognitive benefits, such as greater depth of processing. However, learners often need to be trained to develop active viewing strategies for an efficient use of captioned and subtitled material. Multimedia can offer an even wider range of strategies to learners, who can control access to either captions or subtitles.

339 citations


Patent
Kuansan Wang1
11 May 2004
TL;DR: In this article, a speech input mode dynamically reports partial semantic parses, while audio captioning is still in progress, which is a significant departure from the turn-taking nature of a spoken dialogue.
Abstract: A method and system provide a speech input mode which dynamically reports partial semantic parses, while audio captioning is still in progress. The semantic parses can be evaluated with an outcome immediately reported back to the user. The net effect is that task conventionally performed in the system turn are now carried out in the midst of the user turn thereby presenting a significant departure from the turn-taking nature of a spoken dialogue.

174 citations


Proceedings ArticleDOI
27 Jun 2004
TL;DR: This paper proposes a novel, graph-based approach (GCap) which outperforms previously reported methods for automatic image captioning, and is fast and scales well, with its training and testing time linear to the data set size.
Abstract: Given an image, how do we automatically assign keywords to it? In this paper, we propose a novel, graph-based approach (GCap) which outperforms previously reported methods for automatic image captioning. Moreover, it is fast and scales well, with its training and testing time linear to the data set size. We report auto-captioning experiments on the "standard" Corel image database of 680 MBytes, where GCap outperforms recent, successful auto-captioning methods by up to 10 percentage points in captioning accuracy (50% relative improvement).

131 citations


Proceedings ArticleDOI
27 Jun 2004
TL;DR: This work examines the problem of automatic image captioning and proposes methods to discover correlations between image features and keywords, so that it can automatically find good keywords for a new image.
Abstract: We examine the problem of automatic image captioning. Given a training set of captioned images, we want to discover correlations between image features and keywords, so that we can automatically find good keywords for a new image. We experiment thoroughly with multiple design alternatives on large datasets of various content styles, and our proposed methods achieve up to a 45% relative improvement on captioning accuracy over the state of the art.

127 citations


Patent
12 Nov 2004
TL;DR: In this article, a personalized subtitle system includes a display device, such as a Heads Up Display (HUD) device, worn or carried by a user in a public venue such as movie theater, playhouse, or stadium.
Abstract: A personalized subtitle system. A personalized subtitle system includes a display device, such as a Heads Up Display (HUD) device, worn or carried by a user in a public venue such as a movie theater, playhouse, or stadium. The user utilizes the display device to select and read captioning or subtitle information for a public event such as a movie, play, or sporting event. In this way, subtitles in a variety of languages can be supplied for the public event. In another embodiment, the display device is used in conjunction with a conventional subtitle display system, such as a DVD player for home use. In either embodiment, the user can operate a control panel to select a desired language. Subtitles in the selected language are then displayed to a viewer wearing or using the display device. In this way, a viewer desiring subtitles may have subtitles displayed without disrupting another viewer's enjoyment of a public venue.

58 citations


Patent
13 May 2004
TL;DR: In this article, a remote control device for a TV can be used to select a word in a closed caption text, and if the word is associated with additional content, a list of the content is displayed on the TV.
Abstract: A remote control device for a TV can be used to select a word in a closed caption text, and if the word is associated with additional content, a list of the content is displayed on the TV. The user can select a piece of content from the list. A dictionary meaning of a selected word may also or alternatively be provided. Alternatively, only a subset of closed captioning words need be shown to the viewer, or the presented words may be defined by the content provider independently of closed captioning text.

58 citations


26 Apr 2004
TL;DR: A summary of the Fischlar-News-Stories system in operation is given by following a scenario in which it is used and also outlining how the underlying system realises the functions it offers.
Abstract: The "Fischlar" systems are a family of tools for capturing, analysis, indexing, browsing, searching and summarisation of digital video information. Fischlar-News-Stories, described in this paper, is one of those systems, and provides access to a growing archive of broadcast TV news. Fischlar-News-Stories has several notable features including the fact that it automatically records TV news and segments a broadcast news program into stories, eliminating advertisements and credits at the start/end of the broadcast. Fischlar-News-Stories supports access to individual stories via calendar lookup, text search through closed captions, automatically-generated links between related stories, and personalised access using a personalisation and recommender system based on collaborative filtering. Access to individual news stories is supported either by browsing keyframes with synchronised closed captions, or by playback of the recorded video. One strength of the Fischlar-News-Stories system is that it is actually used, in practice, daily, to access news. Several aspects of the Fischlar systems have been published before, bit in this paper we give a summary of the Fischlar-News-Stories system in operation by following a scenario in which it is used and also outlining how the underlying system realises the functions it offers.

34 citations


Book
01 Jan 2004
TL;DR: In this paper, a brief history of closed captioning is presented, including the Americans with Disabilities Act (ADA), the Telecommunications Act of 1996, and the FCC Rulings.
Abstract: Part I: Overview of Closed Captioning Why Closed Captioning? A Brief History of Captioning Part II: Captioning Law Overview of U.S. Captioning Law The Americans with Disabilities Act The Telecommunications Act of 1996 Other FCC Rulings Part III: Captioning Technology Where the Captions Live: A Line 21 Overview Encoding, Decoding, Troubleshooting Generating Captions in Realtime Part IV: Related Technologies Captions in Internet Streaming Web Site Accessibility Movie Theater Captioning Live Event Captioning Video Description for the Blind Part V: Captioning Around the World Part VII: Appendices, Glossary, and Index

30 citations


Patent
08 Dec 2004
TL;DR: In this paper, a digital television signal receiver automatically processes any analog auxiliary information and provides a television signal having closed caption information suitable for display, regardless of whether the auxiliary information is analog or digital closed captioning.
Abstract: A digital television signal receiver (20) processes a digital television signal and auxiliary information contained in the digital television signal regardless of whether the auxiliary information is analog auxiliary information (36) or digital auxiliary information (34) and produces a television display signal (44) suitable for displaying the processed auxiliary information In the case of the auxiliary information being closed caption information, the digital television signal receiver extracts and processes digital closed captioning information according to standard EIA-708 or extracts and processes analog closed caption information according to standard EIA-608 The television signal processing receiver thus allows the user the option of selecting a fallback method of obtaining captioning for a digital program signal if the digital program signal does not include digital closed captioning Thus, if a digital television signal does not include digital closed caption information, the digital television signal receiver automatically processes any analog auxiliary information and provides a television signal having closed caption information suitable for display

22 citations


Proceedings ArticleDOI
14 Jul 2004
TL;DR: A new highly interactive program to teach American sign language (ASL) for K-3 mathematics by 3D animation to increase the opportunity for deaf children to learn arithmetic via interactive media and the effectiveness of ASL K-8 mathematics teachers.
Abstract: We present a new highly interactive program to teach American sign language (ASL) for K-3 mathematics by 3D animation. We aim at increasing: (1) the opportunity for deaf children to learn arithmetic via interactive media; (2) the effectiveness of (hearing) parents in teaching arithmetic to their deaf children; (3) the effectiveness of ASL K-8 mathematics teachers. Using state-of-the-art 3D computer animation technology we have created a learning tool to be used for K-8 classroom and home learning, and for closed captioning in interactive media and Internet.

DOI
14 Jul 2004
TL;DR: A new highly interactive program to teach American Sign Language (ASL) for K-3 mathematics by 3D animation to increase the opportunity for deaf children to learn arithmetic via interactive media and the effectiveness of ASL K-8 mathematics teachers.
Abstract: We present a new highly interactive program to teach American Sign Language (ASL) for K-3 mathematics by 3D animation. We aim at increasing: (1) the opportunity for deaf children to learn arithmetic via interactive media; (2) the effectiveness of (hearing) parents in teaching arithmetic to their deaf children; (3) the effectiveness of ASL K-8 mathematics teachers. Using state-of-the-art 3D computer animation technology we have created a learning tool to be used for K-8 classroom and home learning, and for closed captioning in interactive media and internet.

Patent
Regis J. Crinon1
25 Oct 2004
TL;DR: In this article, a closed captioning configuration system is described, which can be integrated into a video encoder, a closed captions encoder or both, and can drive captioning for presentations which may differ by frame rate, interlacing, or frame encoding mode.
Abstract: A closed captioning configuration system is described. The system receives parameters of a digital video presentation and computes closed captioning parameters to drive a closed captions encoder, creating closed captions which are compatible with the presentation. In various implementations, the configuration system may be integrated into a video encoder, a closed captions encoder, or both. The configuration system, through analysis of the presentation parameters, can drive captioning for presentations which may differ by frame rate, interlacing, or frame encoding mode, and account for repetition of fields or frames.

Journal ArticleDOI
TL;DR: This paper explores how television fans appropriate video for personal expression and how technology can support such creative appropriation and the implications to digital rights management of this work and the potential for 'view source' television.
Abstract: This paper explores how television fans appropriate video for personal expression and how technology can support such creative appropriation. Televisions do not have an equivalent to a Web browser's view source option; however, programmes can be structured by their transcripts, embedded as closed captions in the signal of most shows. With our talkTV video-editing software, rearranging lines of dialogue automatically creates new scenes, thereby enabling television viewers to become authors and editors. We present a case study of how television fans used talkTV and conclude with a discussion of the implications to digital rights management of this work and the potential for ‘view source’ television.

Patent
Takeshi Kaminosono1
17 Nov 2004
TL;DR: In this article, a serial bus connecting the DVD-IC 20 and the TV-IC 10 is used to determine the status of the closed captioning in a DVD player and a TV receiver to prevent an OSD signal from being transmitted.
Abstract: A television set include a DVD-IC 20 for outputting a composite signal a containing subtitle information and closed caption information, a TV-IC 10 for outputting an OSD signal b based on the closed caption information identified in the composite signal a, and a serial bus connecting the DVD-IC 20 and the TV-IC 10 . The DVD-IC 20 informs the TV-IC 10 of the setting status of subtitle display based on the subtitle information in a DVD player. The TV-IC 10 , if the setting of subtitle display based on the subtitle information in the DVD player is ON, turns OFF the setting of subtitle display based on the closed caption information in the television receiver, thereby prohibiting an OSD signal b from being outputted.

Book
29 Mar 2004
TL;DR: Part I: Overview of Closed Captioning Why closed Captioning?

Patent
16 Jun 2004
TL;DR: In this article, a modular portable image captioning device consisting of an image caption control actuator (38), a first module (22), including a lens (26), a digital image receiving unit (40), and a second module (10) including a display (14) for showing non-captured digital images and a variable distance connection unit (32).
Abstract: The present invention is directed towards a modular portable image captioning device comprising an image captioning control actuator (38), a first module (22) including a lens (26) and a digital image receiving unit (40), a second module (10) including a display (14) for showing non-captured digital images and a variable distance connection unit (32). The first module is separate from and directly connectable to the second module via the variable distance connection unit (32) and arranged to transfer non-captured digital images to the display for letting a user select a view via the display for which a finalised digital image can be captured and stored using the image captioning control actuator. With the invention images that are obscured can still be captured.


Patent
12 Oct 2004
TL;DR: In this article, a transcription system for providing captions to the soundtrack of a program for broadcast, including a broadcast site (22), and a remote captioning site (12) linked by a secure communications channel in a virtual private network (16) and incorporating respective codec devices (22.4, 12.4) to enable real-time transfer of audio data and re-transmission of caption data with minimal delay to broadcast.
Abstract: A transcription system (10) for providing captions to the soundtrack of a program for broadcast, including a broadcast site (22), and a remote captioning site (12) linked by a secure communications channel in a virtual private network (16) and incorporating respective codec devices (22.4, 12.4) to enable real-time transfer of audio data and re-transmission of caption data with minimal delay to broadcast. A transcription method suited for use with the system is also described.


BookDOI
17 May 2004
TL;DR: Information Retrieval techniques are applied to closed caption transcripts to create an application, called the "Similar Program Guide (SPG), which would notify the user of programs being aired on other channels, similar to the one he/she is currently watching orsimilar to the ones in the user's profile.
Abstract: The advent of Digital Television and the "Connected Home" has led to an information overload to the viewer. A lot of research has been conducted towards developing interactive program guide services and products that attempt to simplify the television viewing experience. Closed captions that are part of the transmission signal of the programs aired have been identified as a good source of information [1]. Information Retrieval techniques may be applied to closed caption transcripts to create applications such as the one we describe here, called the "Similar Program Guide (SPG)". Such a guide would notify the user of programs being aired on other channels, similar to the one he/she is currently watching or similar to the ones in the user's profile. Additionally, this guide would allow users to compile their favorite programs and thus maintain user profiles. In this paper we discuss the approaches for development of the application. We use Information Retrieval techniques as the basis for this application.

Patent
09 Feb 2004
TL;DR: In this article, the system has a broadcast listening unit and a computer screen that display alphanumeric characters, and a processing input unit automatically processes a voice of a user into a transcript text piece through the computer screen.
Abstract: The system has a broadcast listening unit (3) and a computer screen (4) that display alphanumeric characters (5). A processing input unit (7) automatically processes a voice of a user into a transcript text piece through the computer screen. Restructuring captioning lines are formed by recomposing the transcript text piece, where the lines are connected to an output device (8).