scispace - formally typeset
Search or ask a question
Topic

Closed captioning

About: Closed captioning is a research topic. Over the lifetime, 3011 publications have been published within this topic receiving 64494 citations. The topic is also known as: CC.


Papers
More filters
Journal ArticleDOI
18 Oct 2021
TL;DR: In this paper, a qualitative account of DHH people's real-time captioning experiences during small-group conversation and future design considerations to better support the groups being captioned, both in person and online, is presented.
Abstract: Real-time captioning is a critical accessibility tool for many d/Deaf and hard of hearing (DHH) people. While the vast majority of captioning work has focused on formal settings and technical innovations, in contrast, we investigate captioning for informal, interactive small-group conversations, which have a high degree of spontaneity and foster dynamic social interactions. This paper reports on semi-structured interviews and design probe activities we conducted with 15 DHH participants to understand their use of existing real-time captioning services and future design preferences for both in-person and remote small-group communication. We found that our participants' experiences of captioned small-group conversations are shaped by social, environmental, and technical considerations (e.g., interlocutors' pre-established relationships, the type of captioning displays available, and how far captions lag behind speech). When considering future captioning tools, participants were interested in greater feedback on non-speech elements of conversation (e.g., speaker identity, speech rate, volume) both for their personal use and to guide hearing interlocutors toward more accessible communication. We contribute a qualitative account of DHH people's real-time captioning experiences during small-group conversation and future design considerations to better support the groups being captioned, both in person and online.?

18 citations

Proceedings ArticleDOI
01 Dec 2020
TL;DR: In this article, a BiGRU-based encoder-decoder architecture was proposed to extract subject-verb embeddings using the subjects and verbs from the audio captions.
Abstract: Audio captioning is a recently proposed task for automatically generating a textual description of a given audio clip. Most existing approaches use the encoder-decoder model without using semantic information. In this study, we propose a bi-directional Gated Recurrent Unit (BiGRU) model based on encoder-decoder architecture using audio and semantic embed-dings. To obtain semantic embeddings, we extract subject-verb embeddings using the subjects and verbs from the audio captions. We use a Multilayer Perceptron classifier to predict subject-verb embeddings of test audio clips for the testing stage. Within the aim of extracting audio features, in addition to log Mel energies, we use a pretrained audio neural network (PANN) as a feature extractor which is used for the first time in the audio captioning task to explore the usability of audio embeddings in the audio captioning task. We combine audio embeddings and semantic embeddings to feed the BiGRU-based encoder-decoder model. Following this, we evaluate our model on two audio captioning datasets: Clotho and AudioCaps. Experimental results show that the proposed BiGRU-based deep model significantly outperforms the state of the art results across different evaluation metrics and inclusion of semantic information enhance the captioning performance.

18 citations

Journal ArticleDOI
TL;DR: Pedagogical implications that captioning support, added or removed, based on learner self-reports, may not be inherently beneficial, as perceptions on the reliance of captioning may be inaccurate.
Abstract: Instructional support has been widely discussed as a strategy to optimize student-learning experiences. This study examines instructional support within the context of a multimedia language-learning environment, with the predominant focus on learners’ perceptions of captioning support for listening comprehension. The study seeks to answer two questions: (1) do learners’ perceptions regarding dependence on captions match their actual reliance on captioning for listening comprehension? and (2) which learners’ perceptions are most influenced by proficiency: low-intermediate, intermediate, or high-intermediate? A total of 139 students from a high school English course in northern Taiwan, all accustomed to multimedia instruction that includes full captions, completed an English language proficiency test as well as a caption reliance test (CRT), and also provided their perceived degree of reliance on captions for English listening comprehension. The results show that overall perceived reliance was significantly...

18 citations

Journal ArticleDOI
TL;DR: This method can utilize image and text data scraped from the internet respectively to improve the performance limited in concepts-decoder framework and can transfer the knowledge learned from web data to the standard dataset.

18 citations

Patent
30 Aug 2017
TL;DR: In this paper, a raw audio waveform including a non-speech sound is received and relevant features are extracted from the raw audio Waveform using a recurrent neural network (RNN) acoustic model.
Abstract: A method, computer readable medium, and system are disclosed for audio captioning. A raw audio waveform including a non-speech sound is received and relevant features are extracted from the raw audio waveform using a recurrent neural network (RNN) acoustic model. A discrete sequence of characters represented in a natural language is generated based on the relevant features, where the discrete sequence of characters comprises a caption that describes the non-speech sound.

18 citations


Network Information
Related Topics (5)
Feature vector
48.8K papers, 954.4K citations
83% related
Object detection
46.1K papers, 1.3M citations
82% related
Convolutional neural network
74.7K papers, 2M citations
82% related
Deep learning
79.8K papers, 2.1M citations
82% related
Unsupervised learning
22.7K papers, 1M citations
81% related
Performance
Metrics
No. of papers in the topic in previous years
YearPapers
2023536
20221,030
2021504
2020530
2019448
2018334