scispace - formally typeset
Journal ArticleDOI

Portable Camera-Based Assistive Text and Product Label Reading From Hand-Held Objects for Blind Persons

Reads0
Chats0
TLDR
A camera-based assistive text reading framework to help blind persons read text labels and product packaging from hand-held objects in their daily lives and a novel text localization algorithm by learning gradient features of stroke orientations and distributions of edge pixels in an Adaboost model.
Abstract
We propose a camera-based assistive text reading framework to help blind persons read text labels and product packaging from hand-held objects in their daily lives. To isolate the object from cluttered backgrounds or other surrounding objects in the camera view, we first propose an efficient and effective motion-based method to define a region of interest (ROI) in the video by asking the user to shake the object. This method extracts moving object region by a mixture-of-Gaussians-based background subtraction method. In the extracted ROI, text localization and recognition are conducted to acquire text information. To automatically localize the text regions from the object ROI, we propose a novel text localization algorithm by learning gradient features of stroke orientations and distributions of edge pixels in an Adaboost model. Text characters in the localized text regions are then binarized and recognized by off-the-shelf optical character recognition software. The recognized text codes are output to blind users in speech. Performance of the proposed text localization algorithm is quantitatively evaluated on ICDAR-2003 and ICDAR-2011 Robust Reading Datasets. Experimental results demonstrate that our algorithm achieves the state of the arts. The proof-of-concept prototype is also evaluated on a dataset collected using ten blind persons to evaluate the effectiveness of the system's hardware. We explore user interface issues and assess robustness of the algorithm in extracting and reading text from different objects with complex backgrounds.

read more

Citations
More filters
Book ChapterDOI

Recognizing Text-Based Traffic Guide Panels with Cascaded Localization Network

TL;DR: Experimental results demonstrate that the proposed framework outperforms multiple recently published text spotting frameworks in real highway scenarios.
Proceedings ArticleDOI

Real time text detection and recognition on hand held objects to assist blind people

TL;DR: This paper presents camera based system which will help blind person for reading text patterns printed on hand held objects and the framework to assist visually impaired persons to read text patterns and convert it into the audio output.
Proceedings ArticleDOI

Unambiguous Text Localization and Retrieval for Cluttered Scenes

TL;DR: A novel recurrent Dense Text Localization Network (DTLN) is proposed to sequentially decode the intermediate convolutional representations of a cluttered scene image into a set of distinct text instance detections, and a Context Reasoning Text Retrieval model is proposed, which jointly encodes text instances and their context information through a recurrent network, and ranks localized text bounding boxes by a scoring function of context compatibility.
Proceedings ArticleDOI

Text recognition and face detection aid for visually impaired person using Raspberry PI

TL;DR: This is a prototype for blind people to recognize the products in real world by extracting the text on image and converting it into speech by Tesseract Optical Character Recognition and e-Speak tool.
Journal ArticleDOI

Making Shopping Easy for People with Visual Impairment Using Mobile Assistive Technologies

TL;DR: This study provides an overview of the various technologies that have been developed in recent years to assist people with visual impairment in shopping tasks and gives an introduction to the latest direction in this area, which will help developers to incorporate such solutions into their research.
References
More filters
Journal ArticleDOI

Robust Real-Time Face Detection

TL;DR: In this paper, a face detection framework that is capable of processing images extremely rapidly while achieving high detection rates is described. But the detection performance is limited to 15 frames per second.
Proceedings ArticleDOI

Robust real-time face detection

TL;DR: A new image representation called the “Integral Image” is introduced which allows the features used by the detector to be computed very quickly and a method for combining classifiers in a “cascade” which allows background regions of the image to be quickly discarded while spending more computation on promising face-like regions.
Proceedings ArticleDOI

Adaptive background mixture models for real-time tracking

TL;DR: This paper discusses modeling each pixel as a mixture of Gaussians and using an on-line approximation to update the model, resulting in a stable, real-time outdoor tracker which reliably deals with lighting changes, repetitive motions from clutter, and long-term scene changes.
Proceedings Article

Experiments with a new boosting algorithm

TL;DR: This paper describes experiments carried out to assess how well AdaBoost with and without pseudo-loss, performs on real learning problems and compared boosting to Breiman's "bagging" method when used to aggregate various classifiers.
Related Papers (5)