scispace - formally typeset
Open AccessJournal ArticleDOI

SUGAMAN: describing floor plans for visually impaired by annotation learning and proximity-based grammar

Reads0
Chats0
TLDR
In this paper, the authors propose a framework called Sugaman (Supervised and Unified framework using Grammar and Annotation Model for Access and Navigation) for describing a floor plan and giving direction for obstacle-free movement within a building.
Abstract
In this study, the authors propose a framework SUGAMAN (Supervised and Unified framework using Grammar and Annotation Model for Access and Navigation). SUGAMAN is a Hindi word meaning ‘easy passage from one place to another’. SUGAMAN synthesises textual description from a given floor plan image, usable by visually impaired to navigate by understanding the arrangement of rooms and furniture. It is the first framework for describing a floor plan and giving direction for obstacle-free movement within a building. The model learns five classes of room categories from 1355 room image samples under a supervised learning paradigm. These learned annotations are fed into a description synthesis framework to yield a holistic description of a floor plan image. Authors demonstrate the performance of various supervised classifiers on room learning and provided a comparative analysis of system generated and human-written descriptions. The contribution of this study includes a novel framework for description generation from document images with graphics while proposing a new feature representing the floor plans, text annotations for a publicly available data set, and an algorithm for door to door obstacle avoidance navigation. This work can be applied to areas like understanding floor plans and design of historical monuments, and retrieval.

read more

Citations
More filters
Journal ArticleDOI

Efficient Multi-Object Detection and Smart Navigation Using Artificial Intelligence for Visually Impaired People

TL;DR: An artificial intelligence-based fully automatic assistive technology to recognize different objects, and auditory inputs are provided to the user in real time, which gives better understanding to the visually impaired person about their surroundings.
Proceedings ArticleDOI

Travelling more independently: A Requirements Analysis for Accessible Journeys to Unknown Buildings for People with Visual Impairments

TL;DR: A survey with 106 people with visual impairments is presented, in which the strategies they use to prepare for a journey to unknown buildings, how they orient themselves in unfamiliar buildings and what materials they use are examined.
Journal ArticleDOI

Traveling More Independently: A Study on the Diverse Needs and Challenges of People with Visual or Mobility Impairments in Unfamiliar Indoor Environments

TL;DR: In this article , the authors present a survey of 125 participants with blindness, low vision, and mobility impairments, and investigate how mobile they are, what strategies they use to prepare a journey to an unknown building, how they orient themselves there, and what materials they use.
Journal ArticleDOI

Knowledge-driven description synthesis for floor plan interpretation

TL;DR: In this paper, the authors proposed two models, description synthesis from image cue (DSIC) and transformer-based description generation (TBDG), for text generation from floor plan images.
Book ChapterDOI

Semantic Segmentation and Topological Mapping of Floor Plans

TL;DR: In this article, a topological mapping method from the floor plan model based on deep learning semantic segmentation is proposed for assistive blind navigation purposes in unknown indoor environments, where disturbances such as image rotation, color transformation and Gaussian noises are taken into consideration in the training to enhance the robustness.
References
More filters
Journal ArticleDOI

Distinctive Image Features from Scale-Invariant Keypoints

TL;DR: This paper presents a method for extracting distinctive invariant features from images that can be used to perform reliable matching between different views of an object or scene and can robustly identify objects among clutter and occlusion while achieving near real-time performance.
Proceedings ArticleDOI

Rapid object detection using a boosted cascade of simple features

TL;DR: A machine learning approach for visual object detection which is capable of processing images extremely rapidly and achieving high detection rates and the introduction of a new image representation called the "integral image" which allows the features used by the detector to be computed very quickly.
Journal ArticleDOI

Multiresolution gray-scale and rotation invariant texture classification with local binary patterns

TL;DR: A generalized gray-scale and rotation invariant operator presentation that allows for detecting the "uniform" patterns for any quantization of the angular space and for any spatial resolution and presents a method for combining multiple operators for multiresolution analysis.
Book ChapterDOI

SURF: speeded up robust features

TL;DR: A novel scale- and rotation-invariant interest point detector and descriptor, coined SURF (Speeded Up Robust Features), which approximates or even outperforms previously proposed schemes with respect to repeatability, distinctiveness, and robustness, yet can be computed and compared much faster.
Book ChapterDOI

Text Categorization with Suport Vector Machines: Learning with Many Relevant Features

TL;DR: This paper explores the use of Support Vector Machines for learning text classifiers from examples and analyzes the particular properties of learning with text data and identifies why SVMs are appropriate for this task.
Related Papers (5)