Institution
International Institute of Information Technology, Hyderabad
Education•Hyderabad, India•
About: International Institute of Information Technology, Hyderabad is a education organization based out in Hyderabad, India. It is known for research contribution in the topics: Computer science & Authentication. The organization has 2048 authors who have published 3677 publications receiving 45319 citations. The organization is also known as: IIIT Hyderabad & International Institute of Information Technology (IIIT).
Topics: Computer science, Authentication, Deep learning, Artificial neural network, Internet security
Papers published on a yearly basis
Papers
More filters
•
01 Feb 2009TL;DR: It is demonstrated that the performance of the proposed method can be far superior to that of commercial OCR systems, and can benefit from synthetically generated training data obviating the need for expensive data collection and annotation.
Abstract: This paper tackles the problem of recognizing characters in images of natural scenes. In particular, we focus on recognizing characters in situations that would traditionally not be handled well by OCR techniques. We present an annotated database of images containing English and Kannada characters. The database comprises of images of street scenes taken in Bangalore, India using a standard camera. The problem is addressed in an object cateogorization framework based on a bag-of-visual-words representation. We assess the performance of various features based on nearest neighbour and SVM classification. It is demonstrated that the performance of the proposed method, using as few as 15 training images, can be far superior to that of commercial OCR systems. Furthermore, the method can benefit from synthetically generated training data obviating the need for expensive data collection and annotation.
520 citations
••
14 Jun 2009TL;DR: It is observed that existing MKL formulations can be extended to learn general kernel combinations subject to general regularization while retaining all the efficiency of existing large scale optimization algorithms.
Abstract: Recent advances in Multiple Kernel Learning (MKL) have positioned it as an attractive tool for tackling many supervised learning tasks. The development of efficient gradient descent based optimization schemes has made it possible to tackle large scale problems. Simultaneously, MKL based algorithms have achieved very good results on challenging real world applications. Yet, despite their successes, MKL approaches are limited in that they focus on learning a linear combination of given base kernels.In this paper, we observe that existing MKL formulations can be extended to learn general kernel combinations subject to general regularization. This can be achieved while retaining all the efficiency of existing large scale optimization algorithms. To highlight the advantages of generalized kernel learning, we tackle feature selection problems on benchmark vision and UCI databases. It is demonstrated that the proposed formulation can lead to better results not only as compared to traditional MKL but also as compared to state-of-the-art wrapper and filter methods for feature selection.
461 citations
••
TL;DR: In this article, the authors perform extensive experiments with multiple deep learning architectures to learn semantic word embeddings to handle the complexity of the natural language constructs and achieve state-of-the-art performance on hate speech detection on Twitter.
Abstract: Hate speech detection on Twitter is critical for applications like controversial event extraction, building AI chatterbots, content recommendation, and sentiment analysis. We define this task as being able to classify a tweet as racist, sexist or neither. The complexity of the natural language constructs makes this task very challenging. We perform extensive experiments with multiple deep learning architectures to learn semantic word embeddings to handle this complexity. Our experiments on a benchmark dataset of 16K annotated tweets show that such deep learning methods outperform state-of-the-art char/word n-gram methods by ~18 F1 points.
382 citations
••
01 Apr 2014TL;DR: A comprehensive dataset of retinal images which include both normal and glaucomatous eyes and manual segmentations from multiple human experts is presented and area and boundary-based evaluation measures are presented to evaluate a method on various aspects relevant to the problem ofglaucoma assessment.
Abstract: Optic nerve head (ONH) segmentation problem has been of interest for automated glaucoma assessment. Although various segmentation methods have been proposed in the recent past, it is difficult to evaluate and compare the performance of individual methods due to a lack of a benchmark dataset. The problem of segmentation involves segmentation of optic disk and cup region within ONH region. Available datasets do not incorporate challenges present in this problem. In this data paper, we present a comprehensive dataset of retinal images which include both normal and glaucomatous eyes and manual segmentations from multiple human experts. Both area and boundary-based evaluation measures are presented to evaluate a method on various aspects relevant to the problem of glaucoma assessment.
369 citations
••
16 Jun 2012TL;DR: This work presents a framework that exploits both bottom-up and top-down cues in the problem of recognizing text extracted from street images, and shows significant improvements in accuracies on two challenging public datasets, namely Street View Text and ICDAR 2003.
Abstract: Scene text recognition has gained significant attention from the computer vision community in recent years. Recognizing such text is a challenging problem, even more so than the recognition of scanned documents. In this work, we focus on the problem of recognizing text extracted from street images. We present a framework that exploits both bottom-up and top-down cues. The bottom-up cues are derived from individual character detections from the image. We build a Conditional Random Field model on these detections to jointly model the strength of the detections and the interactions between them. We impose top-down cues obtained from a lexicon-based prior, i.e. language statistics, on the model. The optimal word represented by the text image is obtained by minimizing the energy function corresponding to the random field model. We show significant improvements in accuracies on two challenging public datasets, namely Street View Text (over 15%) and ICDAR 2003 (nearly 10%).
349 citations
Authors
Showing all 2066 results
Name | H-index | Papers | Citations |
---|---|---|---|
Ravi Shankar | 66 | 672 | 19326 |
Joakim Nivre | 61 | 295 | 17203 |
Aravind K. Joshi | 59 | 249 | 16417 |
Ashok Kumar Das | 56 | 278 | 9166 |
Malcolm F. White | 55 | 172 | 10762 |
B. Yegnanarayana | 54 | 340 | 12861 |
Ram Bilas Pachori | 48 | 182 | 8140 |
C. V. Jawahar | 45 | 479 | 9582 |
Saurabh Garg | 40 | 206 | 6738 |
Himanshu Thapliyal | 36 | 201 | 3992 |
Monika Sharma | 36 | 238 | 4412 |
Ponnurangam Kumaraguru | 33 | 269 | 6849 |
Abhijit Mitra | 33 | 240 | 7795 |
Ramanathan Sowdhamini | 33 | 256 | 4458 |
Helmut Schiessel | 32 | 117 | 3527 |