Author
Teofilo de Campos
Other affiliations: University of Sheffield, University of São Paulo, University of Surrey ...read more
Bio: Teofilo de Campos is an academic researcher from University of Brasília. The author has contributed to research in topics: Computer science & Pose. The author has an hindex of 18, co-authored 62 publications receiving 1830 citations. Previous affiliations of Teofilo de Campos include University of Sheffield & University of São Paulo.
Papers published on a yearly basis
Papers
More filters
•
01 Feb 2009TL;DR: It is demonstrated that the performance of the proposed method can be far superior to that of commercial OCR systems, and can benefit from synthetically generated training data obviating the need for expensive data collection and annotation.
Abstract: This paper tackles the problem of recognizing characters in images of natural scenes. In particular, we focus on recognizing characters in situations that would traditionally not be handled well by OCR techniques. We present an annotated database of images containing English and Kannada characters. The database comprises of images of street scenes taken in Bangalore, India using a standard camera. The problem is addressed in an object cateogorization framework based on a bag-of-visual-words representation. We assess the performance of various features based on nearest neighbour and SVM classification. It is demonstrated that the performance of the proposed method, using as few as 15 training images, can be far superior to that of commercial OCR systems. Furthermore, the method can benefit from synthetically generated training data obviating the need for expensive data collection and annotation.
520 citations
••
Utrecht University1, Case Western Reserve University2, National University of Colombia3, Technical University of Denmark4, Dalle Molle Institute for Artificial Intelligence Research5, Işık University6, French Institute of Health and Medical Research7, PSL Research University8, Curie Institute9, National Taiwan University of Science and Technology10, Konica Minolta11, Panasonic12, University of Central Lancashire13, University of Nice Sophia Antipolis14, University of Surrey15, University of Sheffield16, University of Warwick17, Qatar University18, Princess Alexandra Hospital NHS Trust19
TL;DR: The results from the Assessment of Mitosis Detection Algorithms 2013 (AMIDA13) challenge are described and the top performing method has an error rate that is comparable to the inter-observer agreement among pathologists.
405 citations
••
TL;DR: This work proposes two complementary approaches to account for the spatial layout in an image-independent manner (as is the case of the SP) while the second one adapts to the image content which does not incur an increase of the image signature dimensionality.
126 citations
••
11 Apr 2000TL;DR: A real time system for detection and tracking of facial features in video sequences, using a statistical skin-color model to segment face-candidate regions in the image, based on an efficient template matching scheme.
Abstract: This work presents a real time system for detection and tracking of facial features in video sequences. Such system may be used in visual communication applications, such as teleconferencing, virtual reality, intelligent interfaces, human-machine interaction, surveillance, etc. We have used a statistical skin-color model to segment face-candidate regions in the image. The presence or absence of a face in each region is verified by means of an eye detector, based on an efficient template matching scheme . Once a face is detected, the pupils, nostrils and lip corners are located and these facial features are tracked in the image sequence, performing real time processing.
106 citations
•
25 Jan 2010TL;DR: In this paper, an apparatus, method, and computer program product are provided for generating an image representation, which includes receiving an input digital image, extracting features from the image which are representative of patches of the image, generating (S108) weighting factors for the features based on location relevance data for the image.
Abstract: An apparatus, method, and computer program product are provided for generating an image representation. The method includes receiving (S102) an input digital image, extracting (S110) features from the image which are representative of patches of the image, generating (S108) weighting factors for the features based on location relevance data for the image, and weighting (S114) the extracted features with the weighting factors to form a representation of the image.
91 citations
Cited by
More filters
••
TL;DR: The ImageNet Large Scale Visual Recognition Challenge (ILSVRC) as mentioned in this paper is a benchmark in object category classification and detection on hundreds of object categories and millions of images, which has been run annually from 2010 to present, attracting participation from more than fifty institutions.
Abstract: The ImageNet Large Scale Visual Recognition Challenge is a benchmark in object category classification and detection on hundreds of object categories and millions of images. The challenge has been run annually from 2010 to present, attracting participation from more than fifty institutions. This paper describes the creation of this benchmark dataset and the advances in object recognition that have been possible as a result. We discuss the challenges of collecting large-scale ground truth annotation, highlight key breakthroughs in categorical object recognition, provide a detailed analysis of the current state of the field of large-scale image classification and object detection, and compare the state-of-the-art computer vision accuracy with human accuracy. We conclude with lessons learned in the 5 years of the challenge, and propose future directions and improvements.
30,811 citations
•
3,940 citations
••
14 May 2014TL;DR: It is shown that the data augmentation techniques commonly applied to CNN-based methods can also be applied to shallow methods, and result in an analogous performance boost, and it is identified that the dimensionality of the CNN output layer can be reduced significantly without having an adverse effect on performance.
Abstract: The latest generation of Convolutional Neural Networks (CNN) have achieved impressive results in challenging benchmarks on image recognition and object detection, significantly raising the interest of the community in these methods. Nevertheless, it is still unclear how different CNN methods compare with each other and with previous state-of-the-art shallow representations such as the Bag-of-Visual-Words and the Improved Fisher Vector. This paper conducts a rigorous evaluation of these new techniques, exploring different deep architectures and comparing them on a common ground, identifying and disclosing important implementation details. We identify several useful properties of CNN-based representations, including the fact that the dimensionality of the CNN output layer can be reduced significantly without having an adverse effect on performance. We also identify aspects of deep and shallow methods that can be successfully shared. In particular, we show that the data augmentation techniques commonly applied to CNN-based methods can also be applied to shallow methods, and result in an analogous performance boost. Source code and models to reproduce the experiments in the paper is made publicly available.
3,533 citations
•
[...]
TL;DR: The main focus in MUCKE is on cleaning large scale Web image corpora and on proposing image representations which are closer to the human interpretation of images.
Abstract: MUCKE aims to mine a large volume of images, to structure them conceptually and to use this conceptual structuring in order to improve large-scale image retrieval. The last decade witnessed important progress concerning low-level image representations. However, there are a number problems which need to be solved in order to unleash the full potential of image mining in applications. The central problem with low-level representations is the mismatch between them and the human interpretation of image content. This problem can be instantiated, for instance, by the incapability of existing descriptors to capture spatial relationships between the concepts represented or by their incapability to convey an explanation of why two images are similar in a content-based image retrieval framework. We start by assessing existing local descriptors for image classification and by proposing to use co-occurrence matrices to better capture spatial relationships in images. The main focus in MUCKE is on cleaning large scale Web image corpora and on proposing image representations which are closer to the human interpretation of images. Consequently, we introduce methods which tackle these two problems and compare results to state of the art methods. Note: some aspects of this deliverable are withheld at this time as they are pending review. Please contact the authors for a preview.
2,134 citations