scispace - formally typeset
Search or ask a question

Showing papers on "Object (computer science) published in 2010"


Book ChapterDOI
05 Sep 2010
TL;DR: This paper introduces a method that adapts object models acquired in a particular visual domain to new imaging conditions by learning a transformation that minimizes the effect of domain-induced changes in the feature distribution.
Abstract: Domain adaptation is an important emerging topic in computer vision. In this paper, we present one of the first studies of domain shift in the context of object recognition. We introduce a method that adapts object models acquired in a particular visual domain to new imaging conditions by learning a transformation that minimizes the effect of domain-induced changes in the feature distribution. The transformation is learned in a supervised manner and can be applied to categories for which there are no labeled examples in the new domain. While we focus our evaluation on object recognition tasks, the transform-based adaptation technique we develop is general and could be applied to nonimage data. Another contribution is a new multi-domain object database, freely available for download. We experimentally demonstrate the ability of our method to improve recognition on categories with few or no target domain labels and moderate to large changes in the imaging conditions.

2,624 citations


Proceedings Article
06 Dec 2010
TL;DR: This work focuses on the practically-attractive case when the training images are annotated with dots, and introduces a new loss function, which is well-suited for visual object counting tasks and at the same time can be computed efficiently via a maximum subarray algorithm.
Abstract: We propose a new supervised learning framework for visual object counting tasks, such as estimating the number of cells in a microscopic image or the number of humans in surveillance video frames. We focus on the practically-attractive case when the training images are annotated with dots (one dot per object). Our goal is to accurately estimate the count. However, we evade the hard task of learning to detect and localize individual object instances. Instead, we cast the problem as that of estimating an image density whose integral over any image region gives the count of objects within that region. Learning to infer such density can be formulated as a minimization of a regularized risk quadratic cost function. We introduce a new loss function, which is well-suited for such learning, and at the same time can be computed efficiently via a maximum subarray algorithm. The learning can then be posed as a convex quadratic program solvable with cutting-plane optimization. The proposed framework is very flexible as it can accept any domain-specific visual features. Once trained, our system provides accurate object counts and requires a very small time overhead over the feature extraction step, making it a good candidate for applications involving real-time processing or dealing with huge amount of visual data.

1,098 citations


Book ChapterDOI
05 Sep 2010
TL;DR: A novel method for unsupervised class segmentation on a set of images that alternates between segmenting object instances and learning a class model based on a segmentation energy defined over all images at the same time, which can be optimized efficiently by techniques used before in interactive segmentation.
Abstract: We propose a novel method for unsupervised class segmentation on a set of images. It alternates between segmenting object instances and learning a class model. The method is based on a segmentation energy defined over all images at the same time, which can be optimized efficiently by techniques used before in interactive segmentation. Over iterations, our method progressively learns a class model by integrating observations over all images. In addition to appearance, this model captures the location and shape of the class with respect to an automatically determined coordinate frame common across images. This frame allows us to build stronger shape and location models, similar to those used in object class detection. Our method is inspired by interactive segmentation methods [1], but it is fully automatic and learns models characteristic for the object class rather than specific to one particular object/image. We experimentally demonstrate on the Caltech4, Caltech101, and Weizmann horses datasets that our method (a) transfers class knowledge across images and this improves results compared to segmenting every image independently; (b) outperforms Grabcut [1] for the task of unsupervised segmentation; (c) offers competitive performance compared to the state-of-the-art in unsupervised segmentation and in particular it outperforms the topic model [2].

1,028 citations


Proceedings ArticleDOI
06 Dec 2010
TL;DR: A high-level image representation, called the Object Bank, is proposed, where an image is represented as a scale-invariant response map of a large number of pre-trained generic object detectors, blind to the testing dataset or visual task.
Abstract: Robust low-level image features have been proven to be effective representations for a variety of visual recognition tasks such as object recognition and scene classification; but pixels, or even local image patches, carry little semantic meanings. For high level visual tasks, such low-level image representations are potentially not enough. In this paper, we propose a high-level image representation, called the Object Bank, where an image is represented as a scale-invariant response map of a large number of pre-trained generic object detectors, blind to the testing dataset or visual task. Leveraging on the Object Bank representation, superior performances on high level visual recognition tasks can be achieved with simple off-the-shelf classifiers such as logistic regression and linear SVM. Sparsity algorithms make our representation more efficient and scalable for large scene datasets, and reveal semantically meaningful feature patterns.

1,027 citations


Journal ArticleDOI
TL;DR: The dissertation presented in this article proposes Semantic 3D Object Models as a novel representation of the robot’s operating environment that satisfies these requirements and shows how these models can be automatically acquired from dense 3D range data.
Abstract: Environment models serve as important resources for an autonomous robot by providing it with the necessary task-relevant information about its habitat. Their use enables robots to perform their tasks more reliably, flexibly, and efficiently. As autonomous robotic platforms get more sophisticated manipulation capabilities, they also need more expressive and comprehensive environment models: for manipulation purposes their models have to include the objects present in the world, together with their position, form, and other aspects, as well as an interpretation of these objects with respect to the robot tasks. The dissertation presented in this article (Rusu, PhD thesis, 2009) proposes Semantic 3D Object Models as a novel representation of the robot’s operating environment that satisfies these requirements and shows how these models can be automatically acquired from dense 3D range data.

908 citations


Proceedings ArticleDOI
13 Jun 2010
TL;DR: A new random field model is proposed to encode the mutual context of objects and human poses in human-object interaction activities and it is shown that this mutual context model significantly outperforms state-of-the-art in detecting very difficult objects andhuman poses.
Abstract: Detecting objects in cluttered scenes and estimating articulated human body parts are two challenging problems in computer vision. The difficulty is particularly pronounced in activities involving human-object interactions (e.g. playing tennis), where the relevant object tends to be small or only partially visible, and the human body parts are often self-occluded. We observe, however, that objects and human poses can serve as mutual context to each other – recognizing one facilitates the recognition of the other. In this paper we propose a new random field model to encode the mutual context of objects and human poses in human-object interaction activities. We then cast the model learning task as a structure learning problem, of which the structural connectivity between the object, the overall human pose, and different body parts are estimated through a structure search approach, and the parameters of the model are estimated by a new max-margin algorithm. On a sports data set of six classes of human-object interactions [12], we show that our mutual context model significantly outperforms state-of-the-art in detecting very difficult objects and human poses.

673 citations


Journal ArticleDOI
TL;DR: Whether one-trial object recognition involves working memory and how it involves memory of an episode is discussed, and whether the validity of the novelty preference concept is questioned.

556 citations


Book ChapterDOI
05 Sep 2010
TL;DR: The results demonstrate that incorporating user input drives up recognition accuracy to levels that are good enough for practical applications, while at the same time, computer vision reduces the amount of human interaction required.
Abstract: We present an interactive, hybrid human-computer method for object classification. The method applies to classes of objects that are recognizable by people with appropriate expertise (e.g., animal species or airplane model), but not (in general) by people without such expertise. It can be seen as a visual version of the 20 questions game, where questions based on simple visual attributes are posed interactively. The goal is to identify the true class while minimizing the number of questions asked, using the visual content of the image. We introduce a general framework for incorporating almost any off-the-shelf multi-class object recognition algorithm into the visual 20 questions game, and provide methodologies to account for imperfect user responses and unreliable computer vision algorithms. We evaluate our methods on Birds-200, a difficult dataset of 200 tightly-related bird species, and on the Animals With Attributes dataset. Our results demonstrate that incorporating user input drives up recognition accuracy to levels that are good enough for practical applications, while at the same time, computer vision reduces the amount of human interaction required.

492 citations


Book ChapterDOI
05 Sep 2010
TL;DR: A new descriptor for images is introduced which allows the construction of efficient and compact classifiers with good accuracy on object category recognition, and allows object-category queries to be made against image databases using efficient classifiers such as linear support vector machines.
Abstract: We introduce a new descriptor for images which allows the construction of efficient and compact classifiers with good accuracy on object category recognition. The descriptor is the output of a large number of weakly trained object category classifiers on the image. The trained categories are selected from an ontology of visual concepts, but the intention is not to encode an explicit decomposition of the scene. Rather, we accept that existing object category classifiers often encode not the category per se but ancillary image characteristics; and that these ancillary characteristics can combine to represent visual classes unrelated to the constituent categories' semantic meanings. The advantage of this descriptor is that it allows object-category queries to be made against image databases using efficient classifiers (efficient at test time) such as linear support vector machines, and allows these queries to be for novel categories. Even when the representation is reduced to 200 bytes per image, classification accuracy on object category recognition is comparable with the state of the art (36% versus 42%), but at orders of magnitude lower computational cost.

479 citations


Patent
12 Mar 2010
TL;DR: In this paper, a system and method for automatically providing content associated with captured information is described, in which the system receives input by a user, and automatically provides content or links to the information associated with the input.
Abstract: A system and method for automatically providing content associated with captured information is described. In some examples, the system receives input by a user, and automatically provides content or links to content associated with the input. In some examples, the system receives input via text entry or by capturing text from a rendered document, such as a printed document, an object, an audio stream, and so on.

438 citations


Patent
Imran Chaudhri1
22 Sep 2010
TL;DR: In this paper, a multifunction device displays a plurality of selectable user interface objects on the display, and the device moves a first input object in the plurality of selected user interfaces across the display to a location on display that is proximate to a second input object.
Abstract: A multifunction device displays a plurality of selectable user interface objects on the display. In response to detecting the first input, the device moves a first object in the plurality of selectable user interface objects across the display to a location on the display that is proximate to a second object in the plurality of selectable user interface objects. In response to detecting that the first input meets predefined folder-creation criteria while the first object is proximate to the second object, the device creates a folder that contains the first object and the second object.

Journal ArticleDOI
TL;DR: This work addresses the problem of incorporating different types of contextual information for robust object categorization in computer vision by considering the most common levels of extraction of context and the different levels of contextual interactions.

Proceedings ArticleDOI
13 Jun 2010
TL;DR: This paper introduces a new dataset with images that contain many instances of different object categories and proposes an efficient model that captures the contextual information among more than a hundred ofobject categories and shows that the context model can be applied to scene understanding tasks that local detectors alone cannot solve.
Abstract: There has been a growing interest in exploiting contextual information in addition to local features to detect and localize multiple object categories in an image. Context models can efficiently rule out some unlikely combinations or locations of objects and guide detectors to produce a semantically coherent interpretation of a scene. However, the performance benefit from using context models has been limited because most of these methods were tested on datasets with only a few object categories, in which most images contain only one or two object categories. In this paper, we introduce a new dataset with images that contain many instances of different object categories and propose an efficient model that captures the contextual information among more than a hundred of object categories. We show that our context model can be applied to scene understanding tasks that local detectors alone cannot solve.

Journal ArticleDOI
TL;DR: Object categories with conceptually distinctive exemplars showed less interference in memory as the number of exemplars increased, and observers' capacity to remember visual information in long-term memory depends more on conceptual structure than perceptual distinctiveness.
Abstract: Humans have a massive capacity to store detailed information in visual long-term memory. The present studies explored the fidelity of these visual long-term memory representations and examined how conceptual and perceptual features of object categories support this capacity. Observers viewed 2,800 object images with a different number of exemplars presented from each category. At test, observers indicated which of 2 exemplars they had previously studied. Memory performance was high and remained quite high (82% accuracy) with 16 exemplars from a category in memory, demonstrating a large memory capacity for object exemplars. However, memory performance decreased as more exemplars were held in memory, implying systematic categorical interference. Object categories with conceptually distinctive exemplars showed less interference in memory as the number of exemplars increased. Interference in memory was not predicted by the perceptual distinctiveness of exemplars from an object category, though these perceptual measures predicted visual search rates for an object target among exemplars. These data provide evidence that observers’ capacity to remember visual information in long-term memory depends more on conceptual structure than perceptual distinctiveness.

Journal ArticleDOI
TL;DR: An object class detection approach which fully integrates the complementary strengths offered by shape matchers and can localize object boundaries accurately and does not need segmented examples for training (only bounding-boxes).
Abstract: We present an object class detection approach which fully integrates the complementary strengths offered by shape matchers. Like an object detector, it can learn class models directly from images, and can localize novel instances in the presence of intra-class variations, clutter, and scale changes. Like a shape matcher, it finds the boundaries of objects, rather than just their bounding-boxes. This is achieved by a novel technique for learning a shape model of an object class given images of example instances. Furthermore, we also integrate Hough-style voting with a non-rigid point matching algorithm to localize the model in cluttered images. As demonstrated by an extensive evaluation, our method can localize object boundaries accurately and does not need segmented examples for training (only bounding-boxes).

Proceedings Article
Fangtao Li1, Chao Han1, Minlie Huang1, Xiaoyan Zhu1, Yingju Xia2, Shu Zhang2, Hao Yu2 
23 Aug 2010
TL;DR: This paper proposes a new machine learning framework based on Conditional Random Fields that can employ rich features to jointly extract positive opinions, negative opinions and object features for review sentences and shows that structure-aware models outperform many state-of-the-art approaches to review mining.
Abstract: In this paper, we focus on object feature based review summarization. Different from most of previous work with linguistic rules or statistical methods, we formulate the review mining task as a joint structure tagging problem. We propose a new machine learning framework based on Conditional Random Fields (CRFs). It can employ rich features to jointly extract positive opinions, negative opinions and object features for review sentences. The linguistic structure can be naturally integrated into model representation. Besides linear-chain structure, we also investigate conjunction structure and syntactic tree structure in this framework. Through extensive experiments on movie review and product review data sets, we show that structure-aware models outperform many state-of-the-art approaches to review mining.

Patent
05 Jan 2010
TL;DR: In this article, a first user input identifying a 2D object presented in a user interface can be detected, and a second user input including a 3D gesture input that includes a movement in proximity to a surface, and the 3D object can be presented in the user interface.
Abstract: Three-dimensional objects can be generated based on two-dimensional objects. A first user input identifying a 2D object presented in a user interface can be detected, and a second user input including a 3D gesture input that includes a movement in proximity to a surface can be detected. A 3D object can be generated based on the 2D object according to the first and second user inputs, and the 3D object can be presented in the user interface.

Patent
12 Jan 2010
TL;DR: In this paper, a method of operating a dimensioning system to determine dimensional information for objects is disclosed, where a number of images are acquired and objects in at least one of the acquired images are computationally identified.
Abstract: A method of operating a dimensioning system to determine dimensional information for objects is disclosed. A number of images are acquired. Objects in at least one of the acquired images are computationally identified. One object represented in the at least one of the acquired images is computationally initially selected as a candidate for processing. An indication of the initially selected object is provided to a user. At least one user input indicative of an object selected for processing is received. Dimensional data for the object indicated by the received user input is computationally determined.

Patent
05 Nov 2010
TL;DR: In this article, the authors describe a direct user-interaction method for 3D augmented reality, which comprises displaying a 3D AR environment having a virtual object and a real first and second objects controlled by a user, tracking the position of the objects in 3D using camera images, displaying the virtual object on the first object from the user's viewpoint, and enabling interaction between the second object and the virtual objects when the first and the second objects are touching.
Abstract: Techniques for user-interaction in augmented reality are described. In one example, a direct user-interaction method comprises displaying a 3D augmented reality environment having a virtual object and a real first and second object controlled by a user, tracking the position of the objects in 3D using camera images, displaying the virtual object on the first object from the user's viewpoint, and enabling interaction between the second object and the virtual object when the first and second objects are touching. In another example, an augmented reality system comprises a display device that shows an augmented reality environment having a virtual object and a real user's hand, a depth camera that captures depth images of the hand, and a processor. The processor receives the images, tracks the hand pose in six degrees-of-freedom, and enables interaction between the hand and the virtual object.

Book ChapterDOI
05 Sep 2010
TL;DR: It is shown that a geometric representation of an object occurring in indoor scenes, along with rich scene structure can be used to produce a detector for that object in a single image, and this detector has significantly improved accuracy when compared to the state-of-the-art 2D detectors.
Abstract: In this paper we show that a geometric representation of an object occurring in indoor scenes, along with rich scene structure can be used to produce a detector for that object in a single image. Using perspective cues from the global scene geometry, we first develop a 3D based object detector. This detector is competitive with an image based detector built using state-of-the-art methods; however, combining the two produces a notably improved detector, because it unifies contextual and geometric information. We then use a probabilistic model that explicitly uses constraints imposed by spatial layout - the locations of walls and floor in the image - to refine the 3D object estimates. We use an existing approach to compute spatial layout [1], and use constraints such as objects are supported by floor and can not stick through the walls. The resulting detector (a) has significantly improved accuracy when compared to the state-of-the-art 2D detectors and (b) gives a 3D interpretation of the location of the object, derived from a 2D image. We evaluate the detector on beds, for which we give extensive quantitative results derived from images of real scenes.

Patent
15 Nov 2010
TL;DR: In this paper, the authors proposed a method for providing an augmented reality operations tool to a mobile client positioned in a building, which includes, with a server ( 660 ), receiving ( 720 ) from the client ( 642 ) an augmented augmented reality request for building system equipment ( 612 ) managed by an energy management system (EMS) ( 620 ).
Abstract: A method ( 700 ) for providing an augmented reality operations tool to a mobile client ( 642 ) positioned in a building ( 604 ). The method ( 700 ) includes, with a server ( 660 ), receiving ( 720 ) from the client ( 642 ) an augmented reality request for building system equipment ( 612 ) managed by an energy management system (EMS) ( 620 ). The method ( 700 ) includes transmitting ( 740 ) a data request for the equipment ( 612 ) to the EMS ( 620 ) and receiving ( 750 ) building management data ( 634 ) for the equipment ( 612 ). The method ( 700 ) includes generating ( 760 ) an overlay ( 656 ) with an object created based on the building management data ( 634 ), which may be sensor data, diagnostic procedures, or the like. The overlay ( 656 ) is configured for concurrent display on a display screen ( 652 ) of the client ( 642 ) with a real-time image of the building equipment ( 612 ). The method ( 700 ) includes transmitting ( 770 ) the overlay ( 656 ) to the client ( 642 ).

Proceedings ArticleDOI
13 Jun 2010
TL;DR: This paper presents a new approach for multi-view object class detection which uses a part model which discriminatively learns the object appearance with spatial pyramids from a database of real images, and encodes the 3D geometry of the object class with a generative representation built from adatabase of synthetic models.
Abstract: This paper presents a new approach for multi-view object class detection. Appearance and geometry are treated as separate learning tasks with different training data. Our approach uses a part model which discriminatively learns the object appearance with spatial pyramids from a database of real images, and encodes the 3D geometry of the object class with a generative representation built from a database of synthetic models. The geometric information is linked to the 2D training data and allows to perform an approximate 3D pose estimation for generic object classes. The pose estimation provides an efficient method to evaluate the likelihood of groups of 2D part detections with respect to a full 3D geometry model in order to disambiguate and prune 2D detections and to handle occlusions. In contrast to other methods, neither tedious manual part annotation of training images nor explicit appearance matching between synthetic and real training data is required, which results in high geometric fidelity and in increased flexibility. On the 3D Object Category datasets CAR and BICYCLE [15], the current state-of-the-art benchmark for 3D object detection, our approach outperforms previously published results for viewpoint estimation.

Patent
25 Jun 2010
TL;DR: In this paper, the authors proposed a method for predicting properties of a target object by combining multivariate statistical analysis and principal component analysis in combination with content-based image retrieval for providing two-dimensional attributes of three dimensional objects, for example, via preferential image segmentation using a tree of shapes.
Abstract: Method and apparatus for predicting properties of a target object comprise application of a search manager for analyzing parameters of a plurality of databases for a plurality of objects, the databases comprising an electrical, electromagnetic, acoustic spectral database (ESD), a micro-body assemblage database (MAD) and a database of image data whereby the databases store data objects containing identifying features, source information and information on site properties and context including time and frequency varying data. The method comprises application of multivariate statistical analysis and principal component analysis in combination with content-based image retrieval for providing two-dimensional attributes of three dimensional objects, for example, via preferential image segmentation using a tree of shapes and to predict further properties of objects by means of k-means clustering and related methods. By way of example, one of a criminal activity and a fraudulent activity event, an intrusion event and a fire event and residual objects may be predicted and located and qualified such that, for example, properties of the residual objects may be qualified, for example, via black body radiation and micro-body databases including charcoal assemblages.

Journal ArticleDOI
TL;DR: The goal of this paper is to discover the objects present in the images by analyzing unlabeled data and searching for re-occurring patterns, and a rigorous framework for evaluating unsupervised object discovery methods is proposed.
Abstract: The goal of this paper is to evaluate and compare models and methods for learning to recognize basic entities in images in an unsupervised setting. In other words, we want to discover the objects present in the images by analyzing unlabeled data and searching for re-occurring patterns. We experiment with various baseline methods, methods based on latent variable models, as well as spectral clustering methods. The results are presented and compared both on subsets of Caltech256 and MSRC2, data sets that are larger and more challenging and that include more object classes than what has previously been reported in the literature. A rigorous framework for evaluating unsupervised object discovery methods is proposed.

Patent
16 Sep 2010
TL;DR: In this article, a recognition object condition judging portion determines whether image areas of the recognition object are recognizable, and a recognition result integration portion integrates recognition results of the sub-areas including recognized characters and/or recognized patterns and corresponding reliabilities.
Abstract: A recognition object detecting portion detects a recognition object existing in each of images which are captured in series at different time points by a camera. Images of the detected recognition object are stored in a memory. A recognition object condition judging portion determines whether image areas of the recognition object are recognizable. A movement amount detecting portion detects a movement amount of the recognition object by using the image areas of the recognition object stored in the memory. A sub-area detecting portion detects a sub-area to be recognized, by using the image areas of the recognition object read out from the memory, in accordance with the movement amount of the recognition object. A recognition portion performs recognition processing on the sub-areas. A recognition result integration portion integrates recognition results of the sub-areas including recognized characters and/or recognized patterns and the corresponding reliabilities, respectively.

Book ChapterDOI
10 Sep 2010
TL;DR: This work learns 20 visual attributes and uses them in a zero-shot transfer learning experiment as well as to make visual connections between semantically unrelated object categories.
Abstract: We consider the task of learning visual connections between object categories using the ImageNet dataset, which is a large-scale dataset ontology containing more than 15 thousand object classes. We want to discover visual relationships between the classes that are currently missing (such as similar colors or shapes or textures). In this work we learn 20 visual attributes and use them in a zero-shot transfer learning experiment as well as to make visual connections between semantically unrelated object categories.

Patent
20 Sep 2010
TL;DR: In this paper, an interface for enabling a user to quickly access contact information automatically displays a list of expected contacts that are most likely to be selected by the user when attention is directed to an appropriate object requiring contact information.
Abstract: An interface for enabling a user to quickly access contact information automatically displays a list of expected contacts that are most likely to be selected by the user when attention is directed to an appropriate object requiring contact information. When a contact is selected, the corresponding and appropriate contact information is automatically entered. If a user does not select a listed contact, but instead begins manually typing in the contact information then the interface performs a search and displays a list of unexpected contacts having contact information matching the typed input from the user. Various criteria can be used to identify which contacts will be presented to the user and how they will be presented.

Journal ArticleDOI
TL;DR: A novel 3D shape descriptor that uses a set of panoramic views of a 3D object which describe the position and orientation of the object’s surface in 3D space to increase the retrieval performance by employing a local (unsupervised) relevance feedback technique that shifts the descriptor of an object closer to its cluster centroid in feature space.
Abstract: We present a novel 3D shape descriptor that uses a set of panoramic views of a 3D object which describe the position and orientation of the object's surface in 3D space. We obtain a panoramic view of a 3D object by projecting it to the lateral surface of a cylinder parallel to one of its three principal axes and centered at the centroid of the object. The object is projected to three perpendicular cylinders, each one aligned with one of its principal axes in order to capture the global shape of the object. For each projection we compute the corresponding 2D Discrete Fourier Transform as well as 2D Discrete Wavelet Transform. We further increase the retrieval performance by employing a local (unsupervised) relevance feedback technique that shifts the descriptor of an object closer to its cluster centroid in feature space. The effectiveness of the proposed 3D object retrieval methodology is demonstrated via an extensive consistent evaluation in standard benchmarks that clearly shows better performance against state-of-the-art 3D object retrieval methods.

Patent
Shuqing Zeng1
19 Jan 2010
TL;DR: In this paper, a method for controlling a vehicle operating during a dynamic vehicle event includes monitoring a first input image, monitoring a second input image and determining a dissimilarity measure comparing the first tracked object to the second tracked object.
Abstract: A method for controlling a vehicle operating during a dynamic vehicle event includes monitoring a first input image, monitoring a first tracked object within the first input image in a first tracking cycle, monitoring a second input image, monitoring a second tracked object within the second input image in a second tracking cycle, and determining a dissimilarity measure comparing the first tracked object to the second tracked object. The dissimilarity measure estimates whether the first tracked object and the second tracked object represent a single tracked object proximate to the vehicle. The method further includes associating the first tracked object and the second tracked object based upon the dissimilarity measure, and utilizing the associated objects in a collision preparation system to control operation of the vehicle.

Patent
Alexander Shpunt1, Gerard Medioni1, Daniel Cohen1, Erez Sali1, Ronen Deitch1 
28 Jul 2010
TL;DR: In this article, a first image of the pattern on the object is captured using first image sensor, and this image is processed to generate pattern-based depth data with respect to the object.
Abstract: A method for depth mapping includes projecting a pattern of optical radiation onto an object. A first image of the pattern on the object is captured using a first image sensor, and this image is processed to generate pattern-based depth data with respect to the object. A second image of the object is captured using a second image sensor, and the second image is processed together with another image to generate stereoscopic depth data with respect to the object. The pattern-based depth data is combined with the stereoscopic depth data to create a depth map of the object.