scispace - formally typeset
Search or ask a question
Author

Evan Hildreth

Bio: Evan Hildreth is an academic researcher from Qualcomm. The author has contributed to research in topics: Gesture & Gesture recognition. The author has an hindex of 18, co-authored 32 publications receiving 3450 citations.

Papers
More filters
Patent
24 Jul 2001
TL;DR: In this paper, a method of using stereo vision to interface with a computer is described, which includes capturing a stereo image, processing the stereo image to determine position information of an object in the stereo images, and communicating the position information to the computer.
Abstract: A method of using stereo vision to interface with a computer is provided. The method includes capturing a stereo image, and processing the stereo image to determine position information of an object in the stereo image. The object is controlled by a user. The method also includes communicating the position information to the computer to allow the user to interact with a computer application.

838 citations

Patent
03 Oct 2001
TL;DR: In this paper, a multiple camera tracking system for interfacing with an application program running on a computer is presented, which includes two or more video cameras arranged to provide different viewpoints of a region of interest, and are operable to produce a series of video images.
Abstract: A multiple camera tracking system for interfacing with an application program running on a computer is provided. The tracking system includes two or more video cameras arranged to provide different viewpoints of a region of interest, and are operable to produce a series of video images. A processor is operable to receive the series of video images and detect objects appearing in the region of interest. The processor executes a process to generate a background data set from the video images, generate an image data set for each received video image, compare each image data set to the background data set to produce a difference map for each image data set, detect a relative position of an object of interest within each difference map, and produce an absolute position of the object of interest from the relative positions of the object of interest and map the absolute position to a position indicator associated with the application program.

578 citations

Patent
Evan Hildreth1
06 Jan 2006
TL;DR: In this paper, a method is described for determining a description of motion of a moving mobile camera to determine a user input to an application, based on the determined motion, the application may respond to the user input, for example, by updating a user interface of the application.
Abstract: A method is described for determining a description of motion of a moving mobile camera to determine a user input to an application. The method may involve capturing a series of images from a moving mobile camera and comparing stationary features present in the series of images. Optical flow analysis may be performed on the series of images to determine a description of motion of the moving mobile camera. Based on the determined motion, a user input to an application may be determined and the application may respond to the user input, for example, by updating a user interface of the application.

431 citations

Patent
Evan Hildreth1
23 Sep 2008
TL;DR: In this article, an enhanced interface for voice and video communications is presented, in which a gesture of a user is recognized from a sequence of camera images, and a user interface is provided include a control and a representation of the user.
Abstract: An enhanced interface for voice and video communications, in which a gesture of a user is recognized from a sequence of camera images, and a user interface is provided include a control and a representation of the user. The process also includes causing the representation to interact with the control based on the recognized gesture, and controlling a telecommunication session based on the interaction.

331 citations

Patent
14 Sep 2009
TL;DR: In this paper, an element is initially displayed on an interactive touch-screen display device with an initial orientation relative to the interactive touch screen display device and the user is determined to be interacting with the element displayed on the display device.
Abstract: An element is initially displayed on an interactive touch-screen display device with an initial orientation relative to the interactive touch-screen display device. One or more images of a user of the interactive touch-screen display device are captured. The user is determined to be interacting with the element displayed on the interactive touch-screen display device. In addition, an orientation of the user relative to the interactive touch-screen display device is determined based on at least one captured image of the user of the interactive touch-screen display device. Thereafter, in response to determining that the user is interacting with the displayed element, the initial orientation of the displayed element relative to the interactive touch-screen display device is automatically adjusted based on the determined orientation of the user relative to the interactive touch-screen display device.

283 citations


Cited by
More filters
Patent
12 Nov 2013
TL;DR: In this paper, a variety of technologies by which existing functionality can be improved, and new functionality can also be provided, including visual search capabilities, and determining appropriate actions responsive to different image inputs.
Abstract: Cell phones and other portable devices are equipped with a variety of technologies by which existing functionality can be improved, and new functionality can be provided. Some relate to visual search capabilities, and determining appropriate actions responsive to different image inputs. Others relate to processing of image data. Still others concern metadata generation, processing, and representation. Yet others relate to coping with fixed focus limitations of cell phone cameras, e.g., in reading digital watermark data. Still others concern user interface improvements. A great number of other features and arrangements are also detailed.

2,033 citations

Patent
01 Dec 2003
TL;DR: In this article, a perceptual user interface system includes a tracking component that detects object characteristics of at least one of a plurality of objects within a scene, and tracks the respective object.
Abstract: Architecture for implementing a perceptual user interface. The architecture comprises alternative modalities for controlling computer application programs and manipulating on-screen objects through hand gestures or a combination of hand gestures and verbal commands. The perceptual user interface system includes a tracking component that detects object characteristics of at least one of a plurality of objects within a scene, and tracks the respective object. Detection of object characteristics is based at least in part upon image comparison of a plurality of images relative to a course mapping of the images. A seeding component iteratively seeds the tracking component with object hypotheses based upon the presence of the object characteristics and the image comparison. A filtering component selectively removes the tracked object from the object hypotheses and/or at least one object hypothesis from the set of object hypotheses based upon predetermined removal criteria.

876 citations

Patent
24 Jul 2001
TL;DR: In this paper, a method of using stereo vision to interface with a computer is described, which includes capturing a stereo image, processing the stereo image to determine position information of an object in the stereo images, and communicating the position information to the computer.
Abstract: A method of using stereo vision to interface with a computer is provided. The method includes capturing a stereo image, and processing the stereo image to determine position information of an object in the stereo image. The object is controlled by a user. The method also includes communicating the position information to the computer to allow the user to interact with a computer application.

838 citations

Patent
18 Feb 2003
TL;DR: In this article, three-dimensional position information is used to identify the gesture created by a body part of interest, based on the shape of the body part and its position and orientation.
Abstract: Three-dimensional position information is used to identify the gesture created by a body part of interest. At one or more instances of an interval, the posture of a body part is recognized, based on the shape of the body part and its position and orientation. The posture of the body part over each of the one or more instances in the interval are recognized as a combined gesture. The gesture is classified for determining an input into a related electronic device.

773 citations

Patent
04 Jun 2002
TL;DR: An interactive video display system as mentioned in this paper is a system where a camera is used to detect an object in an interactive area located in front of the display screen, the camera operable to capture three-dimensional information about the object.
Abstract: An interactive video display system. A display screen is for displaying a visual image for presentation to a user. A camera is for detecting an object in an interactive area located in front of the display screen, the camera operable to capture three-dimensional information about the object. A computer system is for directing the display screen to change the visual image in response to the object.

760 citations