scispace - formally typeset
Search or ask a question
Author

Azhar Khan

Bio: Azhar Khan is an academic researcher from Google. The author has contributed to research in topics: Medicine & Nuclear chemistry. The author has an hindex of 13, co-authored 17 publications receiving 2837 citations.

Papers
More filters
Patent
09 May 2006
TL;DR: In this paper, a collection of captured images that form at least a portion of a library of images is used to enable retrieval of the captured images, and an index is generated where the index data is based on recognized information.
Abstract: An embodiment provides for enabling retrieval of a collection of captured images that form at least a portion of a library of images. For each image in the collection, a captured image may be analyzed to recognize information from image data contained in the captured image, and an index may be generated, where the index data is based on the recognized information. Using the index, functionality such as search and retrieval is enabled. Various recognition techniques, including those that use the face, clothing, apparel, and combinations of characteristics may be utilized. Recognition may be performed on, among other things, persons and text carried on objects.

463 citations

Patent
07 Oct 2005
TL;DR: In this article, a collection of captured images that form at least a portion of a library of images is used to enable retrieval of the captured images, and an index is generated where the index data is based on recognized information.
Abstract: An embodiment provides for enabling retrieval of a collection of captured images that form at least a portion of a library of images. For each image in the collection, a captured image may be analyzed to recognize information from image data contained in the captured image, and an index may be generated, where the index data is based on the recognized information. Using the index, functionality such as search and retrieval is enabled. Various recognition techniques, including those that use the face, clothing, apparel, and combinations of characteristics may be utilized. Recognition may be performed on, among other things, persons and text carried on objects.

369 citations

Patent
07 Oct 2005
TL;DR: In this article, a collection of captured images that form at least a portion of a library of images is used to enable retrieval of the captured images, and an index is generated where the index data is based on recognized information.
Abstract: An embodiment provides for enabling retrieval of a collection of captured images that form at least a portion of a library of images. For each image in the collection, a captured image may be analyzed to recognize information from image data contained in the captured image, and an index may be generated, where the index data is based on the recognized information. Using the index, functionality such as search and retrieval is enabled. Various recognition techniques, including those that use the face, clothing, apparel, and combinations of characteristics may be utilized. Recognition may be performed on, among other things, persons and text carried on objects.

301 citations

Patent
07 Oct 2005
TL;DR: In this article, a collection of captured images that form at least a portion of a library of images is used to enable retrieval of the captured images, and an index is generated where the index data is based on recognized information.
Abstract: An embodiment provides for enabling retrieval of a collection of captured images that form at least a portion of a library of images. For each image in the collection, a captured image may be analyzed to recognize information from image data contained in the captured image, and an index may be generated, where the index data is based on the recognized information. Using the index, functionality such as search and retrieval is enabled. Various recognition techniques, including those that use the face, clothing, apparel, and combinations of characteristics may be utilized. Recognition may be performed on, among other things, persons and text carried on objects.

281 citations

Patent
07 Nov 2007
TL;DR: In this article, the authors programmatically analyze each of a plurality of images in order to determine one or more visual characteristics about an item shown in each of the images, and then a search operation is performed to identify items that have a visual characteristic that satisfies at least some of the search criteria.
Abstract: Embodiments programmatically analyze each of a plurality of images in order to determine one or more visual characteristics about an item shown in each of the plurality of images. Data is stored corresponding to the one or more visual characteristics. An interface in is provided for which a user is able to specify one or more search criteria. In response to receiving the one or more search criteria, a search operation is performed to identify one or more items that have a visual characteristic that satisfies at least some of the one or more search criteria.

249 citations


Cited by
More filters
Patent
12 Nov 2013
TL;DR: In this paper, a variety of technologies by which existing functionality can be improved, and new functionality can also be provided, including visual search capabilities, and determining appropriate actions responsive to different image inputs.
Abstract: Cell phones and other portable devices are equipped with a variety of technologies by which existing functionality can be improved, and new functionality can be provided. Some relate to visual search capabilities, and determining appropriate actions responsive to different image inputs. Others relate to processing of image data. Still others concern metadata generation, processing, and representation. Yet others relate to coping with fixed focus limitations of cell phone cameras, e.g., in reading digital watermark data. Still others concern user interface improvements. A great number of other features and arrangements are also detailed.

2,033 citations

Patent
Jong Hwan Kim1
13 Mar 2015
TL;DR: In this article, a mobile terminal including a body; a touchscreen provided to a front and extending to side of the body and configured to display content; and a controller configured to detect one side of a body when it comes into contact with a side of an external terminal, display a first area on the touchscreen corresponding to a contact area of body and the external terminal and a second area including the content.
Abstract: A mobile terminal including a body; a touchscreen provided to a front and extending to side of the body and configured to display content; and a controller configured to detect one side of the body comes into contact with one side of an external terminal, display a first area on the touchscreen corresponding to a contact area of the body and the external terminal and a second area including the content, receive an input of moving the content displayed in the second area to the first area, display the content in the first area, and share the content in the first area with the external terminal.

1,441 citations

Patent
23 Feb 2011
TL;DR: A smart phone senses audio, imagery, and/or other stimulus from a user's environment, and acts autonomously to fulfill inferred or anticipated user desires as discussed by the authors, and can apply more or less resources to an image processing task depending on how successfully the task is proceeding or based on the user's apparent interest in the task.
Abstract: A smart phone senses audio, imagery, and/or other stimulus from a user's environment, and acts autonomously to fulfill inferred or anticipated user desires. In one aspect, the detailed technology concerns phone-based cognition of a scene viewed by the phone's camera. The image processing tasks applied to the scene can be selected from among various alternatives by reference to resource costs, resource constraints, other stimulus information (e.g., audio), task substitutability, etc. The phone can apply more or less resources to an image processing task depending on how successfully the task is proceeding, or based on the user's apparent interest in the task. In some arrangements, data may be referred to the cloud for analysis, or for gleaning. Cognition, and identification of appropriate device response(s), can be aided by collateral information, such as context. A great number of other features and arrangements are also detailed.

1,056 citations

Proceedings ArticleDOI
04 Nov 2009
TL;DR: The results show that the ST-matching algorithm significantly outperform incremental algorithm in terms of matching accuracy for low-sampling trajectories and when compared with AFD-based global algorithm, ST-Matching also improves accuracy as well as running time.
Abstract: Map-matching is the process of aligning a sequence of observed user positions with the road network on a digital map. It is a fundamental pre-processing step for many applications, such as moving object management, traffic flow analysis, and driving directions. In practice there exists huge amount of low-sampling-rate (e.g., one point every 2--5 minutes) GPS trajectories. Unfortunately, most current map-matching approaches only deal with high-sampling-rate (typically one point every 10--30s) GPS data, and become less effective for low-sampling-rate points as the uncertainty in data increases. In this paper, we propose a novel global map-matching algorithm called ST-Matching for low-sampling-rate GPS trajectories. ST-Matching considers (1) the spatial geometric and topological structures of the road network and (2) the temporal/speed constraints of the trajectories. Based on spatio-temporal analysis, a candidate graph is constructed from which the best matching path sequence is identified. We compare ST-Matching with the incremental algorithm and Average-Frechet-Distance (AFD) based global map-matching algorithm. The experiments are performed both on synthetic and real dataset. The results show that our ST-matching algorithm significantly outperform incremental algorithm in terms of matching accuracy for low-sampling trajectories. Meanwhile, when compared with AFD-based global algorithm, ST-Matching also improves accuracy as well as running time.

817 citations

Patent
01 May 2006
TL;DR: In this paper, a method and apparatus for displaying the evolution of an electronic document (e.g., word processing document, portable document format file, spreadsheet, drawing, and the like), containing tracked changes, is disclosed.
Abstract: A method and apparatus for displaying the evolution of an electronic document (e.g. word-processing document, portable-document-format file, spreadsheet, drawing, and the like), containing tracked changes, is disclosed. In accordance with the disclosed method and apparatus, the revisions of an electronic document may be treated as parent frames. In between the parent frames, child frames may be constructed from the combined images of their parent frames. To form a child frame, an image of a parent frame has a degree of translucency as it is combined with the translucent image of the other parent frame, such that every child frame contains traces of the images of both its parent frames. An input device (e.g. keyboard, mouse, touchpad, stylus, voice activation, and the like) may be used to control the frame visible to the user at any one time. The user may use the input device to traverse the frames at various speeds and in either direction, creating the visual illusion that a certain revision of the document is being morphed into its future revisions—or stripped off its changes as it is morphed into previous revisions—depending on the direction chosen by the user.

711 citations