scispace - formally typeset
Search or ask a question

Showing papers by "Kai Li published in 2012"


Proceedings ArticleDOI
05 Jun 2012
TL;DR: It is shown that entropy-based filtering eliminates ambiguous SIFT features that cause most of the false positives, and enables claiming near-duplicity with a single match of the retained high-quality features, and that graph cut can be used for query expansion with a duplicity graph computed offline to substantially improve search quality.
Abstract: In this paper, we propose two techniques for near-duplicate image detection at high confidence and large scale. First, we show that entropy-based filtering eliminates ambiguous SIFT features that cause most of the false positives, and enables claiming near-duplicity with a single match of the retained high-quality features. Second, we show that graph cut can be used for query expansion with a duplicity graph computed offline to substantially improve search quality. Evaluation with web images show that when combined with sketch embedding [6], our methods achieve false positive rate orders of magnitude lower than the standard visual word approach. We demonstrate the proposed techniques with a large-scale image search engine which, using indexing data structure offline computed with a Hadoop cluster, is capable of serving more than 50 million web images with a single commodity server.

42 citations


Patent
Grant Wallace1, Philip N. Shilane, Mark Huang, Edward K. Lee, Kai Li 
13 Jun 2012
TL;DR: In this paper, a computer-implemented method and system for deduplicating sub-chunks in a data storage system selects a data chunk and generates a sketch for the selected data chunk.
Abstract: A computer-implemented method and system for deduplicating sub-chunks in a data storage system selects a data chunk to deduplicate and generates a sketch for the selected data chunk. A similar data chunk is searched for using the sketch. A set of fingerprints corresponding to sub-chunks of the similar data chunk is loaded. The set of fingerprints for the similar data chunk is compared to a set of fingerprints of the selected data chunk and the selected chunk is encoded as a set of references to identical sub-chunks of the similar data chunk and at least one unmatched sub-chunk.

10 citations


Kai Li1, Zhe Wang1
01 Jan 2012
TL;DR: This dissertation shows that it is possible to substantially improve search accuracy and efficiency by leveraging domain specific knowledge of multimodal data in similarity search system designs.
Abstract: Similarity search systems are designed to help people to organize multimedia non-text data and find valuable information. The multimedia data intrinsically has multiple modalities (e.g., visual and audio features from video clips) which can be exploited to construct better search systems. Traditionally, various integration techniques have been used to aggregate multiple modalities. However, such algorithms do not scale well for large datasets. As the multimedia data grows, it is a challenge to build a search system to handle large-scale multimodal data efficiently and provide users with information they need. The goal of this dissertation is to study how to effectively combine multiple modalities to implement similarity search systems for large datasets. I have carried out my study through three similarity search systems each designed for different application. Each system combines multiple modalities to help users find desired information quickly. With VFerret system, I studied how to combine visual features with audio features for effective personal video search. With Image Spam Detection System, I explored several aggregation methods to integrate multiple image spam filters to detect image spams. With my Product Navigation System, I studied how to combine text search with image similarity search to help user find desired products. This thesis has also studied a rank-based model which helps system designers to construct more efficient large-scale multimodal similarity search systems. Although the general solution to using multimodal data in a similarity search system is still unknown, this dissertation shows that it is possible to substantially improve search accuracy and efficiency by leveraging domain specific knowledge of multimodal data. The VFerret system improves search accuracy from an average precision of 0.66 to 0.79 by combining visual and audio features. The Image Spam Detection System significantly lowers the false positive rate from a previous result of 1% to 0.001% while maintaining comparable detection rates by combining multiple image filters intelligently. My Product Navigation System reduces number of user clicks by 60% compared to traditional systems through a new method of combining text search with image similarity search. These results support further adoption and study of multimodal data in similarity search system designs.

3 citations


ReportDOI
20 Dec 2012
TL;DR: This report summarizes the work of the University of Utah, which was a member of the National Fusion Collaboratory (NFC) Project funded by the United States Department of Energy under the Scientific Discovery through Advanced Computing Program (SciDAC) to develop a persistent infrastructure to enable scientific collaboration for magnetic fusion research.
Abstract: This report summarizes the work of the University of Utah, which was a member of the National Fusion Collaboratory (NFC) Project funded by the United States Department of Energy (DOE) under the Scientific Discovery through Advanced Computing Program (SciDAC) to develop a persistent infrastructure to enable scientific collaboration for magnetic fusion research. A five year project that was initiated in 2001, it the NFC built on the past collaborative work performed within the U.S. fusion community and added the component of computer science research done with the USDOE Office of Science, Office of Advanced Scientific Computer Research. The project was itself a collaboration, itself uniting fusion scientists from General Atomics, MIT, and PPPL and computer scientists from ANL, LBNL, and Princeton University, and the University of Utah to form a coordinated team. The group leveraged existing computer science technology where possible and extended or created new capabilities where required. The complete finial report is attached as an addendum. The In the collaboration, the primary technical responsibility of the University of Utah in the collaboration was to develop and deploy an advanced scientific visualization service. To achieve this goal, the SCIRun Problem Solving Environment (PSE) is used on FusionGrid for an advanced scientific visualization service. SCIRun is open source software that gives the user the ability to create complex 3D visualizations and 2D graphics. This capability allows for the exploration of complex simulation results and the comparison of simulation and experimental data. SCIRun on FusionGrid gives the scientist a no-license-cost visualization capability that rivals present day commercial visualization packages. To accelerate the usage of SCIRun within the fusion community, a stand-alone application built on top of SCIRun was developed and deployed. This application, FusionViewer, allows users who are unfamiliar with SCIRun to quickly create visualizations and perform analysis of their simulation data from either the MDSplus data storage environment or from locally stored HDF5 files. More advanced tools for visualization and analysis also were created in collaboration with the SciDAC Center for Extended MHD Modeling. Versions of SCIRun with the FusionViewer have been made available to fusion scientists on the Mac OS X, Linux, and other Unix based platforms and have been downloaded 1163 times. SCIRun has been used with NIMROD, M3D, BOUT fusion simulation data as well as simulation data from other SciDAC application areas (e.g., Astrophysics). The subsequent visualization results - including animations - have been incorporated into invited talks at multiple APS/DPP meetings as well as peer reviewed journal articles. As an example, SCIRun was used for the visualization and analysis of a NIMROD simulation of a disruption that occurred in a DIII-D experiment. The resulting animations and stills were presented as part of invited talks at APS/DPP meetings and the SC04 conference in addition to being highlighted in the NIH/NSF Visualization Research Challenges Report. By achieving its technical goals, the University of Utah played a key role in the successful development of a persistent infrastructure to enable scientific collaboration for magnetic fusion research. Many of the visualization tools developed as part of the NFC continue to be used by Fusion and other SciDAC application scientists and are currently being supported and expanded through follow-on up on SciDAC projects (Visualization and Analytics Center for Enabling Technology, and the Visualization and Analysis in Support of Fusion SAP).

1 citations