Open Access
ImageSeer: Clustering and Searching WWW Images Using Link and Page Layout Analysis
Xiaofei He,Doug Cai,Ji-Rong Wen,Wei-Ying Ma,Hong-Jiang Zhang +4 more
- pp 12
Reads0
Chats0
TLDR
iFind is described, a system for clustering and searching WWW images, which is less sensitive to noisy links than previous methods like PageRank, HITS, and PicASHOW, and hence the image graph can better reflect the semantic relationship between images.Abstract:
Due to the rapid growth of the number of digital images on the Web, there is an increasing demand for effective and efficient method for organizing and retrieving the images available. This paper describes ImageSeer, a system for clustering and searching WWW images. By using a vision-based page segmentation algorithm, a web page is partitioned into blocks, and the textual and link information of an image can be accurately extracted within the block containing that image. The textual information is used for image representation. By extracting the page-to-block, blockto-image, block-to-page relationships through link structure and page layout analysis, we construct an image graph. Our method is less sensitive to noisy links than previous methods like PicASHOW, and hence the image graph can better reflect the semantic relationship between images. With the graph models, we use techniques from spectral graph theory and Markov Chain theory for image ranking, clustering and embedding. Some experimental results are given in the paper.read more
Citations
More filters
Proceedings ArticleDOI
Hierarchical clustering of WWW image search results using visual, textual and link information
TL;DR: Wang et al. as mentioned in this paper proposed a hierarchical clustering method using visual, textual and link analysis to organize the results into different semantic clusters to facilitate users' browsing, which can be applied to image search results.
Journal ArticleDOI
Patch Alignment for Dimensionality Reduction
TL;DR: A new dimensionality reduction algorithm is developed, termed discrim inative locality alignment (DLA), by imposing discriminative information in the part optimization stage, and thorough empirical studies demonstrate the effectiveness of DLA compared with representative dimensionality Reduction algorithms.
Proceedings Article
Document summarization based on data reconstruction
TL;DR: This paper proposes a novel framework named Document Summarization based on Data Reconstruction (DSDR), which generates a summary which consist of those sentences that can best reconstruct the original document.
Proceedings ArticleDOI
Contextual in-image advertising
TL;DR: An innovative contextual advertising system driven by images is proposed, which automatically associates relevant ads with an image rather than the entire text in a Web page and seamlessly inserts the ads in the nonintrusive areas within each individual image.
Journal ArticleDOI
EMR: A Scalable Graph-Based Ranking Model for Content-Based Image Retrieval
TL;DR: This paper proposes a novel scalable graph-based ranking model called Efficient Manifold Ranking (EMR), trying to address the shortcomings of MR from two main perspectives: scalable graph construction and efficient ranking computation.
References
More filters
Book
The Nature of Statistical Learning Theory
TL;DR: Setting of the learning problem consistency of learning processes bounds on the rate of convergence ofLearning processes controlling the generalization ability of learning process constructing learning algorithms what is important in learning theory?
Book
Neural networks for pattern recognition
TL;DR: This is the first comprehensive treatment of feed-forward neural networks from the perspective of statistical pattern recognition, and is designed as a text, with over 100 exercises, to benefit anyone involved in the fields of neural computation and pattern recognition.
Journal ArticleDOI
The anatomy of a large-scale hypertextual Web search engine
Sergey Brin,Lawrence Page +1 more
TL;DR: This paper provides an in-depth description of Google, a prototype of a large-scale search engine which makes heavy use of the structure present in hypertext and looks at the problem of how to effectively deal with uncontrolled hypertext collections where anyone can publish anything they want.
Journal ArticleDOI
Normalized cuts and image segmentation
Jianbo Shi,Jitendra Malik +1 more
TL;DR: This work treats image segmentation as a graph partitioning problem and proposes a novel global criterion, the normalized cut, for segmenting the graph, which measures both the total dissimilarity between the different groups as well as the total similarity within the groups.
Journal Article
The Anatomy of a Large-Scale Hypertextual Web Search Engine.
Sergey Brin,Lawrence Page +1 more
TL;DR: Google as discussed by the authors is a prototype of a large-scale search engine which makes heavy use of the structure present in hypertext and is designed to crawl and index the Web efficiently and produce much more satisfying search results than existing systems.