scispace - formally typeset
S

Shaolei Feng

Researcher at University of Massachusetts Amherst

Publications -  20
Citations -  1314

Shaolei Feng is an academic researcher from University of Massachusetts Amherst. The author has contributed to research in topics: Image retrieval & Hidden Markov model. The author has an hindex of 11, co-authored 20 publications receiving 1290 citations. Previous affiliations of Shaolei Feng include Chinese Academy of Sciences & Princeton University.

Papers
More filters
Proceedings ArticleDOI

Multiple Bernoulli relevance models for image and video annotation

TL;DR: This work shows how it can do both automatic image annotation and retrieval (using one word queries) from images and videos using a multiple Bernoulli relevance model, which significantly outperforms previously reported results on the task of image and video annotation.
Proceedings ArticleDOI

Using Corner Feature Correspondences to Rank Word Images by Similarity

TL;DR: This paper presents an algorithm which compares whole word-images based on their appearance, and recovers correspondences of points of interest in two images, and then uses these correspondences to construct a similarity measure which can be used to rank word- images in order of their closeness to a querying image.
Proceedings ArticleDOI

Statistical models for automatic video annotation and retrieval

TL;DR: A continuous relevance model is applied to the problem of directly retrieving the visual content of videos using text queries and a modified model is proposed - the normalized CRM - which substantially improves performance on a subset of the TREC video dataset.
Proceedings ArticleDOI

A hierarchical, HMM-based automatic evaluation of OCR accuracy for a digital library of books

TL;DR: A hidden Markov model (HMM) based hierarchical alignment algorithm to align OCR output and the ground truth for books is proposed, believed to be the first work to automatically align a whole book without using any book structure information.
Proceedings ArticleDOI

Joint visual-text modeling for automatic retrieval of multimedia documents

TL;DR: A novel framework where individual components are developed to model different relationships between documents and queries and then combined into a joint retrieval framework is proposed, which demonstrates over 14 % improvement in IR performance over the best reported text-only baseline and ranks amongst the best results reported on this corpus.