scispace - formally typeset
Search or ask a question
Author

Tingting Mu

Bio: Tingting Mu is an academic researcher from University of Manchester. The author has contributed to research in topics: Dimensionality reduction & Support vector machine. The author has an hindex of 17, co-authored 78 publications receiving 1120 citations. Previous affiliations of Tingting Mu include Xi'an Jiaotong University & University of Liverpool.


Papers
More filters
Journal ArticleDOI
TL;DR: A novel stochastic multiview hashing algorithm is proposed to facilitate the construction of a large-scale near-duplicate video retrieval system and is compared against various classical and state-of-the-art NDVR systems.
Abstract: Near-duplicate video retrieval (NDVR) has been a significant research task in multimedia given its high impact in applications, such as video search, recommendation, and copyright protection. In addition to accurate retrieval performance, the exponential growth of online videos has imposed heavy demands on the efficiency and scalability of the existing systems. Aiming at improving both the retrieval accuracy and speed, we propose a novel stochastic multiview hashing algorithm to facilitate the construction of a large-scale NDVR system. Reliable mapping functions, which convert multiple types of keyframe features, enhanced by auxiliary information such as video-keyframe association and ground truth relevance to binary hash code strings, are learned by maximizing a mixture of the generalized retrieval precision and recall scores. A composite Kullback–Leibler divergence measure is used to approximate the retrieval scores, which aligns stochastically the neighborhood structures between the original feature and the relaxed hash code spaces. The efficiency and effectiveness of the proposed method are examined using two public near-duplicate video collections and are compared against various classical and state-of-the-art NDVR systems.

99 citations

Journal ArticleDOI
TL;DR: This work proposes a joint optimisation method that learns embeddings that are sensitive to sentiment classification, and reports cross-domain sentiment classification accuracies that are statistically comparable to the current state-of-the-art embedding learning methods for cross- domain sentiment classification.
Abstract: Unsupervised Cross-domain Sentiment Classification is the task of adapting a sentiment classifier trained on a particular domain ( source domain ), to a different domain ( target domain ), without requiring any labeled data for the target domain. By adapting an existing sentiment classifier to previously unseen target domains, we can avoid the cost for manual data annotation for the target domain. We model this problem as embedding learning, and construct three objective functions that capture: (a) distributional properties of pivots (i.e., common features that appear in both source and target domains), (b) label constraints in the source domain documents, and (c) geometric properties in the unlabeled documents in both source and target domains. Unlike prior proposals that first learn a lower-dimensional embedding independent of the source domain sentiment labels, and next a sentiment classifier in this embedding, our joint optimisation method learns embeddings that are sensitive to sentiment classification. Experimental results on a benchmark dataset show that by jointly optimising the three objectives we can obtain better performances in comparison to optimising each objective function separately, thereby demonstrating the importance of task-specific embedding learning for cross-domain sentiment classification. Among the individual objective functions, the best performance is obtained by (c). Moreover, the proposed method reports cross-domain sentiment classification accuracies that are statistically comparable to the current state-of-the-art embedding learning methods for cross-domain sentiment classification.

98 citations

Journal ArticleDOI
TL;DR: A vehicle-to-vehicle (V2V) communication protocol is proposed to realize the context awareness, and a constrained A* ( CA*) algorithm to find the solutions is proposed and it is shown that the CA* algorithm can effectively produce optimal recharging detours.
Abstract: Route planning for fully electric vehicles (FEVs) must take energy efficiency into account due to limited battery capacity and time-consuming recharging. In addition, the planning algorithm should allow for negative energy costs in the road network due to regenerative braking, which is a unique feature of FEVs. In this paper, we propose a framework for energy-driven and context-aware route planning for FEVs. It has two novel aspects: 1) It is context aware, i.e., the framework has access to real-time traffic data for routing cost estimation; and it is energy driven, i.e., both time and energy efficiency are accounted for; which implies a biobjective nature of the optimization. In addition, in the case of insufficient energy on board, an optimal detour via recharge points is computed. Our main contributions to address these issues can be highlighted as follows: A vehicle-to-vehicle (V2V) communication protocol is proposed to realize the context awareness, and we replace the original biobjective form of optimality with two single-objective forms and propose a constrained A* ( CA*) algorithm to find the solutions. The algorithm maintains a Pareto front while it confines its search by energy constraints. The best recharging detour can be also found using the algorithm. We first compared the performance of the CA* algorithm with other algorithms. We then evaluate the impact of the context awareness on road traffic by simulations using a realistic road network regarding different forms of optimality. Finally, we show that the CA* algorithm can effectively produce optimal recharging detours.

92 citations

Journal ArticleDOI
TL;DR: This work shows, using relatively simple image processing and feature extraction, that dynamic foot pressures can be used to identify n = 104 subjects with a CR of 99.6 per cent, a finding that pointedly emphasizes inter-subject pressure pattern uniqueness.
Abstract: Everyone's walking style is unique, and it has been shown that both humans and computers are very good at recognizing known gait patterns. It is therefore unsurprising that dynamic foot pressure pa...

88 citations

Journal ArticleDOI
TL;DR: Improvement in classification accuracy may be gained by using selected combinations of shape, edge-sharpness, and texture features in breast masses computed from 111 regions in mammograms.
Abstract: Breast masses due to benign disease and malignant tumors related to breast cancer differ in terms of shape, edge-sharpness, and texture characteristics. In this study, we evaluate a set of 22 features including 5 shape factors, 3 edge-sharpness measures, and 14 texture features computed from 111 regions in mammograms, with 46 regions related to malignant tumors and 65 to benign masses. Feature selection is performed by a genetic algorithm based on several criteria, such as alignment of the kernel with the target function, class separability, and normalized distance. Fisher’s linear discriminant analysis, the support vector machine (SVM), and our strict two-surface proximal (S2SP) classifier, as well as their corresponding kernel-based nonlinear versions, are used in the classification task with the selected features. The nonlinear classification performance of kernel Fisher’s discriminant analysis, SVM, and S2SP, with the Gaussian kernel, reached 0.95 in terms of the area under the receiver operating characteristics curve. The results indicate that improvement in classification accuracy may be gained by using selected combinations of shape, edge-sharpness, and texture features.

81 citations


Cited by
More filters
Journal ArticleDOI
TL;DR: Machine learning addresses many of the same research questions as the fields of statistics, data mining, and psychology, but with differences of emphasis.
Abstract: Machine Learning is the study of methods for programming computers to learn. Computers are applied to a wide range of tasks, and for most of these it is relatively easy for programmers to design and implement the necessary software. However, there are many tasks for which this is difficult or impossible. These can be divided into four general categories. First, there are problems for which there exist no human experts. For example, in modern automated manufacturing facilities, there is a need to predict machine failures before they occur by analyzing sensor readings. Because the machines are new, there are no human experts who can be interviewed by a programmer to provide the knowledge necessary to build a computer system. A machine learning system can study recorded data and subsequent machine failures and learn prediction rules. Second, there are problems where human experts exist, but where they are unable to explain their expertise. This is the case in many perceptual tasks, such as speech recognition, hand-writing recognition, and natural language understanding. Virtually all humans exhibit expert-level abilities on these tasks, but none of them can describe the detailed steps that they follow as they perform them. Fortunately, humans can provide machines with examples of the inputs and correct outputs for these tasks, so machine learning algorithms can learn to map the inputs to the outputs. Third, there are problems where phenomena are changing rapidly. In finance, for example, people would like to predict the future behavior of the stock market, of consumer purchases, or of exchange rates. These behaviors change frequently, so that even if a programmer could construct a good predictive computer program, it would need to be rewritten frequently. A learning program can relieve the programmer of this burden by constantly modifying and tuning a set of learned prediction rules. Fourth, there are applications that need to be customized for each computer user separately. Consider, for example, a program to filter unwanted electronic mail messages. Different users will need different filters. It is unreasonable to expect each user to program his or her own rules, and it is infeasible to provide every user with a software engineer to keep the rules up-to-date. A machine learning system can learn which mail messages the user rejects and maintain the filtering rules automatically. Machine learning addresses many of the same research questions as the fields of statistics, data mining, and psychology, but with differences of emphasis. Statistics focuses on understanding the phenomena that have generated the data, often with the goal of testing different hypotheses about those phenomena. Data mining seeks to find patterns in the data that are understandable by people. Psychological studies of human learning aspire to understand the mechanisms underlying the various learning behaviors exhibited by people (concept learning, skill acquisition, strategy change, etc.).

13,246 citations

Proceedings ArticleDOI
22 Jan 2006
TL;DR: Some of the major results in random graphs and some of the more challenging open problems are reviewed, including those related to the WWW.
Abstract: We will review some of the major results in random graphs and some of the more challenging open problems. We will cover algorithmic and structural questions. We will touch on newer models, including those related to the WWW.

7,116 citations