scispace - formally typeset
Search or ask a question

Showing papers by "Koji Tsuda published in 2007"


Proceedings ArticleDOI
26 Dec 2007
TL;DR: This work proposes a sequential representation for action classification which retains the global temporal order of motions, and introduces Discriminative Subsequence Mining to find optimal discriminative subsequence patterns.
Abstract: Recent approaches to action classification in videos have used sparse spatio-temporal words encoding local appearance around interesting movements. Most of these approaches use a histogram representation, discarding the temporal order among features. But this ordering information can contain important information about the action itself e.g. consider the sport disciplines of hurdle race and long jump, where the global temporal order of motions (running, jumping) is important to discriminate between the two. In this work we propose to use a sequential representation which retains this temporal order. Further, we introduce Discriminative Subsequence Mining to find optimal discriminative subsequence patterns. In combination with the LPBoost classifier, this amounts to simultaneously learning a classification function and performing feature selection in the space of all possible feature sequences. The resulting classifier linearly combines a small number of interpretable decision functions, each checking for the presence of a single discriminative pattern. The classifier is benchmarked on the KTH action classification data set and outperforms the best known results in the literature.

218 citations


Proceedings Article
01 Jan 2007
TL;DR: This work proposes to combine item set mining and large margin classifiers to select features from the power set of all visual words to derive an interpretable classification rule based on subgraph features that contain more information than the set features.
Abstract: In Web-related applications of image categorization, it is desirable to derive an interpretable classification rule with high accuracy. Using the bag-of-words representation and the linear support vector machine, one can partly fulfill the goal, but the accuracy of linear classifiers is not high and the obtained features are not informative for users. We propose to combine item set mining and large margin classifiers to select features from the power set of all visual words. Our resulting classification rule is easier to browse and simpler to understand, because each feature has richer information. As a next step, each image is represented as a graph where nodes correspond to local image features and edges encode geometric relations between features. Combining graph mining and boosting, we can obtain a classification rule based on subgraph features that contain more information than the set features. We evaluate our algorithm in a web-retrieval ranking task where the goal is to reject outliers from a set of images returned for a keyword query. Furthermore, it is evaluated on the supervised classification tasks with the challenging VOC2005 data set. Our approach yields excellent accuracy in the unsupervised ranking task compared to a recently proposed probabilistic model and competitive results in the supervised classification task.

92 citations


Proceedings ArticleDOI
17 Jun 2007
TL;DR: In this paper, the authors proposed to combine item set mining and large margin classifiers to select features from the power set of all visual words for web-related applications of image categorization.
Abstract: In Web-related applications of image categorization, it is desirable to derive an interpretable classification rule with high accuracy. Using the bag-of-words representation and the linear support vector machine, one can partly fulfill the goal, but the accuracy of linear classifiers is not high and the obtained features are not informative for users. We propose to combine item set mining and large margin classifiers to select features from the power set of all visual words. Our resulting classification rule is easier to browse and simpler to understand, because each feature has richer information. As a next step, each image is represented as a graph where nodes correspond to local image features and edges encode geometric relations between features. Combining graph mining and boosting, we can obtain a classification rule based on subgraph features that contain more information than the set features. We evaluate our algorithm in a web-retrieval ranking task where the goal is to reject outliers from a set of images returned for a keyword query. Furthermore, it is evaluated on the supervised classification tasks with the challenging VOC2005 data set. Our approach yields excellent accuracy in the unsupervised ranking task compared to a recently proposed probabilistic model and competitive results in the supervised classification task.

85 citations


Proceedings ArticleDOI
01 Apr 2007
TL;DR: An efficient algorithm for principal component analysis (PCA) that is applicable when only the inner product with a given vector is needed is proposed, and Krylov subspace learning is shown to work well both in matrix compression and implicit calculation of the innerproduct.
Abstract: We propose an efficient algorithm for principal component analysis (PCA) that is applicable when only the inner product with a given vector is needed We show that Krylov subspace learning works well both in matrix compression and implicit calculation of the inner product by taking full advantage of the arbitrariness of the seed vector We apply our algorithm to a PCA-based change-point detection algorithm, and show that it results in about 50 times improvement in computational time

61 citations


Journal ArticleDOI
01 Sep 2007
TL;DR: This method implements a forward feature selection procedure where, in each iteration, one mutation combination is found by an efficient branch-and-bound search and successfully recovered many mutation associations known in biological literature.
Abstract: Motivation: Human immunodeficiency virus type 1 (HIV-1) evolves in human body, and its exposure to a drug often causes mutations that enhance the resistance against the drug. To design an effective pharmacotherapy for an individual patient, it is important to accurately predict the drug resistance based on genotype data. Notably, the resistance is not just the simple sum of the effects of all mutations. Structural biological studies suggest that the association of mutations is crucial: even if mutations A or B alone do not affect the resistance, a significant change might happen when the two mutations occur together. Linear regression methods cannot take the associations into account, while decision tree methods can reveal only limited associations. Kernel methods and neural networks implicitly use all possible associations for prediction, but cannot select salient associations explicitly. Results: Our method, itemset boosting, performs linear regression in the complete space of power sets of mutations. It implements a forward feature selection procedure where, in each iteration, one mutation combination is found by an efficient branch-and-bound search. This method uses all possible combinations, and salient associations are explicitly shown. In experiments, our method worked particularly well for predicting the resistance of nucleotide reverse transcriptase inhibitors (NRTIs). Furthermore, it successfully recovered many mutation associations known in biological literature. Availability: http://www.kyb.mpg.de/bs/people/hiroto/iboost/ Contact: koji.tsuda@tuebingen.mpg.de Supplementary information: Supplementary data are available at Bioinformatics online.

55 citations


Journal ArticleDOI
TL;DR: A method for consistent inference of network structure is provided, incorporating prior knowledge about sparse connectivity, and is shown to perform qualitatively superior compared to the most well-known method.
Abstract: Background: Identifying large gene regulatory networks is an important task, while the acquisition of data through perturbation experiments (e.g., gene switches, RNAi, heterozygotes) is expensive. It is thus desirable to use an identification method that effectively incorporates available prior knowledge – such as sparse connectivity – and that allows to design experiments such that maximal information is gained from each one. Results: Our main contributions are twofold: a method for consistent inference of network structure is provided, incorporating prior knowledge about sparse connectivity. The algorithm is time efficient and robust to violations of model assumptions. Moreover, we show how to use it for optimal experimental design, reducing the number of required experiments substantially. We employ sparse linear models, and show how to perform full Bayesian inference for these. We not only estimate a single maximum likelihood network, but compute a posterior distribution over networks, using a novel variant of the expectation propagation method. The representation of uncertainty enables us to do effective experimental design in a standard statistical setting: experiments are selected such that the experiments are maximally informative. Conclusion: Few methods have addressed the design issue so far. Compared to the most wellknown one, our method is more transparent, and is shown to perform qualitatively superior. In the former, hard and unrealistic constraints have to be placed on the network structure for mere computational tractability, while such are not required in our method. We demonstrate reconstruction and optimal experimental design capabilities on tasks generated from realistic nonlinear network simulators. The methods described in the paper are available as a Matlab package at http://www.kyb.tuebingen.mpg.de/sparselinearmodel.

52 citations


Proceedings ArticleDOI
Koji Tsuda1
20 Jun 2007
TL;DR: This work proposes an efficient method to select a small number of salient patterns by regularization path tracking, which is considerably more efficient than a simpler approach based on frequent substructure mining.
Abstract: Graph data such as chemical compounds and XML documents are getting more common in many application domains. A main difficulty of graph data processing lies in the intrinsic high dimensionality of graphs, namely, when a graph is represented as a binary feature vector of indicators of all possible subgraph patterns, the dimensionality gets too large for usual statistical methods. We propose an efficient method to select a small number of salient patterns by regularization path tracking. The generation of useless patterns is minimized by progressive extension of the search space. In experiments, it is shown that our technique is considerably more efficient than a simpler approach based on frequent substructure mining.

44 citations


Journal ArticleDOI
20 Nov 2007
TL;DR: This work proposes a novel method for automatic module extraction from protein-protein interaction networks by introducing an additional constraint for modules which accounts for differential expression.
Abstract: Motivation We propose a novel method for automatic module extraction from protein-protein interaction networks. While most previous approaches for module discovery are based on graph partitioning [1], our algorithm can efficiently enumerate all densely connected modules in the network. As currently available interaction data are incomplete, this is a meaningful generalization of clique search techniques [2]. In comparison with partitioning methods, the approach has the following advantages: the user can specify a minimum density for the outcoming modules and has the guarantee that all modules that satisfy this criterion are discovered. Moreover, it provides a natural way to detect overlapping modules. Many proteins are not steadily present in the cell, but are specifically expressed in dependence of cell type, environmental conditions, and developmental state. Therefore we introduce an additional constraint for modules which accounts for differential expression.

9 citations


Journal ArticleDOI
21 Dec 2007
TL;DR: This supplementary issue consists of seven peer-reviewed papers based on the NIPS Workshop on New Problems and Methods in Computational Biology held at Whistler, British Columbia, Canada on December 8, 2006, and contains seven papersbased on a subset of the 13 extended abstracts.
Abstract: December 8, 2006, Whistler, British Columbia, Canada The field of computational biology has seen dramatic growth over the past few years, both in terms of available data, scientific questions and challenges for learning and inference. These new types of scientific and clinical problems require the development of novel supervised and unsupervised learning approaches. In particular, the field is characterized by a diversity of heterogeneous data. The human genome sequence is accompanied by real-valued gene expression data, functional annotation of genes, genotyping information, a graph of interacting proteins, a set of equations describing the dynamics of a system, localization of proteins in a cell, a phylogenetic tree relating species, natural language text in the form of papers describing experiments, partial models that provide priors, and numerous other data sources. This supplementary issue consists of seven peer-reviewed papers based on the NIPS Workshop on New Problems and Methods in Computational Biology held at Whistler, British Columbia, Canada on December 8, 2006. The Neural Information Processing Systems Conference is the premier scientific meeting on neural computation, with session topics spanning artificial intelligence, learning theory, neuroscience, etc. The goal of this workshop was to present emerging problems and machine learning techniques in computational biology, with a particular emphasis on methods for computational learning from heterogeneous data. We received 37 extended abstract submissions, from which 13 were selected for oral presentation. The current supplement contains seven papers based on a subset of the 13 extended abstracts. Submitted manuscripts were rigorously reviewed by at least two referees. The quality of each paper was evaluated with respect to its contribution to biology as well as the novelty of the machine learning methods employed.

5 citations