scispace - formally typeset
Search or ask a question
Author

Deepak Khemani

Bio: Deepak Khemani is an academic researcher from Indian Institute of Technology Madras. The author has contributed to research in topics: Case-based reasoning & Intelligent decision support system. The author has an hindex of 9, co-authored 59 publications receiving 375 citations. Previous affiliations of Deepak Khemani include Indian Institute of Technology Mandi & Indian Institutes of Technology.


Papers
More filters
Journal ArticleDOI
TL;DR: The proposed method is applicable for clustering tasks where clusters are crisp and spherical and determines the number of clusters and the cluster centers in such a way that locally optimal solutions are avoided.

122 citations

Journal ArticleDOI
TL;DR: The survey in this paper presents the current state-of-the-art methods of Simultaneous Localization and Mapping for unmanned ground robots, including important landmark works reported in the literature.
Abstract: This paper presents a survey of Simultaneous Localization And Mapping (SLAM) algorithms for unmanned ground robots. SLAM is the process of creating a map of the environment, sometimes unknown a priori, while at the same time localizing the robot in the same map. The map could be one of different types i.e. metrical, topological, hybrid or semantic. In this paper, the classification of algorithms is done in three classes: (i) Metric map generating approaches, (ii) Qualitative map generating approaches, and (iii) Hybrid map generating approaches. SLAM algorithms for both static and dynamic environments have been surveyed. The algorithms in each class are further divided based on the techniques used. The survey in this paper presents the current state-of-the-art methods, including important landmark works reported in the literature.

20 citations

Journal ArticleDOI
TL;DR: Two clustering algorithms toward the outlined goal of building interpretable and reconfigurable cluster models are developed that outperform the unsupervised decision tree approach for rule-generating clustering and also an approach for generating interpretable models for general clusterings, both by significant margins.
Abstract: Clusters of text documents output by clustering algorithms are often hard to interpret. We describe motivating real-world scenarios that necessitate reconfigurability and high interpretability of clusters and outline the problem of generating clusterings with interpretable and reconfigurable cluster models. We develop two clustering algorithms toward the outlined goal of building interpretable and reconfigurable cluster models. They generate clusters with associated rules that are composed of conditions on word occurrences or nonoccurrences. The proposed approaches vary in the complexity of the format of the rules; RGC employs disjunctions and conjunctions in rule generation whereas RGC-D rules are simple disjunctions of conditions signifying presence of various words. In both the cases, each cluster is comprised of precisely the set of documents that satisfy the corresponding rule. Rules of the latter kind are easy to interpret, whereas the former leads to more accurate clustering. We show that our approaches outperform the unsupervised decision tree approach for rule-generating clustering and also an approach we provide for generating interpretable models for general clusterings, both by significant margins. We empirically show that the purity and f-measure losses to achieve interpretability can be as little as 3 and 5%, respectively using the algorithms presented herein.

19 citations

Journal ArticleDOI
TL;DR: In this paper, a learning algorithm is proposed that minimizes an error term, which reflects the fuzzy classification from the point of view of possibilistic approach, which can encompass different backpropagation learning algorithm based on crisp and constrained fuzzy classification.

18 citations

Book ChapterDOI
01 Sep 2008
TL;DR: An approach to visualize textual case bases by "stacking" similar cases and features close to each other in an image derived from the case-feature matrix and a complexity measure called GAME that exploits regularities in stacked images to evaluate the alignment between problem and solution components of cases are presented.
Abstract: This paper deals with two relatively less well studied problems in Textual CBR, namely visualizing and evaluating complexity of textual case bases. The first is useful in case base maintenance, the second in making informed choices regarding case base representation and tuning of parameters for the TCBR system, and also for explaining the behaviour of different retrieval/classification techniques over diverse case bases. We present an approach to visualize textual case bases by "stacking" similar cases and features close to each other in an image derived from the case-feature matrix. We propose a complexity measure called GAME that exploits regularities in stacked images to evaluate the alignment between problem and solution components of cases. GAME class , a counterpart of GAME in classification domains, shows a strong correspondence with accuracies reported by standard classifiers over classification tasks of varying complexity.

16 citations


Cited by
More filters
Journal ArticleDOI
TL;DR: Machine learning addresses many of the same research questions as the fields of statistics, data mining, and psychology, but with differences of emphasis.
Abstract: Machine Learning is the study of methods for programming computers to learn. Computers are applied to a wide range of tasks, and for most of these it is relatively easy for programmers to design and implement the necessary software. However, there are many tasks for which this is difficult or impossible. These can be divided into four general categories. First, there are problems for which there exist no human experts. For example, in modern automated manufacturing facilities, there is a need to predict machine failures before they occur by analyzing sensor readings. Because the machines are new, there are no human experts who can be interviewed by a programmer to provide the knowledge necessary to build a computer system. A machine learning system can study recorded data and subsequent machine failures and learn prediction rules. Second, there are problems where human experts exist, but where they are unable to explain their expertise. This is the case in many perceptual tasks, such as speech recognition, hand-writing recognition, and natural language understanding. Virtually all humans exhibit expert-level abilities on these tasks, but none of them can describe the detailed steps that they follow as they perform them. Fortunately, humans can provide machines with examples of the inputs and correct outputs for these tasks, so machine learning algorithms can learn to map the inputs to the outputs. Third, there are problems where phenomena are changing rapidly. In finance, for example, people would like to predict the future behavior of the stock market, of consumer purchases, or of exchange rates. These behaviors change frequently, so that even if a programmer could construct a good predictive computer program, it would need to be rewritten frequently. A learning program can relieve the programmer of this burden by constantly modifying and tuning a set of learned prediction rules. Fourth, there are applications that need to be customized for each computer user separately. Consider, for example, a program to filter unwanted electronic mail messages. Different users will need different filters. It is unreasonable to expect each user to program his or her own rules, and it is infeasible to provide every user with a software engineer to keep the rules up-to-date. A machine learning system can learn which mail messages the user rejects and maintain the filtering rules automatically. Machine learning addresses many of the same research questions as the fields of statistics, data mining, and psychology, but with differences of emphasis. Statistics focuses on understanding the phenomena that have generated the data, often with the goal of testing different hypotheses about those phenomena. Data mining seeks to find patterns in the data that are understandable by people. Psychological studies of human learning aspire to understand the mechanisms underlying the various learning behaviors exhibited by people (concept learning, skill acquisition, strategy change, etc.).

13,246 citations

Journal ArticleDOI
01 Jan 2011
TL;DR: Simulation results indicate that ABC algorithm can efficiently be used for multivariate data clustering and is compared with Particle Swarm Optimization (PSO) algorithm and other nine classification techniques from the literature.
Abstract: Artificial Bee Colony (ABC) algorithm which is one of the most recently introduced optimization algorithms, simulates the intelligent foraging behavior of a honey bee swarm. Clustering analysis, used in many disciplines and applications, is an important tool and a descriptive task seeking to identify homogeneous groups of objects based on the values of their attributes. In this work, ABC is used for data clustering on benchmark problems and the performance of ABC algorithm is compared with Particle Swarm Optimization (PSO) algorithm and other nine classification techniques from the literature. Thirteen of typical test data sets from the UCI Machine Learning Repository are used to demonstrate the results of the techniques. The simulation results indicate that ABC algorithm can efficiently be used for multivariate data clustering.

1,007 citations

Journal ArticleDOI
01 Jan 2008
TL;DR: Differential evolution has emerged as one of the fast, robust, and efficient global search heuristics of current interest as mentioned in this paper, which has been applied to the automatic clustering of large unlabeled data sets.
Abstract: Differential evolution (DE) has emerged as one of the fast, robust, and efficient global search heuristics of current interest. This paper describes an application of DE to the automatic clustering of large unlabeled data sets. In contrast to most of the existing clustering techniques, the proposed algorithm requires no prior knowledge of the data to be classified. Rather, it determines the optimal number of partitions of the data "on the run." Superiority of the new method is demonstrated by comparing it with two recently developed partitional clustering techniques and one popular hierarchical clustering algorithm. The partitional clustering algorithms are based on two powerful well-known optimization algorithms, namely the genetic algorithm and the particle swarm optimization. An interesting real-world application of the proposed method to automatic segmentation of images is also reported.

700 citations

Journal ArticleDOI
TL;DR: An up-to-date review of all major nature inspired metaheuristic algorithms employed till date for partitional clustering and key issues involved during formulation of various metaheuristics as a clustering problem and major application areas are discussed.
Abstract: The partitional clustering concept started with K-means algorithm which was published in 1957. Since then many classical partitional clustering algorithms have been reported based on gradient descent approach. The 1990 kick started a new era in cluster analysis with the application of nature inspired metaheuristics. After initial formulation nearly two decades have passed and researchers have developed numerous new algorithms in this field. This paper embodies an up-to-date review of all major nature inspired metaheuristic algorithms employed till date for partitional clustering. Further, key issues involved during formulation of various metaheuristics as a clustering problem and major application areas are discussed.

457 citations

Journal ArticleDOI
TL;DR: This work proposes an alternative to the merge-based approach to clustering by removing the clusters iteratively one by one until the desired number of clusters is reached, and applies local optimization strategy by always removing the cluster that increases the distortion the least.

310 citations