scispace - formally typeset
Search or ask a question

Showing papers in "Expert Systems With Applications in 2005"


Journal ArticleDOI
TL;DR: This paper surveys expert systems (ES) development using a literature review and classification of articles from 1995 to 2004 with a keyword index and article abstract in order to explore how ES methodologies and applications have developed during this period.
Abstract: This paper surveys expert systems (ES) development using a literature review and classification of articles from 1995 to 2004 with a keyword index and article abstract in order to explore how ES methodologies and applications have developed during this period. Based on the scope of 166 articles from 78 academic journals (retrieved from five online database) of ES applications, this paper surveys and classifies ES methodologies using the following eleven categories: rule-based systems, knowledge-based systems, neural networks, fuzzy ESs, object-oriented methodology, case-based reasoning, system architecture, intelligent agent systems, database methodology, modeling, and ontology together with their applications for different research and problem domains. Discussion is presented, indicating the followings future development directions for ES methodologies and applications: (1) ES methodologies are tending to develop towards expertise orientation and ES applications development is a problem-oriented domain. (2) It is suggested that different social science methodologies, such as psychology, cognitive science, and human behavior could implement ES as another kind of methodology. (3) The ability to continually change and obtain new understanding is the driving power of ES methodologies, and should be the ES application of future works.

967 citations


Journal ArticleDOI
TL;DR: This paper applies support vector machines (SVMs) to the bankruptcy prediction problem in an attempt to suggest a new model with better explanatory power and stability, and shows that SVM outperforms the other methods.
Abstract: Bankruptcy prediction has drawn a lot of research interests in previous literature, and recent studies have shown that machine learning techniques achieved better performance than traditional statistical ones. This paper applies support vector machines (SVMs) to the bankruptcy prediction problem in an attempt to suggest a new model with better explanatory power and stability. To serve this purpose, we use a grid-search technique using 5-fold cross-validation to find out the optimal parameter values of kernel function of SVM. In addition, to evaluate the prediction accuracy of SVM, we compare its performance with those of multiple discriminant analysis (MDA), logistic regression analysis (Logit), and three-layer fully connected back-propagation neural networks (BPNs). The experiment results show that SVM outperforms the other methods.

797 citations


Journal ArticleDOI
TL;DR: The results demonstrate that the accuracy and generalization performance of SVM is better than that of BPN as the training set size gets smaller, and the several superior points of the SVM algorithm compared with BPN are investigated.
Abstract: This study investigates the efficacy of applying support vector machines (SVM) to bankruptcy prediction problem. Although it is a well-known fact that the back-propagation neural network (BPN) performs well in pattern recognition tasks, the method has some limitations in that it is an art to find an appropriate model structure and optimal solution. Furthermore, loading as many of the training set as possible into the network is needed to search the weights of the network. On the other hand, since SVM captures geometric characteristics of feature space without deriving weights of networks from the training data, it is capable of extracting the optimal solution with the small training set size. In this study, we show that the proposed classifier of SVM approach outperforms BPN to the problem of corporate bankruptcy prediction. The results demonstrate that the accuracy and generalization performance of SVM is better than that of BPN as the training set size gets smaller. We also examine the effect of the variability in performance with respect to various values of parameters in SVM. In addition, we investigate and summarize the several superior points of the SVM algorithm compared with BPN.

728 citations


Journal ArticleDOI
TL;DR: The obtained results demonstrated that the proposed RNNs employing the Lyapunov exponents can be useful in analyzing long-term EEG signals for early detection of the electroencephalographic changes.
Abstract: There are a number of different quantitative models that can be used in a medical diagnostic decision support system including parametric methods, non-parametric methods and several neural network models. Unfortunately, there is no theory available to guide model selection. The aim of this study is to evaluate the diagnostic accuracy of the recurrent neural networks (RNNs) employing Lyapunov exponents trained with Levenberg-Marquardt algorithm on the electroencephalogram (EEG) signals. An approach based on the consideration that the EEG signals are chaotic signals was used in developing a reliable classification method for electroencephalographic changes. This consideration was tested successfully using the non-linear dynamics tools, like the computation of Lyapunov exponents. We explored the ability of designed and trained Elman RNNs, combined with the Lyapunov exponents, to discriminate the EEG signals (EEG signals recorded from healthy volunteers with eyes open, epilepsy patients in the epileptogenic zone during a seizure-free interval, and epilepsy patients during epileptic seizures). The RNNs achieved accuracy rates which were higher than that of the feedforward neural network models. The obtained results demonstrated that the proposed RNNs employing the Lyapunov exponents can be useful in analyzing long-term EEG signals for early detection of the electroencephalographic changes.

500 citations


Journal ArticleDOI
TL;DR: The principle interest of this work is to benchmark the performance of the proposed hybrid IDS architecture by using KDD Cup 99 Data Set, the benchmark dataset used by IDS researchers.
Abstract: In this paper, we propose a novel Intrusion Detection System (IDS) architecture utilizing both anomaly and misuse detection approaches. This hybrid Intrusion Detection System architecture consists of an anomaly detection module, a misuse detection module and a decision support system combining the results of these two detection modules. The proposed anomaly detection module uses a Self-Organizing Map (SOM) structure to model normal behavior. Deviation from the normal behavior is classified as an attack. The proposed misuse detection module uses J.48 decision tree algorithm to classify various types of attacks. The principle interest of this work is to benchmark the performance of the proposed hybrid IDS architecture by using KDD Cup 99 Data Set, the benchmark dataset used by IDS researchers. A rule-based Decision Support System (DSS) is also developed for interpreting the results of both anomaly and misuse detection modules. Simulation results of both anomaly and misuse detection modules based on the KDD 99 Data Set are given. It is observed that the proposed hybrid approach gives better performance over individual approaches.

460 citations


Journal ArticleDOI
TL;DR: An information gain technique used in machine learning for data mining to evaluate the predictive relationships of numerous financial and economic variables is introduced and shows that the trading strategies guided by the classification models generate higher risk-adjusted profits than the buy-and-hold strategy.
Abstract: It has been widely accepted by many studies that non-linearity exists in the financial markets and that neural networks can be effectively used to uncover this relationship. Unfortunately, many of these studies fail to consider alternative forecasting techniques, the relevance of input variables, or the performance of the models when using different trading strategies. This paper introduces an information gain technique used in machine learning for data mining to evaluate the predictive relationships of numerous financial and economic variables. Neural network models for level estimation and classification are then examined for their ability to provide an effective forecast of future values. A cross-validation technique is also employed to improve the generalization ability of several models. The results show that the trading strategies guided by the classification models generate higher risk-adjusted profits than the buy-and-hold strategy, as well as those guided by the level-estimation based forecasts of the neural network and linear regression models.

426 citations


Journal ArticleDOI
TL;DR: The proposed hybrid approach outperforms the results using discriminant analysis, logistic regression, artificial neural networks and MARS and hence provides an alternative in handling credit scoring tasks.
Abstract: The objective of the proposed study is to explore the performance of credit scoring using a two-stage hybrid modeling procedure with artificial neural networks and multivariate adaptive regression splines (MARS). The rationale under the analyses is firstly to use MARS in building the credit scoring model, the obtained significant variables are then served as the input nodes of the neural networks model. To demonstrate the effectiveness and feasibility of the proposed modeling procedure, credit scoring tasks are performed on one bank housing loan dataset using cross-validation approach. As the results reveal, the proposed hybrid approach outperforms the results using discriminant analysis, logistic regression, artificial neural networks and MARS and hence provides an alternative in handling credit scoring tasks.

380 citations


Journal ArticleDOI
TL;DR: The neighbor-weighted K-nearest neighbor algorithm, i.e. NWKNN, is proposed, which achieves significant classification performance improvement on imbalanced corpora.
Abstract: Text categorization or classification is the automated assigning of text documents to pre-defined classes based on their contents. Many of classification algorithms usually assume that the training examples are evenly distributed among different classes. However, unbalanced data sets often appear in many practical applications. In order to deal with uneven text sets, we propose the neighbor-weighted K-nearest neighbor algorithm, i.e. NWKNN. The experimental results indicate that our algorithm NWKNN achieves significant classification performance improvement on imbalanced corpora.

354 citations


Journal ArticleDOI
TL;DR: In this paper, genetic programming (GP) is used to build credit scoring models and it is concluded that GP can provide better performance than other models.
Abstract: Credit scoring models have been widely studied in the areas of statistics, machine learning, and artificial intelligence (AI) Many novel approaches such as artificial neural networks (ANNs), rough sets, or decision trees have been proposed to increase the accuracy of credit scoring models Since an improvement in accuracy of a fraction of a percent might translate into significant savings, a more sophisticated model should be proposed to significantly improving the accuracy of the credit scoring mode In this paper, genetic programming (GP) is used to build credit scoring models Two numerical examples will be employed here to compare the error rate to other credit scoring models including the ANN, decision trees, rough sets, and logistic regression On the basis of the results, we can conclude that GP can provide better performance than other models

332 citations


Journal ArticleDOI
TL;DR: The research findings demonstrate that both random forests techniques provide better fit for the estimation and validation sample compared to ordinary linear regression and logistic regression models, and suggest that past customer behavior is more important to generate repeat purchasing and favorable profitability evolutions.
Abstract: In an era of strong customer relationship management (CRM) emphasis, firms strive to build valuable relationships with their existing customer base. In this study, we attempt to better understand three important measures of customer outcome: next buy, partial-defection and customers' profitability evolution. By means of random forests techniques we investigate a broad set of explanatory variables, including past customer behavior, observed customer heterogeneity and some typical variables related to intermediaries. We analyze a real-life sample of 100,000 customers taken from the data warehouse of a large European financial services company. Two types of random forests techniques are employed to analyze the data: random forests are used for binary classification, whereas regression forests are applied for the models with linear dependent variables. Our research findings demonstrate that both random forests techniques provide better fit for the estimation and validation sample compared to ordinary linear regression and logistic regression models. Furthermore, we find evidence that the same set of variables have a different impact on buying versus defection versus profitability behavior. Our findings suggest that past customer behavior is more important to generate repeat purchasing and favorable profitability evolutions, while the intermediary's role has a greater impact on the customers' defection proneness. Finally, our results demonstrate the benefits of analyzing different customer outcome variables simultaneously, since an extended investigation of the next buy-partial-defection-customer profitability triad indicates that one cannot fully understand a particular outcome without understanding the other related behavioral outcome variables.

315 citations


Journal ArticleDOI
TL;DR: A three-way comparison of prediction accuracy involving nonlinear regression, NNs and CART models using a continuous dependent variable and a set of dichotomous and categorical predictor variables is performed.
Abstract: Numerous articles comparing performances of statistical and Neural Networks (NNs) models are available in the literature, however, very few involved Classification and Regression Tree (CART) models in their comparative studies. We perform a three-way comparison of prediction accuracy involving nonlinear regression, NNs and CART models using a continuous dependent variable and a set of dichotomous and categorical predictor variables. A large dataset on smokers is used to run these models. Different prediction accuracy measuring procedures are used to compare performances of these models. The outcomes of predictions are discussed and the outcomes of this research are compared with the results of similar studies.

Journal ArticleDOI
TL;DR: A novel method of analysis of EEG signals using wavelet transform, and classification using ANN, which selected the error back-propagation neural network as a classifier to discriminate the alertness level of a subject.
Abstract: Electrophysiological recordings are considered a reliable method of assessing a person's alertness. Sleep medicine is asked to offer objective methods to measure daytime alertness, tiredness and sleepiness. As EEG signals are non-stationary, the conventional method of frequency analysis is not highly successful in recognition of alertness level. This paper deals with a novel method of analysis of EEG signals using wavelet transform, and classification using ANN. EEG signals were decomposed into the frequency sub-bands using wavelet transform and a set of statistical features was extracted from the sub-bands to represent the distribution of wavelet coefficients. Then these statistical features were used as an input to an ANN with three discrete outputs: alert, drowsy and sleep. The error back-propagation neural network is selected as a classifier to discriminate the alertness level of a subject. EEG signals were obtained from 30 healthy subjects. The group consisted of 14 females and 16 males with ages ranging from 18 to 65 years and a mean age of 33.5 years, and a Body Mass Index (BMI) of 32.4+/-7.3kg/m^2. Alertness level and classification properties of ANN were tested using the data recorded in 12 healthy subjects, whereby the EEG recordings were not used to train the ANN. The statistics were used as a measure of potential applicability of the ANN. The accuracy of the ANN was 95+/-3% alert, 93+/-4% drowsy and 92+/-5% sleep.

Journal ArticleDOI
TL;DR: This paper presents an application of the analytic hierarchy process (AHP) used to select the most appropriate tool to support knowledge management (KM) and adopts a multi-criteria approach that can be used to analyse and compare KM tools in the software market.
Abstract: This paper presents an application of the analytic hierarchy process (AHP) used to select the most appropriate tool to support knowledge management (KM). This method adopts a multi-criteria approach that can be used to analyse and compare KM tools in the software market. The method is based on pairwise comparisons between several factors that affect the selection of the most appropriate KM tool. An AHP model is formulated and applied to a real case of assisting decision-makers in a leading communications company in Hong Kong to evaluate a suitable KM tool. We believe that the application shown can be of use to managers and that, because of its ease of implementation, others can benefit from this approach.

Journal ArticleDOI
TL;DR: Wang et al. as mentioned in this paper explored a hybrid collaborative filtering method, collaborative filtering based on item and user techniques, by combining collaborative filtering on items and user on user together for multiple-interests and multiple-content recommendation.
Abstract: Recommender systems apply knowledge discovery techniques to the problem of making personalized recommendations for products or services during a live interaction. These systems, especially collaborative filtering based on user, are achieving widespread success on the Web. The tremendous growth in the amount of available information and the kinds of commodity to Web sites in recent years poses some key challenges for recommender systems. One of these challenges is ability of recommender systems to be adaptive to environment where users have many completely different interests or items have completely different content (We called it as Multiple interests and Multiple-content problem). Unfortunately, the traditional collaborative filtering systems can not make accurate recommendation for the two cases because the predicted item for active user is not consist with the common interests of his neighbor users. To address this issue we have explored a hybrid collaborative filtering method, collaborative filtering based on item and user techniques, by combining collaborative filtering based on item and collaborative filtering based on user together. Collaborative filtering based on item and user analyze the user-item matrix to identify similarity of target item to other items, generate similar items of target item, and determine neighbor users of active user for target item according to similarity of other users to active user based on similar items of target item. In this paper we firstly analyze limitation of collaborative filtering based on user and collaborative filtering based on item algorithms respectively and emphatically make explain why collaborative filtering based on user is not adaptive to Multiple-interests and Multiple-content recommendation. Based on analysis, we present collaborative filtering based on item and user for Multiple-interests and Multiple-content recommendation. Finally, we experimentally evaluate the results and compare them with collaborative filtering based on user and collaborative filtering based on item, respectively. The experiments suggest that collaborative filtering based on item and user provide better recommendation quality than collaborative filtering based on user and collaborative filtering based on item dramatically.

Journal ArticleDOI
TL;DR: The findings suggest that the BP network is a better choice when a target vector is available than the Kohonen self-organizing feature map in the area of bankruptcy prediction.
Abstract: In this study, two learning paradigms of neural networks, supervised versus unsupervised, are compared using their representative types. The back-propagation (BP) network and the Kohonen self-organizing feature map, selected as the representative type for supervised and unsupervised neural networks, respectively, are compared in terms of prediction accuracy in the area of bankruptcy prediction. Discriminant analysis and logistic regression are also performed to provide performance benchmarks. The findings suggest that the BP network is a better choice when a target vector is available.

Journal ArticleDOI
TL;DR: This study integrates customer behavioral variables, demographic variables, and transaction database to establish a method of mining changes in customer behavior that can assist managers in developing better marketing strategies.
Abstract: During the past decade, there have been a variety of significant developments in data mining techniques. Some of these developments are implemented in customized service to develop customer relationship. Customized service is actually crucial in retail markets. Marketing managers can develop long-term and pleasant relationships with customers if they can detect and predict changes in customer behavior. In the dynamic retail market, understanding changes in customer behavior can help managers to establish effective promotion campaigns. This study integrates customer behavioral variables, demographic variables, and transaction database to establish a method of mining changes in customer behavior. For mining change patterns, two extended measures of similarity and unexpectedness are designed to analyze the degree of resemblance between patterns at different time periods. The proposed approach for mining changes in customer behavior can assist managers in developing better marketing strategies.

Journal ArticleDOI
TL;DR: A mathematical programming model is suggested that considers the change in suppliers' supply capabilities and customer needs over a period in time and design a model which not only maximizes revenue but also satisfies customer needs.
Abstract: We propose an effective supplier selection method to maintain a continuous supply-relationship with suppliers. Costs have been sharply increasing and profit decreasing as the global competition among companies has increased and customer demands have diversified in the current business environment. Many other functions are now outsourced globally to strengthen competition. As a result, one of the issues is how to select good suppliers which can maintain a continuous supply-relationship. We suggest a mathematical programming model that considers the change in suppliers' supply capabilities and customer needs over a period in time. We design a model which not only maximizes revenue but also satisfies customer needs. The suggested model is applied to supplier selection and management of the agriculture industry in Korea.

Journal ArticleDOI
Osman Kulak1
TL;DR: A fuzzy information axiom approach is introduced and used in the selection of material handling equipment in a real case and it is used for the most proper equipment among the alternatives of the same type using the informationAxiom of axiomatic design principles.
Abstract: Effective use of labor, providing system flexibility, increasing productivity, decreasing lead times and costs are some of the most important factors influencing selection of material handling equipment. In this study, a decision support system (FUMAHES: fuzzy multi-attribute material handling equipment selection) considering these factors for material handling equipment selection is developed. FUMAHES consists of a database, a rule-based system and multi-attribute decision making modules. This database includes detailed data about equipment types and their properties. The rule-based system module provides rules, which are utilized by inference engine for determining the most proper material handling equipment type. Ultimately, a final decision is made for the most proper equipment among the alternatives of the same type using the information axiom of axiomatic design principles. Evaluation of alternatives is made for the cases of both complete and incomplete information. This paper also introduces a fuzzy information axiom approach and uses it in the selection of material handling equipment in a real case.

Journal ArticleDOI
TL;DR: The purpose of this study is to reduce effectively the time and cost of design under the premise to manufacture an accurate new product by using the Case-Based Reasoning (CBR) algorithm to construct the new BOM.
Abstract: Product variation and customization is a trend in current market-oriented manufacturing environment. Companies produce products in order to satisfy customer's needs. In the customization environment, the R&D sector in an enterprise should be able to offer differentiation in product selection after they take the order. Such product differentiation should meet the requirement of cost and manufacturing procedure. In the light of this, how to generate an accurate bill of material (BOM) that meets the customer's needs and gets ready for the production is an important issue in the intensely competitive market. The purpose of this study is to reduce effectively the time and cost of design under the premise to manufacture an accurate new product. In this study, the Case-Based Reasoning (CBR) algorithm was used to construct the new BOM. Retrieving previous cases that resemble the current problem can save a lot of time in figuring out the problem and offer a correct direction for designers. When solving a new problem, CBR technique can quickly help generate a right BOM that fits the present situation.

Journal ArticleDOI
TL;DR: This study explores the CRM system success model that consists of CRM initiatives: process fit, customer information quality, and system support; intrinsic success: efficiency and customer satisfaction; and extrinsic success: profitability.
Abstract: As the market competition becomes keen, constructing a customer relationship management system is coming to the front for winning over new customers, developing service and products for customer satisfaction and retaining existing customers However, decisions for CRM implementation have been hampered by inconsistency between information technology and marketing strategies, and the lack of conceptual bases necessary to develop the success measures Using a structural equation analysis, this study explores the CRM system success model that consists of CRM initiatives: process fit, customer information quality, and system support; intrinsic success: efficiency and customer satisfaction; and extrinsic success: profitability These constructs underlie much of the existing literature on information system success and customer satisfaction perspectives We found the empirical support for CRM implementation decision-making from 253 respondents of 14 companies which have implemented the CRM system These findings should be of great interest to both researchers and practitioners

Journal ArticleDOI
TL;DR: This study used clustering techniques to preprocess the input samples with the objective of indicating unrepresentative samples into isolated and inconsistent clusters, and used neural networks to construct the credit scoring model.
Abstract: Unrepresentative data samples are likely to reduce the utility of data classifiers in practical application. This study presents a hybrid mining approach in the design of an effective credit scoring model, based on clustering and neural network techniques. We used clustering techniques to preprocess the input samples with the objective of indicating unrepresentative samples into isolated and inconsistent clusters, and used neural networks to construct the credit scoring model. The clustering stage involved a class-wise classification process. A self-organizing map clustering algorithm was used to automatically determine the number of clusters and the starting points of each cluster. Then, the K-means clustering algorithm was used to generate clusters of samples belonging to new classes and eliminate the unrepresentative samples from each class. In the neural network stage, samples with new class labels were used in the design of the credit scoring model. The proposed method demonstrates by two real world credit data sets that the hybrid mining approach can be used to build effective credit scoring models.

Journal ArticleDOI
TL;DR: A novel method of analysis of EEG signals using discrete wavelet transform, and classification using ANN, where the DWN-based classifier outperformed the FEBANN based counterpart and within the same group, theDWN- based classifier was more accurate than theFEBANN-basedclassifier.
Abstract: Epileptic seizures are manifestations of epilepsy. Careful analyses of the electroencephalograph (EEG) records can provide valuable insight and improved understanding of the mechanisms causing epileptic disorders. The detection of epileptiform discharges in the EEG is an important component in the diagnosis of epilepsy. Wavelet transform is particularly effective for representing various aspects of non-stationary signals such as trends, discontinuities, and repeated patterns where other signal processing approaches fail or are not as effective. Through wavelet decomposition of the EEG records, transient features are accurately captured and localized in both time and frequency context. This paper deals with a novel method of analysis of EEG signals using discrete wavelet transform, and classification using ANN. EEG signals were decomposed into the frequency sub-bands using wavelet transform. Then these sub-band frequencies were used as an input to an ANN with two discrete outputs: normal and epileptic. In this study, FEBANN and DWN based classifiers were developed and compared in relation to their accuracy in classification of EEG signals. The comparisons between the developed classifiers were primarily based on analysis of the ROC curves as well as a number of scalar performance measures pertaining to the classification. The DWN-based classifier outperformed the FEBANN based counterpart. Within the same group, the DWN-based classifier was more accurate than the FEBANN-based classifier.

Journal ArticleDOI
TL;DR: MLP and RBF which are one each of neural networks procedures, performed better than other techniques in predicting hypertension, and QUEST had a lesser performance than other technique.
Abstract: Hypertension is a leading cause of heart disease and stroke. In this study, performance of classification techniques is compared in order to predict the risk of essential hypertension disease. A retrospective analysis was performed in 694 subjects (452 patients and 242 controls). We compared performances of three decision trees, four statistical algorithms, and two neural networks. Predictor variables were age, sex, family history of hypertension, smoking habits, lipoprotein (a), triglyceride, uric acid, total cholesterol, and body mass index (BMI). Classification techniques were grouped using hierarchical cluster analysis (HCA). The data points appeared to cluster in three groups. The first cluster included MLP and RBF. Furthermore CART which was more similar than other techniques linked this cluster. The second cluster included FDA/MARS (degree=1), LR and QUEST, but FDA/MARS (degree=1) and LR was more similar than QUEST. The third cluster included FDA/MARS (degree=2), CHAID and FDA, but FDA/MARS (degree=2) and CHAID was more similar than FDA. MLP and RBF which are one each of neural networks procedures, performed better than other techniques in predicting hypertension. QUEST had a lesser performance than other techniques.

Journal ArticleDOI
TL;DR: It is reported that index fund could improve its performance greatly with the proposed GA portfolio scheme, which will be demonstrated for index fund designed to track Korea Stock Price Index (KOSPI) 200.
Abstract: Using genetic algorithm (GA), this study proposes a portfolio optimization scheme for index fund management. Index fund is one of popular strategies in portfolio management that aims at matching the performance of the benchmark index such as the SP Gruber, M. J. (1996). Another puzzle: the growth in actively managed mutual funds. Journal of Finance, 51(3), 783-810; Malkiel, B. (1995). Returns from investing in equity mutual funds 1971 to 1991. Journal of Finance, 50, 549-572]. The main objective of this paper is to report that index fund could improve its performance greatly with the proposed GA portfolio scheme, which will be demonstrated for index fund designed to track Korea Stock Price Index (KOSPI) 200.

Journal ArticleDOI
TL;DR: Two techniques are proposed based on wavelet analysis and fuzzy-neural approaches that can be very helpful for radiologist in detection and diagnosing abnormalities earlier and faster than traditional screening programs.
Abstract: The high incidence of breast cancer in women has increased significantly in the recent years. The most familiar breast tumors types are mass and microcalcification. Mammograms-breast X-ray-are considered the most reliable method in early detection of breast cancer. Computer-aided diagnosis system can be very helpful for radiologist in detection and diagnosing abnormalities earlier and faster than traditional screening programs. Several techniques can be used to accomplish this task. In this paper, two techniques are proposed based on wavelet analysis and fuzzy-neural approaches. These techniques are mammography classifier based on globally processed image and mammography classifier based on locally processed image (region of interest). The system is classified normal from abnormal, mass for microcalcification and abnormal severity (benign or malignant). The evaluation of the system is carried out on Mammography Image Analysis Society (MIAS) dataset. The accuracy achieved is satisfied.

Journal ArticleDOI
TL;DR: This paper presents a multi-agent-based system, called the survey-based profitable customers segmentation system that executes the customer satisfaction survey and conducts the mining of customer Satisfaction survey, socio-demographic and accounting database through the integrated uses of business intelligence tools.
Abstract: For the success of CRM, it is important to target the most profitable customers of a company. Many CRM researches have been performed to calculate customer profitability and develop a comprehensive model of it. Most of them, however, had some limitations and accordingly the customer segmentation based on the customer profitability model is still underutilized. This paper aims at providing an easy, efficient and more practical alternative approach based on the customer satisfaction survey for the profitable customers segmentation. We present a multi-agent-based system, called the survey-based profitable customers segmentation system that executes the customer satisfaction survey and conducts the mining of customer satisfaction survey, socio-demographic and accounting database through the integrated uses of business intelligence tools such as DEA (Data Envelopment Analysis), Self-Organizing Map (SOM) neural network and C4.5 for the profitable customers segmentation. A case study on a Motor company's profitable customer segmentation is illustrated.

Journal ArticleDOI
TL;DR: A new crossover mechanism named dominated gene crossover will be introduced to enhance the performance of genetic search, and eliminate the problem of determining optimal crossover rate in multi-factory and multi-product environment.
Abstract: This paper proposes an adaptive genetic algorithm for distributed scheduling problems in multi-factory and multi-product environment. Distributed production strategy enables factories to be more focused on their core product types, to achieve better quality, to reduce production cost, and to reduce management risk. However, when comparing with single-factory production, scheduling problems involved in multi-factory one are more complicated, since different jobs distributed to different factories will have different production scheduling, consequently affect the performance of the supply chain. Distributed scheduling problems deal with the assignment of jobs to suitable factories and determine their production scheduling accordingly. In this paper, a new crossover mechanism named dominated gene crossover will be introduced to enhance the performance of genetic search, and eliminate the problem of determining optimal crossover rate. A number of experiments have been carried out. For the comparison purpose, five multi-factory models have been solved by different well known optimization approaches. The results indicate that significant improvement could be obtained by the proposed algorithm.

Journal ArticleDOI
TL;DR: The proposed CF (collaborative filtering)-based recommender system is versatile and can be applied to a variety of e-commerce sites as long as the navigational and behavioral patterns of customers can be captured.
Abstract: In this article, a novel CF (collaborative filtering)-based recommender system is developed for e-commerce sites. Unlike the conventional approach in which only binary purchase data are used, the proposed approach analyzes the data captured from the navigational and behavioral patterns of customers, estimates the preference levels of a customer for the products which are clicked but not purchased, and CF is conducted using the preference levels for making recommendations. This also compares with the existing works on clickstream data analysis in which the navigational and behavioral patterns of customers are analyzed for simple relationships with the target variable. The effectiveness of the proposed approach is assessed using an experimental e-commerce site. It is found among other things that the proposed approach outperforms the conventional approach in almost all cases considered. The proposed approach is versatile and can be applied to a variety of e-commerce sites as long as the navigational and behavioral patterns of customers can be captured.

Journal ArticleDOI
TL;DR: In this paper, a decision table based on the causesymptom matrix is used as a probabilistic method for diagnosing abnormal vibration for rotating machinery and a decision tree is used for the acquisition of structured knowledge in the form of concepts.
Abstract: This paper proposes an expert system called VIBEX (VIBration EXpert) to aid plant operators in diagnosing the cause of abnormal vibration for rotating machinery. In order to automatize the diagnosis, a decision table based on the cause-symptom matrix is used as a probabilistic method for diagnosing abnormal vibration. Also a decision tree is used as the acquisition of structured knowledge in the form of concepts is introduced to build a knowledge base which is indispensable for vibration expert systems. The decision tree is a technique used for building knowledge-based systems by the inductive inference from examples and plays a role itself as a vibration diagnostic tool. The proposed system has been successfully implemented on Microsoft Windows environment and is written in Microsoft Visual Basic and Visual C++. To validate the system performance, the diagnostic system was tested with some examples using the two diagnostic methods.

Journal ArticleDOI
Sung-Hwan Min1, Ingoo Han1
TL;DR: This paper suggests a methodology for detecting a user's time-variant pattern in order to improve the performance of collaborative filtering recommendations and detects changes in customer behavior using the customer data at different periods of time and improves theperformance of recommendations using information on changes.
Abstract: Due to the explosion of e-commerce, recommender systems are rapidly becoming a core tool to accelerate cross-selling and strengthen customer loyalty. There are two prevalent approaches for building recommender systems-content-based recommending and collaborative filtering. So far, collaborative filtering recommender systems have been very successful in both information filtering and e-commerce domains. However, the current research on recommendation has paid little attention to the use of time-related data in the recommendation process. Up to now there has not been any study on collaborative filtering to reflect changes in user interest. This paper suggests a methodology for detecting a user's time-variant pattern in order to improve the performance of collaborative filtering recommendations. The methodology consists of three phases of profiling, detecting changes, and recommendations. The proposed methodology detects changes in customer behavior using the customer data at different periods of time and improves the performance of recommendations using information on changes.