scispace - formally typeset
Journal ArticleDOI

Comparison of random forest, support vector machine and back propagation neural network for electronic tongue data classification: Application to the recognition of orange beverage and Chinese vinegar

Reads0
Chats0
TLDR
Results suggest that RF may be a promising pattern recognition method for E-tongue data processing, because it can deal with classification problems of unbalanced, multiclass and small sample data without data preprocessing procedures.
Abstract
Random forest (RF) has been proposed on the basis of classification and regression trees (CART) with “ensemble learning” strategy by Breiman in 2001. In this paper, RF is introduced and investigated for electronic tongue (E-tongue) data processing. The experiments were designed for type and brand recognition of orange beverage and Chinese vinegar by an E-tongue with seven potentiometric sensors and an Ag/AgCl reference electrode. Principal component analysis (PCA) was used to visualize the distribution of total samples of each data set. Back propagation neural network (BPNN) and support vector machine (SVM), as comparative methods, were also employed to deal with four data sets. Five-fold cross-validation (CV) with twenty replications was applied during modeling and an external testing set was employed to validate the prediction performance of models. The average correct rates (CR) on CV sets of the four data sets performed by BPNN, SVM and RF were 86.68%, 66.45% and 99.07%, respectively. RF has been proved to outperform BPNN and SVM, and has some advantages in such cases, because it can deal with classification problems of unbalanced, multiclass and small sample data without data preprocessing procedures. These results suggest that RF may be a promising pattern recognition method for E-tongues.

read more

Citations
More filters
Journal ArticleDOI

A Systematic Review on Imbalanced Data Challenges in Machine Learning: Applications and Solutions

TL;DR: A comparative analysis of the approaches from the reference of data pre-processing, algorithmic and hybrid paradigms for contemporary imbalance data analysis techniques, and their comparative study in lieu of different data distribution and their application areas are presented.
Journal ArticleDOI

Comparison of random forest, artificial neural networks and support vector machine for intelligent diagnosis of rotating machinery:

TL;DR: Random forest has been proven to outperform the comparative classifiers in terms of recognition accuracy, stability and robustness to features, especially with a small training set, and the user-friendly parameters in random forest offer great convenience for practical engineering.
Journal ArticleDOI

Comparing fully convolutional networks, random forest, support vector machine, and patch-based deep convolutional neural networks for object-based wetland mapping using images from small unmanned aircraft system

TL;DR: This study assessed the deep learning classifiers using different training sample sizes and compared their performance with traditional classifiers to indicate that DCNN may produce inferior performance compared to conventional classifiers when the training sample size is small, but it tends to show substantially higher accuracy than the conventional classifier when theTraining sample size becomes large.
Journal ArticleDOI

Multi-categorical deep learning neural network to classify retinal images: A pilot study employing small database.

TL;DR: It is found that the transfer learning incorporated with ensemble classifiers can improve the classification performance in order to detect multi-categorical retinal diseases.
Journal ArticleDOI

Applications of electronic nose (e-nose) and electronic tongue (e-tongue) in food quality-related properties determination: A review

TL;DR: E-nose or e-tongue combining pattern recognition algorithms are very powerful analytical tools, which are relatively low-cost, rapid, and accurate in determining the quality-related properties of foods.
References
More filters
Journal ArticleDOI

Random Forests

TL;DR: Internal estimates monitor error, strength, and correlation and these are used to show the response to increasing the number of features used in the forest, and are also applicable to regression.
Journal ArticleDOI

Support-Vector Networks

TL;DR: High generalization ability of support-vector networks utilizing polynomial input transformations is demonstrated and the performance of the support- vector network is compared to various classical learning algorithms that all took part in a benchmark study of Optical Character Recognition.

Statistical learning theory

TL;DR: Presenting a method for determining the necessary and sufficient conditions for consistency of learning process, the author covers function estimates from small data pools, applying these estimations to real-life problems, and much more.
Journal ArticleDOI

Bagging predictors

Leo Breiman
TL;DR: Tests on real and simulated data sets using classification and regression trees and subset selection in linear regression show that bagging can give substantial gains in accuracy.

Classification and Regression by randomForest

TL;DR: random forests are proposed, which add an additional layer of randomness to bagging and are robust against overfitting, and the randomForest package provides an R interface to the Fortran programs by Breiman and Cutler.
Related Papers (5)