scispace - formally typeset
Search or ask a question
Journal ArticleDOI

Real-time analysis and regulation of EPB shield steering using Random Forest

01 Oct 2019-Automation in Construction (Elsevier)-Vol. 106, pp 102860
TL;DR: The results indicate that the predicted settlements show great agreement with the measured settlements and the consistency of the face pressure and grout filling values calculated by the PSO-RF and the grid search method demonstrates the feasibility and applicability of proposed hybrid algorithm.
About: This article is published in Automation in Construction.The article was published on 2019-10-01. It has received 122 citations till now. The article focuses on the topics: Settlement (structural) & Hybrid algorithm.
Citations
More filters
Journal ArticleDOI
TL;DR: An overview of some soft computing techniques as well as their applications in underground excavations is presented and a case study is adopted to compare the predictive performances ofsoft computing techniques including eXtreme Gradient Boosting, Multivariate Adaptive Regression Splines, and Support Vector Machine in estimating the maximum lateral wall deflection induced by braced excavation.
Abstract: Soft computing techniques are becoming even more popular and particularly amenable to model the complex behaviors of most geotechnical engineering systems since they have demonstrated superior predictive capacity, compared to the traditional methods. This paper presents an overview of some soft computing techniques as well as their applications in underground excavations. A case study is adopted to compare the predictive performances of soft computing techniques including eXtreme Gradient Boosting (XGBoost), Multivariate Adaptive Regression Splines (MARS), Artificial Neural Networks (ANN), and Support Vector Machine (SVM) in estimating the maximum lateral wall deflection induced by braced excavation. This study also discusses the merits and the limitations of some soft computing techniques, compared with the conventional approaches available.

287 citations

Journal ArticleDOI
TL;DR: Parametric investigation indicates that the relationships between Cα and the four input variables in the proposed RF models harmonize with the physical explanation, and these three proposed models demonstrably outperform empirical methods, featuring as they do lower levels of prediction error.

106 citations

Journal ArticleDOI
TL;DR: The results indicate that proposed hybrid intelligent algorithm with the integration of the meta-heuristic algorithm particle swarm optimization (PSO) and ML can effectively determine the global optimum hyper-parameters of ML algorithms.

100 citations

Journal ArticleDOI
Renpeng Chen1, Pin Zhang1, Huaina Wu1, Zhi-Teng Wang1, Zhi-Quan Zhong 
TL;DR: The results indicated that ML algorithms have great potential to predict tunneling-induced settlement, compared with the traditional multivariate linear regression method.
Abstract: Predicting the tunneling-induced maximum ground surface settlement is a complex problem since the settlement depends on plenty of intrinsic and extrinsic factors. This study investigates the efficiency and feasibility of six machine learning (ML) algorithms, namely, back-propagation neural network, wavelet neural network, general regression neural network (GRNN), extreme learning machine, support vector machine and random forest (RF), to predict tunneling-induced settlement. Field data sets including geological conditions, shield operational parameters, and tunnel geometry collected from four sections of tunnel with a total of 3.93 km are used to build models. Three indicators, mean absolute error, root mean absolute error, and coefficient of determination the (R2) are used to demonstrate the performance of each computational model. The results indicated that ML algorithms have great potential to predict tunneling-induced settlement, compared with the traditional multivariate linear regression method. GRNN and RF algorithms show the best performance among six ML algorithms, which accurately recognize the evolution of tunneling-induced settlement. The correlation between the input variables and settlement is also investigated by Pearson correlation coefficient.

78 citations

Journal ArticleDOI
TL;DR: The results indicate that ML models outperform empirical prediction formulations with lower prediction error and the predicted correlations between input and output variables using five ML models show great agreement with the physical explanation.
Abstract: Compression index Cc is an essential parameter in geotechnical design for which the effectiveness of correlation is still a challenge. This paper suggests a novel modelling approach using machine learning (ML) technique. The performance of five commonly used machine learning (ML) algorithms, i.e. back-propagation neural network (BPNN), extreme learning machine (ELM), support vector machine (SVM), random forest (RF) and evolutionary polynomial regression (EPR) in predicting Cc is comprehensively investigated. A database with a total number of 311 datasets including three input variables, i.e. initial void ratio e0, liquid limit water content wL, plasticity index Ip, and one output variable Cc is first established. Genetic algorithm (GA) is used to optimize the hyper-parameters in five ML algorithms, and the average prediction error for the 10-fold cross-validation (CV) sets is set as the fitness function in the GA for enhancing the robustness of ML models. The results indicate that ML models outperform empirical prediction formulations with lower prediction error. RF yields the lowest error followed by BPNN, ELM, EPR and SVM. If the ranges of input variables in the database are large enough, BPNN and RF models are recommended to predict Cc. Furthermore, if the distribution of input variables is continuous, RF model is the best one. Otherwise, EPR model is recommended if the ranges of input variables are small. The predicted correlations between input and output variables using five ML models show great agreement with the physical explanation.

70 citations

References
More filters
Journal ArticleDOI
01 Aug 1996
TL;DR: Tests on real and simulated data sets using classification and regression trees and subset selection in linear regression show that bagging can give substantial gains in accuracy.
Abstract: Bagging predictors is a method for generating multiple versions of a predictor and using these to get an aggregated predictor. The aggregation averages over the versions when predicting a numerical outcome and does a plurality vote when predicting a class. The multiple versions are formed by making bootstrap replicates of the learning set and using these as new learning sets. Tests on real and simulated data sets using classification and regression trees and subset selection in linear regression show that bagging can give substantial gains in accuracy. The vital element is the instability of the prediction method. If perturbing the learning set can cause significant changes in the predictor constructed, then bagging can improve accuracy.

16,118 citations

01 Jan 2007
TL;DR: random forests are proposed, which add an additional layer of randomness to bagging and are robust against overfitting, and the randomForest package provides an R interface to the Fortran programs by Breiman and Cutler.
Abstract: Recently there has been a lot of interest in “ensemble learning” — methods that generate many classifiers and aggregate their results. Two well-known methods are boosting (see, e.g., Shapire et al., 1998) and bagging Breiman (1996) of classification trees. In boosting, successive trees give extra weight to points incorrectly predicted by earlier predictors. In the end, a weighted vote is taken for prediction. In bagging, successive trees do not depend on earlier trees — each is independently constructed using a bootstrap sample of the data set. In the end, a simple majority vote is taken for prediction. Breiman (2001) proposed random forests, which add an additional layer of randomness to bagging. In addition to constructing each tree using a different bootstrap sample of the data, random forests change how the classification or regression trees are constructed. In standard trees, each node is split using the best split among all variables. In a random forest, each node is split using the best among a subset of predictors randomly chosen at that node. This somewhat counterintuitive strategy turns out to perform very well compared to many other classifiers, including discriminant analysis, support vector machines and neural networks, and is robust against overfitting (Breiman, 2001). In addition, it is very user-friendly in the sense that it has only two parameters (the number of variables in the random subset at each node and the number of trees in the forest), and is usually not very sensitive to their values. The randomForest package provides an R interface to the Fortran programs by Breiman and Cutler (available at http://www.stat.berkeley.edu/ users/breiman/). This article provides a brief introduction to the usage and features of the R functions.

14,830 citations

Journal ArticleDOI
Tin Kam Ho1
TL;DR: A method to construct a decision tree based classifier is proposed that maintains highest accuracy on training data and improves on generalization accuracy as it grows in complexity.
Abstract: Much of previous attention on decision trees focuses on the splitting criteria and optimization of tree sizes. The dilemma between overfitting and achieving maximum accuracy is seldom resolved. A method to construct a decision tree based classifier is proposed that maintains highest accuracy on training data and improves on generalization accuracy as it grows in complexity. The classifier consists of multiple trees constructed systematically by pseudorandomly selecting subsets of components of the feature vector, that is, trees constructed in randomly chosen subspaces. The subspace method is compared to single-tree classifiers and other forest construction methods by experiments on publicly available datasets, where the method's superiority is demonstrated. We also discuss independence between trees in a forest and relate that to the combined classification accuracy.

5,984 citations

Journal ArticleDOI
TL;DR: In this paper, the three primary sources of geotechnical uncertainties are inherent variability, measurem, and measurem uncertainties, and the three main sources of variability are measurem and inherent variability.
Abstract: Geotechnical variability is a complex attribute that results from many disparate sources of uncertainties. The three primary sources of geotechnical uncertainties are inherent variability, measurem...

1,663 citations

Journal ArticleDOI
TL;DR: In this paper, closed form solutions for obtaining the strain field in an initially isotropic and homogeneous incompressible soil due to near-surface ground loss are presented for some typical problems, such as soft ground tunnelling and pile driving.
Abstract: Closed form solutions are presented for obtaining the strain field in an initially isotropic and homogeneous incompressible soil due to near-surface ground loss. This problem fits into a category of cases in which the imposed boundary conditions are only or mainly in terms of displacements (strain-controlled problems). In these cases there is a possibility of eliminating the stresses and obtaining the strains by using only the incompressibility condition. The presence of the top free surface is considered by means of a virtual image technique and some results for the elastic half-space. The results are simple, especially for the movements of the soil surface. The application to some typical problems, such as soft ground tunnelling and pile driving or extraction, shows that the calculated movements agree well with the experimental observations and compare favourably with commonly used numerical methods. Some hints are given for extending the method to other cases, such as to compressible materials. Des sol...

503 citations