scispace - formally typeset
Search or ask a question
Author

Boxin Tang

Bio: Boxin Tang is an academic researcher from Simon Fraser University. The author has contributed to research in topics: Orthogonal array & Fractional factorial design. The author has an hindex of 23, co-authored 65 publications receiving 2410 citations. Previous affiliations of Boxin Tang include University of Memphis & University of Michigan.


Papers
More filters
Journal ArticleDOI
Boxin Tang1
TL;DR: It is proved that when used for integration, the sampling scheme with OA-based Latin hypercubes offers a substantial improvement over Latin hypercube sampling.
Abstract: In this article, we use orthogonal arrays (OA's) to construct Latin hypercubes. Besides preserving the univariate stratification properties of Latin hypercubes, these strength r OA-based Latin hypercubes also stratify each r-dimensional margin. Therefore, such OA-based Latin hypercubes provide more suitable designs for computer experiments and numerical integration than do general Latin hypercubes. We prove that when used for integration, the sampling scheme with OA-based Latin hypercubes offers a substantial improvement over Latin hypercube sampling.

768 citations

01 Jan 1999
TL;DR: In this paper, a generalized resolution criterion is defined and used for assessing non-regular fractional factorials, notably Plackett-Burman designs, which is intended to capture projection properties, complementing that of Webb (1964) whose concept of resolution concerns the estimability of lower order fractional fractional factors under the assumption that higher order effects are negligible.
Abstract: Resolution has been the most widely used criterion for comparing regular fractional factorials since it was introduced in 1961 by Box and Hunter. In this pa- per, we examine how a generalized resolution criterion can be defined and used for assessing nonregular fractional factorials, notably Plackett-Burman designs. Our generalization is intended to capture projection properties, complementing that of Webb (1964) whose concept of resolution concerns the estimability of lower order ef- fects under the assumption that higher order effects are negligible. Our generalized resolution provides a fruitful criterion for ranking different designs while Webb's resolution is mainly useful as a classification rule. An additional advantage of our approach is that the idea leads to a natural generalization of minimum aberration. Examples are given to illustrate the usefulness of the new criteria.

188 citations

Journal ArticleDOI
TL;DR: In this article, a general method for constructing supersaturated designs is proposed and shown to produce designs with n runs and m = k(n − 1) factors that achieve the lower bound for Es2 and are thus optimal with respect to the Es2 criterion.
Abstract: A lower bound for the Es2 value of an arbitrary supersaturated design is derived. A general method for constructing supersaturated designs is proposed and shown to produce designs with n runs and m = k(n — 1) factors that achieve the lower bound for Es2 and are thus optimal with respect to the Es2 criterion. Within the class of designs given by the construction method, further discrimination can be made by minimizing the pairwise correlations and using the generalized D and A criteria proposed by Wu (1993). Efficient designs of 12, 16, 20 and 24 runs are constructed by following this approach. Nous derivons une borne inferieure pour la valeur Es2 d'un schema arbitraire supersature et nous proposons une methode generale de construction de schemas supersatures. Nous demontrons que cette methode produit des schemas avec n iterations et m = k(n — 1) facteurs qui atteignent la borne inferieure pour Es2 et sont donc optimaux relativement au critere Es2. Au sein de la classe de schemas donnee par la methode de construction, une discrimination supplementaire peut ětre faite en minimisant les correlations par paire et en utilisant les criteres generalises D et A proposes par Wu (1993). Nous construisons des schemas efficaces de 12, 16, 20 et 24 iterations en suivant cette approche.

159 citations

Journal ArticleDOI
TL;DR: This article considers the problem of classifying and ranking designs that are based on Hadamard matrices and finds that generalized aberration performs quite well under these familiar criteria.
Abstract: Deng and Tang (1999) and Tang and Deng (1999) proposed and justified two criteria of generalized minimum aberration for general two-level fractional factorial designs. The criteria are defined using a set of values called J characteristics. In this article, we examine the practical use of the criteria in design selection. Specifically, we consider the problem of classifying and ranking designs that are based on Hadamard matrices. A theoretical result on J characteristics is developed to facilitate the computation. The issue of design selection is further studied by linking generalized aberration with the criteria of efficiency and estimation capacity. Our studies reveal that generalized aberration performs quite well under these familiar criteria.

103 citations

Journal ArticleDOI
TL;DR: In this paper, the authors introduce a method for constructing a rich class of designs that are suitable for use in computer experiments, including Latin hypercube designs and two-level fractional factorial designs.
Abstract: We introduce a method for constructing a rich class of designs that are suitable for use in computer experiments. The designs include Latin hypercube designs and two-level fractional factorial designs as special cases and fill the vast vacuum between these two familiar classes of designs. The basic construction method is simple, building a series of larger designs based on a given small design. If the base design is orthogonal, the resulting designs are orthogonal; likewise, if the base design is nearly orthogonal, the resulting designs are nearly orthogonal. We present two generalizations of our basic construction method. The first generalization improves the projection properties of the basic method; the second generalization gives rise to designs that have smaller correlations. Sample constructions are presented and properties of these designs are discussed.

103 citations


Cited by
More filters
Journal ArticleDOI
TL;DR: This paper presents a meta-modelling framework for estimating Output from Computer Experiments-Predicting Output from Training Data and Criteria Based Designs for computer Experiments.
Abstract: Many scientific phenomena are now investigated by complex computer models or codes A computer experiment is a number of runs of the code with various inputs A feature of many computer experiments is that the output is deterministic--rerunning the code with the same inputs gives identical observations Often, the codes are computationally expensive to run, and a common objective of an experiment is to fit a cheaper predictor of the output to the data Our approach is to model the deterministic output as the realization of a stochastic process, thereby providing a statistical basis for designing experiments (choosing the inputs) for efficient prediction With this model, estimates of uncertainty of predictions are also available Recent work in this area is reviewed, a number of applications are discussed, and we demonstrate our methodology with an example

6,583 citations

Book
02 Sep 2008
TL;DR: This chapter discusses the design and exploration of a Surrogate-based kriging model, and some of the techniques used in that process, as well as some new approaches to designing models based on the data presented.
Abstract: Preface. About the Authors. Foreword. Prologue. Part I: Fundamentals. 1. Sampling Plans. 1.1 The 'Curse of Dimensionality' and How to Avoid It. 1.2 Physical versus Computational Experiments. 1.3 Designing Preliminary Experiments (Screening). 1.3.1 Estimating the Distribution of Elementary Effects. 1.4 Designing a Sampling Plan. 1.4.1 Stratification. 1.4.2 Latin Squares and Random Latin Hypercubes. 1.4.3 Space-filling Latin Hypercubes. 1.4.4 Space-filling Subsets. 1.5 A Note on Harmonic Responses. 1.6 Some Pointers for Further Reading. References. 2. Constructing a Surrogate. 2.1 The Modelling Process. 2.1.1 Stage One: Preparing the Data and Choosing a Modelling Approach. 2.1.2 Stage Two: Parameter Estimation and Training. 2.1.3 Stage Three: Model Testing. 2.2 Polynomial Models. 2.2.1 Example One: Aerofoil Drag. 2.2.2 Example Two: a Multimodal Testcase. 2.2.3 What About the k -variable Case? 2.3 Radial Basis Function Models. 2.3.1 Fitting Noise-Free Data. 2.3.2 Radial Basis Function Models of Noisy Data. 2.4 Kriging. 2.4.1 Building the Kriging Model. 2.4.2 Kriging Prediction. 2.5 Support Vector Regression. 2.5.1 The Support Vector Predictor. 2.5.2 The Kernel Trick. 2.5.3 Finding the Support Vectors. 2.5.4 Finding . 2.5.5 Choosing C and epsilon. 2.5.6 Computing epsilon : v -SVR 71. 2.6 The Big(ger) Picture. References. 3. Exploring and Exploiting a Surrogate. 3.1 Searching the Surrogate. 3.2 Infill Criteria. 3.2.1 Prediction Based Exploitation. 3.2.2 Error Based Exploration. 3.2.3 Balanced Exploitation and Exploration. 3.2.4 Conditional Likelihood Approaches. 3.2.5 Other Methods. 3.3 Managing a Surrogate Based Optimization Process. 3.3.1 Which Surrogate for What Use? 3.3.2 How Many Sample Plan and Infill Points? 3.3.3 Convergence Criteria. 3.3.4 Search of the Vibration Isolator Geometry Feasibility Using Kriging Goal Seeking. References. Part II: Advanced Concepts. 4. Visualization. 4.1 Matrices of Contour Plots. 4.2 Nested Dimensions. Reference. 5. Constraints. 5.1 Satisfaction of Constraints by Construction. 5.2 Penalty Functions. 5.3 Example Constrained Problem. 5.3.1 Using a Kriging Model of the Constraint Function. 5.3.2 Using a Kriging Model of the Objective Function. 5.4 Expected Improvement Based Approaches. 5.4.1 Expected Improvement With Simple Penalty Function. 5.4.2 Constrained Expected Improvement. 5.5 Missing Data. 5.5.1 Imputing Data for Infeasible Designs. 5.6 Design of a Helical Compression Spring Using Constrained Expected Improvement. 5.7 Summary. References. 6. Infill Criteria With Noisy Data. 6.1 Regressing Kriging. 6.2 Searching the Regression Model. 6.2.1 Re-Interpolation. 6.2.2 Re-Interpolation With Conditional Likelihood Approaches. 6.3 A Note on Matrix Ill-Conditioning. 6.4 Summary. References. 7. Exploiting Gradient Information. 7.1 Obtaining Gradients. 7.1.1 Finite Differencing. 7.1.2 Complex Step Approximation. 7.1.3 Adjoint Methods and Algorithmic Differentiation. 7.2 Gradient-enhanced Modelling. 7.3 Hessian-enhanced Modelling. 7.4 Summary. References. 8. Multi-fidelity Analysis. 8.1 Co-Kriging. 8.2 One-variable Demonstration. 8.3 Choosing X c and X e . 8.4 Summary. References. 9. Multiple Design Objectives. 9.1 Pareto Optimization. 9.2 Multi-objective Expected Improvement. 9.3 Design of the Nowacki Cantilever Beam Using Multi-objective, Constrained Expected Improvement. 9.4 Design of a Helical Compression Spring Using Multi-objective, Constrained Expected Improvement. 9.5 Summary. References. Appendix: Example Problems. A.1 One-Variable Test Function. A.2 Branin Test Function. A.3 Aerofoil Design. A.4 The Nowacki Beam. A.5 Multi-objective, Constrained Optimal Design of a Helical Compression Spring. A.6 Novel Passive Vibration Isolator Feasibility. References. Index.

2,335 citations

Journal ArticleDOI
TL;DR: The multi-objective optimal design of a liquid rocket injector is presented to highlight the state of the art and to help guide future efforts.

2,152 citations

Journal ArticleDOI
TL;DR: This paper surveys their existing application in engineering design, and addresses the dangers of applying traditional statistical techniques to approximate deterministic computer analysis codes, along with recommendations for the appropriate use of statistical approximation techniques in given situations.
Abstract: The use of statistical techniques to build approximations of expensive computer analysis codes pervades much of today’s engineering design. These statistical approximations, or metamodels, are used to replace the actual expensive computer analyses, facilitating multidisciplinary, multiobjective optimization and concept exploration. In this paper, we review several of these techniques, including design of experiments, response surface methodology, Taguchi methods, neural networks, inductive learning and kriging. We survey their existing application in engineering design, and then address the dangers of applying traditional statistical techniques to approximate deterministic computer analysis codes. We conclude with recommendations for the appropriate use of statistical approximation techniques in given situations, and how common pitfalls can be avoided.

1,886 citations