scispace - formally typeset
Search or ask a question
Author

Gordon D. Booth

Bio: Gordon D. Booth is an academic researcher from Iowa State University. The author has contributed to research in topics: Matérn covariance function & Data matrix (multivariate statistics). The author has an hindex of 5, co-authored 5 publications receiving 16816 citations.

Papers
More filters
Journal ArticleDOI
TL;DR: In this paper, the authors compared trichloroacetic acid (TCA) and 1,2,4-triazol-3-ylamine [aminotriazole (BSI) or amitrole (ISO) with HTO as a reference standard in permeation studies.
Abstract: Water penetrated through isolated leaf cuticles of dwarf orange (Citrus mitis Blanco, ‚Calamondin’) as undissociated molecules because both [18O] water (1H218O) and HTO (1H3H16O) permeated at the same rate. HTO penetrated to 3 to 21% of the theoretical equilibrium value (TEV) in an unstirred system within 10 days for astomatous cuticles and 50 to 60% of TEV for stomatous cuticles. The permeability coefficient (k) of HTO through astomatous cuticles at 25°C was 6.8 × 10−7 cm s−1. Two highly water-soluble 14C-labelled compounds, trichloroacetic acid (TCA) and 1,2,4-triazol-3-ylamine [aminotriazole (BSI) or amitrole (ISO)], and two nearly water-insoluble 14C-labelled compounds, 1-naphthyl methylcarbamate (carbaryl) and 2,6-dichloro-4-nitroaniline (dicloran), were compared to HTO as a reference standard in permeation studies. All four organic molecules permeated without decomposing. The relative k values for TCA, aminotriazole, carbaryl, HTO and dicloran were 0.32, 0.47, 0.71, 1.0, and 1.5 respectively. Although this suggested that the permeation of organic molecules may be inversely related to water solubility, this could not be established with certainty due to large variations in the data. The k values were obtained for 12 other organic compounds through a variety of biological and model membranes or were calculated from the literature. Any relationships between k and various molecular characteristics were unclear because a wide variety of cuticle sources and experimental design was used by different investigators working in this area. The calculation of k is considered essential in all permeability studies so that comparisons can be made between laboratories.

19 citations

Journal ArticleDOI
TL;DR: In this article, a sample survey is used to compare two levels of each factor, and a procedure to determine the optimal first phase sample size is given, where these allocations require recourse to programming algorithms, approximate solutions are given.
Abstract: In this paper it is assumed that, using a sample survey, two factors are to be studied, comparisons between the “levels” of the factors are of greatest interest, and there is “interaction” between the factors. Attention is concentrated on situations in which only two levels of each factor are to be compared, but extensions to more complex surveys are discussed. Assuming independent sampling, optimal sample size allocations are obtained. Where these allocations require recourse to programming algorithms, approximate solutions are given. If independent sampling is not feasible, a double sampling procedure is suggested. To indicate how sub-sampling from the first phase sample is to be carried out, a sampling rule (possessing optimal conditional precision properties) is derived. Then, a procedure to determine the optimal first phase sample size is given. Finally, it is demonstrated that this double sampling procedure can be applied to estimation of the (finite) population mean when double sampling wi...

5 citations


Cited by
More filters
Journal ArticleDOI
TL;DR: In this paper, it was shown that a simple FDR controlling procedure for independent test statistics can also control the false discovery rate when test statistics have positive regression dependency on each of the test statistics corresponding to the true null hypotheses.
Abstract: Benjamini and Hochberg suggest that the false discovery rate may be the appropriate error rate to control in many applied multiple testing problems. A simple procedure was given there as an FDR controlling procedure for independent test statistics and was shown to be much more powerful than comparable procedures which control the traditional familywise error rate. We prove that this same procedure also controls the false discovery rate when the test statistics have positive regression dependency on each of the test statistics corresponding to the true null hypotheses. This condition for positive dependency is general enough to cover many problems of practical interest, including the comparisons of many treatments with a single control, multivariate normal test statistics with positive correlation matrix and multivariate $t$. Furthermore, the test statistics may be discrete, and the tested hypotheses composite without posing special difficulties. For all other forms of dependency, a simple conservative modification of the procedure controls the false discovery rate. Thus the range of problems for which a procedure with proven FDR control can be offered is greatly increased.

9,335 citations

Book
01 Jan 2015
TL;DR: This book offers a complete blueprint for structuring projects to achieve rapid completion with high engineering productivity during the research and development phase to ensure that high quality products can be made quickly and at the lowest possible cost.
Abstract: From the Publisher: Phadke was trained in robust design techniques by Genichi Taguchi, the mastermind behind Japanese quality manufacturing technologies and the father of Japanese quality control. Taguchi's approach is currently under consideration to be adopted as a student protocol with the US govrnment. The foreword is written by Taguchi. This book offers a complete blueprint for structuring projects to achieve rapid completion with high engineering productivity during the research and development phase to ensure that high quality products can be made quickly and at the lowest possible cost. Some topics covered are: orthogonol arrays, how to construct orthogonal arrays, computer-aided robutst design techniques, dynamic systems design methods, and more.

3,928 citations

Book
01 Apr 2004
TL;DR: In this paper, the authors present a method for sensitivity analysis of a fish population model using Monte Carlo filtering and variance-based methods, which is based on the Bayesian uncertainty estimation.
Abstract: PREFACE. 1. A WORKED EXAMPLE. 1.1 A simple model. 1.2 Modulus version of the simple model. 1.3 Six--factor version of the simple model. 1.4 The simple model 'by groups'. 1.5 The (less) simple correlated--input model. 1.6 Conclusions. 2. GLOBAL SENSITIVITY ANALYSIS FOR IMPORTANCE ASSESSMENT. 2.1 Examples at a glance. 2.2 What is sensitivity analysis? 2.3 Properties of an ideal sensitivity analysis method. 2.4 Defensible settings for sensitivity analysis. 2.5 Caveats. 3. TEST CASES. 3.1 The jumping man. Applying variance--based methods. 3.2 Handling the risk of a financial portfolio: the problem of hedging. Applying Monte Carlo filtering and variance--based methods. 3.3 A model of fish population dynamics. Applying the method of Morris. 3.4 The Level E model. Radionuclide migration in the geosphere. Applying variance--based methods and Monte Carlo filtering. 3.5 Two spheres. Applying variance based methods in estimation/calibration problems. 3.6 A chemical experiment. Applying variance based methods in estimation/calibration problems. 3.7 An analytical example. Applying the method of Morris. 4. THE SCREENING EXERCISE. 4.1 Introduction. 4.2 The method of Morris. 4.3 Implementing the method. 4.4 Putting the method to work: an analytical example. 4.5 Putting the method to work: sensitivity analysis of a fish population model. 4.6 Conclusions. 5. METHODS BASED ON DECOMPOSING THE VARIANCE OF THE OUTPUT. 5.1 The settings. 5.2 Factors Prioritisation Setting. 5.3 First--order effects and interactions. 5.4 Application of Si to Setting 'Factors Prioritisation'. 5.5 More on variance decompositions. 5.6 Factors Fixing (FF) Setting. 5.7 Variance Cutting (VC) Setting. 5.8 Properties of the variance based methods. 5.9 How to compute the sensitivity indices: the case of orthogonal input. 5.9.1 A digression on the Fourier Amplitude Sensitivity Test (FAST). 5.10 How to compute the sensitivity indices: the case of non--orthogonal input. 5.11 Putting the method to work: the Level E model. 5.11.1 Case of orthogonal input factors. 5.11.2 Case of correlated input factors. 5.12 Putting the method to work: the bungee jumping model. 5.13 Caveats. 6. SENSITIVITY ANALYSIS IN DIAGNOSTIC MODELLING: MONTE CARLO FILTERING AND REGIONALISED SENSITIVITY ANALYSIS, BAYESIAN UNCERTAINTY ESTIMATION AND GLOBAL SENSITIVITY ANALYSIS. 6.1 Model calibration and Factors Mapping Setting. 6.2 Monte Carlo filtering and regionalised sensitivity analysis. 6.2.1 Caveats. 6.3 Putting MC filtering and RSA to work: the problem of hedging a financial portfolio. 6.4 Putting MC filtering and RSA to work: the Level E test case. 6.5 Bayesian uncertainty estimation and global sensitivity analysis. 6.5.1 Bayesian uncertainty estimation. 6.5.2 The GLUE case. 6.5.3 Using global sensitivity analysis in the Bayesian uncertainty estimation. 6.5.4 Implementation of the method. 6.6 Putting Bayesian analysis and global SA to work: two spheres. 6.7 Putting Bayesian analysis and global SA to work: a chemical experiment. 6.7.1 Bayesian uncertainty analysis (GLUE case). 6.7.2 Global sensitivity analysis. 6.7.3 Correlation analysis. 6.7.4 Further analysis by varying temperature in the data set: fewer interactions in the model. 6.8 Caveats. 7. HOW TO USE SIMLAB. 7.1 Introduction. 7.2 How to obtain and install SIMLAB. 7.3 SIMLAB main panel. 7.4 Sample generation. 7.4.1 FAST. 7.4.2 Fixed sampling. 7.4.3 Latin hypercube sampling (LHS). 7.4.4 The method of Morris. 7.4.5 Quasi--Random LpTau. 7.4.6 Random. 7.4.7 Replicated Latin Hypercube (r--LHS). 7.4.8 The method of Sobol'. 7.4.9 How to induce dependencies in the input factors. 7.5 How to execute models. 7.6 Sensitivity analysis. 8. FAMOUS QUOTES: SENSITIVITY ANALYSIS IN THE SCIENTIFIC DISCOURSE. REFERENCES. INDEX.

2,297 citations

Book
01 Jan 1980
TL;DR: The writer really shows how the simple words can maximize how the impression of this book is uttered directly for the readers.
Abstract: Every word to utter from the writer involves the element of this life. The writer really shows how the simple words can maximize how the impression of this book is uttered directly for the readers. Even you have known about the content of randomization tests so much, you can easily do it for your better connection. In delivering the presence of the book concept, you can find out the boo site here.

1,999 citations

Journal ArticleDOI
TL;DR: In this paper, the Fourier amplitude sensitivity test (FAST) has been extended to include all the interaction terms involving a factor and the main effect of the factor's main effect.
Abstract: A new method for sensitivity analysis (SA) of model output is introduced. It is based on the Fourier amplitude sensitivity test (FAST) and allows the computation of the total contribution of each input factor to the output's variance. The term “total” here means that the factor's main effect, as well as all the interaction terms involving that factor, are included. Although computationally different, the very same measure of sensitivity is offered by the indices of Sobol'. The main advantages of the extended FAST are its robustness, especially at low sample size, and its computational efficiency. The computational aspects of the extended FAST are described. These include (1) the definition of new sets of parametric equations for the search-curve exploring the input space, (2) the selection of frequencies for the parametric equations, and (3) the procedure adopted to estimate the total contributions. We also address the limitations of other global SA methods and suggest that the total-effect indices are id...

1,652 citations