scispace - formally typeset
Search or ask a question
Author

William W. Cooper

Bio: William W. Cooper is an academic researcher from University of Texas at Austin. The author has contributed to research in topics: Data envelopment analysis & Linear programming. The author has an hindex of 79, co-authored 254 publications receiving 76641 citations. Previous affiliations of William W. Cooper include Harvard University & Carnegie Mellon University.


Papers
More filters
Journal ArticleDOI
TL;DR: Return to scale (RTS) in data envelopment analysis (DEA) for each of the presently available types of models is discussed for each type of model and relations between the results for the different models are established.

352 citations

Journal ArticleDOI
TL;DR: In this paper, DEA is extended to deal with zero inputs and outputs and zero virtual multipliers (shadow prices) by partitioning DMUs into six classes via primal and dual representation theorems by means of which restrictions to positive observed values for all inputs and output are eliminated along with positivity conditions imposed on the variables which are usually accomplished by recourse to nonarchimedian concepts.
Abstract: DEA (Data Envelopment Analysis) attempts to identify sources and estimate amounts of inefficiencies contained in the outputs and inputs generated by managed entities called DMUs (Decision Making Units). Explicit formulation of underlying functional relations with specified parametric forms relating inputs to outputs is not required. An overall (scalar) measure of efficiency is obtained for each DMU from the observed magnitudes of its multiple inputs and outputs without requiring use of a priori weights or relative value assumptions and, in addition, sources and amounts of inefficiency are estimated for each input and each output for every DMU. Earlier theory is extended so that DEA can deal with zero inputs and outputs and zero virtual multipliers (shadow prices). This is accomplished by partitioning DMUs into six classes via primal and dual representation theorems by means of which restrictions to positive observed values for all inputs and outputs are eliminated along with positivity conditions imposed on the variables which are usually accomplished by recourse to nonarchimedian concepts. Three of the six classes are scale inefficient and two of the three scale efficient classes are also technically (zero waste) efficient.

326 citations

Journal ArticleDOI
TL;DR: In this article, the authors develop theory and algorithms for a multiplicative data envelope analysis (DEA) model employing virtual outputs and inputs as does the CCR ratio method for efficiency analysis.
Abstract: This paper develops theory and algorithms for a “multiplicative” Data Envelopment Analysis (DEA) model employing virtual outputs and inputs as does the CCR ratio method for efficiency analysis. The frontier production function results here are of piecewise log-linear rather than piecewise linear form.

254 citations

Journal ArticleDOI
TL;DR: The newer analytical methods described in this paper make it possible to determineranges within which all data may be varied for any DMU before a reclassification from efficient to inefficient status occurs.
Abstract: This paper surveys recently developed analytical methods for studying the sensitivity of DEA results to variations in the data. The focus is on the stability of classification of DMUs (Decision Making Units) into efficient and inefficient performers. Early work on this topic concentrated on developing solution methods and algorithms for conducting such analyses after it was noted that standard approaches for conducting sensitivity analyses in linear programming could not be used in DEA. However, some of the recent work we cover has bypassed the need for such algorithms. Evolving from early work that was confined to studying data variations in only one input or output for only one DMU at a time, the newer methods described in this paper make it possible to determine ranges within which all data may be varied for any DMU before a reclassification from efficient to inefficient status (or vice versa) occurs. Other coverage involves recent extensions which include methods for determining ranges of data variation that can be allowed when all data are varied simultaneously for all DMUs. An initial section delimits the topics to be covered. A final section suggests topics for further research.

252 citations


Cited by
More filters
Journal ArticleDOI
TL;DR: A nonlinear (nonconvex) programming model provides a new definition of efficiency for use in evaluating activities of not-for-profit entities participating in public programs and methods for objectively determining weights by reference to the observational data for the multiple outputs and multiple inputs that characterize such programs.

25,433 citations

Journal ArticleDOI
TL;DR: The CCR ratio form introduced by Charnes, Cooper and Rhodes, as part of their Data Envelopment Analysis approach, comprehends both technical and scale inefficiencies via the optimal value of the ratio form, as obtained directly from the data without requiring a priori specification of weights and/or explicit delineation of assumed functional forms of relations between inputs and outputs as mentioned in this paper.
Abstract: In management contexts, mathematical programming is usually used to evaluate a collection of possible alternative courses of action en route to selecting one which is best. In this capacity, mathematical programming serves as a planning aid to management. Data Envelopment Analysis reverses this role and employs mathematical programming to obtain ex post facto evaluations of the relative efficiency of management accomplishments, however they may have been planned or executed. Mathematical programming is thereby extended for use as a tool for control and evaluation of past accomplishments as well as a tool to aid in planning future activities. The CCR ratio form introduced by Charnes, Cooper and Rhodes, as part of their Data Envelopment Analysis approach, comprehends both technical and scale inefficiencies via the optimal value of the ratio form, as obtained directly from the data without requiring a priori specification of weights and/or explicit delineation of assumed functional forms of relations between inputs and outputs. A separation into technical and scale efficiencies is accomplished by the methods developed in this paper without altering the latter conditions for use of DEA directly on observational data. Technical inefficiencies are identified with failures to achieve best possible output levels and/or usage of excessive amounts of inputs. Methods for identifying and correcting the magnitudes of these inefficiencies, as supplied in prior work, are illustrated. In the present paper, a new separate variable is introduced which makes it possible to determine whether operations were conducted in regions of increasing, constant or decreasing returns to scale in multiple input and multiple output situations. The results are discussed and related not only to classical single output economics but also to more modern versions of economics which are identified with "contestable market theories."

14,941 citations

Book
31 Jul 1985
TL;DR: The book updates the research agenda with chapters on possibility theory, fuzzy logic and approximate reasoning, expert systems, fuzzy control, fuzzy data analysis, decision making and fuzzy set models in operations research.
Abstract: Fuzzy Set Theory - And Its Applications, Third Edition is a textbook for courses in fuzzy set theory. It can also be used as an introduction to the subject. The character of a textbook is balanced with the dynamic nature of the research in the field by including many useful references to develop a deeper understanding among interested readers. The book updates the research agenda (which has witnessed profound and startling advances since its inception some 30 years ago) with chapters on possibility theory, fuzzy logic and approximate reasoning, expert systems, fuzzy control, fuzzy data analysis, decision making and fuzzy set models in operations research. All chapters have been updated. Exercises are included.

7,877 citations

Journal ArticleDOI
01 May 1981
TL;DR: This chapter discusses Detecting Influential Observations and Outliers, a method for assessing Collinearity, and its applications in medicine and science.
Abstract: 1. Introduction and Overview. 2. Detecting Influential Observations and Outliers. 3. Detecting and Assessing Collinearity. 4. Applications and Remedies. 5. Research Issues and Directions for Extensions. Bibliography. Author Index. Subject Index.

4,948 citations

Book
30 Nov 1999
TL;DR: In this article, the basic CCR model and DEA models with restricted multipliers are discussed. But they do not consider the effect of non-discretionary and categorical variables.
Abstract: List of Tables. List of Figures. Preface. 1. General Discussion. 2. The Basic CCR Model. 3. The CCR Model and Production Correspondence. 4. Alternative DEA Models. 5. Returns to Scale. 6. Models with Restricted Multipliers. 7. Discretionary, Non-Discretionary and Categorical Variables. 8. Allocation Models. 9. Data Variations. Appendices. Index.

4,395 citations