Open AccessBook
Perceptual Computing: Aiding People in Making Subjective Judgments
Jerry M. Mendel,Dongrui Wu +1 more
Reads0
Chats0
TLDR
Perceptual Computing explains how to implement CWW to aid in the important area of making subjective judgments, using a methodology that propagates random and linguistic uncertainties into the subjective judgment in a way that can be modeled and observed by the judgment maker.Abstract:
Explains for the first time how "computing with words" can aid in making subjective judgments Lotfi Zadeh, the father of fuzzy logic, coined the phrase "computing with words" (CWW) to describe a methodology in which the objects of computation are words and propositions drawn from a natural language. Perceptual Computing explains how to implement CWW to aid in the important area of making subjective judgments, using a methodology that leads to an interactive devicea "Perceptual Computer"that propagates random and linguistic uncertainties into the subjective judgment in a way that can be modeled and observed by the judgment maker. This book focuses on the three components of a Perceptual Computerencoder, CWW engines, and decoderand then provides detailed applications for each. It uses interval type-2 fuzzy sets (IT2 FSs) and fuzzy logic as the mathematical vehicle for perceptual computing, because such fuzzy sets can model first-order linguistic uncertainties whereas the usual kind of fuzzy sets cannot. Drawing upon the work on subjective judgments that Jerry Mendel and his students completed over the past decade, Perceptual Computing shows readers how to: Map word-data with its inherent uncertainties into an IT2 FS that captures these uncertainties Use uncertainty measures to quantify linguistic uncertainties Compare IT2 FSs by using similarity and rank Compute the subsethood of one IT2 FS in another such set Aggregate disparate data, ranging from numbers to uniformly weighted intervals to nonuniformly weighted intervals to words Aggregate multiple-fired IF-THEN rules so that the integrity of word IT2 FS models is preserved Free MATLAB-based software is also available online so readers can apply the methodology of perceptual computing immediately, and even try to improve upon it. Perceptual Computing is an important go-to for researchers and students in the fields of artificial intelligence and fuzzy logic, as well as for operations researchers, decision makers, psychologists, computer scientists, and computational intelligence experts.read more
Citations
More filters
Journal ArticleDOI
Enhanced Karnik--Mendel Algorithms
Dongrui Wu,Jerry M. Mendel +1 more
TL;DR: Methods to reduce their computational cost are proposed in this paper and, on average, the enhanced KM algorithms can save about two iterations, which corresponds to more than a 39% reduction in computation time.
Journal ArticleDOI
An overview on the 2-tuple linguistic model for computing with words in decision making: Extensions, applications and challenges
Luis Martínez,Francisco Herrera +1 more
TL;DR: After a decade of extensive and intensive successful use of the 2-tuple linguistic representation model in computing with words for different fields, it is the right moment to overview the model, its extensions, specific methodologies, applications and discuss challenges in the topic.
Journal ArticleDOI
A Historical Account of Types of Fuzzy Sets and Their Relationships
Humberto Bustince,Edurne Barrenechea,Miguel Pagola,Javier Fernández,Zeshui Xu,Benjamin Bedregal,Javier Montero,Hani Hagras,Francisco Herrera,Bernard De Baets +9 more
TL;DR: The definition and basic properties of the different types of fuzzy sets that have appeared up to now in the literature are reviewed and the relationships between them are analyzed.
Journal ArticleDOI
Personalized individual semantics in computing with words for supporting linguistic group decision making. An application on consensus reaching
TL;DR: In this article, a personalized individual semantics (PIS) model is proposed to personalize individual semantics by means of an interval numerical scale and the 2-tuple linguistic model, and a new CW framework is defined, such a CW framework allows us to deal with PIS to facilitate CW keeping the idea that words mean different things to different people.
Journal ArticleDOI
Managing consistency and consensus in group decision making with hesitant fuzzy linguistic preference relations
Zhibin Wu,Jiuping Xu +1 more
TL;DR: This paper develops separate consistency and consensus processes to deal with HFLPR individual rationality and group rationality and introduces a possibility distribution approach and a 2-tuple linguistic model to aid the consistency improvement process in a given H FLPR.
References
More filters
Book
System Identification: Theory for the User
TL;DR: Das Buch behandelt die Systemidentifizierung in dem theoretischen Bereich, der direkte Auswirkungen auf Verstaendnis and praktische Anwendung der verschiedenen Verfahren zur IdentifIZierung hat.
Journal ArticleDOI
A Fuzzy-Set-Theoretic Interpretation of Linguistic Hedges
TL;DR: A basic idea suggested in this paper is that a linguistic hedge such as very, more or less, much, essentially, etc. may be viewed as an operator which acts on the fuzzy set representing the meaning of its operand.
Journal ArticleDOI
Probability-Possibility Transformations, Triangular Fuzzy Sets, and Probabilistic Inequalities
TL;DR: New properties of this transformation of a probability distribution into a possibility distribution are described, by relating it with the well-known probability inequalities of Bienaymé-Chebychev and Camp-Meidel.
Proceedings ArticleDOI
Introduction to type-2 fuzzy logic systems
N.N. Karnik,Jerry M. Mendel +1 more
TL;DR: A robust fuzzy logic system is introduced, one that can handle rule uncertainties and make use of type-2 fuzzy sets for this purpose, and a new operation that is called type-reduction is introduced.
Book
Uncertainty and Information : Foundations of Generalized Information Theory
TL;DR: This chapter discusses the Hartley Measure, a measure of uncertainty based on the Shannon Entropy model, which was developed in the second half of the 1990s to help clarify the role of uncertainty in evidence-based decision-making.