On Advanced Computing with Words Using the

0 downloads 0 Views 11MB Size Report
Jul 27, 2013 - set solutions are decoded into natural language words using Jaccard's ... Advanced Computing with Words, Generalized Extension Principle, ...
1

On Advanced Computing with Words Using the Generalized Extension Principle for Type-1 Fuzzy Sets Mohammad Reza Rajati and Jerry M. Mendel

Signal and Image Processing Institute Ming Hsieh Department of Electrical Engineering University of Southern California Los Angeles, CA 90089 Email: [email protected], [email protected] Abstract In this paper, we propose and demonstrate an effective methodology for implementing the Generalized Extension Principle to solve Advanced Computing with Words (ACWW) problems. Such problems involve implicit assignments of linguistic truth, probability, and possibility. To begin, we establish vocabularies of the words involved in the problems, and then collect data from subjects about the words after which fuzzy set models for the words are obtained by using the Interval Approach (IA) or the Enhanced Interval Approach (EIA). Next, the solutions of the ACWW problems, which involve the fuzzy set models of the words, are formulated using the Generalized Extension Principle. Because the solutions to those problems involve complicated functional optimization problems that cannot be solved analytically, we then develop a numerical method for their solution. Finally, the resulting fuzzy set solutions are decoded into natural language words using Jaccard’s similarity measure. We explain how ACWW problems can solve some potential prototype Engineering problems and connect the methodology of this paper with Perceptual Computing. Index Terms Advanced Computing with Words, Generalized Extension Principle, Perceptual Computing, Type-1 Fuzzy Sets, Validation, Zadeh’s Challenge Problems

July 27, 2013

DRAFT

2

I. I NTRODUCTION Computing with Words (CWW or CW) is a methodology of computation whose objects are words rather than numbers [19], [25], [47], [49]. Words are drawn from natural languages, and are modeled by fuzzy sets. Basic Computing with Words deals with simple assignments of attributes through IF-THEN rules. According to Zadeh [49], Advanced Computing with Words (ACWW) involves problems in which the carriers of information are numbers, intervals, and words. Assignment of attributes may be implicit, and one generally deals with assignments of linguistic truth, probability, and possibility constraints through complicated natural language statements. Moreover, world knowledge is necessary for solving ACWW problems [20]. Modeling words is subsequently an important task for solving ACWW problems. Mendel [19] argues that since words mean different things to different people, a first order uncertainty model for a word should be an interval type-2 fuzzy set (IT2 FS). Interestingly, Zadeh [49] anticipates that in the future, fuzzy sets of higher type will play a central role in ACWW. Therefore, it is plausible to examine the solutions to ACWW problems when words are modeled by IT2 FSs. In this paper, however, we use IT2 FS models of words only to establish T1 FS models of them, for reasons that are given in Section II-A. CWW has been applied successfully to hierarchical and distributed decision making [8], [22], Perceptual Reasoning and Perceptual Computing [21], [23], [40], and decision support [9], [15]. There have been extensive attempts to implement the approach in more realistic settings, e.g. [12], [24], [28], and attempts to formalize the paradigm of CWW [3], [4], [13]. Despite the extensive literature on CWW, there have only been a few attempts to deal with ACWW problems to the best of our knowledge [27], [29], [30], [33], [34]. Zadeh has introduced a set of challenge problems for ACWW in his recent works, and has proposed solutions to them [46], [49]. The solutions to many of the problems utilize the Generalized Extension Principle (GEP), which results in complicated optimization problems. In this paper, we establish a methodology to carry out the computations of the GEP for solving ACWW problems. We are focusing on solving one of Zadeh’s famous ACWW problems, but later demonstrate that some of his other famous ACWW problems can also be solved in exactly the same manner. Historically, Zadeh’s ACWW challenge problems have been formulated in terms of everyday

July 27, 2013

DRAFT

3

reasoning examples involving attributes like height, age, distance, etc (e.g., Probably John is tall. What is the probability that John is short?1 ) In this paper, we also demonstrate how ACWW can address more realistic problems dealing with subjective judgments. It has been demonstrated in [6], [7], [10], [26], [37], [38], [42], [53] that subjective, fuzzy, and linguistic probabilities can be used to model the subjective assessments of safety, reliability, and risk. Some examples that address product safety, network security, and network trust are: It is somewhat improbable that cars of model X are unsafe. What is the probability that they are safe? It is very probable that the network provided by Company Y is highly secure. What is the probability that it is somewhat insecure? Probably the online auction of website Z is pretty trustworthy. What is the probability that it is extremely trustworthy? It is somewhat likely that the risk associated with investment in real estate is risky. What is the average risk associated with investment in real estate? In all of these examples the numeric probability of a fuzzy event (e.g., unsafe, highly secure) is implicitly constrained by a linguistic probability (e.g., somewhat probable, somewhat likely), and therefore they are ACWW problems. These examples suggest that the methodologies for solving Zadeh’s ACWW problems can also be applied to solve more realistic problems. We show how to do this in Section V. The rest of this paper is organized as follows: In Section II, we describe a famous ACWW problem, its variations and their solutions that are obtained by using the GEP; in Section III, we first establish fuzzy set models of the words that are involved in the ACWW problems of Section II, and then implement numerical solutions to the problems when type-1 fuzzy set word models are used; in Section IV, we provide discussions about how to validate solutions of ACWW problems that are solved in this paper; in Section V, we solve an Engineering ACWW problem, using the methodology of our earlier sections; in Section VI, we show how some of Zadeh’s other ACWW problems can be solved using the same methodology as described in Section III; in Section VII, we present a high level discussion describing how ACWW problems can be formulated by investigating the linguistic description of the problems; in Section VIII, we 1

This question can also be stated as: What is the probability that (the height of) John is short?

July 27, 2013

DRAFT

4

investigate the relationship between the methodology of this paper with Perceptual Computing; and, finally in Section IX, we present some conclusions, as well as some directions for future research. II. P ROBLEM D ESCRIPTION A. The PJS and PJW Problems Among Zadeh’s many ACWW problems is the following famous Probability that John is short (PJS) problem: Probably John is tall. What is the probability that John is short? This problem involves a linguistic probability (probably) and an implicit assignment of that linguistic probability to the probability that John is tall. The probability that “John is tall”, PT all , is calculated as2 [43]: Z

b

pH (h)µT all (h)dh

PT all =

(1)

a

in which a and b are the minimum and maximum possible heights of men and pH is the probability distribution function of heights, where: Z b pH (h)dh = 1

(2)

a

The probability of the fuzzy event “Short” is calculated as: Z b PShort = pH (h)µShort (h)dh

(3)

a

To derive the soft constraint imposed on PShort by the fact that PT all is constrained by “Probably,” one needs to use the framework of the GEP, which is an important tool for propagation of possibilistic constraints, and was originally introduced in [45]. Assume that f (·) and g(·) are real functions: f, g : U1 × U2 × · · · × Un −→ V 2

(4)

We assume that “Probably John is tall” is equivalent to “It is probable that John is tall”.

July 27, 2013

DRAFT

5

Moreover, assume that: f (X1 , X2 , · · · , Xn ) is A

g(X1 , X2 , · · · , Xn ) is B where A and B are T1 FSs. Then A induces B as follows:

µB (v) =

  

µA (f (u1 , u2 , · · · , un ))

sup

∃v = g(u1 , u2 , · · · , un )

u1 ,u2 ,··· ,un |v=g(u1 ,u2 ,··· ,un )

(5) 6 ∃v = g(u1 , u2 , · · · , un )

  0

The GEP basically extends the function g(f −1 (·)) : V → V to T1 FSs, where f −1 is the pre-image of the function f (·). In the PJS problem, f = PT all , and: f : X[a,b] −→ R

(6)

where X[a,b] is the space of probability distribution functions on [a, b]. Also, g = PShort , and: g : X[a,b] −→ R

(7)

The GEP then implies that the soft constraint on the probability that “John is short” is:

µPShort (v) =     

sup

µP robable

R pH |v= ab pH (h)µShort (h)dh Rb a pH (h)dh=1

    0

R

b a

 Rb pH (h)µT all (h)dh ∃pH ∈ X[a,b] s.t. v = a pH (h)µShort (h)dh 6 ∃pH ∈ X[a,b] s.t. v =

Rb a

pH (h)µShort (h)dh (8)

Note that (8) cannot be solved by using the α-cut decomposition theorem, since B(α) = g(f −1 (A(α))) (f = PT all and g = PShort ), but the relation f −1 (·) cannot be derived explicitly. July 27, 2013

DRAFT

6

In this paper, we also study variations of the PJS problem, that include: Probably John is tall. What is the probability that John is W ?. This is called the “PJW problem”. In this problem W represents any height word. Similar to (8), the GEP yields the soft constraint PW on the probability that “John is W ”: µPW (v) =    sup  p |v=R b p (h)µ H

    0

R ab a

H

µP robable

R

b a

pH (h)µT all (h)dh



∃pH ∈ X[a,b] s.t. v =

Rb

6 ∃pH ∈ X[a,b] s.t. v =

Rb

a

pH (h)µW (h)dh

W (h)dh

pH (h)dh=1 a

pH (h)µW (h)dh (9)

Note that (8) and (9) are difficult functional optimizations that have to be carried out over X[a,b] , the space of probability distributions over [a, b]. In this paper, a methodology for performing the optimizations is offered. Although our solutions to the PJW problems use T1 FSs, those FSs are obtained from IT2 FS models for reasons that are given in Section III. The readers may be wondering why we have not presented our solutions to the PJW problem entirely within the framework of IT2 FSs since we are strong advocates for using such FSs in CWW problems. We have not done this for the following reasons: 1. To-date, although Zadeh formulated the PJS problem and gave the structure of its solution using the GEP for T1 FSs, no numerical or linguistic solutions were provided even for T1 FSs. We believe that we are the first to do so in this paper. 2. All of the basic concepts for solving the PJW problem are more easily explained and understood using T1 FSs. 3. To-date, how to state and solve the GEP for IT2 FSs is an open research topic, one that we are researching and will report on in the future. 4. We strongly believe that an IT2 solution to a problem needs to be compared with T1 solutions, so that one can: (a) see if an IT2 solution is needed, and (b) observe similarities and differences between the T1 and IT2 solutions. The results that we have presented in this paper will serve as a useful baseline for such comparisons.

July 27, 2013

DRAFT

7

B. An Interesting Special Case When the probability distribution of the height of the population of which John is a member is known, the PJW problem becomes a special problem. Assume that the probability distribution of heights is known to be p∗H . Then the answer to the question “What is the probability that John is W ” is a numerical value: ∗ PW

Z =

b

p∗H (h)µW (h)dh

(10)

a

and there is no need for the information “Probably John is tall.” This means that this piece of information in the PJW and PJS problems compensates for the imprecision of knowledge about the probability distribution of heights. However, if the problem is to be solved given that “ProbRb ∗ ably John is tall,” we can still use (9) to derive a MF for PW . Since PT∗all = a p∗H (h)µT all (h)dh is also a numerical value, (9) reduces to: µPW∗ (v) =    µP robable R b p∗ (h)µT all (h)dh a H  0

v=

Rb a

p∗H (h)µW (h)dh

(11)

otherwise

∗ Equation (11) says that µPW∗ is a fuzzy singleton at PW . Obviously, the membership value of Rb the fuzzy singleton depends on the compatibility of Probable with a p∗H (h)µT all (h)dh (which

is exactly µP robable (PT∗all )). The less this compatibility is, then the smaller is the membership value for this fuzzy singleton, which can be interpreted as less confidence in the solution. In the  R b ∗ extreme case when µP robable a pH (h)µT all (h)dh = 0, because µPW∗ (v) = 0 ∀v, we obtain the ∗ empty set as PW , which reflects a total incompatibility of a human’s world knowledge (assumed

for the problem) with reality. It is worth noting that the empty set can be interpreted as the word Undefined [50]. III. I MPLEMENTATION OF T HE S OLUTION TO T HE PJW P ROBLEM In this section, we implement the solution to the PJW problem which includes the PJS problem. We model the words involved in the problem, and use a methodology to approximate the solution to the optimization problem of (9).

July 27, 2013

DRAFT

8

A. Modeling Words To begin, we established the following vocabularies of linguistic heights and linguistic probabilities: Heights ={Very short, Short, Moderately short, Medium, Moderately tall, Tall, Very tall} and, Probabilities ={Extremely improbable, Very improbable, Improbable, Somewhat improbable, Tossup, Somewhat probable, Probable, Very probable, Extremely probable}. Next, we modeled all of these words as FSs. Recall, that there are at least two types of uncertainty associated with a word [18], [19]: intrauncertainty, the uncertainty an individual has about the meaning of a word,3 and inter-uncertainty, the uncertainty a group of people have about the meaning of a word. In other words, words mean different things to different people, and this fact calls for (at least) using IT2 FSs as models of words [18], [19]. In order to synthesize IT2 FS models of words, we began by collecting data from a group of subjects and then used either the Interval Approach (IA) [14] or Enhanced Interval Approach (EIA) [41]. We collected data from 48 subjects using the Amazon Mechanical Turk website [1] for the above vocabularies of linguistic heights and linguistic probabilities. We used the EIA [41] to obtain IT2 FS models of the words from that data. The IT2 FS footprints of uncertainties (FOUs) have nine parameters (see Fig. 1); (q, r, s, t) determine the upper membership function (UMF) and (q1 , r1 , s1 , t1 , h) determine the lower membership function (LMF), where h is the height of the lower membership function.

1

h

q

Fig. 1.

q

1

r

r1

s1

s

t

1

t

FOU of an IT2 FS described by nine parameters.

The vocabulary of linguistic heights modeled by IT2 FSs is depicted in Fig. 2. We assumed that 3

This is related to the uncertainty associated with unsharpness of classes [48].

July 27, 2013

DRAFT

9

the minimum possible height is a = 139.7 cm and the maximum possible height is b = 221 cm.4 The parameters of the FOUs of the linguistic height words are given in Table I.

Very short

Short

Moderately short

1

1

1

0.5

0.5

0.5

0

150

200

0

Medium

150

0

200

1

1

0.5

0.5

0.5

150

200

0

150

200

200

Tall

1

0

150

Moderately tall

0

150

200

Very tall 1 0.5 0

150

Fig. 2.

200

Vocabulary of IT2 FSs representing linguistic heights.

TABLE I M EMBERSHIP FUNCTION PARAMETERS OF THE HEIGHT WORDS DEPICTED IN F IG . 2

Very short Short Moderately short Medium Moderately tall Tall Very tall

UMF parameters (139.70, 139.70, 140.92, 153.17) (141.62, 150.00, 158.75, 168.11) (149.24, 157.48, 162.56, 170.80) (154.32, 163.83, 171.45, 178.42) (166.89, 175.00, 181.61, 190.59) (174.12, 185.00, 193.04, 204.14) (176.58, 207.70, 221.00, 221.00)

LMF parameters (139.70,139.70, 140.28, 143.95, 1) (151.35, 153.85, 153.85, 156.04, 0.45) (159.79, 160.84, 160.84 161.21, 0.52) (166.06, 168.53, 168.53, 169.83, 0.46) (173.96, 177.91, 177.91, 181.04, 0.59) (185.86, 188.99, 188.99, 192.07, 0.44) (193.67, 219.07, 221, 221, 1)

To make sure that the vocabulary of height words provides an appropriate partitioning of the space of heights, we calculated the pairwise Jaccard similarities between the height words. e and B e is calculated as: Jaccard similarity [5], [39] between IT2 FSs A R R min(µAe(u), µBe (u))du + U min(µAe(u), µBe (u))du U e B) e =R R (12) sJ (A, max(µ (u), µ (u))du + max(µAe(u), µBe (u))du e e A B U U 4

This range corresponds to [40 700 , 60 900 ], and is inspired by the range [140 cm, 220 cm] that Zadeh uses in [51].

July 27, 2013

DRAFT

10

TABLE II PAIRWISE SIMILARITIES BETWEEN THE HEIGHT WORDS DEPICTED IN F IG . 2 Very short 1.000 0.129 0.016 0 0 0 0

Very short Short Moderately short Medium Moderately tall Tall Very tall

Moderately Medium short 0.016 0 0.451 0.166 1.000 0.336 0.336 1.000 0.015 0.149 0 0.014 0 0.001

Short 0.129 1.000 0.451 0.166 0.001 0 0

Moderately tall 0 0.001 0.015 0.149 1.000 0.222 0.0422

Tall 0 0 0 0.014 0.222 1.000 0.165

Very tall 0 0 0 0.001 0.042 0.165 1.000

Pairwise similarities between the height words are shown in Table II. Observe that the words have pairwise similarities that are less than 0.5, indicating that this vocabulary provides a good partitioning of the universe of discourse. Similar information for the vocabulary of linguistic probabilities [32] is given in Fig. 3 and Tables III and IV. In Table IV, observe that the probability words also have pairwise similarities less than 0.5, indicating that this vocabulary provides a good partitioning of the universe of discourse. Extremely Improbable Very Improbable

Improbable

1

1

1

0.5

0.5

0.5

0

0

0.5

0

1

0

Somewhat Improbable

0.5

1

Tossup

0

1

1

1

0.5

0.5

0.5

0

0

0.5

1

0

0

Probable

0.5

1

Very Probable

0

1

1

0.5

0.5

0.5

0

Fig. 3.

0.5

1

0

0

0.5

0.5

1

0

0.5

1

Extremely Probable

1

0

0

Somewhat Probable

1

0

0

0.5

1

Vocabulary of IT2 FSs representing linguistic probabilities.

Because Zadeh’s solutions [49] involve type-1 FS models of words, we chose those models from our interval type-2 FS models by using two kinds of embedded T1 FSs. e over U is a T1 FS over U for Definition 1. An embedded type-1 fuzzy set Ae of an IT2 FS A July 27, 2013

DRAFT

11

TABLE III M EMBERSHIP FUNCTION PARAMETERS OF THE PROBABILITY WORDS DEPICTED IN F IG . 3

Extremely improbable Very improbable Improbable Somewhat improbable Tossup Somewhat probable Probable Very probable Extremely probable

UMF parameters (0, 0 , 0.0183 , 0.1316) (0.0293, 0.1000, 0.1250, 0.1707) (0.0586, 0.1750, 0.2500, 0.3414) (0.0982, 0.2500, 0.3000, 0.4518) (0.3586 , 0.5000 , 0.5500 , 0.6414 ) (0.4793, 0.5500, 0.6000, 0.6707) (0.5086, 0.6500, 0.7000, 0.7914) (0.7189, 0.8250, 0.9000, 0.9811) (0.8684, 0.9772, 1.0000, 1.0000)

LMF parameters (0, 0, 0.0046, 0.0627, 1.0000) (0.0896, 0.1167, 0.1167, 0.1604, 0.7643) (0.1896, 0.2200, 0.2200, 0.2604, 0.5757) (0.2293, 0.2750, 0.2750, 0.3207, 0.6464) (0.4896, 0.5083, 0.5083, 0.5141, 0.4107) (0.5293, 0.5750, 0.5750, 0.6207,0.6464) (0.6293, 0.6750, 0.6750, 0.7207, 0.6464) (0.8293, 0.8700, 0.8700, 0.9207, 0.5757 ) (0.9405, 0.9954, 1.0000, 1.0000, 1.0000 )

TABLE IV PAIRWISE SIMILARITIES BETWEEN THE PROBABILITY WORDS DEPICTED IN F IG . 3

Extremely improbable Very improbable Improbable Somewhat improbable Tossup Somewhat probable Probable Very probable Extremely probable

Very Extremely Somewhat Tossup Improbable improbable improbable improbable 1.0000 0.1502 0.0394 0.0063 0 0.1502 1.0000 0.1432 0.0405 0 0.0394 0.1432 1.0000 0.4091 0 0.0063 0.0405 0.4091 1.0000 0.0382 0 0 0 0.0382 1.0000 0 0 0 0 0.3369 0 0 0 0 0.1150 0 0 0 0 0 0 0 0 0 0

Somewhat probable 0 0 0 0 0.3369 1.0000 0.2179 0 0

Probable 0 0 0 0 0.1150 0.2179 1.0000 0.0353 0

Very probable 0 0 0 0 0 0 0.0353 1.0000 0.1242

Extremely probable 0 0 0 0 0 0 0 0.1242 1.0000

which ∀t ∈ U, µAe(t) ≤ µAe (t) ≤ µAe(t). e over U is an embedded Definition 2. A normal embedded type-1 fuzzy set Ane of an IT2 FS A e that is normal. i.e. supt µAne (t) = 1. An UMF is a normal embedded T1 FS. T1 FS of A Our first T1 FS is the UMF from each of the IT2 FS FOUs (see Figs. 2 and 3, and Tables I and III). Our second T1 FS is obtained from the Fig. 1 and Fig. 3 FOU parameters, as shown in Fig. 4, and is also a normal embedded T1 FS, sort of an average T1 FS. In the sequel we refer to this T1 FS as a “normal embedded T1 FS.” The vocabulary of height words modeled by normal embedded T1 FSs is shown in Fig. 5 and their parameters are given in Table V. We computed the pairwise similarities for T1 FS models and observed that they were less than 0.5, indicating that this vocabulary provides a good partitioning of the universe of discourse for heights. Comparable MFs of normal embedded T1 FS models of our nine probability words are given in Fig. 6, and Table VI, respectively. We also observed that our probability words have pairwise similarities less than 0.5, indicating that this vocabulary provides a good partitioning of the July 27, 2013

DRAFT

12

Left shoulder

Interior

Right shoulder

1

Fig. 4.

Normal embedded type-1 fuzzy sets (dashed) for left-shoulder, interior, and right-shoulder FOUs.

Very short

Short

Medium

1

1

1

0.8

0.8

0.8

0.8

0.6

0.6

0.6

0.6

0.4

0.4

0.4

0.4

0.2

0.2

0.2

0.2

0

150

200

0

150

Moderately tall

Fig. 5.

Moderately short

1

200

0

Tall 1

1

0.8

0.8

0.8

0.6

0.6

0.6

0.4

0.4

0.4

0.2

0.2

0.2

0

0

200

150

200

200

0

150

200

Very tall

1

150

150

0

150

200

Vocabulary of linguistic heights modeled by normal embedded type-1 fuzzy sets.

TABLE V M EMBERSHIP FUNCTION PARAMETERS OF THE HEIGHT WORDS DEPICTED IN F IG . 5

Very short Short Moderately short Medium Moderately tall Tall Very tall

July 27, 2013

Trapezoidal membership function parameters (139.7000, 139.7000, 140.5964, 148.5572) (141.6237, 150.0000, 156.3023, 162.0711) (149.2437, 157.4800, 161.6994, 166.0017) (154.3237, 163.8300, 169.9909, 174.1224) (166.8934, 175.0000, 179.7609, 185.8129) (174.1176 185.0000 191.0140 198.1066) (185.1236 213.3897 221.0000 221.0000)

DRAFT

13

universe of discourse for probability.

Extremely Improbable Very Improbable 1

1

0.5

0.5

0.5

0

0

0.5

0

1

0

Somewhat Improbable

0.5

1

Tossup

0

1

1

0.5

0.5

0.5

0

0.5

1

0

0

Probable

0.5

1

Very Probable

0

1

1

0.5

0.5

0.5

0

0.5

1

0

0

0.5

0.5

1

0

0.5

1

Extremely Probable

1

0

0

Somewhat Probable

1

0

Fig. 6.

Improbable

1

1

0

0

0.5

1

Vocabulary of normal embedded type-1 fuzzy set models of linguistic probabilities.

TABLE VI M EMBERSHIP FUNCTION PARAMETERS OF THE PROBABILITY WORDS IN F IG . 6

Extremely improbable Very improbable Improbable Somewhat improbable Tossup Somewhat probable Probable Very probable Extremely probable

Trapezoidal membership function parameters (0, 0, 0.0114, 0.0972) (0.0293, 0.1000, 0.1208, 0.1655) (0.0586, 0.1750, 0.2350, 0.3009) (0.0982, 0.2500, 0.2875, 0.3862) (0.3586, 0.5000, 0.5292, 0.5778) (0.4793, 0.5500, 0.5875, 0.6457) (0.5086, 0.6500, 0.6875, 0.7561) (0.7189, 0.8250, 0.8850, 0.9509) (0.9044, 0.9863, 1.0000, 1.0000)

B. Approximate Solution to The Optimization Problem Next, we solve the optimization problem of (9). It is a functional optimization problem, and cannot be solved analytically. Instead, our approach is to: 1. Choose the family (families) of probability distributions pertinent to the problem. 2. Choose the ranges of the parameters of the families of probability distributions. 3. Discretize the ranges of parameters of the probability distributions.

July 27, 2013

DRAFT

14

4. Construct a pool of probability distributions having all possible combinations of parameters, and for all of its members: 4.1. Choose a specific pH from the pool (again, note that

Rb a

pH (h)dh = 1, because

pH (h) is a probability distribution function on [a, b]). Rb 4.2. Compute v = a pH (h)µW (h)dh. Rb 4.3. Compute a pH (h)µT all (h)dh.  R b 4.4. Compute γ(v) = µP robable a pH (h)µT all (h)dh . 5. Construct a scatter plot of γ(v) versus v. 6. Detect an envelope of γ(v), namely µPW (v). The envelope detection plays the role of taking the sup. One can imagine different ways of detecting the envelope. We used the following algorithm: 6.1 Divide the space of possible v’s, which is [0, 1], into N bins. 6.2 For each bin: i) Search for all the (v, γ(v)) pairs whose v value falls in the bin. ii) Compute γ ∗ , the maximum of the γ(v)’s associated with the pairs found in the previous step. If there are no pairs whose v’s fall in the bin, γ ∗ = 0. iii) For v’s that are members of the bin, set µPW (v) = γ ∗ . Implementation of solutions using a huge number of distributions involves an enormous computational burden, and it is impossible to carry out the optimization over all possible distributions. One needs to incorporate some additional world knowledge about the type of probability distribution of heights into the solution of the PJW problem, since one should only use probability distributions for heights of males that make sense. More generally, each ACWW problem has a real world domain associated with it; hence, it is either explicitly or implicitly constrained by that domain. Therefore, when probability distributions are needed, they should be selected pertinent to that domain. It is shown in [36] that the distribution of the height of Americans is a mixture of Gaussian distributions, since heights of both American men and women obey Gaussian distributions. This N suggests that the optimization problem in (9) can be carried out on X[a,b] , the space of all normal

distributions over [a, b].

July 27, 2013

DRAFT

15

The probability density function of a Gaussian probability distribution is formulated as: f (x|α, β) = √

(x−α)2 1 − e 2β2 2πβ

(13)

Because x ∈ [a, b], we normalize each probability distribution f (x|α, β) by F (b|α, β)−F (a|α, β), Rx so as to make (13) a probability distribution on [a, b], where F (x|α, β) = −∞ f (λ|α, β)dλ is the cumulative distribution function of f (x|α, β); so, for each distribution, we construct: g(x|α, β) ≡ f (x|α, β)/(F (b|α, β) − F (a|α, β))I[a,b] (x)

(14)

where I[a,b] (·) is the indicator function of the interval [a, b]. In this paper, we chose 200 equally spaced points in the intervals [160, 183.8] and [5, 10] respectively5 as candidates for α, and β, which led to 40, 000 Gaussian distributions (normalized over [a, b]), and then implemented the above algorithm. The γ(v) versus v scatter plots for the case of using UMF T1 FSs are depicted in Fig. 7. It is not surprising that the scatter plots for the words Very short, Short, and Moderately short are left-shoulders, because, the a priori knowledge that “Probably John is tall” intuitively suggests that the probability of short sounding height words must be close to zero. Note that in Fig. 7(a), the scatter plot is so narrow that is not visible. As we will see, it gives a very narrow fuzzy set, which is close to a singleton at zero. The scatter plot associated with Tall has points that are exactly on the MF of Probable. The scatter plot associated with Very tall has an interior shape and results in an interior MF. Similar plots were obtained for normal embedded T1 FSs. The envelopes for the scatter plots associated with UMFs and normal embedded T1 FSs are depicted in Figs. 8 and 9, respectively. We then calculated the Jaccard’s similarity of the solutions with each linguistic probability, so as to translate the results into natural language words. Those similarities are summarized in Tables VII and VIII. The boldface numbers signify the words with the highest similarities to each PW . All of our solutions to the problem “What is the probability that John is W ,” given “Probably John is tall,” are summarized in Table IX. They were obtained by choosing the word (in Tables 5

The smallest average height among world countries is 160 cm and the largest average height is 183.8 [2]. For standard deviations, we could not find a source which reports all the standard deviations of heights for each country. The standard deviation of height of Americans is 7-7.5 cm [36]; therefore, we intuitively chose the range [5, 10] as the range of standard deviations of heights. Note that this information comprises additional world knowledge for solving the problem.

July 27, 2013

DRAFT

16

PShort

0.8

0.8

0.8

0.6

0.6

0.6

γ( v)

1

0.4

0.4

0.4

0.2

0.2

0.2

0 0

0.2

0.4

0.6

0.8

0 0

1

0.2

0.4

0.6

0.8

0 0

1

v

(a)

(b)

(c)

P

Medium

0.8

0.6

0.6

γ( v)

0.8

0.6

γ( v)

0.8

0.4

0.4

0.4

0.2

0.2

0.2

0.4

0.6

0.8

1

0.2

0.4

0.6

0.8

1

0.6

0.8

1

Tall

1

0 0

0.6

P

Moderately tall

1

0.2

0.4

v

1

0 0

0.2

v

P

γ( v)

PModerately short

1

γ( v)

γ( v)

PVery short 1

0.8

1

0 0

0.2

0.4

v

v

v

(d)

(e)

(f)

PVery tall 1

γ( v)

0.8 0.6 0.4 0.2 0 0

0.2

0.4

0.6

0.8

1

v

(g) Fig. 7.

Scatter plots for each PW using UMFs as T1 FS models of words, when distributions are Gaussian.

TABLE VII S IMILARITIES BETWEEN THE T1 FS S DEPICTED IN F IG . 8 AND UMF S FOR LINGUISTIC PROBABILITY WORDS PW when W is: Very short Short Moderately short Medium heigt Moderately tall Tall Very tall

Extremely improbable 0.00048936 0.12922 0.31151 0.15951 0 0 1.4184e-05

Very Somewhat Improbable improbable improbable 0 0 0 0.0016615 0 0 0.029157 0.00086374 0 0.61401 0.32412 0.14217 0 0 0 0 0 0 0.013219 0.35196 0.28151

Tossup 0 0 0 0 0.082425 0.1317 0

Somewhat probable 0 0 0 0 0.15721 0.28426 0

Probable 0 0 0 0 0.22015 0.95698 0

Very probable 0 0 0 0 0.025553 0.019543 0

Extremely probable 0 0 0 0 0 0 0

TABLE VIII S IMILARITIES BETWEEN THE T1 FS S DEPICTED IN F IG . 9 AND NORMAL EMBEDDED T1 FS S FOR LINGUISTIC PROBABILITY WORDS

PW when W is: Very short Short Moderately short Medium heigt Moderately tall Tall Very tall

July 27, 2013

Extremely Very Improbable Somewhat Tossup improbable improbable improbable 1.9484e-05 0 0 0 0 0.010599 0 0 0 0 0.061676 0 0 0 0 0.66285 0.20875 0.039784 0.0057136 0 0 0 0 0.019676 0.12411 0 0 0 0 0.14113 0.22658 0.15348 0.03044 0.0035016 0

Very Somewhat Probable Extremely probable probable probable 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0.064178 0.024577 0 0 0.31176 0.84129 0.0057427 0 0 0 0 0

DRAFT

17

PShort 0.8

0.8

0.4

0 0

0.4 0.2

0.2

0.4

0.6

0.8

0 0

1

0.6

W

0.6

W

0.6

µP (v)

0.8

µP (v)

1

0.2

0.4 0.2

0.2

0.4

v

0.6

0.8

0 0

1

(a)

(b) Moderately tall 0.8

0.4 0.2

0.4

0.6

0.8

0 0

1

0.6

0.8

1

0.6

W

0.6

W

W

µP (v)

0.8

µP (v)

0.8

0.2

1

Tall

1

0.4

0.8

P

1

0.6

0.6

(c)

P

Medium

0.2

0.4

v

1

0 0

0.2

v

P

µP (v)

PModerately short

1

W

µP (v)

PVery short 1

0.4 0.2

0.2

0.4

v

0.6

0.8

1

0 0

0.2

0.4

v

(d)

v

(e)

(f)

P

Very tall

1

0.6

W

µP (v)

0.8

0.4 0.2 0 0

0.2

0.4

0.6

0.8

1

v

(g) Fig. 8. The detected envelopes µPW (v) for plots in Fig. 7. UMFs were used as T1 FS models of words and probability distributions were Gaussian.

VII and VIII) that has the largest similarity. So, for example, when UMFs are used as T1 FS models of words, then the linguistic solutions to the PJW problems are: “It is extremely improbable that John is very short.” “It is extremely improbable that John is short.” “It is extremely improbable that John is moderately short.” “It is very improbable that John has medium height.” “It is somewhat improbable that John is moderately tall.” “It is probable that John is tall.” (which means that the algorithm works correctly, since we have already assumed that “Probably John is tall”.) “It is improbable that John is very tall.” Note that in Fig. 8 and 9, some of the fuzzy sets are so narrow that are not visible. Consequently, their similarities with the word extremely improbable are very small. One can linguisJuly 27, 2013

DRAFT

18

PShort 0.8

0.8

0.4

0 0

0.4 0.2

0.2

0.4

0.6

0.8

0 0

1

0.6

W

0.6

W

0.6

µP (v)

0.8

µP (v)

1

0.2

0.4 0.2

0.2

0.4

v

0.6

0.8

0 0

1

(a)

(b) Moderately tall 0.8

0.4 0.2

0.4

0.6

0.8

0 0

1

0.6

0.8

1

0.6

W

0.6

W

W

µP (v)

0.8

µP (v)

0.8

0.2

1

Tall

1

0.4

0.8

P

1

0.6

0.6

(c)

P

Medium

0.2

0.4

v

1

0 0

0.2

v

P

µP (v)

PModerately short

1

W

µP (v)

PVery short 1

0.4 0.2

0.2

0.4

v

0.6

0.8

1

0 0

0.2

0.4

v

(d)

v

(e)

(f)

P

Very tall

1

0.6

W

µP (v)

0.8

0.4 0.2 0 0

0.2

0.4

0.6

0.8

1

v

(g) Fig. 9. The detected envelopes µPW (v) when normal embedded type-1 fuzzy set models of words and Gaussian distributions were used.

tically interpret those narrow fuzzy probabilities as around zero. Observe, from Table IX that four of the linguistic solutions are the same, regardless of which T1 FS model was used for all the words; however, the fact that three of the solutions are different suggests that results for this ACWW problem can be sensitive to the T1 FS that is used to model a word. TABLE IX S UMMARY OF THE SOLUTIONS TO THE PROBLEM “W HAT IS THE PROBABILITY THAT J OHN IS W , GIVEN “P ROBABLY J OHN IS TALL ” W UMF normal embedded T1 FSs Very short Extremely improbable Extremely improbable Short Extremely improbable Extremely improbable Moderately short Extremely improbable Extremely improbable Medium height Very improbable Extremely improbable Moderately tall Probable Tossup Tall Probable Probable Very tall Improbable Extremely improbable

July 27, 2013

DRAFT

19

Comparing Figs. 8 and 9, observe that normal embedded T1 FSs provide narrower MFs for the words involved in the PJW problem as compared to UMF T1 FSs. Because the GEP propagates uncertainty, solutions corresponding to normal embedded T1 FS are therefore narrower (i.e., less uncertain) than solutions using UMF T1 FSs. A reviewer of this paper suggested that the MFs may be determined only by using the endpoints of the intervals that are used as ranges for the means and standard deviations of the Gaussian probability distributions. We examined v = PT all and µP robable (PV ery

tall )

for m = 160, 183.6

and σ = 5, 10 and obtained the following (v, γ(v)) pairs: (0, 0), (0.0067, 0), (0.2314, 0.5055), (0.2704, 0.4566). These four points do not describe the shape of the fuzzy set in Fig. 8(g); hence, it is not possible to determine solution MFs by using only the endpoints of the intervals that are used as ranges of the parameters of the probability distributions.6 IV. O N C ORRECTNESS OF T HE R ESULTS A natural question to ask at this point is: How does one know that the solution obtained for our ACWW problem is correct? In [19], Mendel claims that a Turing-type test is presently the only way to validate a CWW solution. Such a test requires that a human provides an answer to the ACWW problem; however, we have found that ACWW problems are so complicated that a human might be unable to do this. Consequently, we know of no general way to answer this question. Instead, we fall back on a very simple idea, namely we consider one PJW problem for which we surely know what the correct answer is, and see if the GEP and our algorithm for solving (9) provide that answer. That PJW problem is: “Probably John is W . What is the probability that John is W ?.” Common sense tells us the answer is “Probable,” i.e., It is probable that John is W . The solution to this problem using T1 FSs and the GEP is (see (9)): 6 A somewhat analogous situation occurs when one computes the centroid of an IT2 FS [11]. The two endpoints of the centroid are not associated with the the center of gravity of the lower and upper MFs. Instead, each is associated with a T1 FS that includes a portion of both the lower and upper MFs.

July 27, 2013

DRAFT

20

µPW (v) =    sup   p |v=R b p (h)µ H

R ab

    0

a

H

µP robable

R b a

 pH (h)µW (h)dh

∃pH ∈ X[a,b] s.t. v =

Rb

6 ∃pH ∈ X[a,b] s.t. v =

Rb

a

pH (h)µW (h)dh

W (h)dh

pH (h)dh=1 a

pH (h)µW (h)dh (15)

Because Z sup R pH |v= ab pH (h)µW (h)dh Rb a pH (h)dh=1

=

sup

µP robable

b

 pH (h)µW (h)dh

a

µP robable (v) = µP robable (v)

(16)

R pH |v= ab pH (h)µW (h)dh Rb a pH (h)dh=1

(15) simplifies to:  Rb  µ ∃pH ∈ X[a,b] s.t. v = a pH (h)µW (h)dh P robable (v) µPW (v) = Rb  0 6 ∃pH ∈ X[a,b] s.t. v = a pH (h)µW (h)dh

(17)

From (17), it is obvious that the membership function of PW is always equal to that of P robable, except for those v’s for which no probability distribution is found that satisfies v = Rb pH (h)µW (h)dh, which, by the second line of (17), means that µPW (v) = 0. The latter occurs a when W ’s have membership functions that have small overlaps with all pH ’s, so the integral Rb µ (h)pH (h)dh has an upper bound very smaller than 1. Therefore, the solution using T1 FSs a W gives the word “Probable” as the solution, except for those v’s for which no pH is found so that Rb Rb v = a µW (h)pH (h)dh (i.e., v’s that are greater than the upper bound of a µW (h)pH (h)dh), for which µPW is set to zero by (17), resulting in a “trimmed” version of “Probable.” The similarity of the output with “Probable” is not exactly 1, due to the approximation enforced by the envelope detection algorithm and possible “trimming”. We could stop with theoretical result (17), but we want to verify that our computational method also leads to the right-hand side of (17). In order to demonstrate (17), we calculated the solution to the PJW problem “Probably John is W . What is the probability that John is W ?”. We used UMF T1 FS word models, and divided the intervals [160, 183.8] and [5, 10] into 200 equally spaced points, so that there were July 27, 2013

DRAFT

21

200 × 200 = 40000 Gaussian probability distributions. The solutions for each W are shown in Fig. 10, and their similarities with the vocabulary of linguistic probabilities modeled by UMFs are given in Table X. Observe that five of the seven solutions map to the word “Probable,” which is the result that was expected. For the words Very short and Very tall, we got zero MFs over the whole [0, 1] range, due to the fact that the overlaps of the selected pH ’s with µW were small, so the maximum v = Rb µ (h)pH (h)dh was so small that µP robable (v) = 0 for all of them. (Note that µP robable (v) = 0 a W Rb for v 6∈ [0.5086, 0.7914].) Also, for the v’s that are not equal to a µW (h)pH (h)dh for any pH , the GEP in (17) and our algorithm sets µPW = 0. One can interpret this phenomenon as the discrepancy between the FS models of the extreme words “Very short,” and “Very tall” and the world knowledge about the ranges of the parameters of the probability distributions, i.e. the possibility assigned by the extreme words to height values that are close to the endpoints of the interval of heights is large, while the probability densities that are assigned to those points by the probability distributions are very small. This may suggest that, for not having such a discrepancy, one should establish a new scale for heights based on the ranges for means and standard deviations (e.g., one could say that the smallest height to be considered is the smallest mean height (160 cm) minus twice the smallest conceivable standard deviation (2 × 5 cm), which gives 150 cm, and the largest height that has to be considered is the largest average height (183.6 cm) plus twice the largest conceivable standard deviation (2×10 cm) which gives 203.6 cm. Collecting data from subjects about heights on the new scale [150, 203.6] might solve the discrepancy problem, because then even extreme words have reasonable amounts of overlap with the probability distributions. However, neglecting people with heights less than 150 cm or more than 193.6 cm, does not seem to be a reasonable decision. We leave further investigation about this problem to future research. If we use a wider range for the means of probability distributions, e.g. [142.7, 201] , all of the seven solutions do map into “Probable.” This suggests (somewhat surprisingly) that besides the world knowledge the validation process needs to be taken into account for determining the ranges of parameters of probability distributions. How to validate the GEP approach for solving ACWW problems in a more general way is an important and open research question, one that we leave for future research.

July 27, 2013

DRAFT

22

PShort

0.8

0.8

0.8

0.6 0.4

µ P W (v)

1

0.2

0.6 0.4 0.2

0 0

0.2

0.4

v

0.6

0.8

0 0

1

0.6 0.4 0.2

0.2

0.4

(a)

v

0.6

0.8

0 0

1

Medium

0.8

µ P W (v)

0.8

µ P W (v)

0.8

0.2

0.6 0.4 0.2

0.4

v

0.6

0.8

0 0

1

0.8

1

0.6

0.8

1

Tall

1

0.4

0.6

(c)

Moderately tall

0.6

v

P

1

0.2

0.4

P

1

0 0

0.2

(b)

P

µ P W (v)

PModerately short

1

µ P W (v)

µ P W (v)

PVery short 1

0.6 0.4 0.2

0.2

0.4

(d)

v

0.6

0.8

1

0 0

0.2

0.4

(e)

v

(f)

P

Very tall

1

µ P W (v)

0.8 0.6 0.4 0.2 0 0

0.2

0.4

v

0.6

0.8

1

(g) Fig. 10. The detected envelopes µPef (v), which are solutions to the PJW problem “Probably John is W . What is the probability W that John is W ?,” using numeric probability of IT2 fuzzy events. TABLE X S IMILARITIES BETWEEN THE WORDS IN F IG . 10 AND THE UMF S OF THE LINGUISTIC PROBABILITY WORDS OF F IG .3 Extremely Very PW when Improbable Somewhat Tossup improbable improbable improbable W is: Very short 0 0 0 0 0 Short 0 0 0 0 0.14182 Moderately short 0 0 0 0 0.12963 Medium heigt 0 0 0 0 0.12848 Moderately tall 0 0 0 0 0.12848 Tall 0 0 0 0 0.1317 Very tall 0 0 0 0 0

Somewhat probable 0 0.3138 0.27841 0.27516 0.27517 0.28425 0

Probable 0 0.83324 0.98437 1 1 0.95704 0

Very Extremely probable probable 0 0 0.43199 0 0.033291 0 0.04135 0 0.041343 0 0.019574 0 0 0

V. A N E NGINEERING ACWW P ROBLEM In this section we solve a specific problem about product reliability to demonstrate how ACWW problems for real world applications can be solved by the methodology of this paper. Consider the following statement, which provides the subjective judgment about the reliability of a product, and a related question: Probably product X is highly reliable. What is the probability that the reliability of X is R? July 27, 2013

DRAFT

23

R represents a word describing reliability, and can be one of the words: None to very little, Very low, Low, More or less low, From fair to more or less high, More or less high, High, Extremely high. These words constitute the user-friendly eight-word vocabulary introduced in [23, Ch. 7]. Their FOUs are depicted in Fig. 11, and the parameters of the MFs are given in Table XI. The pairwise Jaccard similarities between those words are given in Table XII. Note that the threshold of pairwise similarity for an appropriate partitioning of the space was set to 0.6 in [23, Ch. 7], therefore this vocabulary represents a good partitioning of the space. We chose a [0, 10] scale, and interpreted reliability as “time to failure,” so “high reliability” corresponds to a large time to failure. This scale is a “hypothetical scale,” and can be re-scaled to any appropriate time scale. None to very little

Very low 1

1

0.5

0.5

0.5

0

0

5

10

More or less low

0

0

5

10

0

1

0.5

0.5

0.5

5

10

0

0

High 1

0.5

0.5

0

5

10

10

0

0

5

10

Extremely high

1

0

5

5

More or less high

1

0

0

From Fair to more or less high

1

0

Fig. 11.

Low

1

10

0

0

5

10

Vocabulary of IT2 FSs representing linguistic reliability [23, Fig. 7.5].

TABLE XI M EMBERSHIP FUNCTIONS OF THE WORDS DEPICTED IN F IG . 11 [23, TABLE 7.13]

None to very little Very low Low More or less low From fair to more or less high More or less high High Extremely high

July 27, 2013

UMF parameters (0, 0, 0.22, 3.16) (0, 0, 1.37, 3.95) (0.38, 1.63, 3.00, 4.62) (0.38, 2.25, 4.00, 5.92) (2.33, 5.11, 7.00, 9.59) (4.38, 6.25, 8.00, 9.62) (4.73, 8.82, 10, 10) (7.10, 9.80, 10, 10)

LMF parameters (0, 0, 0.02, 0.33, 1) (0, 0, 0.14, 1.82, 1) (1.90, 2.24, 2.24, 2.51, 0.31) (2.99, 3.31, 3.31, 3.81, 0.32) (5.79, 6.31, 6.31, 7.21, 0.43) (6.90, 7.21, 7.21, 7.60, 0.29) (7.68, 9.82, 10, 10, 1) (9.74, 9.98, 10, 10, 1)

DRAFT

24

TABLE XII PAIRWISE SIMILARITIES BETWEEN THE WORDS DEPICTED IN F IG . 11

None to very little Very low Low More or less low From Fair to more or less high More or less high High Extremely high

None to very little Very low Low More or less low From fair to more or less high More or less high High Extremely high 1 0.50973 0.2415 0.1671 0.0090779 0 0 0 0.50973 1 0.34288 0.2405 0.029751 0 0 0 0.2415 0.34288 1 0.59678 0.082946 0.0012906 0 0 0.1671 0.2405 0.59678 1 0.18822 0.044303 0.01457 0 0.0090779 0.029751 0.082946 0.18822 1 0.5464 0.2342 0.098566 0 0 0.0012906 0.044303 0.5464 1 0.35195 0.16251 0 0 0 0.01457 0.2342 0.35195 1 0.37394 0 0 0 0 0.098566 0.16251 0.37394 1

As in Section III, we used the UMFs of IT2 FS models of reliability words as the T1 FS models of those words (see Table XI and Fig. 11). In the reliability literature, time to failure is modeled by a variety of distributions including exponential, Gaussian, and Weibull. Here, we assume that the distribution of time to failure (as a measure of reliability) is Weibull, whose probability density function is: x

β

f (x|α, β) = βα−β xβ−1 e−( α ) I[0,∞) (x)

(18)

where I[0,∞) (·) is the indicator function of the interval [0, ∞). In all of the simulations of this section, we chose7 α ∈ [3, 8] and β ∈ [0.01, 10], and normalized the Weibull distributions by F (10|α, β) − F (0|α, β), where F (·|α, β) is the cdf of a Weibull distribution with the parameters α and β, to make them finite probability distributions on [0, 10]. We divided each of the intervals pertinent to α and β into 200 equally spaced points, obtaining 40, 000 Weibull probability distributions. Solutions from the GEP using UMFs were obtained employing the same methodology as was applied to the PJW problem; the detected envelopes are shown in Fig. 12. We then calculated the Jaccard’s similarity of the solutions with each linguistic probability, so as to translate the results into natural language words. Those similarities are summarized in Table XIII.

All of our solutions to the problem ‘What is the probability that the reliability of product X is R” given “Probably product X has high reliability”, are summarized in Table XIV. They were obtained by choosing the word (in Table XIII) that has the largest similarity; thus, the linguistic 7 Since we normalize the probability distributions and make them finite on [0, 10], for whatever set of parameters that we select, the mean of those truncated probability distributions stays within [0, 10]; however, we showed through numerical simulation that when α ∈ [3, 8] and β ∈ [0.01, 10], the means of the probability distribustions that we chose are in the range [0.0578, 7.6106] and their standard deviations are in the range [0.3434, 2.7998].

July 27, 2013

DRAFT

25

PVery low 0.8

0.8

0.4

0.4 0.2

0 0

0.2

0.4

0.6

0.8

0.2

0.4

v

0.6

0.8

0 0

1

(a)

(b) From fair to more or less high 0.8

0.4 0.2

0.4

0.6

0.8

0.2

0.4

v

0.6

0.8

0 0

1

0.2

0.4

v

(d)

0.6

1

v

(e)

(f)

P

P

High

Extremely high

1

0.8

0.8

µP (v)

1

0.6

R

0.6

R

µP (v)

0.8

0.4 0.2

0 0

1

0.6

R

0.6

R

R

µP (v)

0.8

µP (v)

0.8

0.2

1

More or less high

1

0.4

0.8

P

1

0.6

0.6

(c)

P

More or less low

0.2

0.4

v

1

0 0

0.2

v

P

0.4 0.2 0 0

0.4 0.2

0.2

0.4

0.6

0.8

1

0 0

0.2

0.4

v

(g) Fig. 12.

0.4 0.2

0 0

1

0.6

R

0.6

R

0.6

µP (v)

0.8

µP (v)

1

0.2

µP (v)

PLow

1

R

µP (v)

PNone to very little 1

0.6

0.8

1

v

(h)

The detected envelopes µPR when UMF fuzzy set models of words and Weibull distributions were used.

TABLE XIII S IMILARITIES BETWEEN THE WORDS DEPICTED IN F IG . 12 AND LINGUISTIC PROBABILITY WORDS Extremely Very PR when Improbable Somewhat Tossup improbable improbable improbable R is: None to very little 0.0037471 0 0 0 0 Very low 0.021429 1.1487e-6 0 0 0 Low 0.10147 0.0071337 0.00026397 0 0 More or less low 0.49782 0.12358 0.025898 0.0075834 0 From fair to more or less high 0 0 0 0 0 More or less high 0 0 0 0 0.00074049 High 0 0 0 0 0.15177 Extremely high 0.0060577 0.036147 0.32735 0.25393 0

Very Extremely Somewhat Probable probable probable probable 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 7.6892e-07 0.086602 0.068662 0.001794 0.0046997 0.049525 0.2617 0.078856 0.34437 0.72777 0 0 0 0 0 0

solutions to this problem are: “It is extremely improbable that X’s reliability is none to very little.” “It is extremely improbable that X has very low reliability.” “It is extremely improbable that X has low reliability.” “It is extremely improbable that X has more or less low reliability.” “It is probable that X’s reliability is from fair to more or less high.”

July 27, 2013

DRAFT

26

TABLE XIV S UMMARY OF THE SOLUTIONS TO THE PROBLEM “W HAT IS THE PROBABILITY THAT THE RELIABILITY OF PRODUCT X R” GIVEN “P ROBABLY PRODUCT X HAS HIGH RELIABILITY ” PR when R is: None to very little Very low Low More or less low From fair to more or less high More or less high High Extremely high

IS

UMF Extremely improbable Extremely improbable Extremely improbable Extremely improbable Probable Very probable Probable Improbable

“It is very probable that X’s reliability is more or less high.” “It is probable that X’s reliability is high.” (which means that the algorithm works correctly, since we have already assumed that “Probably X has high reliability”.) “It is improbable that X’s reliability is extremely high.” Comment: In some of the solutions there is a “spike” close to v = 0. Actually, when the distributions used for solving the reliability problems (or the PJW problems) have “tails” (like Gaussian and Weibull distributions), the membership function of PR (or PW ), at least for some R’s (W ’s), are non-zero for v’s that are very close to zero. When µR is far from e.g. µHigh (R = F rom f air to more or less high satisfies this), and when a probability Rb distribution has a large amount of overlap with µHigh , then a µHigh (λ)p(λ)dλ can be large, so Rb that µP robable ( a µHigh (λ)p(λ)dλ) 6= 0. However, since µR (h) is in the region of the tail of pH , Rb v = a µR (λ)p(λ)dλ is very small, but is non zero; hence, µPR (v) is non-zero for very small values of v. One may therefore expect such spikes to appear in many solutions to the reliability and PJW problems that mainly look like right-shoulders or interiors (but are actually non-convex fuzzy sets due to the spikes). Those spikes are so small that, in many solutions, they are not visible. More importantly they do not contribute very much to similarity calculations, because they have very small areas. VI. OTHER Z ADEH ACWW C HALLENGE P ROBLEMS Zadeh has proposed some other challenge problems for ACWW [49], some of which can also be solved by the GEP. In this section, we present three of them, so that the reader will see that the methodology we have presented for the PJW problems can in principle also be used to solve the other problems.

July 27, 2013

DRAFT

27

A. Tall Swedes Problem (AHS) The tall Swedes problem8 is about the average height of Swedes (AHS), and is: Most Swedes are tall. What is the average height of Swedes? This problem involves a linguistic quantifier (Most) and an implicit assignment of the linguistic quantifier to the portion of tall Swedes. The portion of Swedes who are tall is equivalent to the probability of the fuzzy event Tall, and is calculated as: Z b PT all = µT all (h)pH (h)dh

(19)

a

in which a and b are the minimum and maximum possible heights and pH is the probability distribution function of heights which clearly satisfies: Z b pH (h)dh = 1

(20)

a

The soft constraint “Most” is assigned to PT all . On the other hand, the average height of Swedes is calculated as: Z

b

pH (h)hdh

AH =

(21)

a

To derive the soft constraint imposed on AH by the fact that PT all is constrained by “Most”, one needs to use the framework of the GEP In the Tall Swedes problem, f = PT all , where f : X[a,b] −→ R

(22)

in which X[a,b] is the space of all possible probability distribution functions on [a, b]; and, g = AH, where g : X[a,b] −→ R

(23)

The GEP then implies that the soft constraint on the average height of Swedes is computed, 8

We do not use the acronym “TSP” because it is already widely used for the famous Traveling Salesman Problem.

July 27, 2013

DRAFT

28

as: Z µAH (v) =

µM ost

sup R pH |v= ab pH (h)h(h)dh Rb a pH (h)dh=1

b

 pH (h)µT all (h)dh

(24)

a

Comparing (24) and (9), it is clear that our methodology for solving (9) can also be used to solve (24). It is worth noting that other approaches to solving the tall Swedes problem have been offered in [27], [34]. We are interested in converting Zadeh’s challenge problems to analogous engineering problems. An example of such a problem for the AHS problem would be the following: Most of the products of Company X have somewhat short life-times. What is the average life-time that is expected from the products of Company X? B. Robert’s Problem (RP) The RP is: Usually Robert leaves his office at about 5 pm. Usually it takes Robert about an hour to get home from work. What is the probability that Robert is at home at (or before) 6:15 pm? The time of arrival of Robert, Z, the time of departure of Robert, X, and his travel time, Y are three random variables that satisfy:

Z =X +Y The probability density function of Z is calculated as: Z +∞ pZ (v) = pX (λ)pY (λ − v)dλ ≡ pX ∗ pY

(25)

(26)

−∞

The probability of the fuzzy event A ≡ About 5 pm, PA is: Z b1 PA = pX (λ)µA (λ)dλ

(27)

a1

where a1 and a2 are respectively the earliest and latest possible times that Robert leaves the

July 27, 2013

DRAFT

29

office. The probability of the fuzzy event B ≡ About an hour, PB is: Z b2 pY (λ)µB (λ)dλ PB =

(28)

a2

where a2 and b2 are respectively the smallest and the largest possible amounts of time that it takes Robert to get home from work. The probability that Robert is at home before t = 6 : 15 pm, PHome is: Z

t

pZ (λ)dλ ≡ w

PHome =

(29)

t0

where t0 is the earliest time for Robert to get home. PA and PB are both constrained by “usually”. We want to find the constraint on PHome , i.e.:

Z

b1

pX (λ)µA (λ)dλ is U sually a1

and Z b2 pY (λ)µB (λ)dλ is U sually a2

Z

t

(pX ∗ pY )(λ)dλ is Π

PHome = t0

Then: µΠ (w) =

sup

R w= tt pZ (λ)dλ 0 pZ =pX ∗pY Rb pX | a 1 pX (λ)dλ=1 R 1 pY | ab2 pY (λ)dλ=1

 Z min µU sually

b1

 Z pX (λ)µA (λ)dλ , µU sually

a1

b2

 pY (λ)µB (λ)dλ

a2

2

(30) The following algorithm can be used to solve the RP: 1. Choose the families of probability distributions pertinent to pX and pY . 2. Choose the ranges of the parameters of the families of the probability distributions. 3. Discretize the ranges of the parameters of the probability distributions. 4. For pX , construct a pool of probability distributions having all possible combinations of July 27, 2013

DRAFT

30

parameters. 5. For pY , construct a pool of probability distributions having all possible combinations of parameters. For all of the members of the two pools: 5.1. Choose two specific probability distributions pX and pY from the pools (Note Rb Rb that a11 pX (λ)dλ = 1 and a22 pY (λ)dλ = 1, because pX and pY are probability distribution functions on [a1 , b1 ] and [a2 , b2 ], respectively). 5.2. Compute pZ = pX ∗ pY . Rt 5.3. Compute w = t0 pZ (λ)dλ. Rb Rb 5.4. Compute γ(w) = min(µU sually ( a11 pX (λ)µA (λ)dλ), µU sually ( a22 pY (λ)µB (λ)dλ)). 6. Construct a scatter plot of γ(w) versus w. 7. Detect an envelope of γ(w), namely µΠ (w). An equivalent engineering ACWW problem for Robert’s problem would be [31]: Usually Product I lasts for about 5 years and is then replaced by a refurbished one. Usually, the refurbished Product I lasts for about 2 years. What is the probability that a new Product I is not needed until the eighth year? C. Swedes and Italians Problem (SIP) The SIP is: Most Swedes are much taller than most Italians. What is the difference in the average height of Swedes and the average height of Italians? Zadeh formulates the SIP using generalized constraints, as follows. Assume that the population of Swedes is represented by {S1 , S2 , · · · , Sm } and the population of Italians is represented by {I1 , I2 , · · · , In }. The height of Si is denoted by xi , i = 1, · · · , m, and the height of Ij is denoted by yj , j = 1, · · · , n. Let:  

x ≡ (x1 , x2 , · · · , xm )



y ≡ (y1 , y2 , · · · , yn )

(31)

M uch taller is defined as a fuzzy relation on HSm ×HIn , in which HSm and HIn are respectively the spaces of all possible heights of Swedes and Italians, and the degree to which Si is much

July 27, 2013

DRAFT

31

taller than Ij is hij , (i = 1, · · · , m; j = 1, · · · , n), i.e.: hij ≡ µM uch

taller (xi , yj )

(32)

The cardinality ci of the set of Italians in relation to whom a Swede Si is much taller can be calculated using the following Σ-count [44]:

ci =

n X

µM uch

taller (xi , yj )

=

j=1

n X

hij

(33)

j=1

The proportion of Italians in relation to whom Si is much taller, ρi , is then: ρi ≡

ci n

(34)

Using a T1 FS model for the linguistic quantifier M ost, the degree ui to which a Swede Si , is much taller than most Italians, is: ui = µM ost (ρi )

(35)

The proportion of the m Swedes who are much taller than most Italians can be derived via the division of the Σ-count of those Swedes by m: m

v=

1 X ui m i=1

(36)

Consequently, the degree to which v belongs to the linguistic quantifier M ost is determined by:

M (x, y) = µM ost (v)

(37)

in which the fact that v is a function of x and y is emphasized in the argument of M . The difference in average height of Swedes and the average height of Italians, d, is calculated as:

d=

July 27, 2013

1 X 1X xi − yj m i n j

(38)

DRAFT

32

To derive the linguistic constraint imposed on d by (37), one exploits the GEP. Zadeh’s approach states that there is a soft constraint “Most” on v, the Σ-count of Swedes who are much taller than most Italians, given by (36), and requires the calculation of the soft constraint on d, given by (38). Therefore, in (5), f (x, y) = v and: f : HSm × HIn −→ R Also, g(x, y) = d = 1/m

P

i

xi − 1/n

P

j

(39)

yj , and:

g : HSm × HIn −→ R

(40)

The GEP implies that the soft constraint D on the difference in average heights, d, is characterized by the following membership function: µD (d) =

=

sup (x,y)∈Hm ×Hn I S 1 P x −1 P y d= m i i n j j

sup

µM ost

m 1 X µM ost m i=1

n 1X µM uch n j=1

!! taller (xi , yj )

M (x, y)

(41)

(x,y)∈Hm ×Hn I S 1 P x −1 P y d= m i i n j j

in which (x, y) belongs to HSm ×HIn , the space of all possible heights that Swedes and Italians can have. The sup is taken over this space since we have no information on the height distributions among these two nationalities. The problem stated in (41) can be solved using the following algorithm, which is in the spirit of our methodology for solving the PJW problem: 1. Choose N1 different x’s and N2 different y’s for which xi ∈ HS and yj ∈ HI . 2. Construct a pool of all possible pairs (x, y), and for all of its members, repeat 2.1-2.3: 2.1. Choose a specific (x, y) from the pool. P P 2.2. Compute d = m1 i xi − n1 j yj .  P  P n 1 2.3. Compute γ(d) = µM ost m1 m µ i=1 M ost n j=1 µM uch

taller (xi , yj )

 .

3. Construct a scatter plot of γ(d) versus d. 4. Detect an envelope of γ(d), namely µD (d). An equivalent engineering ACWW problem for SIP would be: Most of the products of Company X have much longer life-times than most of the products of July 27, 2013

DRAFT

33

Company Y. What is the average difference between the life-times of the products of Company X and Company Y? VII. D ISCUSSION In this section, we briefly discuss how one can solve an ACWW problem in general. This issue is not an easy one, because ACWW problems can conceivably be very complicated, due to the complicated nature of natural languages; hence, more research needs to be done to recognize different categories of ACWW problems involving implicit assignments of probability, truth, possibility, etc. Nevertheless, inspired by ACWW problems that have been discussed in this paper, we attempt to demonstrate a fairly general framework for solving ACWW problems that involve linguistic probabilities, namely: 1. Determine the linguistic probability words, linguistic quantifiers, and linguistic usuality words in the problem. 2. From the linguistic description of the problem, determine which quantities (i.e., numeric probabilities) are constrained by the words that were found in the previous step. This may be a challenging task, since those words may be implicitly assigned to the quantity. (e.g., Most is assigned to the portion of Swedes who are tall, Probable is assigned to the probability that John is tall, etc.) 3. Formulate the numeric probabilities in the previous step: 3.1. Use definite integrals involving indicator functions of non-fuzzy events or membership functions of fuzzy events and the (unknown) probability distribution functions pertinent to those events (continuous case); or, 3.2. Use the fraction of the cardinality of the fuzzy event over the cardinality of the whole population (discrete case). Like the continuous case, for which the probability density functions might be unknown, some quantities related to the population (e.g., the average height of a population or the variance of the height of a population) may be unknown. 4. Formulate the quantity on which a fuzzy constraint must be calculated, in terms of the unknowns (e.g., probability distributions) of Steps 3.1. or 3.2. This quantity may be an average, a numeric probability, or a function of some averages, etc. To do this, the calculi of fuzzy sets is needed. July 27, 2013

DRAFT

34

5. Apply the GEP: knowing the soft constraints on the quantities formulated in Step 3, determine the soft constraint on the quantity that was formulated in Step 4. The sup of the GEP is taken over all possible unknowns of Steps 3.1. or 3.2. For example, if a probability distribution is not known, or if the height of the individuals in a population is unknown, the sup is taken over all admissible probability distributions or admissible heights of individuals. The word “admissible” is crucial and implies that additional problem-specific world knowledge is needed. We leave it to the reader to return to our earlier sections to see how these five steps were implemented for specific ACWW problems. VIII. T HE R ELATIONSHIP BETWEEN P ERCEPTUAL C OMPUTING AND A DVANCED C OMPUTING WITH W ORDS USING THE GEP In this section, we investigate the relationship between ACWW and Perceptual Computing [23]. The architecture of a Perceptual Computer (Per-C for short) is illustrated in Fig. 13 [16]– [18]. The Per-C has three major subsystems: encoder, which constructs fuzzy set models of words that are used by the CWW engine; CWW engine, which, based on the knowledge pertinent to a specific problem and the calculi of fuzzy sets, computes a fuzzy set at its output, that is used by the decoder; and, decoder, which transforms the FS output of the engine into a recommendation (a word) that is comprehensible to humans. Data that support the recommendation may also be provided. In this paper, we used the Interval Approach (IA) [14] or the Enhanced Interval Approach (EIA) [41] for establishing IT2 FS models of words. We then used UMFSs or normal embedded T1 FSs extracted from the IT2 FS models of words, as T1 FS models of them; hence, we used the IA or EIA as encoders. To compute the solutions to ACWW problems using those fuzzy set models, we used the GEP; hence, the GEP acts as a new9 CWW engine. We then used Jaccard’s similarity measure as a decoder to transform the fuzzy set provided by the GEP, into a word. We can therefore conclude that ACWW using the GEP fits within the framework of Perceptual Computing. 9

Other CWW engines are: novel weighted averages and IF-THEN rules.

July 27, 2013

DRAFT

35

Fig. 13.

The perceptual computer that uses FS models for words.

IX. C ONCLUSIONS Zadeh believes that the GEP is one of the main aggregation tools for CWW, especially when dealing with probability constraints. Unfortunately, analytic solutions involving the GEP are presently hopeless, and existing numerical algorithms (e.g., α-cut decomposition) are not directly applicable. In this paper, we solved some of Zadeh’s challenge problems that involve linguistic probabilities using a novel algorithm for implementing the GEP. To the best of our knowledge, our algorithm is the first attempt to actually implement Zadeh’s solution to some of his ACWW problems. The applicability of our algorithm has been demonstrated by solving some specific ACWW problems. An important prerequisite for using the GEP for problems involving linguistic probabilities, is knowing the type of probability distributions for a specific problem. This additional “world knowledge” is a luxury that may not always be available. A limitation of our algorithm for implementing the GEP is its “exhaustive search” on the space of the parameters involved in the problem, especially when the number of parameters (of the probability distributions) proliferates.10 Implementing the GEP without exhaustive search is certainly very worthy of future research. We also believe that there is more than one way to solve ACWW problems, and have presented different solutions to Zadeh’s ACWW problems in [29]–[31], [33], [34] using syllogisms and Novel Weighted Averages [35]. Those solutions do not need any world knowledge about probability distributions, but they need world knowledge about the domains of variables involved in the problems (e.g., time and height). This is at the expense of obtaining lower and upper probabilities for a problem, instead of getting only one fuzzy probability. Interestingly, those soloutions also use the Extension Principle, but in a different way from the GEP. They involve 10 An example of this occurs when we need the distribution of heights of people (men and women), which is a mixture of Gaussians (kf (x|α1 , β1 ) + (1 − k)f (x|α2 , β2 )), and hence, has five parameters.

July 27, 2013

DRAFT

36

the challenge of choosing appropriate compatibility measures [29], and may require multiple computational methodologies for fusion of inconsistent information [31]. How to validate a particular solution to an ACWW problem is an important and interesting issue. It is worth reminding the readers that most ACWW problems are so complicated that a human cannot provide an answer to them; hence, validation by means of a Turing Test is not possible. In this paper, we established a method to investigate the correctness of solutions by checking whether the algorithm yields the correct solution to a problem whose solution is intuitively known. We believe that establishing methodologies for validating the solutions to an ACWW problem is an open and very important research topic. Since the GEP is the main tool for manipulating Z-numbers [52], our methodology may also contribute to developing viable frameworks for computing with Z-numbers. ACKNOWLEDGMENT The authors would like to thank Prof. Lotfi A. Zadeh for explaining his solutions to his ACWW challenge problems to us and his continuing support and enthusiasm. They would also like to thank the reviewers of the paper for constructive feedback and comments which improved the present paper. Mohammad Reza Rajati would also like to sincerely acknowledge the generous support of Annenberg Fellowship Program and the 2011 Summer Research Institute Fellowship Program of the University of Southern California. R EFERENCES [1] Amazon Mechanical Turk. [Online]. Available: https://www.mturk.com/mturk/ [2] Human height. [Online]. Available: https://en.wikipedia.org/wiki/Human height [3] Y. Cao and G. Chen, “A fuzzy petri-nets model for computing with words,” IEEE Transactions on Fuzzy Systems, vol. 18, no. 3, pp. 486–499, 2010. [4] Y. Cao, M. Ying, and G. Chen, “Retraction and generalized extension of computing with words,” IEEE Transactions on Fuzzy Systems, vol. 15, no. 6, pp. 1238–1250, 2007. [5] V. Cross and T. Sudkamp, Similarity and compatibility in fuzzy set theory: Assessment and applications. Physica Verlag, 2002. [6] Y. Ding and A. Lisnianski, “Fuzzy universal generating functions for multi-state system reliability assessment,” Fuzzy Sets and Systems, vol. 159, no. 3, pp. 307–324, 2008. [7] J. Dunyak, I. Saad, and D. Wunsch, “A theory of independent fuzzy probability for system reliability,” IEEE Transactions on Fuzzy Systems, vol. 7, no. 3, pp. 286–294, 1999. [8] S. Han and J. Mendel, “A new method for managing the uncertainties in evaluating multi-person multi-criteria location choices, using a perceptual computer,” Annals of Operations Research, vol. 195, no. 1, pp. 277–309, 2012.

July 27, 2013

DRAFT

37

[9] F. Herrera, S. Alonso, F. Chiclana, and E. Herrera-Viedma, “Computing with words in decision making: foundations, trends and prospects,” Fuzzy Optimization and Decision Making, vol. 8, no. 4, pp. 337–364, 2009. [10] I. Karimi and E. H¨ullermeier, “Risk assessment system of natural hazards: A new approach based on fuzzy probability,” Fuzzy Sets and Systems, vol. 158, no. 9, pp. 987–999, 2007. [11] N. N. Karnik and J. M. Mendel, “Centroid of a type-2 fuzzy set,” Information Sciences, vol. 132, no. 1, pp. 195–220, 2001. [12] E. S. Khorasani, P. Patel, S. Rahimi, and D. Houle, “An inference engine toolkit for computing with words,” Journal of Ambient Intelligence and Humanized Computing, pp. 1–20, 2012. [13] E. S. Khorasani, S. Rahimi, and W. Calvert, “Formalization of generalized constraint language: A crucial prelude to computing with words,” IEEE Transactions on Systems, Man, and Cybernetics-Part B: Cybernetics. [14] F. Liu and J. M. Mendel, “Encoding words into interval type-2 fuzzy sets using an interval approach,” IEEE Transactions on Fuzzy Systems, vol. 16, no. 6, pp. 1503–1521, 2008. [15] L. Martınez, D. Ruan, and F. Herrera, “Computing with words in decision support systems: An overview on models and applications,” International Journal of Computational Intelligence Systems, vol. 3, no. 4, pp. 382–395, 2010. [16] J. M. Mendel, “The perceptual computer: An architecture for computing with words,” in Proceedings of The 10th IEEE International Conference on Fuzzy Systems, FUZZ-IEEE 2001., vol. 1, 2001, pp. 35–38. [17] ——, “An architecture for making judgments using computing with words,” International Journal of Applied Mathematics and Computer Science, vol. 12, no. 3, pp. 325–336, 2002. [18] ——, “Computing with words and its relationships with fuzzistics,” Information Sciences, vol. 177, no. 4, pp. 988–1006, 2007. [19] ——, “Computing with words: Zadeh, Turing, Popper and Occam,” IEEE Computational Intelligence Magazine, vol. 2, no. 4, pp. 10–17, 2007. [20] J. M. Mendel, J. Lawry, and L. A. Zadeh, “Foreword to the special section on computing with words,” IEEE Transactions on Fuzzy Systems, vol. 18, no. 3, pp. 437–440, 2010. [21] J. M. Mendel and D. Wu, “Perceptual reasoning for perceptual computing,” IEEE Transactions on Fuzzy Systems, vol. 16, no. 6, pp. 1550–1564, 2008. [22] ——, “Computing with words for hierarchical and distributed decision making,,” in Computational Intelligence in Complex Decision Systems.

Paris, France: Atlantis, 2009.

[23] ——, Perceptual computing: Aiding people in making subjective judgments.

Wiley-IEEE Press, 2010.

[24] ——, “Challenges for perceptual computer applications and how they were overcome,” IEEE Computational Intelligence Magazine, vol. 7, no. 3, pp. 36–47, 2012. [25] J. M. Mendel, L. A. Zadeh, E. Trillas, R. R. Yager, J. Lawry, H. Hagras, and S. Guadarrama, “What computing with words means to me [discussion forum],” IEEE Computational Intelligence Magazine, vol. 5, no. 1, pp. 20–26, 2010. [26] B. M¨oller, W. Graf, and M. Beer, “Safety assessment of structures in view of fuzzy randomness,” Computers & Structures, vol. 81, no. 15, pp. 1567–1582, 2003. [27] V. Nov´ak and P. Murinov´a, “Intermediate quantifiers of natural language and their syllogisms,” in Proceedings of World Conference on Soft Computing, 2011, pp. 178–184. [28] P. Patel, E. Khorasani, and S. Rahimi, “An API for generalized constraint language based expert system,” in Proceedings of 2012 Annual Meeting of the North American Fuzzy Information Processing Society (NAFIPS), 2012, pp. 1–6.

July 27, 2013

DRAFT

38

[29] M. R. Rajati and J. M. Mendel, “Lower and upper probability calculations using compatibility measures for solving zadeh’s challenge problems,” in 2012 IEEE International Conference on Fuzzy Systems (FUZZ-IEEE), 2012, pp. 1–8. [30] ——, “Solving Zadeh’s Swedes and Italians challenge problem,” in Proceedings of 2012 Annual Meeting of the North American Fuzzy Information Processing Society (NAFIPS), 2012, pp. 1–6. [31] ——, “Advanced computing with words using syllogistic reasoning and arithmetic operations on linguistic belief structures,” in Proceedings of 2013 IEEE International Conference on Fuzzy Systems (IEEE-FUZZ 2013), 2013, paper F-1449. [32] ——, “Modeling linguistic probabilities and linguistic quantifiers using interval type-2 fuzzy sets,” in Proceedings of 2013 Joint IFSA World Congress and Annual Meeting of the North American Fuzzy Information Processing Society (NAFIPS), 2013. [33] M. R. Rajati, J. M. Mendel, and D. Wu, “Solving Zadeh’s Magnus challenge problem on linguistic probabilities via linguistic weighted averages,” in Proceedings of 2011 IEEE International Conference on Fuzzy Systems (IEEE FUZZ 2011), 2011, pp. 2177–2184. [34] M. R. Rajati, D. Wu, and J. M. Mendel, “On solving Zadeh’s tall Swedes problem,” in Proceedings of 2011 World Conference on Soft Computing, 2011. [35] M. R. Rajati and J. M. Mendel, “Novel weighted averages versus normalized sums in computing with words,” Information Sciences, vol. 235, pp. 130–149, 2013. [36] M. Schilling, A. Watkins, and W. Watkins, “Is human height bimodal?” The American Statistician, vol. 56, no. 3, pp. 223–229, 2002. [37] F. Tatari, M. Akbarzadeh-T, and A. Sabahi, “Fuzzy-probabilistic multi agent system for breast cancer risk assessment and insurance premium assignment,” Journal of Biomedical Informatics, 2012. [38] J. Wang, “A subjective methodology for safety analysis of safety requirements specifications,” IEEE Transactions on Fuzzy Systems, vol. 5, no. 3, pp. 418–430, 1997. [39] D. Wu and J. M. Mendel, “A comparative study of ranking methods, similarity measures and uncertainty measures for interval type-2 fuzzy sets,” Information Sciences, vol. 179, no. 8, pp. 1169–1192, 2009. [40] ——, “Perceptual reasoning for perceptual computing: A similarity-based approach,” IEEE Transactions on Fuzzy Systems, vol. 17, no. 6, pp. 1397–1411, 2009. [41] D. Wu, J. M. Mendel, and S. Coupland, “Enhanced interval approach for encoding words into interval type-2 fuzzy sets and its convergence analysis,” IEEE Transactions on Fuzzy Systems, vol. 20, no. 3, pp. 499–513, 2012. [42] R. R. Yager, “Perception-based granular probabilities in risk modeling and decision making,” IEEE Transactions on Fuzzy Systems, vol. 14, no. 2, pp. 329–339, 2006. [43] L. A. Zadeh, “Probability measures of fuzzy events,” J. Math. Anal. Appl, vol. 23, no. 2, pp. 421–427, 1968. [44] ——, “A computational approach to fuzzy quantifiers in natural languages,” Computers & Mathematics with Applications, vol. 9, no. 1, pp. 149–184, 1983. [45] ——, “Fuzzy logic= computing with words,” IEEE Transactions on Fuzzy Systems, vol. 4, no. 2, pp. 103–111, 1996. [46] ——, “From computing with numbers to computing with words,” Annals of the New York Academy of Sciences, vol. 929, no. 1, pp. 221–252, 2001. [47] ——, “From computing with numbers to computing with words– from manipulation of measurements to manipulation of perceptions,” in Logic, Thought and Action, ser. Logic, Epistemology, and the Unity of Science, D. Vanderveken, Ed. Springer Netherlands, 2005, vol. 2, pp. 507–544.

July 27, 2013

DRAFT

39

[48] ——, “A summary and update of fuzzy logic,” in Proceedings of 2010 IEEE International Conference on Granular Computing (GrC), 2010, pp. 42–44. [49] ——, Computing with words: Principal concepts and ideas, ser. Studies in Fuzziness and Soft Computing.

Springer

Berlin Heidelberg New York, 2012, vol. 277. [50] ——, “The concept of a linguistic variable and its application to approximate reasoning–II,” Information sciences, vol. 8, no. 4, pp. 301–357, 1975. [51] ——, “Precisiated natural language (PNL),” AI magazine, vol. 25, no. 3, p. 74, 2004. [52] ——, “A note on Z-numbers,” Information Sciences, vol. 181, no. 14, pp. 2923–2932, 2011. [53] J. Zhou, “Reliability assessment method for pressure piping containing circumferential defects based on fuzzy probability,” International Journal of Pressure Vessels and Piping, vol. 82, no. 9, pp. 669–678, 2005.

July 27, 2013

DRAFT

Suggest Documents