[IEEE 2009 International Conference on Electronic Commerce and Business Intelligence, ECBI -...

4
A A A A New New New New Method Method Method Method of of of of Ascertaining Ascertaining Ascertaining Ascertaining Attribute Attribute Attribute Attribute Weight Weight Weight Weight Based Based Based Based on on on on Rough Rough Rough Rough Sets Sets Sets Sets Theor Theor Theor Theory y y y BAO Xin-zhong LIU Ying-wen School of Economics and Management School of Economics and Management University of Science and Technology Beijing University of Science and Technology Beijing Beijing 10083, China Beijing 10083, China [email protected] Abstract---- ---- ---- ----Attribute Attribute Attribute Attribute importance importance importance importance represented represented represented represented by by by by rough rough rough rough set set set set is is is is studied studied studied studied deeply. deeply. deeply. deeply. Aiming Aiming Aiming Aiming at at at at the the the the shortcomings shortcomings shortcomings shortcomings of of of of existing existing existing existing methods methods methods methods of of of of attribute attribute attribute attribute weight weight weight weight assignment assignment assignment assignment based based based based on on on on rough rough rough rough set, set, set, set, by by by by considering considering considering considering both both both both the the the the overall overall overall overall significance significance significance significance degree degree degree degree and and and and the the the the individual individual individual individual significance significance significance significance degree degree degree degree of of of of each each each each attribute attribute attribute attribute in in in in the the the the conditional conditional conditional conditional attributes, attributes, attributes, attributes, a new new new new method method method method of of of of attribute attribute attribute attribute weight weight weight weight assignment assignment assignment assignment based based based based on on on on rough rough rough rough set set set set is is is is proposed, proposed, proposed, proposed, which which which which can can can can improve improve improve improve the the the the ability ability ability ability of of of of generalization generalization generalization generalization and and and and interpretation. interpretation. interpretation. interpretation. Key Key Key Key words: words: words: words: weight; weight; weight; weight; rough rough rough rough sets; sets; sets; sets; attribute attribute attribute attribute significance significance significance significance degree degree degree degree; ; ; decision decision decision decision table table table table I. PREFACE Weight is vitally important in the process of judging and decision making, which reflects the status and function of every attribute. Methods of attribute weight assignment that are normally used include expert evaluation method, fuzzy Statistics method, method of ranking by dual comparison and so on. The disadvantage of these methods is that attribute weight assignment is limited by experts’ experience. Now, some academicians did researches on methods of attribute weight assignment based on rough set. Rough sets theorem was raised by Professor Pawlak from Poland Warsaw Institute of Technology in 1980s, which is a theoretical method researching on the expression, study and generalization of incomplete and uncertain knowledge and data. Since it does not need prior information, rough set has been applied successfully in various fields such as expert system, machine study and mode identifying [2- 4]. Literature [5] provides the method of conditional attribute weight assignment in decision table, which has been quoted in different areas by many academicians [6-8]. This thesis analyzed the shortcomings of existing methods of attribute weight assignment based on rough set, proposed a new method of attribute weight assignment based on rough set and proved its rationality. II. ANALYSIS OF THE EXISTING METHODS OF ATTRIBUTE WEIGHT ASSIGNMENT BASED ON ROUGH SET The significance degree of attribute (indicator) in the conditional attributes varies from one attribute to another, so we should ascertain the significance degree of each attribute at first. While applying the rough sets theory, we take out an attribute and then consider how the classification will change without this attribute. If the classification changes relatively large after removing the attribute, we can conclude that the strength of this attribute is large and the attribute is very important; otherwise, the strength of this attribute is small and the attribute is of little importance.[4-5]. According to this feature, literature [5] defines the importance of attribute (indicator) as follows. Dependence on knowledge: Definition Definition Definition Definition 1[1]: In decision table , (,,,,) S UCDV f = the dependence of decision attribute D ( ) B D γ to conditional attribute sets (indicators) B C is defined as follows: ( ) | ( )|/| | B B D POS D U γ = Definition Definition Definition Definition 2[7]: In decision table , (,,,,) S UCDV f = the significance degree of conditional attribute (indicator) is defined as follows: c . {} () ( ) ( ) C C c Sig c D D g g - = - The weight of conditional attribute 0 () W c (indicator) is defined as : c 0 () () () a C Sig c W c Sig a Î = å Note: The definition illustrates that the greater is, the more important the () Sig c conditional attribute (indicator) is, so as to c the weight of the attribute. 2009 International Conference on Electronic Commerce and Business Intelligence 978-0-7695-3661-3/09 $25.00 © 2009 IEEE DOI 10.1109/ECBI.2009.129 507 2009 International Conference on Electronic Commerce and Business Intelligence 978-0-7695-3661-3/09 $25.00 © 2009 IEEE DOI 10.1109/ECBI.2009.129 507

Transcript of [IEEE 2009 International Conference on Electronic Commerce and Business Intelligence, ECBI -...

Page 1: [IEEE 2009 International Conference on Electronic Commerce and Business Intelligence, ECBI - Beijing, China (2009.06.6-2009.06.7)] 2009 International Conference on Electronic Commerce

AAAA NewNewNewNewMethodMethodMethodMethod ofofofof AscertainingAscertainingAscertainingAscertaining AttributeAttributeAttributeAttribute WeightWeightWeightWeight BasedBasedBasedBased ononononRoughRoughRoughRough SetsSetsSetsSets TheorTheorTheorTheoryyyy

BAO Xin-zhong LIU Ying-wenSchool of Economics and Management School of Economics and Management

University of Science and Technology Beijing University of Science and Technology BeijingBeijing 10083, China Beijing 10083, China

[email protected]

Abstract----------------AttributeAttributeAttributeAttribute importanceimportanceimportanceimportance representedrepresentedrepresentedrepresented bybybybyroughroughroughrough setsetsetset isisisis studiedstudiedstudiedstudied deeply.deeply.deeply.deeply. AimingAimingAimingAiming atatatat thethethetheshortcomingsshortcomingsshortcomingsshortcomings ofofofof existingexistingexistingexisting methodsmethodsmethodsmethods ofofofof attributeattributeattributeattributeweightweightweightweight assignmentassignmentassignmentassignment basedbasedbasedbased onononon roughroughroughrough set,set,set,set, bybybybyconsideringconsideringconsideringconsidering bothbothbothboth thethethethe overalloveralloveralloverall significancesignificancesignificancesignificance degreedegreedegreedegreeandandandand thethethethe individualindividualindividualindividual significancesignificancesignificancesignificance degreedegreedegreedegree ofofofof eacheacheacheachattributeattributeattributeattribute inininin thethethethe conditionalconditionalconditionalconditional attributes,attributes,attributes,attributes, aaaa newnewnewnewmethodmethodmethodmethod ofofofof attributeattributeattributeattribute weightweightweightweight assignmentassignmentassignmentassignment basedbasedbasedbased ononononroughroughroughrough setsetsetset isisisis proposed,proposed,proposed,proposed, whichwhichwhichwhich cancancancan improveimproveimproveimprove thethethetheabilityabilityabilityability ofofofof generalizationgeneralizationgeneralizationgeneralization andandandand interpretation.interpretation.interpretation.interpretation.

KeyKeyKeyKey words:words:words:words: weight;weight;weight;weight; roughroughroughrough sets;sets;sets;sets; attributeattributeattributeattributesignificancesignificancesignificancesignificance degreedegreedegreedegree;;;; decisiondecisiondecisiondecision tabletabletabletable

I. PREFACEWeight is vitally important in the process of

judging and decision making, which reflects thestatus and function of every attribute. Methods ofattribute weight assignment that are normallyused include expert evaluation method, fuzzyStatistics method, method of ranking by dualcomparison and so on. The disadvantage of thesemethods is that attribute weight assignment islimited by experts’ experience. Now, someacademicians did researches on methods ofattribute weight assignment based on rough set.

Rough sets theorem was raised by ProfessorPawlak from Poland Warsaw Institute ofTechnology in 1980s, which is a theoreticalmethod researching on the expression, study andgeneralization of incomplete and uncertainknowledge and data. Since it does not need priorinformation, rough set has been appliedsuccessfully in various fields such as expertsystem, machine study and mode identifying [2-4]. Literature [5] provides the method ofconditional attribute weight assignment indecision table, which has been quoted in differentareas by many academicians [6-8]. This thesisanalyzed the shortcomings of existing methods ofattribute weight assignment based on rough set,proposed a new method of attribute weightassignment based on rough set and proved itsrationality.

II. ANALYSIS OF THE EXISTING METHODS OFATTRIBUTE WEIGHT ASSIGNMENT BASED ON

ROUGH SETThe significance degree of attribute (indicator)

in the conditional attributes varies from oneattribute to another, so we should ascertain thesignificance degree of each attribute at first.While applying the rough sets theory, we take outan attribute and then consider how theclassification will change without this attribute. Ifthe classification changes relatively large afterremoving the attribute, we can conclude that thestrength of this attribute is large and the attributeis very important; otherwise, the strength of thisattribute is small and the attribute is of littleimportance.[4-5]. According to this feature,literature [5] defines the importance of attribute(indicator) as follows.

Dependence on knowledge:

DefinitionDefinitionDefinitionDefinition 1111[1]: In decision table ,( , , , , )S UCDV f=

the dependence of decision attribute D( )B Dγ

to conditional attribute sets (indicators) B C⊆

is defined as follows:( ) | ( ) | / | |B BD POS D Uγ =

DefinitionDefinitionDefinitionDefinition 2222[7]: In decision table ,( , , , , )S UCDV f=

the significance degree of conditional attribute(indicator) is defined as follows:c

.{ }( ) ( ) ( )C C cSig c D Dg g -= -

The weight of conditional attribute0( )W c

(indicator) is defined as :c0

( )( )( )

a C

Sig cW cSig a

Î

Note: The definition illustrates that the

greater is, the more important the( )Sig cconditional attribute (indicator) is, so as tocthe weight of the attribute.

2009 International Conference on Electronic Commerce and Business Intelligence

978-0-7695-3661-3/09 $25.00 © 2009 IEEE

DOI 10.1109/ECBI.2009.129

507

2009 International Conference on Electronic Commerce and Business Intelligence

978-0-7695-3661-3/09 $25.00 © 2009 IEEE

DOI 10.1109/ECBI.2009.129

507

Page 2: [IEEE 2009 International Conference on Electronic Commerce and Business Intelligence, ECBI - Beijing, China (2009.06.6-2009.06.7)] 2009 International Conference on Electronic Commerce

In order to explain the shortage of thedefinition better, take an example as follows.

Example 1: The decision table 1

Table 1: The decision table 1

From the decision table 1, we can easily get:U/{a,b,c}={{x1,x2},{x3},{x4},{x5},{x6,x7},

{x8,x9}, {x10,x11}}U/D={{x1,x5,x6,x8,x11},{x2,x3,x4,x7,x9,x1

0}}Therefore:

POS(a,b,c)(D)={x3,x4,x5};POS(a,b)(D)={x3,x4,x5}POS(b,c)(D)={x3,x4,x5};POS(a,c)(D)={x3,x4,x5}.

As a result:

.{ , , } { , , }( ) | ( ) | / | | 3 / 11a b c a b cD POS D Ug = =

Similar that:

.{ , } { , } { , }( ) ( ) ( ) 3 / 11a b a c b cD D Dg g g= = =

Then: ;{ , , } { , }( ) ( ) ( ) 0a b c b cSig a D Dg g= - =

;{ , , } { , }( ) ( ) ( ) 0a b c a cSig b D Dg g= - =

.{ , , } { , }( ) ( ) ( ) 0a b c a bSig c D Dg g= - =

From the example above, we can see that theweight of conditional attributes can not becalculated accurately in that event. Conditionalattributes (a, b, and c) are unnecessary in thedecision Table 1 in respect to simple attribute.Aiming at solving this problem, literature [7]simplified the decision table and figured out theimportance of the attributes. In this way, theimportance of at least one attribute is not zero,and then we can figured out the weight with theformula of ascertaining weight in literature [5].Although this method can guarantee that at leastone weight not to be zero, it ignores the practicalsignificance of the attributes which occupies noweight and the reduced ones. In order to solvethis problem, we offer a new method ofascertaining weight.

III. Analysis of significance degree ofattribute weight based on rough sets

a b c dx1 0 1 2 0x2 0 1 2 1x3 1 0 1 1x4 0 0 1 1x5 1 3 1 0x6 2 1 2 0x7 2 1 2 1x8 2 1 1 0x9 2 1 1 1x10 0 2 2 1x11 0 2 2 0

Definition 3: In decision table ,( , , , , )S UCDV f=

compared to the conditional attribute sets( ), the conditionalC 1 2/ { , , , }mU C C C C= ⋯

entropy of decision attribute sets( | )I D C

( ) is defined as follows:D 1 2/ { , , , }kU D D D D= ⋯

2

21 1

| | | || |( | ) 1

| | | || |

m kj i j ii

i j i i

D C D CCI D CC CU= =

∩ ∩⎛ ⎞= −⎜ ⎟

⎝ ⎠∑ ∑

Definition4: In decisiontable ,for , denote( , , , , )S U C DV f= , ( )P Q C D∀ ⊆ ∪

, ,if1 2/ { , ,..., }hU P P P P= 1 2/ { , ,..., }SU Q Q Q Q=

, such that ,then/ /i jP U P Q U Q∀ ∈ ⇒ ∃ ∈ i jP Q⊆

called as the refinement of , denoted/U Q /U P

by / /U P U Q°

Theorem 1: In decision table ,( , , , , )S U C DV f=

for , there was .( )Q P C D∀ ⊆ ⊆ ∪ / /U P U Q°

Proof:

Denoted , ,1 2/ { , , , }hU P P P P= ⋯ 1 2/ { , , , }sU Q Q Q Q= ⋯ ∀

,due to ,then[ ] /i PP x U P= ∈ Q P⊆

,|{][ PayxP pi ∈∀== jQaxfayf ⊆= )},(),(

.)},(),(,|{][ axfayfQayx Q =∈∀==

According to the randomicity of , we caniP

get ./ /U P U Q°

Theorem 2: In domain of discourse , letU

,1 2/ { , , , }tU B B B B= ⋯

1 2/ { , , , }kU D D D D= ⋯

,1 2 1 1 1 1/ { , , , , , , , , , , }p p q q t p qU B B B B B B B B B B− + − +′ = ∪⋯ ⋯ ⋯

then .( | ) ( | )I D B I D B′ ≥

The equality holds if and only if:|)||(||| qjqpj BDBBD ∩−∩

=0(j=1,2,...,k)|)||(|||0 qjqpj BDBBD ∩−∩∧=

Proof:( | ) ( | )I D B I D B′ −

2

21

| | | ( ) | | ( ) |1

| | | || |

kP q j P q j P q

j P q P q

B B D B B D B BB B B BU =

⎛ ⎞∪ ∩ ∪ ∩ ∪= −⎜ ⎟⎜ ⎟∪ ∪⎝ ⎠

∑22

2 21 1

| | | | | | | | | || |1 1

| | | | | | | || | | |

k kj P j P q j q j qP

j jP P q q

D B D B B D B D BBB B B BU U= =

⎛ ⎞∩ ∩ ∩ ∩⎛ ⎞− − − −⎜ ⎟⎜ ⎟ ⎜ ⎟⎝ ⎠ ⎝ ⎠

∑ ∑

21

1 (| | | |)(| | | | | |)| |

k

j P j q P q j P j qj

D B D B B B D B D BU =

= ∩ + ∩ ∪ − ∩ − ∩∑

|)||(|||||

11

2 pjppj

k

jBDBBD

U∩−∩− ∑

=

508508

Page 3: [IEEE 2009 International Conference on Electronic Commerce and Business Intelligence, ECBI - Beijing, China (2009.06.6-2009.06.7)] 2009 International Conference on Electronic Commerce

|)||(|||||

11

2 qjqqj

k

jBDBBD

U∩−∩− ∑

=

Let ,| |, | |, | |, | |j j P j j q P qx D B y D B x B y B= ∩ = ∩ = =

then ,and0, 0j jx x y y≥ ≥ ≥ ≥ ( | ) ( | )I D B I D B′ −

21

1 ( )( )| |

k

j j j jj

x y x y x yU =

= + + − −∑

2 21 1

1 1( ) ( )| | | |

k k

j j j jj jx x x y y y

U U= =

− − − −∑ ∑

21

1 ( ( ) ( ) ) 0| |

k

j j j jj

x y y x x yU =

= − + − ≥∑

(for )0, 0j jx x y y≥ ≥ ≥ ≥

The equality holds if and only if :

For , ,{1,2, , }j k∀ ∈ ⋯ ( ) 0 ( ) 0j j j jx y y x x y− = ∧ − =

| |(| | | |) 0 | |(| | | |) 0j P q j q j q p j pD B B D B D B B D B∩ − ∩ = ∧ ∩ − ∩ =

Theorem 3: In decision table S=(U,C,D,V,f),, , thenB P C∀ ⊆ ⊆ 1 2/ { , , , }kU D D D D= ⋯

.( | ) ( | )I D P I D B≤

The equality holds if and only if:

For , if ,, /i jX X U P∀ ∈ ( ) /i jX X Y U B∪ ⊆ ∈

there were :|)||(||| jrjir XDXXD ∩−∩

|)||(|||0 jrjir XDXXD ∩−∩∧==0 (r=1,2,…,k)

Proof: Let , ,1 2/ { , , , }sU P P P P= ⋯ 1 2/ { , , , }tU B B B B= ⋯according to the probative process of theorem 1,

we can easily get that: for any ,there/iB U B∈

exists , which causes1 2, , , /j j jhP P P U P∈⋯

.1 2i j j jhB P P P= ∪ ∪ ∪⋯

According to theorem 2, we can easily getthat: .( | ) ( | )I D P I D B≤

The equality holds if and only if:

For , if ,, /i jX X U P∀ ∈ ( ) /i jX X Y U B∪ ⊆ ∈

there were

: |)||(||| jrjir XDXXD ∩−∩

|)||(|||0 jrjir XDXXD ∩−∩∧==0(r=1,2,…,k)Note: The theorem illustrates that the

information entropy of the conditional attributessets is monotonically descending. So we candefine the importance of attributes as follows:

Definition5: In decision table S=(U,C,D,V,f),for , the significance of attribute (index)c C∀ ∈

is defined as :c

.( ) ( | ( { })) ( | )NewSig c I D C c I D C= − −

Theorem 4: In decision table S=(U,C,D,V,f),

for , NewSig(c)=0, then Sig(c)=0.c C∀ ∈

Note: Theorem 4 shows that if thesignificance degree of a certain attribute is zeroaccording to the definition of new attribute'ssignificance degree, then it is also zero with theoriginal attribute's significance degree. But theconverse proposition of theorem 4 does not hold,so we illustrate this with the decision table 1.

According to the foregoing calculation indecision table 1, we can get that : Sig(a)=Sig(b)=Sig(c)=0.

In decision table 1:U/{a,b,c}={{x1,x2},{x3},{x4},{x5},{x6,.x7},{x

8,x9},{x10,x11}}U/D={{x1,x5,x6,x8,x11},{x2,x3,x4,x7,x9,x10}},then:

2

21 1

| | | || |( | ) 1

| | | || |

m kj i j ii

i j i i

D C D CCI D CC CU= =

∩ ∩⎛ ⎞= −⎜ ⎟

⎝ ⎠∑ ∑ 11

121=

U/{a,b}={{x1,x2},{x3},{x4},{x5},{x6,.x7,x8,x9},{x10,x11}}( | ( { }))I D C c−

+⎟⎠⎞

⎜⎝⎛ −+−⎟

⎠⎞

⎜⎝⎛= )

211(

21)

211(

21

112 2

⎟⎠⎞

⎜⎝⎛ −+−⎟

⎠⎞

⎜⎝⎛ )

111(

11)

101(

10

111 2

+ +⎟⎠⎞

⎜⎝⎛ −+−⎟

⎠⎞

⎜⎝⎛ )

111(

11)

101(

10

111 2

⎟⎠⎞

⎜⎝⎛ −+−⎟

⎠⎞

⎜⎝⎛ )

101(

10)

111(

11

111 2

+⎟⎠⎞

⎜⎝⎛ −+−⎟

⎠⎞

⎜⎝⎛ )

421(

42)

421(

42

114 2

⎟⎠⎞

⎜⎝⎛ −+−⎟

⎠⎞

⎜⎝⎛+ )

211(

21)

211(

21

112 2

.15121

=

Thus:( ) ( |( { })) ( | ) 15/121 11/121 4/121 0Sig c I D C c I D C= − − = − = >

Similar that: ,( ) 4 /121 0Sig b = >

.( ) 6 /121 0Sig a = >

The example shows that the significance ofeach conditional attribute is not zero according tothe new significance degree. This example andtheorem 4 show that the new definition ofattribute’s significance degree is more reasonable.But when we calculating the importance degreeof attributes in whichever decision table with thenew definition ,we cannot guarantee that everyattribute’s significance degree is note zero .Inorder to solve this problem, we give a newdefinition of attribute weight as follows.

IV.A new method of attribute weight assignmentbased on rough set

509509

Page 4: [IEEE 2009 International Conference on Electronic Commerce and Business Intelligence, ECBI - Beijing, China (2009.06.6-2009.06.7)] 2009 International Conference on Electronic Commerce

Definition 6:In decision table S=(U,C,D,V,f),,then conditional attribute (index) c’sc C∀ ∈

weight W(c) is defined as :

.

( ) ( | { })( ){ ( ) ( | { })}

a C

NewSig c I D cW cNewSig a I D a

+=

+∑Note: In this definition, NewSig(c) represents

conditional attribute c’s importance degree in thewhole set, and I(D|{c}) represents conditionalattribute c’ s importance degree in the system.

NewSig(c) represents the variation degree ofattribute set after removing attribute c, so it iseasy to cause some conditional attributes to beunnecessary when the number of attributes in thedecision table is large. The attribute’ssignificance degree is zero when applying themethod of weight assignment in literature [7].Weight represents the certain attribute’srelatively importance degree in the whole and theattribute’s objective weight should not be zero.This definition introduced the definition ofweight into the combination of NewSig(c)andI(D|{c}). On one hand, it is the overallinvestigation of the significance degree ofattribute in the whole set and in itself. On theother hand, it makes every attribute’s weight notto be zero and therefore makes weightassignment to be more reasonable.

Theorem 5: In decision table S=(U,C,D,V,f),

, there at least exists , suchc C∀ ∈ ,i jx x U∈

that and ,( , } ( , }i jf x c f x c= ( , } ( , }i jf x D f x D≠

then W(c)>0.Proof: Let { | ( , } ( , }i iD x x U f x D f x D= ∈ ∧ =

,{ | ( , } ( , }j jD x x U f x D f x D= ∈ ∧ =

due to , then and .( , } ( , }i jf x D f x D≠ i jD D≠ ,i i j jx D x D∈ ∈

Due to , then ,( , } ( , }i jf x c f x c= { }[ ]i j cx x∈

thus .Due to ,{ }[ ]i i j cx D x∈ ∩ i i i jx D x D∈ ∧ ∉

then , therefore{ } { }| [ ] | | [ ] |i j c j cD x x∩ <

.

2{ } { } { }2

{ } { }

| [ ] | | [ ] | | [ ] |1 0

| [ ] | | [ ] || |j c i j c i j c

j c j c

x D x D xx xU

⎛ ⎞∩ ∩− >⎜ ⎟⎜ ⎟

⎝ ⎠Then I(D|{c})>0, thus W(c)>0..According to the upper probative process, in a

decision table which contains at least twodecision variable: for any conditional attribute c,there exist more than two objects based on cwhich cannot be distinguished, but their decisionattribute values are different. That is, more thantwo objects’ values based on c are the same, butdifferent in their decision attribute. It shows that,in that decision table, we cannot get assured

decision result with a single attribute c. Underthis condition, all results of attribute weight arenot zero according to our method of attributeweight assignment based on rough set.

V. Conclusion

Shortcomings of existing methods of attributeweight assignment based on rough set have beendiscussed. The significant degree of rough setattributes has been analyzed. A new method ofattribute weight assignment based on rough sethas been proposed and justified. The new method,by considering both the overall significancedegree and the individual significance degree ofeach attribute in the conditional attributes,improves the ability of generalization andinterpretation.

References[1][1][1][1] Pawlak Z. Rough sets: probability versusdeterministic approach[J]. Int. J. Man- machineStudies, 1988, 29: 81~95[2][2][2][2] Hu X H, Cercone N. Learning in relationaldata bases: a rough set approach[J].Computational Inteligence, 1995, 11( 2) :323~338[3][3][3][3] Swiniarski S, Hargis L. Rough set as a frontend of neural-networks texture classifiers[J].Neuro -computing, 2001, 36(1): 85~102[4][4][4][4] Zhang W X, Wu W Z, Liang J Y. Thetheory and method of rough sets[M]. Beijing;Science Press, 2000.[5][5][5][5] Cao X Y, Liang J G. The Method ofAscertaining Weight Based on Rough SetsTheory[J]. Chinese Journal of ManagementScience, 2002, 10(5): 98~100[6][6][6][6] Jiang Y J, Chen J, Ruan X Y. Fuzzysimilarity-based rough set method for case-basedreasoning and its application in tool selection[J].International Journal of Machine Tools &Manufacture, 2006, 46(1):107~113[7][7][7][7] Zhou A F, Chen Z Y. How to Choose theSC Partner Based on Tough Set[J]. LogisticsTechnology, 2007,26(8): 178∼181[8][8][8][8] Huang D R, Song J, Hu B J. A New Methodof Fault Diagnosis Based on AttributeSignificance of Multi-field Experts Database[J].Compute Science, 2007,34(4):231~233

510510