In
computer science
Computer science is the study of computation, information, and automation. Computer science spans Theoretical computer science, theoretical disciplines (such as algorithms, theory of computation, and information theory) to Applied science, ...
, a rough set, first described by
Polish
Polish may refer to:
* Anything from or related to Poland, a country in Europe
* Polish language
* Polish people, people from Poland or of Polish descent
* Polish chicken
* Polish brothers (Mark Polish and Michael Polish, born 1970), American twin ...
computer scientist
Zdzisław I. Pawlak, is a formal approximation of a
crisp set
Crisp may refer to:
Foods
* Potato chips, known in British English as a potato crisp, a thin slice of a potato deep fried or baked until crispy
* Chip (snack), or crisp in British English
* Crisp (dessert), a type of American dessert, usually inc ...
(i.e., conventional set) in terms of a pair of sets which give the ''lower'' and the ''upper'' approximation of the original set. In the standard version of rough set theory described in Pawlak (1991),
the lower- and upper-approximation sets are crisp sets, but in other variations, the approximating sets may be
fuzzy set
Fuzzy or Fuzzies may refer to:
Music
* Fuzzy (band), a 1990s Boston indie pop band
* Fuzzy (composer), Danish composer Jens Vilhelm Pedersen (born 1939)
* Fuzzy (album), ''Fuzzy'' (album), 1993 debut album of American rock band Grant Lee Buffalo
...
s.
Definitions
The following section contains an overview of the basic framework of rough set theory, as originally proposed by
Zdzisław I. Pawlak, along with some of the key definitions. More formal properties and boundaries of rough sets can be found in and cited references. The initial and basic theory of rough sets is sometimes referred to as ''"Pawlak Rough Sets"'' or ''"classical rough sets"'', as a means to distinguish it from more recent extensions and generalizations.
Information system framework
Let
be an information system (
attribute–value system
An attribute–value system is a basic knowledge representation framework comprising a table with columns designating "attributes" (also known as "properties", "predicates", "features", " dimensions", "characteristics", "fields", "headers" or "inde ...
), where
is a non-empty, finite set of objects (the universe) and
is a non-empty, finite set of attributes such that
for every
.
is the set of values that attribute
may take. The information table assigns a value
from
to each attribute
and object
in the universe
.
With any
there is an associated
equivalence relation
In mathematics, an equivalence relation is a binary relation that is reflexive, symmetric, and transitive. The equipollence relation between line segments in geometry is a common example of an equivalence relation. A simpler example is equ ...
:
:
The relation
is called a
''-indiscernibility relation''. The partition of
is a family of all
equivalence class
In mathematics, when the elements of some set S have a notion of equivalence (formalized as an equivalence relation), then one may naturally split the set S into equivalence classes. These equivalence classes are constructed so that elements ...
es of
and is denoted by
(or
).
If
, then
and
are ''indiscernible'' (or indistinguishable) by attributes from
.
The equivalence classes of the
-indiscernibility relation are denoted
.
Example: equivalence-class structure
For example, consider the following information table:
:
When the full set of attributes
is considered, we see that we have the following seven equivalence classes:
:
Thus, the two objects within the first equivalence class,
, cannot be distinguished from each other based on the available attributes, and the three objects within the second equivalence class,
, cannot be distinguished from one another based on the available attributes. The remaining five objects are each discernible from all other objects.
It is apparent that different attribute subset selections will in general lead to different indiscernibility classes. For example, if attribute
alone is selected, we obtain the following, much coarser, equivalence-class structure:
:
Definition of a ''rough set''
Let
be a target set that we wish to represent using attribute subset
; that is, we are told that an arbitrary set of objects
comprises a single class, and we wish to express this class (i.e., this subset) using the equivalence classes induced by attribute subset
. In general,
cannot be expressed exactly, because the set may include and exclude objects which are indistinguishable on the basis of attributes
.
For example, consider the target set
, and let attribute subset
, the full available set of features. The set
cannot be expressed exactly, because in
, objects
are indiscernible. Thus, there is no way to represent any set
which ''includes''
but ''excludes'' objects
and
.
However, the target set
can be ''approximated'' using only the information contained within
by constructing the
-lower and
-upper approximations of
:
:
:
Lower approximation and positive region
The
''-lower approximation'', or ''positive region'', is the union of all equivalence classes in
which are contained by (i.e., are subsets of) the target set – in the example,
, the union of the two equivalence classes in
which are contained in the target set. The lower approximation is the complete set of objects in
that can be ''positively'' (i.e., unambiguously) classified as belonging to target set
.
Upper approximation and negative region
The
''-upper approximation'' is the union of all equivalence classes in
which have non-empty intersection with the target set – in the example,
, the union of the three equivalence classes in
that have non-empty intersection with the target set. The upper approximation is the complete set of objects that in
that ''cannot'' be positively (i.e., unambiguously) classified as belonging to the ''complement'' (
) of the target set
. In other words, the upper approximation is the complete set of objects that are ''possibly'' members of the target set
.
The set
therefore represents the ''negative region'', containing the set of objects that can be definitely ruled out as members of the target set.
Boundary region
The ''boundary region'', given by set difference
, consists of those objects that can neither be ruled in nor ruled out as members of the target set
.
In summary, the lower approximation of a target set is a ''conservative'' approximation consisting of only those objects which can positively be identified as members of the set. (These objects have no indiscernible "clones" which are excluded by the target set.) The upper approximation is a ''liberal'' approximation which includes all objects that might be members of target set. (Some objects in the upper approximation may not be members of the target set.) From the perspective of
, the lower approximation contains objects that are members of the target set with certainty (probability = 1), while the upper approximation contains objects that are members of the target set with non-zero probability (probability > 0).
The rough set
The tuple
composed of the lower and upper approximation is called a ''rough set''; thus, a rough set is composed of two crisp sets, one representing a ''lower boundary'' of the target set
, and the other representing an ''upper boundary'' of the target set
.
The ''accuracy'' of the rough-set representation of the set
can be given
by the following:
:
That is, the accuracy of the rough set representation of
,
,
, is the ratio of the number of objects which can ''positively'' be placed in
to the number of objects that can ''possibly'' be placed in
– this provides a measure of how closely the rough set is approximating the target set. Clearly, when the upper and lower approximations are equal (i.e., boundary region empty), then
, and the approximation is perfect; at the other extreme, whenever the lower approximation is empty, the accuracy is zero (regardless of the size of the upper approximation).
Objective analysis
Rough set theory is one of many methods that can be employed to analyse uncertain (including vague) systems, although less common than more traditional methods of
probability
Probability is a branch of mathematics and statistics concerning events and numerical descriptions of how likely they are to occur. The probability of an event is a number between 0 and 1; the larger the probability, the more likely an e ...
,
statistics
Statistics (from German language, German: ', "description of a State (polity), state, a country") is the discipline that concerns the collection, organization, analysis, interpretation, and presentation of data. In applying statistics to a s ...
,
entropy
Entropy is a scientific concept, most commonly associated with states of disorder, randomness, or uncertainty. The term and the concept are used in diverse fields, from classical thermodynamics, where it was first recognized, to the micros ...
and
Dempster–Shafer theory
The theory of belief functions, also referred to as evidence theory or Dempster–Shafer theory (DST), is a general framework for reasoning with uncertainty, with understood connections to other frameworks such as probability, possibility and ...
. However a key difference, and a unique strength, of using classical rough set theory is that it provides an objective form of analysis.
Unlike other methods, as those given above, classical rough set analysis requires no additional information, external parameters, models, functions, grades or subjective interpretations to determine set membership – instead it only uses the information presented within the given data.
More recent adaptations of rough set theory, such as dominance-based, decision-theoretic and fuzzy rough sets, have introduced more subjectivity to the analysis.
Definability
In general, the upper and lower approximations are not equal; in such cases, we say that target set
is ''undefinable'' or ''roughly definable'' on attribute set
. When the upper and lower approximations are equal (i.e., the boundary is empty),
, then the target set
is ''definable'' on attribute set
. We can distinguish the following special cases of undefinability:
* Set
is ''internally'' ''undefinable'' if
and
. This means that on attribute set
, there are ''no'' objects which we can be certain belong to target set
, but there ''are'' objects which we can definitively exclude from set
.
* Set
is ''externally undefinable'' if
and
. This means that on attribute set
, there ''are'' objects which we can be certain belong to target set
, but there are ''no'' objects which we can definitively exclude from set
.
* Set
is ''totally undefinable'' if
and
. This means that on attribute set
, there are ''no'' objects which we can be certain belong to target set
, and there are ''no'' objects which we can definitively exclude from set
. Thus, on attribute set
, we cannot decide whether any object is, or is not, a member of
.
Reduct and core
An interesting question is whether there are attributes in the information system (attribute–value table) which are more important to the knowledge represented in the equivalence class structure than other attributes. Often, we wonder whether there is a subset of attributes which can, by itself, fully characterize the knowledge in the database; such an attribute set is called a ''reduct''.
Formally, a reduct is a subset of attributes
such that
*
=
, that is, the equivalence classes induced by the reduced attribute set
are the same as the equivalence class structure induced by the full attribute set
.
* the attribute set
is ''minimal'', in the sense that
for any attribute
; in other words, no attribute can be removed from set
without changing the equivalence classes
.
A reduct can be thought of as a ''sufficient'' set of features – sufficient, that is, to represent the category structure. In the example table above, attribute set
is a reduct – the information system projected on just these attributes possesses the same equivalence class structure as that expressed by the full attribute set:
:
Attribute set
is a reduct because eliminating any of these attributes causes a collapse of the equivalence-class structure, with the result that
.
The reduct of an information system is ''not unique'': there may be many subsets of attributes which preserve the equivalence-class structure (i.e., the knowledge) expressed in the information system. In the example information system above, another reduct is
, producing the same equivalence-class structure as
.
The set of attributes which is common to all reducts is called the ''core'': the core is the set of attributes which is possessed by ''every'' reduct, and therefore consists of attributes which cannot be removed from the information system without causing collapse of the equivalence-class structure. The core may be thought of as the set of ''necessary'' attributes – necessary, that is, for the category structure to be represented. In the example, the only such attribute is
; any one of the other attributes can be removed singly without damaging the equivalence-class structure, and hence these are all ''dispensable''. However, removing
by itself ''does'' change the equivalence-class structure, and thus
is the ''indispensable'' attribute of this information system, and hence the core.
It is possible for the core to be empty, which means that there is no indispensable attribute: any single attribute in such an information system can be deleted without altering the equivalence-class structure. In such cases, there is no ''essential'' or necessary attribute which is required for the class structure to be represented.
Attribute dependency
One of the most important aspects of database analysis or data acquisition is the discovery of attribute dependencies; that is, we wish to discover which variables are strongly related to which other variables. Generally, it is these strong relationships that will warrant further investigation, and that will ultimately be of use in predictive modeling.
In rough set theory, the notion of dependency is defined very simply. Let us take two (disjoint) sets of attributes, set
and set
, and inquire what degree of dependency obtains between them. Each attribute set induces an (indiscernibility) equivalence class structure, the equivalence classes induced by
given by
, and the equivalence classes induced by
given by
.
Let
, where
is a given equivalence class from the equivalence-class structure induced by attribute set
. Then, the ''dependency'' of attribute set
on attribute set
,
, is given by
:
That is, for each equivalence class
in
, we add up the size of its lower approximation by the attributes in
, i.e.,
. This approximation (as above, for arbitrary set
) is the number of objects which on attribute set
can be positively identified as belonging to target set
. Added across all equivalence classes in
, the numerator above represents the total number of objects which – based on attribute set
– can be positively categorized according to the classification induced by attributes
. The dependency ratio therefore expresses the proportion (within the entire universe) of such classifiable objects. The dependency
"can be interpreted as a proportion of such objects in the information system for which it suffices to know the values of attributes in
to determine the values of attributes in
".
Another, intuitive, way to consider dependency is to take the partition induced by
as the target class
, and consider
as the attribute set we wish to use in order to "re-construct" the target class
. If
can completely reconstruct
, then
depends totally upon
; if
results in a poor and perhaps a random reconstruction of
, then
does not depend upon
at all.
Thus, this measure of dependency expresses the degree of ''functional'' (i.e., deterministic) dependency of attribute set
on attribute set
; it is ''not'' symmetric. The relationship of this notion of attribute dependency to more traditional information-theoretic (i.e., entropic) notions of attribute dependence has been discussed in a number of sources, e.g. Pawlak, Wong, & Ziarko (1988),
Yao & Yao (2002),
Wong, Ziarko, & Ye (1986),
and Quafafou & Boussouf (2000).
Rule extraction
The category representations discussed above are all ''extensional'' in nature; that is, a category or complex class is simply the sum of all its members. To represent a category is, then, just to be able to list or identify all the objects belonging to that category. However, extensional category representations have very limited practical use, because they provide no insight for deciding whether novel (never-before-seen) objects are members of the category.
What is generally desired is an ''intensional'' description of the category, a representation of the category based on a set of ''rules'' that describe the scope of the category. The choice of such rules is not unique, and therein lies the issue of
inductive bias
The inductive bias (also known as learning bias) of a learning algorithm is the set of assumptions that the learner uses to predict outputs of given inputs that it has not encountered.
Inductive bias is anything which makes the algorithm learn o ...
. See
Version space
Version space learning is a logical approach to machine learning, specifically binary classification. Version space learning algorithms search a predefined space of hypotheses, viewed as a set of logical sentences. Formally, the hypothesis space i ...
and
Model selection
Model selection is the task of selecting a model from among various candidates on the basis of performance criterion to choose the best one.
In the context of machine learning and more generally statistical analysis, this may be the selection of ...
for more about this issue.
There are a few rule-extraction methods. We will start from a rule-extraction procedure based on Ziarko & Shan (1995).
Decision matrices
Let us say that we wish to find the minimal set of consistent rules (
logical implication
Logic is the study of correct reasoning. It includes both formal and informal logic. Formal logic is the study of deductively valid inferences or logical truths. It examines how conclusions follow from premises based on the structure of ...
s) that characterize our sample system. For a set of ''condition'' attributes
and a decision attribute
, these rules should have the form
, or, spelled out,
:
where
are legitimate values from the domains of their respective attributes. This is a form typical of
association rules
Association rule learning is a rule-based machine learning method for discovering interesting relations between variables in large databases. It is intended to identify strong rules discovered in databases using some measures of interestingness.Pi ...
, and the number of items in
which match the condition/antecedent is called the ''support'' for the rule. The method for extracting such rules given in is to form a ''decision matrix'' corresponding to each individual value
of decision attribute
. Informally, the decision matrix for value
of decision attribute
lists all attribute–value pairs that ''differ'' between objects having
and
.
This is best explained by example (which also avoids a lot of notation). Consider the table above, and let
be the decision variable (i.e., the variable on the right side of the implications) and let
be the condition variables (on the left side of the implication). We note that the decision variable
takes on two different values, namely
. We treat each case separately.
First, we look at the case
, and we divide up
into objects that have
and those that have
. (Note that objects with
in this case are simply the objects that have
, but in general,
would include all objects having any value for
''other than''
, and there may be several such classes of objects (for example, those having
).) In this case, the objects having
are
while the objects which have
are
. The decision matrix for
lists all the differences between the objects having
and those having
; that is, the decision matrix lists all the differences between
and
. We put the "positive" objects (
) as the rows, and the "negative" objects
as the columns.
:
To read this decision matrix, look, for example, at the intersection of row
and column
, showing
in the cell. This means that ''with regard to'' decision value
, object
differs from object
on attributes
and
, and the particular values on these attributes for the positive object
are
and
. This tells us that the correct classification of
as belonging to decision class
rests on attributes
and
; although one or the other might be dispensable, we know that ''at least one'' of these attributes is ''in''dispensable.
Next, from each decision matrix we form a set of
Boolean expressions, one expression for each row of the matrix. The items within each cell are aggregated disjunctively, and the individuals cells are then aggregated conjunctively. Thus, for the above table we have the following five Boolean expressions:
:
Each statement here is essentially a highly specific (probably ''too'' specific) rule governing the membership in class
of the corresponding object. For example, the last statement, corresponding to object
, states that all the following must be satisfied:
# Either
must have value 2, or
must have value 0, or both.
#
must have value 0.
# Either
must have value 2, or
must have value 0, or both.
# Either
must have value 2, or
must have value 0, or
must have value 0, or any combination thereof.
#
must have value 0.
It is clear that there is a large amount of redundancy here, and the next step is to simplify using traditional
Boolean algebra
In mathematics and mathematical logic, Boolean algebra is a branch of algebra. It differs from elementary algebra in two ways. First, the values of the variable (mathematics), variables are the truth values ''true'' and ''false'', usually denot ...
. The statement
corresponding to objects
simplifies to
, which yields the implication
:
Likewise, the statement
corresponding to objects
simplifies to
. This gives us the implication
:
The above implications can also be written as the following rule set:
:
It can be noted that each of the first two rules has a ''support'' of 1 (i.e., the antecedent matches two objects), while each of the last two rules has a support of 2. To finish writing the rule set for this knowledge system, the same procedure as above (starting with writing a new decision matrix) should be followed for the case of
, thus yielding a new set of implications for that decision value (i.e., a set of implications with
as the consequent). In general, the procedure will be repeated for each possible value of the decision variable.
LERS rule induction system
The data system LERS (Learning from Examples based on Rough Sets)
may induce rules from inconsistent data, i.e., data with conflicting objects. Two objects are conflicting when they are characterized by the same values of all attributes, but they belong to different concepts (classes). LERS uses rough set theory to compute lower and upper approximations for concepts involved in conflicts with other concepts.
Rules induced from the lower approximation of the concept ''certainly'' describe the concept, hence such rules are called ''certain''. On the other hand, rules induced from the upper approximation of the concept describe the concept ''possibly'', so these rules are called ''possible''. For rule induction LERS uses three algorithms: LEM1, LEM2, and IRIM.
The LEM2 algorithm of LERS is frequently used for rule induction and is used not only in LERS but also in other systems, e.g., in RSES.
LEM2 explores the search space of
attribute–value pairs. Its input data set is a lower or upper approximation of a concept, so its input data set is always consistent. In general, LEM2 computes a local covering and then converts it into a rule set. We will quote a few definitions to describe the LEM2 algorithm.
The LEM2 algorithm is based on an idea of an attribute–value pair block. Let
be a nonempty lower or upper approximation of a concept represented by a decision-value pair
. Set
''depends'' on a set
of attribute–value pairs
if and only if
:
Set
is a ''minimal complex'' of
if and only if
depends on
and no proper subset
of
exists such that
depends on
. Let
be a nonempty collection of nonempty sets of attribute–value pairs. Then
is a ''local covering'' of
if and only if the following three conditions are satisfied:
each member
of
is a minimal complex of
,
:
:
is minimal, i.e.,
has the smallest possible number of members.
For our sample information system, LEM2 will induce the following rules:
:
Other rule-learning methods can be found, e.g., in Pawlak (1991),
Stefanowski (1998),
Bazan et al. (2004),
etc.
Incomplete data
Rough set theory is useful for rule induction from incomplete data sets. Using this approach we can distinguish between three types of missing attribute values: ''lost values'' (the values that were recorded but currently are unavailable), ''attribute-concept values'' (these missing attribute values may be replaced by any attribute value limited to the same concept), and ''"do not care" conditions'' (the original values were irrelevant). A ''concept'' (''class'') is a set of all objects classified (or diagnosed) the same way.
Two special data sets with missing attribute values were extensively studied: in the first case, all missing attribute values were lost,
in the second case, all missing attribute values were "do not care" conditions.
In attribute-concept values interpretation of a missing attribute value, the missing attribute value may be replaced by any value of the attribute domain restricted to the concept to which the object with a missing attribute value belongs.
For example, if for a patient the value of an attribute Temperature is missing, this patient is sick with flu, and all remaining patients sick with flu have values high or very-high for Temperature when using the interpretation of the missing attribute value as the attribute-concept value, we will replace the missing attribute value with high and very-high. Additionally, the ''characteristic relation'', (see, e.g., ) enables to process data sets with all three kind of missing attribute values at the same time: lost, "do not care" conditions, and attribute-concept values.
Applications
Rough set methods can be applied as a component of hybrid solutions in
machine learning
Machine learning (ML) is a field of study in artificial intelligence concerned with the development and study of Computational statistics, statistical algorithms that can learn from data and generalise to unseen data, and thus perform Task ( ...
and
data mining
Data mining is the process of extracting and finding patterns in massive data sets involving methods at the intersection of machine learning, statistics, and database systems. Data mining is an interdisciplinary subfield of computer science and ...
. They have been found to be particularly useful for
rule induction
Rule induction is an area of machine learning in which formal rules are extracted from a set of observations. The rules extracted may represent a full scientific model of the data, or merely represent local patterns in the data.
Data mining in ...
and
feature selection
In machine learning, feature selection is the process of selecting a subset of relevant Feature (machine learning), features (variables, predictors) for use in model construction. Feature selection techniques are used for several reasons:
* sim ...
(semantics-preserving
dimensionality reduction
Dimensionality reduction, or dimension reduction, is the transformation of data from a high-dimensional space into a low-dimensional space so that the low-dimensional representation retains some meaningful properties of the original data, ideally ...
). Rough set-based data analysis methods have been successfully applied in
bioinformatics
Bioinformatics () is an interdisciplinary field of science that develops methods and Bioinformatics software, software tools for understanding biological data, especially when the data sets are large and complex. Bioinformatics uses biology, ...
,
economics
Economics () is a behavioral science that studies the Production (economics), production, distribution (economics), distribution, and Consumption (economics), consumption of goods and services.
Economics focuses on the behaviour and interac ...
and finance, medicine, multimedia, web and
text mining
Text mining, text data mining (TDM) or text analytics is the process of deriving high-quality information from text. It involves "the discovery by computer of new, previously unknown information, by automatically extracting information from differe ...
, signal and image processing,
software engineering
Software engineering is a branch of both computer science and engineering focused on designing, developing, testing, and maintaining Application software, software applications. It involves applying engineering design process, engineering principl ...
, robotics, and engineering (e.g. power systems and
control engineering
Control engineering, also known as control systems engineering and, in some European countries, automation engineering, is an engineering discipline that deals with control systems, applying control theory to design equipment and systems with d ...
). Recently the three regions of rough sets are interpreted as regions of acceptance, rejection and deferment. This leads to three-way decision making approach with the model which can potentially lead to interesting future applications.
History
The idea of rough set was proposed by
Pawlak (1981) as a new mathematical tool to deal with vague concepts. Comer, Grzymala-Busse, Iwinski, Nieminen, Novotny, Pawlak, Obtulowicz, and Pomykala have studied algebraic properties of rough sets. Different algebraic semantics have been developed by P. Pagliani, I. Duntsch, M. K. Chakraborty, M. Banerjee and A. Mani; these have been extended to more generalized rough sets by D. Cattaneo and A. Mani, in particular. Rough sets can be used to represent
ambiguity
Ambiguity is the type of meaning (linguistics), meaning in which a phrase, statement, or resolution is not explicitly defined, making for several interpretations; others describe it as a concept or statement that has no real reference. A com ...
,
vagueness
In linguistics and philosophy, a vague predicate is one which gives rise to borderline cases. For example, the English adjective "tall" is vague since it is not clearly true or false for someone of middling height. By contrast, the word " prime" ...
and general
uncertainty
Uncertainty or incertitude refers to situations involving imperfect or unknown information. It applies to predictions of future events, to physical measurements that are already made, or to the unknown, and is particularly relevant for decision ...
.
Extensions and generalizations
Since the development of rough sets, extensions and generalizations have continued to evolve. Initial developments focused on the relationship - both similarities and difference - with
fuzzy sets
Fuzzy or Fuzzies may refer to:
Music
* Fuzzy (band), a 1990s Boston indie pop band
* Fuzzy (composer), Danish composer Jens Vilhelm Pedersen (born 1939)
* Fuzzy (album), ''Fuzzy'' (album), 1993 debut album of American rock band Grant Lee Buffalo
...
. While some literature contends these concepts are different, other literature considers that rough sets are a generalization of fuzzy sets - as represented through either fuzzy rough sets or rough fuzzy sets. Pawlak (1995) considered that fuzzy and rough sets should be treated as being complementary to each other, addressing different aspects of uncertainty and vagueness.
Three notable extensions of classical rough sets are:
*
Dominance-based rough set approach
The dominance-based rough set approach (DRSA) is an extension of rough set theory for multi-criteria decision analysis (MCDA), introduced by Greco, Matarazzo and Słowiński.
Greco, S., Matarazzo, B., Słowiński, R.: Rough sets theory for multi- ...
(DRSA) is an extension of rough set theory for
multi-criteria decision analysis
Multiple-criteria decision-making (MCDM) or multiple-criteria decision analysis (MCDA) is a sub-discipline of operations research that explicitly evaluates multiple conflicting criteria in decision making (both in daily life and in settings ...
(MCDA), introduced by Greco, Matarazzo and Słowiński (2001).
The main change in this extension of classical rough sets is the substitution of the indiscernibility relation by a ''dominance'' relation, which permits the formalism to deal with inconsistencies typical in consideration of criteria and preference-ordered decision classes.
*
Decision-theoretic rough sets
In the mathematical theory of decisions, decision-theoretic rough sets (DTRS) is a probabilistic extension of rough set classification. First created in 1990 by Dr. Yiyu Yao, the extension makes use of loss functions to derive \textstyle \alpha ...
(DTRS) is a probabilistic extension of rough set theory introduced by Yao, Wong, and Lingras (1990).
It utilizes a Bayesian decision procedure for minimum risk decision making. Elements are included into the lower and upper approximations based on whether their conditional probability is above thresholds
and
. These upper and lower thresholds determine region inclusion for elements. This model is unique and powerful since the thresholds themselves are calculated from a set of six loss functions representing classification risks.
*
Game-theoretic rough sets (GTRS) is a game theory-based extension of rough set that was introduced by Herbert and Yao (2011).
It utilizes a game-theoretic environment to optimize certain criteria of rough sets based classification or decision making in order to obtain effective region sizes.
Rough membership
Rough sets can be also defined, as a generalisation, by employing a rough membership function instead of objective approximation. The rough membership function expresses a conditional probability that
belongs to
given
. This can be interpreted as a degree that
belongs to
in terms of information about
expressed by
.
Rough membership primarily differs from the fuzzy membership in that the membership of union and intersection of sets cannot, in general, be computed from their constituent membership as is the case of fuzzy sets. In this, rough membership is a generalization of fuzzy membership. Furthermore, the rough membership function is grounded more in probability than the conventionally held concepts of the fuzzy membership function.
Other generalizations
Several generalizations of rough sets have been introduced, studied and applied to solving problems. Here are some of these generalizations:
*Rough multisets
*Fuzzy rough sets extend the rough set concept through the use of fuzzy equivalence classes
*Alpha rough set theory (α-RST) - a generalization of rough set theory that allows approximation using of fuzzy concepts
*Intuitionistic fuzzy rough sets
*Generalized rough fuzzy sets
*Rough intuitionistic fuzzy sets
*Soft rough fuzzy sets and soft fuzzy rough sets
*Composite rough sets
See also
*
Algebraic semantics
*
Alternative set theory
In mathematical logic, an alternative set theory is any of the alternative mathematical approaches to the concept of set and any alternative to the de facto standard set theory described in axiomatic set theory by the axioms of Zermelo–Fraenke ...
*
Analog computer
An analog computer or analogue computer is a type of computation machine (computer) that uses physical phenomena such as Electrical network, electrical, Mechanics, mechanical, or Hydraulics, hydraulic quantities behaving according to the math ...
*
Description logic
Description logics (DL) are a family of formal knowledge representation languages. Many DLs are more expressive than propositional logic but less expressive than first-order logic. In contrast to the latter, the core reasoning problems for DLs are ...
*
Fuzzy logic
Fuzzy logic is a form of many-valued logic in which the truth value of variables may be any real number between 0 and 1. It is employed to handle the concept of partial truth, where the truth value may range between completely true and completely ...
*
Fuzzy set theory
Fuzzy or Fuzzies may refer to:
Music
* Fuzzy (band), a 1990s Boston indie pop band
* Fuzzy (composer), Danish composer Jens Vilhelm Pedersen (born 1939)
* ''Fuzzy'' (album), 1993 debut album of American rock band Grant Lee Buffalo
* "Fuzzy", a ...
*
Granular computing
Granular computing is an emerging computing paradigm of Data processing, information processing that concerns the processing of complex information entities called "information granulation, granules", which arise in the process of data abstractio ...
*
Near sets
*
Rough fuzzy hybridization
*
Type-2 fuzzy sets and systems Type-2 fuzzy sets and systems generalize standard fuzzy sets, type-1 fuzzy sets and fuzzy sets and systems, systems so that more uncertainty can be handled. From the beginning of fuzzy sets, criticism was made about the fact that the membership func ...
*
Decision-theoretic rough sets
In the mathematical theory of decisions, decision-theoretic rough sets (DTRS) is a probabilistic extension of rough set classification. First created in 1990 by Dr. Yiyu Yao, the extension makes use of loss functions to derive \textstyle \alpha ...
*
Version space
Version space learning is a logical approach to machine learning, specifically binary classification. Version space learning algorithms search a predefined space of hypotheses, viewed as a set of logical sentences. Formally, the hypothesis space i ...
*
Dominance-based rough set approach
The dominance-based rough set approach (DRSA) is an extension of rough set theory for multi-criteria decision analysis (MCDA), introduced by Greco, Matarazzo and Słowiński.
Greco, S., Matarazzo, B., Słowiński, R.: Rough sets theory for multi- ...
References
Further reading
* Gianpiero Cattaneo and Davide Ciucci, "Heyting Wajsberg Algebras as an Abstract Environment Linking Fuzzy and Rough Sets" in J.J. Alpigini et al. (Eds.): RSCTC 2002, LNAI 2475, pp. 77–84, 2002.
*
* Pawlak, Zdzisław ''Rough Sets'' Research Report PAS 431, Institute of Computer Science, Polish Academy of Sciences (1981)
*
*
*
*
*}
*
*
*
*Zhang J., Wong J-S, Pan Y, Li T. (2015). A parallel matrix-based method for computing approximations in incomplete information systems, IEEE Transactions on Knowledge and Data Engineering, 27(2): 326-339
*Burgin M. (1990). Theory of Named Sets as a Foundational Basis for Mathematics, In Structures in mathematical theories: Reports of the San Sebastian international symposium, September 25–29, 1990 (http://www.blogg.org/blog-30140-date-2005-10-26.html)
*Burgin, M. (2004). Unified Foundations of Mathematics, Preprint Mathematics LO/0403186, p39. (electronic edition: https://arxiv.org/ftp/math/papers/0403/0403186.pdf)
*Burgin, M. (2011), Theory of Named Sets, Mathematics Research Developments, Nova Science Pub Inc,
*Chen H., Li T., Luo C., Horng S-J., Wang G. (2015). A decision-theoretic rough set approach for dynamic data mining. IEEE Transactions on Fuzzy Systems, 23(6): 1958-1970
*Chen H., Li T., Luo C., Horng S-J., Wang G. (2014). A rough set-based method for updating decision rules on attribute values' coarsening and refining, IEEE Transactions on Knowledge and Data Engineering, 26(12): 2886-2899
*Chen H., Li T., Ruan D., Lin J., Hu C, (2013) A rough-set based incremental approach for updating approximations under dynamic maintenance environments. IEEE Transactions on Knowledge and Data Engineering, 25(2): 274-284
External links
The International Rough Set SocietyRough set tutorialRough Set Exploration SystemRough Sets in Data Warehousing
{{Authority control
Systems of set theory
Theoretical computer science
Approximations