7.2
CiteScore
3.7
Impact Factor
Generic selectors
Exact matches only
Search in title
Search in content
Post Type Selectors
Search in posts
Search in pages
Filter by Categories
ABUNDANCE ESTIMATION IN AN ARID ENVIRONMENT
Case Study
Editorial
Invited review
Letter to the Editor
Original Article
REVIEW
Review Article
SHORT COMMUNICATION
7.2
CiteScore
3.7
Impact Factor
Generic selectors
Exact matches only
Search in title
Search in content
Post Type Selectors
Search in posts
Search in pages
Filter by Categories
ABUNDANCE ESTIMATION IN AN ARID ENVIRONMENT
Case Study
Editorial
Invited review
Letter to the Editor
Original Article
REVIEW
Review Article
SHORT COMMUNICATION
View/Download PDF

Translate this page into:

ORIGINAL ARTICLE
23 (
1
); 41-45
doi:
10.1016/j.jksus.2010.06.005

Near set theory and its generalization

Department of Mathematics, Faculty of Science, Tanta University, Egypt

*Corresponding author. Tel.: +20 565344419 via-marei@yahoo.com (E.A. Marei)

Disclaimer:
This article was originally published by Elsevier and was migrated to Scientific Scholar after the change of Publisher.

Available online 20 June 2010

Abstract

In this work we aim to use general topological structures as tools for near approximation space in information systems. General relations for granules form a subbase for some topological spaces. Theses topologies are applied for obtaining near lower and upper approximations. We apply it to obtain a topological structure which opens up the way for applying rich amount of topological facts and methods in the process of granular computing.

Keywords

Information system
Approximation space
Topological space
Rough set
Feature
Near set
1

1 Introduction

Rough set theory, proposed by Pawlak (1991, 1982), can be seen as a new mathematical approach to vagueness. The rough set philosophy is founded on the assumption that with every object of the universe of discourse we associate some information (data, knowledge) Abu-Donia et al. (2007). For example, if objects are patients suffering from a certain disease, symptoms of the disease form information about patients. Objects characterized by the same information are indiscernible (similar) in view of the available information about them. The indiscernibility relation generated in this way is the mathematical basis of rough set theory. This understanding of indiscernibility is related to the idea of Gottfried Wilhelm Leibniz that objects are indiscernible if and only if all available functionals take on identical values (Leibnizs Law of Indiscernibility: The Identity of Indiscernibles) (Ariew et al., 1989). However, in the rough set approach, indiscernibility is defined relative to a given set of functionals (attributes). Any set of all indiscernible (similar) objects is called an elementary set, and forms a basic granule (atom) of knowledge about the universe. Any union of some elementary sets is referred to as a crisp (precise) set. A set which is not crisp is called rough (imprecise, vague).

Consequently, each rough set has boundary region cases, i.e., objects which cannot with certainty be classified either as members of the set or of its complement. Obviously crisp sets have no boundary region elements at all. This means that boundary region cases cannot be properly classified by employing available knowledge.

Thus, the assumption that objects can be seen only through the information available about them leads to the view that knowledge has a granular structure. Due to the granularity of knowledge, some objects of interest cannot be discerned and appear as the same (or similar). As a consequence, vague concepts, in contrast to precise concepts, cannot be characterized in terms of information about their elements.

Ultimately, there is interest in selecting the probe functions (Peters, 2007) that lead to descriptions of objects that are minimally near each other. This is an essential idea in the near set approach (Peters, 2007; Peters et al., 2007) and differs markedly from the minimum description length (MDL) proposed in 1983 by Jorma Rissanen. MDL depends on the identification of possible data models and possible probability models. By contrast, NDP deals with a set X that is the domain of a description used to identify similar objects. The term similar is used here to denote the presence of objects that have descriptions that match each other to some degree.

The near set approach leads to partitions of ensembles of sample objects with measurable information content and an approach to feature selection. The proposed feature selection method considers combinations of n probe functions taken r at a time in searching for those combinations of probe functions that lead to partitions of a set of objects that has the highest information content.

In this work, we assume that any vague concept is replaced by a pair of precise concepts, called the lower and the upper approximations of the vague concept. The lower approximation consists of all objects which surely belong to the concept and the upper approximation contains all objects which possibly belong to the concept. The difference between the upper and the lower approximation constitutes the boundary region of the vague concept. These approximations are two basic operations in rough set theory. There is a chance to be useful in the analysis of sample data. The proposed approach does not depend on the joint probability of finding a feature value for an input vectors that belong to the same class. In addition, the proposed approach to measuring the information content of families of neighborhoods differs from the rough set. The near set approach does not depend on preferential ordering of value sets of functions representing object features. The contribution of this research is the introduction of a generalization of near set approach to feature selection.

2

2 Preliminaries

Rough set theory expresses vagueness, not by means of membership, but employing a boundary region of a set. If the boundary region of a set is empty, it means that the set is crisp, otherwise the set is rough (inexact). Nonempty boundary region of a set means that our knowledge about the set is not sufficient to define the set precisely.

Suppose we are given a set of objects U called the universe and an indiscernibility relation E U × U , representing our lack of knowledge about elements of U. For the sake of simplicity we assume that E is an equivalence relation and X be a subset of U, we want to characterize the set X with respect to E. To this end we will need the basic concepts of rough set theory given below (Pawlak, 1982).

The equivalence class of E determined by element x is: [ x ] E = { x X : E ( x ) = E ( x ) } . Hence E-lower, upper approximations and boundary region of X are: E ̲ ( X ) = { [ x ] E : X U , [ x ] E X } ; E ¯ ( X ) = { [ x ] E : X U , [ x ] E X ϕ } ; BND E ( X ) = E ¯ ( X ) - E ̲ ( X ) . It is easily seen that approximations are in fact interior and closure operations in a topology generated by the indiscernibility relation (Abd El-Monsef et al., 2010).

The rough membership function is a degree that x belongs to X in view of information about expressed by E. It defined as (Pawlak and Skowron, 1994): μ X E ( x ) : U [ 0 , 1 ] , μ X E ( x ) = X [ x ] E [ x ] E , where | * | denotes the cardinality of * .

A rough set can also be characterized numerically by the accuracy measure of an approximation (Pawlak, 1991) which is defined as: α E ( X ) = E ̲ ( X ) E ¯ ( X ) .

Obviously, 0 α E ( X ) 1 . If α E ( X ) = 1 , X is crisp with respect to E (X is precise with respect to E), and otherwise, if α E ( X ) < 1 , X is rough with respect to E (X is vague with respect to E).

Underlying the study of near set theory is an interest in classifying sample objects by means of probe functions associated with object features. More recently, the term feature is defined as the form, fashion or shape (of an object).

Let F denotes a set of features for objects in a set X. For any feature a F , we associate a function f a that maps X to some set V f a (range of f a ).

The value of f a ( x ) is a measurement associated with feature a of an object x X . The function f a is called a probe function (Pavel, 1993).

The following concepts introduced by Peters (2007) and Peters et al. (2006).

GAS = ( U , F , N r , ν B ) is a generalized approximation space, where U is a universe of objects, F is a set of functions representing object features, N r is a neighborhood family function defined as N r ( F ) = A P r ( F ) [ x ] A , where P r ( F ) = { A F : A = r , 1 r F } . And ν B r is an overlap function defined by ν B r : P ( U ) × P ( U ) [ 0 , 1 ] , ν B r ( Y , N r ( B ) * X ) = Y N r ( B ) * X N r ( B ) * X , where N r ( B ) * X ϕ , Y is a member of the family of neighborhoods N r ( B ) and ν B r ( Y , N r ( B ) * X ) is equal to 1, if N r ( B ) * X = ϕ .

The overlap function ν B r maps a pair of sets to a number in [ 0 , 1 ] representing the degree of overlap between the sets of objects with features B r .

N r ( B ) -lower, upper approximations and boundary region of a set X with respect to r features from the probe functions B are defined as: N r ( B ) * X = x : [ x ] B r X [ x ] B r ; N r ( B ) * X = x : [ x ] B r X ϕ [ x ] B r ; BND N r ( B ) X = N r ( B ) * X - N r ( B ) * X . Peters introduces the following concepts:

Objects x and x are minimally near each other if f B such that f ( x ) = f ( x ) . A set X is near to X if x X , x X such that x and x are near objects. A set X is termed a near set relative to a chosen family of neighborhoods N r ( B ) if | BND N r ( B ) X | 0 .

3

3 Generalized near set theory

In the following, we use a general relation to deduce a new approach to near set theory, consequently we obtain a new general near lower (upper) approximation for any near set. Also we introduce a modification of some concepts.

Definition 3.1

Let f B be a general relation on a nonempty set X. A special neighborhood of an element x X is ( x ) f r = { y X : f ( y ) - f ( x ) r } , where | * | is the absolute value of * and r is the length of neighborhood with respect to the feature f.

Remark 3.1

We will replace the equivalence class in the classical approximations defined by Peters by the special neighborhood defined in Definition 3.1.

Definition 3.2

Let B F be a set of functions representing features of x , x X . Objects x and x are minimally near each other if f B such that | f ( x ) - f ( x ) | r , where r is the length of a special neighborhood defined in Definition 3.1, with respect to the feature f. Denoted by xN f x .

Definition 3.3

Let Y , Y X and B F . A set Y is near to Y if x Y , x Y and f B such that xN f x . Denoted by YN f Y .

Remark 3.2

We can determine the degree of nearness K between two sets Y and Y as the following: K = ϕ i B : YN ϕ i Y B .

Theorem 3.1

Let x , y X and f B . Then x is near to y if x ( y ) f r or y ( x ) f r .

Proof

From Definitions 3.1 and 3.2, we get the proof.  □

Theorem 3.2

Any subset of X is near to X.

Proof

The proof is obvious. □

Postulation 3.1

Every set X is called near set, near to itself, as every element x X is near to itself.

Definition 3.4

Let ( X , τ ϕ i ) be topological spaces, where ϕ i B , 1 i | B | . Then the lower and upper approximations for any subset A X with respect to the feature ϕ i are defined as: N ̲ ϕ i ( A ) = int ϕ i ( A ) and N ¯ ϕ i ( A ) = cl ϕ i ( A ) , where int ϕ i ( cl ϕ i ) is the interior (closure) with respect to the topology τ ϕ i , whose subbase is the family of special neighborhoods defined in Definition 3.1.

Definition 3.5

Let ( X , τ ϕ i ) be topological spaces, where ϕ i B , 1 i | B | . A new near lower and upper approximations for any subset A X with respect to one feature of the probe functions B are defined as apr ̲ 1 ( A ) = ϕ i B N ̲ ϕ i ( A ) and apr ¯ 1 ( A ) = ϕ i B N ¯ ϕ i ( A ) . Consequently b apr 1 ( A ) = apr ¯ 1 ( A ) - apr ̲ 1 ( A ) .

Remark 3.3

The new near lower and upper approximations with respect to two features of the probe functions B will be defined as apr ̲ 2 ( A ) = ϕ i , ϕ j B N ̲ ϕ i ϕ j ( A ) and apr ¯ 2 ( A ) = ϕ i , ϕ j B N ¯ ϕ i ϕ j ( A ) , where N ̲ ϕ i ϕ j ( A ) = int ϕ i ϕ j ( A ) and N ¯ ϕ i ϕ j ( A ) = cl ϕ i ϕ j ( A ) . Consequently, apr ̲ B ( A ) = ϕ 1 , ϕ 2 , , ϕ B B N ̲ ϕ i ϕ j ϕ B ( A ) ; apr ¯ B ( A ) = ϕ 1 , ϕ 2 , , ϕ B B N ¯ ϕ i ϕ j ϕ B ( A ) .

Definition 3.6

Let ( X , τ ϕ i ) be topological spaces, where ϕ i B , 1 i | B | . The accuracy measure of any subset A X with respect to i features is defined as: α i ( A ) = apr ̲ i ( A ) apr ¯ i ( A ) , A ϕ .

Remark 3.4

0 α i ( A ) 1 , α i ( A ) means the degree of exactness of any subset A X . If α i ( A ) = 1 , then A is exact set with respect to i features.

Theorem 3.3

For any subset A X , apr ̲ i ( A ) and b apr i ( A ) are near to apr ¯ i ( A ) .

Proof

Obvious. □

Remark 3.5

A set A with a boundary | b apr i ( A ) | 0 is a near set.

Theorem 3.4

Every rough set is a near set but not every near set is a rough set.

Proof

There are two cases to consider

  1. | b apr i ( A ) | > 0 . Given a set A X that has been approximated with a nonempty boundary, this means A is a rough set as well as a near set.

  2. | b apr i ( A ) | = 0 . Given a set A X that has been approximated with an empty boundary, this means A is a near set but not a rough set. □

Definition 3.7

Let ( X , τ ϕ i ) be topological spaces, where ϕ i B , 1 i | B | . The new generalized lower rough coverage of any subset Y of the family of neighborhoods with respect to the probe functions B is defined as ν i ( Y , apr ̲ i ( D ) ) = Y apr ̲ i ( D ) apr ̲ i ( D ) , where D is the decision class, means the acceptable objects (Peters, 2007), apr ̲ i ( D ) ϕ and if apr ̲ i ( D ) = ϕ then ν i ( Y , apr ̲ i ( D ) ) = 1 .

Remark 3.6

0 ν i 1 . It measures the degree that the subset Y coverers the acceptable objects or sure region ( apr ̲ i ( D ) ).

Now, we give an example to explain these definitions.

Example 3.1

Let s , a , r be three features defined on a nonempty set X = { x 1 , x 2 , x 3 , x 4 } as in Table 1.

Table 1 Values features for the objects.
s a r
x 1 0.51 1.2 0.53
x 2 0.56 3.1 2.35
x 3 0.72 2.8 0.72
x 4 0.77 1.9 0.95

If the length of the neighborhood of the feature s (resp a and r) equals to 0.2 (resp 0.9 and 0.5), then N 1 ( B ) = { ξ ( s 0.2 ) , ξ ( a 0.9 ) , ξ ( r 0.5 ) } , where ξ ( s 0.2 ) = { { x 1 , x 2 } , { x 1 , x 2 , x 3 } , { x 2 , x 3 , x 4 } , { x 3 , x 4 } } ; ξ ( a 0.9 ) = { { x 1 , x 4 } , { x 2 , x 3 } , { x 2 , x 3 , x 4 } , { x 1 , x 3 , x 4 } } ; ξ ( r 0.5 ) = { { x 1 , x 3 , x 4 } , { x 2 } } . Hence, τ s 0.2 = { { x 2 } , { x 3 } , { x 1 , x 2 } , { x 2 , x 3 } , { x 3 , x 4 } , { x 1 , x 2 , x 3 } , { x 2 , x 3 , x 4 } , X , ϕ } ; τ a 0.9 = { { x 3 } , { x 4 } , { x 3 , x 4 } , { x 2 , x 3 } , { x 1 , x 4 } , { x 2 , x 3 , x 4 } , { x 1 , x 3 , x 4 } , X , ϕ } ; τ r 0.5 = { { x 2 } , { x 1 , x 3 , x 4 } , X , ϕ } . Also, we get : N 2 ( B ) = { ξ ( s 0.2 , a 0.9 ) , ξ ( s 0.2 , r 0.5 ) , ξ ( a 0.9 , r 0.5 ) } , where ξ ( s 0.2 , a 0.9 ) = { { x 1 } , { x 2 , x 3 } , { x 2 , x 3 , x 4 } , { x 3 , x 4 } } ; ξ ( s 0.2 , r 0.5 ) = { { x 1 } , { x 2 } , { x 3 , x 4 } } ; ξ ( a 0.9 , r 0.5 ) = { { x 1 , x 4 } , { x 2 } , { x 3 , x 4 } , { x 1 , x 3 , x 4 } } . Hence, τ s 0.2 a 0.9 = { { x 1 } , { x 3 } , { x 2 , x 3 } , { x 3 , x 4 } , { x 1 , x 3 } , { x 2 , x 3 , x 4 } , { x 1 , x 2 , x 3 } , { x 1 , x 3 , x 4 } , X , ϕ } ; τ s 0.2 r 0.5 = { { x 1 } , { x 2 } , { x 3 , x 4 } , { x 1 , x 2 } , { x 1 , x 3 , x 4 } , { x 2 , x 3 , x 4 } , X , ϕ } ; τ a 0.9 r 0.5 = { { x 2 } , { x 4 } , { x 1 , x 4 } , { x 3 , x 4 } , { x 2 , x 4 } , { x 1 , x 3 , x 4 } , { x 1 , x 2 , x 4 } , { x 2 , x 3 , x 4 } , X , ϕ } . Also, we find that τ s 0.2 r 0.5 τ s 0.2 r 0.5 a 0.9 . Consequently the reduct of these features is the features { s , r } , so the feature { a } can be canceled.

Now the following example deduces a comparison between the classical and new general near approaches by using the accuracy measures.

Example 3.2

As in Example 3.1 we get Table 2, where Q ( X ) is a family of subsets of X.

Table 2 Comparison between classical and our new near approaches.
Q ( X ) α 1 α 2 α 3 α 1 α 2
{ x 1 } 0 1 3 1 0 1
{ x 2 } 1 4 1 3 1 1 1
{ x 3 } 0 0 0 1 1
{ x 4 } 0 0 0 1 1
{ x 1 , x 2 } 1 2 1 2 1 1 1
{ x 1 , x 3 } 0 1 4 1 3 1 2 1
{ x 1 , x 4 } 1 2 1 2 1 3 1 1
{ x 2 , x 3 } 1 2 1 2 1 3 1 1
{ x 2 , x 4 } 1 4 1 4 1 3 2 3 1
{ x 3 , x 4 } 1 2 1 2 1 1 1
{ x 1 , x 2 , x 3 } 3 4 3 4 1 2 1 1
{ x 1 , x 2 , x 4 } 3 4 3 4 1 2 1 1
{ x 1 , x 3 , x 4 } 3 4 3 4 1 1 1
{ x 2 , x 3 , x 4 } 3 4 3 4 1 3 4 1

Note that, α 2 = α 3 for any subset of X, as τ s 0.2 r 0.5 τ a 0.9 s 0.2 r 0.5 .

Remark 3.7

From Table 2, we note that the classical approximations of near sets is more strong than the classical approximations of rough sets, but when we use our generalized approximations of near sets, we find that many sets will be completely exact.

So our topological approach to near sets is the best of our study. Hence we can consider that, our approximations is the start point to apply of our life applications in many fields of science.

4

4 Conclusion

This research is as a start point of many works, which aims to improve the lower and upper approximations of near sets introduced by J.F. Peters. This has led to a refinement of the generalized approximation space model to include families of neighborhoods. These works will be useful to a decision making in many fields of our life.

References

  1. , , , . Near approximations in topological spaces. International Journal of Mathematical Analysis. 2010;4(6):279-290.
    [Google Scholar]
  2. , , , . Finite information systems. Applied Mathematics and Information Sciences. 2007;1(1):13-21.
    [Google Scholar]
  3. , , , eds. Philosophical Essays. Indianapolis: Hackett Publishing Company; .
  4. , . Fundamentals of Pattern Recognition (2nd ed.). New York: Marcel Dekker, Inc.; .
  5. , , . Rough membership functions. In: , , , eds. Advances in the Dempster–Shafer Theory of Evidence. New York, NY: John Wiley and Sons; . p. :251-271.
    [Google Scholar]
  6. , . Rough sets: theoretical aspects of reasoning about data.System Theory, Knowledge Engineering and Problem Solving. Vol vol. 9. Dordrecht, The Netherlands: Kluwer Academic Publishers; .
  7. , . Rough sets. International Journal of Computer and Information Sciences. 1982;11:341-356.
    [Google Scholar]
  8. , . Near sets. Special theory about nearness of objects. Fundamenta Informaticae. 2007;76:1-28.
    [Google Scholar]
  9. Peters, J.F., 2007. Classification of objects by means of features. In: Proc. IEEE Symposium Series on Foundations of Computational Intelligence (IEEE SCCI 2007), Honolulu, Hawaii, pp. 1–8.
  10. , , , . Nearness of objects: extension of approximation space model. Fundamenta Informaticae. 2007;79:1-24.
    [Google Scholar]
  11. Peters, J.F., Skowron, A., Stepaniuk, J., 2006. Nearness in approximation spaces. In: Lindemann, G., Schlilngloff, H., et al. (Eds.), Proc. Concurrency, Specification and Programming (CS,P2006). Informatik-Berichte Nr. 206, Humboldt-Universitat zu Berlin, pp. 434–445.
Show Sections