scholarly journals Covering-Based Rough Sets on Eulerian Matroids

2013 ◽  
Vol 2013 ◽  
pp. 1-8 ◽  
Author(s):  
Bin Yang ◽  
Ziqiong Lin ◽  
William Zhu

Rough set theory is an efficient and essential tool for dealing with vagueness and granularity in information systems. Covering-based rough set theory is proposed as a significant generalization of classical rough sets. Matroid theory is a vital structure with high applicability and borrows extensively from linear algebra and graph theory. In this paper, one type of covering-based approximations is studied from the viewpoint of Eulerian matroids. First, we explore the circuits of an Eulerian matroid from the perspective of coverings. Second, this type of covering-based approximations is represented by the circuits of Eulerian matroids. Moreover, the conditions under which the covering-based upper approximation operator is the closure operator of a matroid are presented. Finally, a matroidal structure of covering-based rough sets is constructed. These results show many potential connections between covering-based rough sets and matroids.

2012 ◽  
Vol 548 ◽  
pp. 735-739
Author(s):  
Hong Mei Nie ◽  
Jia Qing Zhou

Rough set theory has been proposed by Pawlak as a useful tool for dealing with the vagueness and granularity in information systems. Classical rough set theory is based on equivalence relation. The covering rough sets are an improvement of Pawlak rough set to deal with complex practical problems which the latter one can not handle. This paper studies covering-based generalized rough sets. In this setting, we investigate common properties of classical lower and upper approximation operations hold for the covering-based lower and upper approximation operations and relationships among some type of covering rough sets.


Author(s):  
ZHIMING ZHANG ◽  
JINGFENG TIAN

Intuitionistic fuzzy (IF) rough sets are the generalization of traditional rough sets obtained by combining the IF set theory and the rough set theory. The existing research on IF rough sets mainly concentrates on the establishment of lower and upper approximation operators using constructive and axiomatic approaches. Less effort has been put on the attribute reduction of databases based on IF rough sets. This paper systematically studies attribute reduction based on IF rough sets. Firstly, attribute reduction with traditional rough sets and some concepts of IF rough sets are reviewed. Then, we introduce some concepts and theorems of attribute reduction with IF rough sets, and completely investigate the structure of attribute reduction. Employing the discernibility matrix approach, an algorithm to find all attribute reductions is also presented. Finally, an example is proposed to illustrate our idea and method. Altogether, these findings lay a solid theoretical foundation for attribute reduction based on IF rough sets.


2014 ◽  
Vol 2014 ◽  
pp. 1-7
Author(s):  
Hui Li ◽  
Yanfang Liu ◽  
William Zhu

Rough set theory provides an effective tool to deal with uncertain, granular, and incomplete knowledge in information systems. Matroid theory generalizes the linear independence in vector spaces and has many applications in diverse fields, such as combinatorial optimization and rough sets. In this paper, we construct a matroidal structure of the generalized rough set based on a tolerance relation. First, a family of sets are constructed through the lower approximation of a tolerance relation and they are proved to satisfy the circuit axioms of matroids. Thus we establish a matroid with the family of sets as its circuits. Second, we study the properties of the matroid including the base and the rank function. Moreover, we investigate the relationship between the upper approximation operator based on a tolerance relation and the closure operator of the matroid induced by the tolerance relation. Finally, from a tolerance relation, we can get a matroid of the generalized rough set based on the tolerance relation. The matroid can also induce a new relation. We investigate the connection between the original tolerance relation and the induced relation.


Author(s):  
S. Arjun Raj ◽  
M. Vigneshwaran

In this article we use the rough set theory to generate the set of decision concepts in order to solve a medical problem.Based on officially published data by International Diabetes Federation (IDF), rough sets have been used to diagnose Diabetes.The lower and upper approximations of decision concepts and their boundary regions have been formulated here.


Author(s):  
B. K. Tripathy

Granular Computing has emerged as a framework in which information granules are represented and manipulated by intelligent systems. Granular Computing forms a unified conceptual and computing platform. Rough set theory put forth by Pawlak is based upon single equivalence relation taken at a time. Therefore, from a granular computing point of view, it is single granular computing. In 2006, Qiang et al. introduced a multi-granular computing using rough set, which was called optimistic multigranular rough sets after the introduction of another type of multigranular computing using rough sets called pessimistic multigranular rough sets being introduced by them in 2010. Since then, several properties of multigranulations have been studied. In addition, these basic notions on multigranular rough sets have been introduced. Some of these, called the Neighborhood-Based Multigranular Rough Sets (NMGRS) and the Covering-Based Multigranular Rough Sets (CBMGRS), have been added recently. In this chapter, the authors discuss all these topics on multigranular computing and suggest some problems for further study.


Author(s):  
Benjamin Griffiths

Rough Set Theory (RST), since its introduction in Pawlak (1982), continues to develop as an effective tool in data mining. Within a set theoretical structure, its remit is closely concerned with the classification of objects to decision attribute values, based on their description by a number of condition attributes. With regards to RST, this classification is through the construction of ‘if .. then ..’ decision rules. The development of RST has been in many directions, amongst the earliest was with the allowance for miss-classification in the constructed decision rules, namely the Variable Precision Rough Sets model (VPRS) (Ziarko, 1993), the recent references for this include; Beynon (2001), Mi et al. (2004), and Slezak and Ziarko (2005). Further developments of RST have included; its operation within a fuzzy environment (Greco et al., 2006), and using a dominance relation based approach (Greco et al., 2004). The regular major international conferences of ‘International Conference on Rough Sets and Current Trends in Computing’ (RSCTC, 2004) and ‘International Conference on Rough Sets, Fuzzy Sets, Data Mining and Granular Computing’ (RSFDGrC, 2005) continue to include RST research covering the varying directions of its development. This is true also for the associated book series entitled ‘Transactions on Rough Sets’ (Peters and Skowron, 2005), which further includes doctoral theses on this subject. What is true, is that RST is still evolving, with the eclectic attitude to its development meaning that the definitive concomitant RST data mining techniques are still to be realised. Grzymala-Busse and Ziarko (2000), in a defence of RST, discussed a number of points relevant to data mining, and also made comparisons between RST and other techniques. Within the area of data mining and the desire to identify relationships between condition attributes, the effectiveness of RST is particularly pertinent due to the inherent intent within RST type methodologies for data reduction and feature selection (Jensen and Shen, 2005). That is, subsets of condition attributes identified that perform the same role as all the condition attributes in a considered data set (termed ß-reducts in VPRS, see later). Chen (2001) addresses this, when discussing the original RST, they state it follows a reductionist approach and is lenient to inconsistent data (contradicting condition attributes - one aspect of underlying uncertainty). This encyclopaedia article describes and demonstrates the practical application of a RST type methodology in data mining, namely VPRS, using nascent software initially described in Griffiths and Beynon (2005). The use of VPRS, through its relative simplistic structure, outlines many of the rudiments of RST based methodologies. The software utilised is oriented towards ‘hands on’ data mining, with graphs presented that clearly elucidate ‘veins’ of possible information identified from ß-reducts, over different allowed levels of missclassification associated with the constructed decision rules (Beynon and Griffiths, 2004). Further findings are briefly reported when undertaking VPRS in a resampling environment, with leave-one-out and bootstrapping approaches adopted (Wisnowski et al., 2003). The importance of these results is in the identification of the more influential condition attributes, pertinent to accruing the most effective data mining results.


Author(s):  
Malcolm J. Beynon

Rough set theory (RST), since its introduction in Pawlak (1982), continues to develop as an effective tool in classification problems and decision support. In the majority of applications using RST based methodologies, there is the construction of ‘if .. then ..’ decision rules that are used to describe the results from an analysis. The variation of applications in management and decision making, using RST, recently includes discovering the operating rules of a Sicilian irrigation purpose reservoir (Barbagallo, Consoli, Pappalardo, Greco, & Zimbone, 2006), feature selection in customer relationship management (Tseng & Huang, 2007) and decisions that insurance companies make to satisfy customers’ needs (Shyng, Wang, Tzeng, & Wu, 2007). As a nascent symbolic machine learning technique, the popularity of RST is a direct consequence of its set theoretical operational processes, mitigating inhibiting issues associated with traditional techniques, such as within-group probability distribution assumptions (Beynon & Peel, 2001). Instead, the rudiments of the original RST are based on an indiscernibility relation, whereby objects are grouped into certain equivalence classes and inference taken from these groups. Characteristics like this mean that decision support will be built upon the underlying RST philosophy of “Let the data speak for itself” (Dunstch & Gediga, 1997). Recently, RST was viewed as being of fundamental importance in artificial intelligence and cognitive sciences, including decision analysis and decision support systems (Tseng & Huang, 2007). One of the first developments on RST was through the variable precision rough sets model (VPRSß), which allows a level of mis-classification to exist in the classification of objects, resulting in probabilistic rules (see Ziarko, 1993; Beynon, 2001; Li and Wang, 2004). VPRSß has specifically been applied as a potential decision support system with the UK Monopolies and Mergers Commission (Beynon & Driffield, 2005), predicting bank credit ratings (Griffiths & Beynon, 2005) and diffusion of medicaid home care programs (Kitchener, Beynon, & Harrington, 2004). Further developments of RST include extended variable precision rough sets (VPRSl,u), which infers asymmetric bounds on the possible classification and mis-classification of objects (Katzberg & Ziarko, 1996), dominance-based rough sets, which bases their approach around a dominance relation (Greco, Matarazzo, & Slowinski, 2004), fuzzy rough sets, which allows the grade of membership of objects to constructed sets (Greco, Inuiguchi, & Slowinski, 2006), and probabilistic bayesian rough sets model that considers an appropriate certainty gain function (Ziarko, 2005). A literal presentation of the diversity of work on RST can be viewed in the annual volumes of the Transactions on Rough Sets (most recent year 2006), also the annual conferences dedicated to RST and its developments (see for example, RSCTC, 2004). In this article, the theory underlying VPRSl,u is described, with its special case of VPRSß used in an example analysis. The utilisation of VPRSl,u, and VPRSß, is without loss of generality to other developments such as those referenced, its relative simplicity allows the non-proficient reader the opportunity to fully follow the details presented.


2011 ◽  
pp. 129-151
Author(s):  
Theresa Beaubouef ◽  
Frederick E Petry

This chapter discusses ways in which rough set theory can enhance databases by allowing for the management of uncertainty. Rough sets can be integrated into an underlying database model, relational or object oriented, and also used in design and querying of databases. Because rough sets are a versatile theory, they can also be combined with other theories. The authors discuss the rough relational database model, the rough object oriented database model, and fuzzy set and intuitionistic set extensions to each of these models. Comparisons and benefits of the various approaches are discussed, illustrating the usefulness and versatility of rough sets for uncertainty management in databases.


2009 ◽  
pp. 1127-1150
Author(s):  
Theresa Beaubouef ◽  
Frederick E. Petry

This chapter discusses ways in which rough-set theory can enhance databases by allowing for the management of uncertainty. Rough sets can be integrated into an underlying database model, relational or object oriented, and also used in the design and uerying of databases, because roughsets are a versatile theory, theories. The authors discuss the rough relational databases model, the rough object-oriented database model, and fuzzy set and intuitionistic set extensions to each of these models. Comparisons and benefits of the various approaches are discussed, illustrating the usefulness and versatility of rough sets for uncertainty management in databases.


Sign in / Sign up

Export Citation Format

Share Document