scholarly journals A Modified Rough Set Approach to Incomplete Information Systems

2007 ◽  
Vol 2007 ◽  
pp. 1-13 ◽  
Author(s):  
E. A. Rady ◽  
M. M. E. Abd El-Monsef ◽  
W. A. Abd El-Latif

The key point of the tolerance relation or similarity relation presented in the literature is to assign a “null” value to all missing attribute values. In other words, a “null” value may be equal to any value in the domain of the attribute values. This may cause a serious effect in data analysis and decision analysis because the missing values are just “missed” but they do exist and have an influence on the decision. In this paper, we will introduce the modified similarity relation denoted by MSIM that is dependent on the number of missing values with respect to the number of the whole defined attributes for each object. According to the definition of MSIM, many problems concerning the generalized decisions are solved. This point may be used in scaling in statistics in a wide range. Also, a new definition of the discernibility matrix, deduction of the decision rules, and reducts in the present of the missing values are obtained.

Author(s):  
Hemant Rana ◽  
Manohar Lal

Handling of missing attribute values are a big challenge for data analysis. For handling this type of problems, there are some well known approaches, including Rough Set Theory (RST) and classification via clustering. In the work reported here, RSES (Rough Set Exploration System) one of the tools based on RST approach, and WEKA (Waikato Environment for Knowledge Analysis), a data mining tool—based on classification via clustering—are used for predicting learning styles from given data, which possibly has missing values. The results of the experiments using the tools show that the problem of missing attribute values is better handled by RST approach as compared to the classification via clustering approach. Further, in respect of missing values, RSES yields better decision rules, if the missing values are simply ignored than the rules obtained by assigning some values in place of missing attribute values.


Author(s):  
Héctor Oscar Nigro ◽  
Sandra Elizabeth González Císaro

Today’s technology allows storing vast quantities of information from different sources in nature. This information has missing values, nulls, internal variation, taxonomies, and rules. We need a new type of data analysis that allows us represent the complexity of reality, maintaining the internal variation and structure (Diday, 2003). In Data Analysis Process or Data Mining, it is necessary to know the nature of null values - the cases are by absence value, null value or default value -, being also possible and valid to have some imprecision, due to differential semantic in a concept, diverse sources, linguistic imprecision, element resumed in Database, human errors, etc (Chavent, 1997). So, we need a conceptual support to manipulate these types of situations. As we are going to see below, Symbolic Data Analysis (SDA) is a new issue based on a strong conceptual model called Symbolic Object (SO). A “SO” is defined by its “intent” which contains a way to find its “extent”. For instance, the description of habitants in a region and the way of allocating an individual to this region is called “intent”, the set of individuals, which satisfies this intent, is called “extent” (Diday 2003). For this type of analysis, different experts are needed, each one giving their concepts.


Complexity ◽  
2019 ◽  
Vol 2019 ◽  
pp. 1-17
Author(s):  
Zhaohao Wang ◽  
Xiaoping Zhang

How to effectively deal with missing values in incomplete information systems (IISs) according to the research target is still a key issue for investigating IISs. If the missing values in IISs are not handled properly, they will destroy the internal connection of data and reduce the efficiency of data usage. In this paper, in order to establish effective methods for filling missing values, we propose a new information system, namely, a fuzzy set-valued information system (FSvIS). By means of the similarity measures of fuzzy sets, we obtain several binary relations in FSvISs, and we investigate the relationship among them. This is a foundation for the researches on FSvISs in terms of rough set approach. Then, we provide an algorithm to fill the missing values in IISs with fuzzy set values. In fact, this algorithm can transform an IIS into an FSvIS. Furthermore, we also construct an algorithm to fill the missing values in IISs with set values (or real values). The effectiveness of these algorithms is analyzed. The results showed that the proposed algorithms achieve higher correct rate than traditional algorithms, and they have good stability. Finally, we discuss the importance of these algorithms for investigating IISs from the viewpoint of rough set theory.


2011 ◽  
Vol 187 ◽  
pp. 251-256
Author(s):  
Lei Wang ◽  
Tian Rui Li ◽  
Jun Ye

The essence of the rough set theory (RST) is to deal with the inconsistent problems by two definable subsets which are called the lower and upper approximations respectively. Asymmetric Similarity relation based Rough Sets (ASRS) model is one kind of extensions of the classical rough set model in incomplete information systems. In this paper, we propose a new matrix view of ASRS model and give the matrix representation of the lower and upper approximations of a concept under ASRS model. According to this matrix view, a new method is obtained for calculation of the lower and upper approximations under ASRS model. An example is given to illustrate processes of calculating the approximations of a concept based on the matrix point of view.


Author(s):  
JUNHONG WANG ◽  
JIYE LIANG ◽  
YUHUA QIAN ◽  
CHUANGYIN DANG

Rough set theory is a relatively new mathematical tool for computer applications in circumstances characterized by vagueness and uncertainty. In this paper, we address uncertainty of rough sets for incomplete information systems. An axiom definition of knowledge granulation for incomplete information systems is obtained, under which a measure of uncertainty of a rough set is proposed. This measure has some nice properties such as equivalence, maximum and minimum. Furthermore, we prove that the uncertainty measure is effective and suitable for measuring roughness and accuracy of rough sets for incomplete information systems.


2019 ◽  
Vol 2019 ◽  
pp. 1-9
Author(s):  
Jiucheng Xu ◽  
Yun Wang ◽  
Keqiang Xu ◽  
Tianli Zhang

To select more effective feature genes, many existing algorithms focus on the selection and study of evaluation methods for feature genes, ignoring the accurate mapping of original information in data processing. Therefore, for solving this problem, a new model is proposed in this paper: rough uncertainty metric model. First, the fuzzy neighborhood granule of the sample is constructed by combining the fuzzy similarity relation with the neighborhood radius in the rough set, and the rough decision is defined by using the fuzzy similarity relation and the decision equivalence class. Then, the fuzzy neighborhood granule and the rough decision are introduced into the conditional entropy, and the rough uncertainty metric model is proposed; in the meantime, the definition of measuring the significance of feature genes and the proof of some related theorems are given. To make this model tolerate noises in data, this paper introduces a variable precision model and discusses the selection of parameters. Finally, based on the rough uncertainty metric model, we design a feature genes selection algorithm and compare it with some existing similar algorithms. The experimental results show that the proposed algorithm can select the smaller feature genes subset with higher classification accuracy and verify that the model proposed in this paper is more effective.


Author(s):  
R. FÉLIX ◽  
T. USHIO

Rough set based methods have been applied successfully in many real world applications such as data mining, knowledge discovery, machine learning, and control. The rough set theory is used to deal with imperfect data and to eliminate dispensable, superfluous and redundant information as to obtain a simplified set of decision rules. Thus, several approaches and methods have been proposed to find minimal coverings, from which the decision rules can be induced. In many of these approaches, an improvement in the utilization of computational resources is encouraged. In this paper, a binary encoding for attribute sets and a discernibility matrix is proposed. Such a binary representation of sets and sets operations in the implementation of algorithms provides a machine-oriented approach to the utilization of computational memory and allow parallel processing among groups of attributes. The discernibility matrix is reduced to its minimal size through the identification of main patterns in order to eliminate redundancies. Bit-wise operations replace sets operations, thus the search for minimal coverings is performed in an efficient way. Resulting improvement is shown in the analysis of medium-sized data sets using two generic methods to obtain minimal coverings.


2011 ◽  
Vol 50-51 ◽  
pp. 180-184
Author(s):  
Jin Peng Wang ◽  
Bao Xiang Liu ◽  
Zhen Dong Li ◽  
Li Chao Feng

Concept lattice and rough set are powerful tools for data analyzing and processing, has been successfully applied to many fields. However, the decision information is incomplete in many information systems. In this paper, the definition of incomplete concept lattice has been proposed, and some relation established between imperfect concept lattice and rough set. As is very important that the paper gives a new attributes reduction algorithm about incomplete concept lattice aims at the matter of the inefficient of reduction strategy based on discernibility matrix. Comparing with the attributes reduction for incomplete concept lattice which based on discernibility matrix, this reduction algorithm, reduces the spatial-temporal complexity.


Sign in / Sign up

Export Citation Format

Share Document