• Documents
  • Authors
  • Tables
  • Log in
  • Sign up
  • MetaCart
  • DMCA
  • Donate

CiteSeerX logo

Advanced Search Include Citations
Advanced Search Include Citations

A Treatise on Rough Sets (2005)

by Zdzislaw Pawlak
Add To MetaCart

Tools

Sorted by:
Results 1 - 10 of 793
Next 10 →

Wrappers for Feature Subset Selection

by Ron Kohavi, George H. John - AIJ SPECIAL ISSUE ON RELEVANCE , 1997
"... In the feature subset selection problem, a learning algorithm is faced with the problem of selecting a relevant subset of features upon which to focus its attention, while ignoring the rest. To achieve the best possible performance with a particular learning algorithm on a particular training set, a ..."
Abstract - Cited by 1569 (3 self) - Add to MetaCart
In the feature subset selection problem, a learning algorithm is faced with the problem of selecting a relevant subset of features upon which to focus its attention, while ignoring the rest. To achieve the best possible performance with a particular learning algorithm on a particular training set, a feature subset selection method should consider how the algorithm and the training set interact. We explore the relation between optimal feature subset selection and relevance. Our wrapper method searches for an optimal feature subset tailored to a particular algorithm and a domain. We study the strengths and weaknesses of the wrapper approach andshow a series of improved designs. We compare the wrapper approach to induction without feature subset selection and to Relief, a filter approach to feature subset selection. Significant improvement in accuracy is achieved for some datasets for the two families of induction algorithms used: decision trees and Naive-Bayes.

Learning in the Presence of Concept Drift and Hidden Contexts

by Gerhard Widmer, M. Kubat - Machine Learning , 1996
"... . On-line learning in domains where the target concept depends on some hidden context poses serious problems. A changing context can induce changes in the target concepts, producing what is known as concept drift. We describe a family of learning algorithms that flexibly react to concept drift and c ..."
Abstract - Cited by 285 (1 self) - Add to MetaCart
. On-line learning in domains where the target concept depends on some hidden context poses serious problems. A changing context can induce changes in the target concepts, producing what is known as concept drift. We describe a family of learning algorithms that flexibly react to concept drift and can take advantage of situations where contexts reappear. The general approach underlying all these algorithms consists of (1) keeping only a window of currently trusted examples and hypotheses; (2) storing concept descriptions and re-using them when a previous context reappears; and (3) controlling both of these functions by a heuristic that constantly monitors the system's behavior. The paper reports on experiments that test the systems' performance under various conditions such as different levels of noise and different extent and rate of concept drift. Keywords: Incremental concept learning, on-line learning, context dependence, concept drift, forgetting 1. Introduction The work presen...

The Power of Decision Tables

by Ron Kohavi - Proceedings of the European Conference on Machine Learning , 1995
"... . We evaluate the power of decision tables as a hypothesis space for supervised learning algorithms. Decision tables are one of the simplest hypothesis spaces possible, and usually they are easy to understand. Experimental results show that on artificial and real-world domains containing only discre ..."
Abstract - Cited by 160 (5 self) - Add to MetaCart
. We evaluate the power of decision tables as a hypothesis space for supervised learning algorithms. Decision tables are one of the simplest hypothesis spaces possible, and usually they are easy to understand. Experimental results show that on artificial and real-world domains containing only discrete features, IDTM, an algorithm inducing decision tables, can sometimes outperform state-of-the-art algorithms such as C4.5. Surprisingly, performance is quite good on some datasets with continuous features, indicating that many datasets used in machine learning either do not require these features, or that these features have few values. We also describe an incremental method for performing crossvalidation that is applicable to incremental learning algorithms including IDTM. Using incremental cross-validation, it is possible to cross-validate a given dataset and IDTM in time that is linear in the number of instances, the number of features, and the number of label values. The time for incre...
(Show Context)

Citation Context

...he table are correctly classified and the structures are not used for making predictions. The rough sets community has been using the hypothesis space of decision tables for a few years (Pawlak 1987, =-=Pawlak 1991-=-, Slowinski 1992). Researchers in the field of rough sets suggest using the degrees-of-dependency of a feature on the label (called fl) to determine which features should be included in a decision tab...

A probabilistic extension to ontology language owl

by Zhongli Ding, Yun Peng - In Proceedings of the 37th Hawaii International Conference On System Sciences (HICSS-37), Big Island , 2004
"... With the development of the semantic web activity, ontologies become widely used to represent the conceptualization of a domain. However, none of the existing ontology languages provides a means to capture uncertainty about the concepts, properties and instances in a domain. Probability theory is a ..."
Abstract - Cited by 112 (3 self) - Add to MetaCart
With the development of the semantic web activity, ontologies become widely used to represent the conceptualization of a domain. However, none of the existing ontology languages provides a means to capture uncertainty about the concepts, properties and instances in a domain. Probability theory is a natural choice for dealing with uncertainty. Incorporating probability theory into existing ontology languages will provide these languages additional expressive power to quantify the degree of the overlap or inclusion between two concepts, support probabilistic queries such as finding the most probable concept that a given description belongs to, and make more accurate semantic integration possible. One approach to provide such a probabilistic extension to ontology languages is to use Bayesian networks, a widely used graphic model for knowledge representation under uncertainty. In this paper, we present our on-going research on extending OWL, an ontology language recently proposed by W3C’s Semantic Web Activity. First, the language is augmented to allow additional probabilistic markups, so probabilities can be attached with individual concepts and properties in an OWL ontology. Secondly, a set of translation rules is defined to convert this probabilistically annotated OWL ontology into a Bayesian network. Our probabilistic extension to OWL has clear semantics: the Bayesian network obtained will be associated with a joint probability distribution over the application domain. General Bayesian network inference procedures (e.g., belief propagation or junction tree) can be used to compute P(C | e): the degree of the overlap or inclusion between a concept C and a concept represented by a description e. We also provide a similarity measure that can be used to find the most probable concept that a given description belongs to. 1.

GRID RESOURCE MANAGEMENT -- State of the art and future trends

by Jarek Nabrzyski, Jennifer M. Schopf, Jan Weglarz
"... ..."
Abstract - Cited by 88 (0 self) - Add to MetaCart
Abstract not found

Rough sets: some extensions,”

by Zdzisław Pawlak , Andrzej Skowron - Information Sciences, , 2007
"... Abstract In this article, we present some extensions of the rough set approach and we outline a challenge for the rough set based research. ..."
Abstract - Cited by 87 (6 self) - Add to MetaCart
Abstract In this article, we present some extensions of the rough set approach and we outline a challenge for the rough set based research.
(Show Context)

Citation Context

...ecisively in the absence of certainty. Bertrand Russell (1950). An Inquiry into Meaning and Truth. George Allen and Unwin, London; W.W. Norton, New York. 1. Introduction We use notation introduced in =-=[47]-=-. The reader is also referred to the literature cited in [47]. The basic notions of rough sets and approximation spaces were introduced during the early 1980s (see, e.g., [42–44]). In this section, we...

The art of granular computing:

by Yiyu Yao - Proceeding of the International Conference on Rough Sets and Emerging Intelligent Systems Paradigms, , 2007
"... Abstract: This paper has two purposes. One is to present a critical examination of the rise of granular computing and the other is to suggest a triarchic theory of granular computing. By examining the reasons, justifications, and motivations for the rise of granular computing, we may be able to ful ..."
Abstract - Cited by 74 (20 self) - Add to MetaCart
Abstract: This paper has two purposes. One is to present a critical examination of the rise of granular computing and the other is to suggest a triarchic theory of granular computing. By examining the reasons, justifications, and motivations for the rise of granular computing, we may be able to fully appreciate its scope, goal and potential values. The results enable us to formulate a triarchic theory in the light of research results from many disciplines. The three components of the theory are labeled as the philosophy, the methodology, and the computation. The integration of the three offers a unified view of granular computing as a way of structured thinking, a method of structured problem solving, and a paradigm of structured information processing, focusing on hierarchical granular structures. The triarchic theory is an important effort in synthesizing the various theories and models of granular computing. Key words: Triarchic theory of granular computing; systems theory; structured thinking, problem solving and information processing. CLC number: Document code: A Introduction Although granular computing, as a separate field of study, started a decade ago [1], its basic philosophy, ideas, principles, methodologies, theories and tools has, in fact, long been used either explicitly or implicitly across many branches of natural and social sciences The answers, at least partial answers, to these questions may be obtained by drawing and synthesizing results from well-established disciplines, including philosophy, psychology, neuroscience, cognitive science, education, artificial intelligence, computer programming, and many more. Previously, I argued that granular computing represents an idea converged from many branches of natural and social sciences Human-Inspired Computing Research on understanding the human brain and natural intelligence is closely related to the field of artificial intelligence (AI) and information technology (IT). The results have led to a computational view for explaining how the mind works

Mining Association Rules with Weighted Items

by Chun Hing Cai , 1998
"... ..."
Abstract - Cited by 65 (0 self) - Add to MetaCart
Abstract not found

Global discretization of continuous attributes as preprocessing for machine learning

by Michal R. Chmielewski, Jerzy W. Grzymala-busse - International Journal of Approximate Reasoning , 1996
"... Abstract. Real-life data usually are presented in databases by real numbers. On the other hand, most inductive learning methods require small number of attribute values. Thus it is necessary to convert input data sets with continuous attributes into input data sets with discrete attributes. Methods ..."
Abstract - Cited by 62 (4 self) - Add to MetaCart
Abstract. Real-life data usually are presented in databases by real numbers. On the other hand, most inductive learning methods require small number of attribute values. Thus it is necessary to convert input data sets with continuous attributes into input data sets with discrete attributes. Methods of discretization restricted to single continuous attributes will be called local, while methods that simultaneously convert all continuous attributes will be called global. In this paper, a method of transforming any local discretization method into a global one is presented. A global discretization method, based on cluster analysis, is presented and compared experimentally with three known local methods, transformed into global. Experiments include ten-fold cross validation and leaving-one-out methods for ten real-life data sets.
(Show Context)

Citation Context

...ent. We need a measure of consistency for inconsistent data sets. Our measure, called a level of consistency, is based on rough set theory, a tool to deal with uncertainty, introduced by Z. Pawlak in =-=[12]-=-. Let U denote the set of all examplessof the data set. Let P denote a nonempty subset of the set of all variables, i.e., attributes and a decision. Obviously, set P defines an equivalence relation ℘ ...

Information granulation and rough set approximation

by Y. Y. Yao - International Journal of Intelligent Systems , 2001
"... Information granulation and concept approximation are some of the fundamental issues of granular computing. Granulation of a universe involves grouping of similar elements into granules to form coarse-grained views of the universe. Approximation of concepts, represented by subsets of the universe, d ..."
Abstract - Cited by 52 (19 self) - Add to MetaCart
Information granulation and concept approximation are some of the fundamental issues of granular computing. Granulation of a universe involves grouping of similar elements into granules to form coarse-grained views of the universe. Approximation of concepts, represented by subsets of the universe, deals with the descriptions of concepts using granules. In the context of rough set theory, this paper examines the two related issues. The granulation structures used by standard rough set theory and the corresponding approximation structures are reviewed. Hierarchical granulation and approximation structures are studied, which results in stratified rough set approximations. A nested sequence of granulations induced by a set of nested equivalence relations leads to a nested sequence of rough set approximations. A multi-level granulation, characterized by a special class of equivalence relations, leads to a more general approximation structure. The notion of neighborhood systems is also explored. 1
(Show Context)

Citation Context

...eir lower approximations, nor obtain the upper approximation of the intersection of some sets from their upper approximations. Additional properties of rough set approximations can be found in Pawlak =-=[14, 15]-=-, and Yao and Lin [34]. The accuracy of rough set approximation is defined as [14]: α(X) = |apr(X)| , (4) |apr(X)| where | · | denotes the cardinality of a set. For the empty set ∅, we define α(∅) = 1...

Powered by: Apache Solr
  • About CiteSeerX
  • Submit and Index Documents
  • Privacy Policy
  • Help
  • Data
  • Source
  • Contact Us

Developed at and hosted by The College of Information Sciences and Technology

© 2007-2019 The Pennsylvania State University