Results 1  10
of
1,316
Rough Sets.
 Int. J. of Information and Computer Sciences
, 1982
"... Abstract. This article presents some general remarks on rough sets and their place in general picture of research on vagueness and uncertainty concepts of utmost interest, for many years, for philosophers, mathematicians, logicians and recently also for computer scientists and engineers particular ..."
Abstract

Cited by 793 (13 self)
 Add to MetaCart
Abstract. This article presents some general remarks on rough sets and their place in general picture of research on vagueness and uncertainty concepts of utmost interest, for many years, for philosophers, mathematicians, logicians and recently also for computer scientists and engineers particularly those working in such areas as AI, computational intelligence, intelligent systems, cognitive science, data mining and machine learning. Thus this article is intended to present some philosophical observations rather than to consider technical details or applications of rough set theory. Therefore we also refrain from presentation of many interesting applications and some generalizations of the theory.
Correlationbased feature selection for machine learning
, 1998
"... A central problem in machine learning is identifying a representative set of features from which to construct a classification model for a particular task. This thesis addresses the problem of feature selection for machine learning through a correlation based approach. The central hypothesis is that ..."
Abstract

Cited by 318 (3 self)
 Add to MetaCart
A central problem in machine learning is identifying a representative set of features from which to construct a classification model for a particular task. This thesis addresses the problem of feature selection for machine learning through a correlation based approach. The central hypothesis is that good feature sets contain features that are highly correlated with the class, yet uncorrelated with each other. A feature evaluation formula, based on ideas from test theory, provides an operational definition of this hypothesis. CFS (Correlation based Feature Selection) is an algorithm that couples this evaluation formula with an appropriate correlation measure and a heuristic search strategy. CFS was evaluated by experiments on artificial and natural datasets. Three machine learning algorithms were used: C4.5 (a decision tree learner), IB1 (an instance based learner), and naive Bayes. Experiments on artificial datasets showed that CFS quickly identifies and screens irrelevant, redundant, and noisy features, and identifies relevant features as long as their relevance does not strongly depend on other features. On natural domains, CFS typically eliminated well over half the features. In most cases, classification accuracy using the reduced feature set equaled or bettered accuracy using the complete feature set.
Feature Subset Selection Using A Genetic Algorithm
, 1997
"... : Practical pattern classification and knowledge discovery problems require selection of a subset of attributes or features (from a much larger set) to represent the patterns to be classified. This is due to the fact that the performance of the classifier (usually induced by some learning algorithm) ..."
Abstract

Cited by 279 (7 self)
 Add to MetaCart
: Practical pattern classification and knowledge discovery problems require selection of a subset of attributes or features (from a much larger set) to represent the patterns to be classified. This is due to the fact that the performance of the classifier (usually induced by some learning algorithm) and the cost of classification are sensitive to the choice of the features used to construct the classifier. Exhaustive evaluation of possible feature subsets is usually infeasible in practice because of the large amount of computational effort required. Genetic algorithms, which belong to a class of randomized heuristic search techniques, offer an attractive approach to find nearoptimal solutions to such optimization problems. This paper presents an approach to feature subset selection using a genetic algorithm. Some advantages of this approach include the ability to accommodate multiple criteria such as accuracy and cost of classification into the feature selection process and to find fe...
Interval propagation to reason about sets: definition and implementation of a practical language
 CONSTRAINTS
, 1997
"... Local consistency techniques have been introduced in logic programming in order to extend the application domain of logic programming languages. The existing languages based on these techniques consider arithmetic constraints applied to variables ranging over nite integer domains. This makes difficu ..."
Abstract

Cited by 119 (8 self)
 Add to MetaCart
Local consistency techniques have been introduced in logic programming in order to extend the application domain of logic programming languages. The existing languages based on these techniques consider arithmetic constraints applied to variables ranging over nite integer domains. This makes difficult a natural and concise modelling as well as an efficient solving of a class of NPcomplete combinatorial search problems dealing with sets. To overcome these problems, we propose a solution which consists in extending the notion of integer domains to that of set domains (sets of sets). We specify a set domain by an interval whose lower and upper bounds are known sets, ordered by set inclusion. We define the formal and practical framework of a new constraint logic programming language over set domains, called Conjunto. Conjunto comprises the usual set operation symbols ([ � \ � n), and the set inclusion relation (). Set expressions built using the operation symbols are interpreted as relations (s [ s1 = s2,...). In addition, Conjunto provides us with a set of constraints called graduated constraints (e.g. the set cardinality) which map sets onto arithmetic terms. This allows us to handle optimization problems by applying a cost function to the quantifiable, i.e., arithmetic, terms which are associated to set terms. The constraint solving in Conjunto is based on local consistency techniques using interval reasoning which are extended to handle set constraints. The main contribution of this paper concerns the formal definition of the language and its design and implementation as a practical language.
Data Mining in Soft Computing Framework: A Survey
 IEEE Transactions on Neural Networks
, 2001
"... The present article provides a survey of the available literature on data mining using soft computing. A categorization has been provided based on the different soft computing tools and their hybridizations used, the data mining function implemented, and the preference criterion selected by the mode ..."
Abstract

Cited by 109 (3 self)
 Add to MetaCart
(Show Context)
The present article provides a survey of the available literature on data mining using soft computing. A categorization has been provided based on the different soft computing tools and their hybridizations used, the data mining function implemented, and the preference criterion selected by the model. The utility of the different soft computing methodologies is highlighted. Generally fuzzy sets are suitable for handling the issues related to understandability of patterns, incomplete/noisy data, mixed media information and human interaction, and can provide approximate solutions faster. Neural networks are nonparametric, robust, and exhibit good learning and generalization capabilities in datarich environments. Genetic algorithms provide efficient search algorithms to select a model, from mixed media data, based on some preference criterion/objective function. Rough sets are suitable for handling different types of uncertainty in data. Some challenges to data mining and the application of soft computing methodologies are indicated. An extensive bibliography is also included.
Rough sets: some extensions,”
 Information Sciences,
, 2007
"... Abstract In this article, we present some extensions of the rough set approach and we outline a challenge for the rough set based research. ..."
Abstract

Cited by 87 (6 self)
 Add to MetaCart
(Show Context)
Abstract In this article, we present some extensions of the rough set approach and we outline a challenge for the rough set based research.
Rough Mereology: A New Paradigm For Approximate Reasoning
, 1996
"... We are concerned with formal models of reasoning under uncertainty. Many approaches to this problem are known in the literature e.g. DempsterShafer theory, bayesianbased reasoning, belief networks, fuzzy logics etc. We propose rough mereology as a foundation for approximate reasoning about complex ..."
Abstract

Cited by 81 (35 self)
 Add to MetaCart
We are concerned with formal models of reasoning under uncertainty. Many approaches to this problem are known in the literature e.g. DempsterShafer theory, bayesianbased reasoning, belief networks, fuzzy logics etc. We propose rough mereology as a foundation for approximate reasoning about complex objects. Our notion of a complex object includes approximate proofs understood as schemes constructed to support our assertions about the world on the basis of our incomplete or uncertain knowledge. 1 Introduction We present a formal model of approximate reasoning about processes of synthesis of complex systems. First ideas of this approach have been presented in [15], [24], [25], [27], [28], [29], [30], [31]. Our research has been stimulated by the demand for solutions of the following groups of problems, estimated in [1] to be crucial for the progress in the area of automated design and manufacturing. These groups of problems are concerned with the treatment of: Group 1. Poorly defined...
The art of granular computing:
 Proceeding of the International Conference on Rough Sets and Emerging Intelligent Systems Paradigms,
, 2007
"... Abstract: This paper has two purposes. One is to present a critical examination of the rise of granular computing and the other is to suggest a triarchic theory of granular computing. By examining the reasons, justifications, and motivations for the rise of granular computing, we may be able to ful ..."
Abstract

Cited by 74 (20 self)
 Add to MetaCart
(Show Context)
Abstract: This paper has two purposes. One is to present a critical examination of the rise of granular computing and the other is to suggest a triarchic theory of granular computing. By examining the reasons, justifications, and motivations for the rise of granular computing, we may be able to fully appreciate its scope, goal and potential values. The results enable us to formulate a triarchic theory in the light of research results from many disciplines. The three components of the theory are labeled as the philosophy, the methodology, and the computation. The integration of the three offers a unified view of granular computing as a way of structured thinking, a method of structured problem solving, and a paradigm of structured information processing, focusing on hierarchical granular structures. The triarchic theory is an important effort in synthesizing the various theories and models of granular computing. Key words: Triarchic theory of granular computing; systems theory; structured thinking, problem solving and information processing. CLC number: Document code: A Introduction Although granular computing, as a separate field of study, started a decade ago [1], its basic philosophy, ideas, principles, methodologies, theories and tools has, in fact, long been used either explicitly or implicitly across many branches of natural and social sciences The answers, at least partial answers, to these questions may be obtained by drawing and synthesizing results from wellestablished disciplines, including philosophy, psychology, neuroscience, cognitive science, education, artificial intelligence, computer programming, and many more. Previously, I argued that granular computing represents an idea converged from many branches of natural and social sciences HumanInspired Computing Research on understanding the human brain and natural intelligence is closely related to the field of artificial intelligence (AI) and information technology (IT). The results have led to a computational view for explaining how the mind works
Dynamic Reducts as a Tool for Extracting Laws from Decisions Tables
, 1994
"... . We apply rough set methods and boolean reasoning for knowledge discovery from decision tables. It is not always possible to extract general laws from experimental data by computing first all reducts [12] of a decision table and next decision rules on the basis of these reducts. We investigate a pr ..."
Abstract

Cited by 74 (13 self)
 Add to MetaCart
. We apply rough set methods and boolean reasoning for knowledge discovery from decision tables. It is not always possible to extract general laws from experimental data by computing first all reducts [12] of a decision table and next decision rules on the basis of these reducts. We investigate a problem how information about the reduct set changes in a random sampling process of a given decision table could be used to generate these laws. The reducts stable in the process of decision table sampling are called dynamic reducts. Dynamic reducts define the set of attributes called the dynamic core. This is the set of attributes included in all dynamic reducts. The set of decision rules can be computed from the dynamic core or from the best dynamic reducts. We report the results of experiments with different data sets, e.g. market data, medical data, textures and handwritten digits. The results are showing that dynamic reducts can help to extract laws from decision tables. Key words: evol...
SemanticsPreserving Dimensionality Reduction: Rough and FuzzyRough Based Approaches
 IEEE Transactions on Knowledge and Data Engineering
, 2004
"... Abstract—Semanticspreserving dimensionality reduction refers to the problem of selecting those input features that are most predictive of a given outcome; a problem encountered in many areas such as machine learning, pattern recognition, and signal processing. This has found successful application ..."
Abstract

Cited by 72 (12 self)
 Add to MetaCart
(Show Context)
Abstract—Semanticspreserving dimensionality reduction refers to the problem of selecting those input features that are most predictive of a given outcome; a problem encountered in many areas such as machine learning, pattern recognition, and signal processing. This has found successful application in tasks that involve data sets containing huge numbers of features (in the order of tens of thousands), which would be impossible to process further. Recent examples include text processing and Web content classification. One of the many successful applications of rough set theory has been to this feature selection area. This paper reviews those techniques that preserve the underlying semantics of the data, using crisp and fuzzy rough setbased methodologies. Several approaches to feature selection based on rough set theory are experimentally compared. Additionally, a new area in feature selection, feature grouping, is highlighted and a rough setbased feature grouping technique is detailed. Index Terms—Dimensionality reduction, feature selection, feature transformation, rough selection, fuzzyrough selection. 1