Results 1 - 10
of
15
Bagging, Boosting, and C4.5
- In Proceedings of the Thirteenth National Conference on Artificial Intelligence
, 1996
"... Breiman's bagging and Freund and Schapire's boosting are recent methods for improving the predictive power of classifier learning systems. Both form a set of classifiers that are combined by voting, bagging by generating replicated bootstrap samples of the data, and boosting by adjusting ..."
Abstract
-
Cited by 326 (1 self)
- Add to MetaCart
(Show Context)
Breiman's bagging and Freund and Schapire's boosting are recent methods for improving the predictive power of classifier learning systems. Both form a set of classifiers that are combined by voting, bagging by generating replicated bootstrap samples of the data, and boosting by adjusting the weights of training instances. This paper reports results of applying both techniques to a system that learns decision trees and testing on a representative collection of datasets. While both approaches substantially improve predictive accuracy, boosting shows the greater benefit. On the other hand, boosting also produces severe degradation on some datasets. A small change to the way that boosting combines the votes of learned classifiers reduces this downside and also leads to slightly better results on most of the datasets considered. Introduction Designers of empirical machine learning systems are concerned with such issues as the computational cost of the learning method and the accuracy and ...
Simplifying Decision Trees: A Survey
, 1996
"... Induced decision trees are an extensively-researched solution to classification tasks. For many practical tasks, the trees produced by tree-generation algorithms are not comprehensible to users due to their size and complexity. Although many tree induction algorithms have been shown to produce simpl ..."
Abstract
-
Cited by 47 (6 self)
- Add to MetaCart
Induced decision trees are an extensively-researched solution to classification tasks. For many practical tasks, the trees produced by tree-generation algorithms are not comprehensible to users due to their size and complexity. Although many tree induction algorithms have been shown to produce simpler, more comprehensible trees (or data structures derived from trees) with good classification accuracy, tree simplification has usually been of secondary concern relative to accuracy and no attempt has been made to survey the literature from the perspective of simplification. We present a framework that organizes the approaches to tree simplification and summarize and critique the approaches within this framework. The purpose of this survey is to provide researchers and practitioners with a concise overview of tree-simplification approaches and insight into their relative capabilities. In our final discussion, we briefly describe some empirical findings and discuss the application of tree i...
Feature Generation Using General Constructor Functions
- MACHINE LEARNING
, 2002
"... Most classification algorithms receive as input a set of attributes of the classified objects. In many cases, however, the supplied set of attributes is not sufficient for creating an accurate, succinct and comprehensible representation of the target concept. To overcome this problem, researchers ha ..."
Abstract
-
Cited by 43 (6 self)
- Add to MetaCart
Most classification algorithms receive as input a set of attributes of the classified objects. In many cases, however, the supplied set of attributes is not sufficient for creating an accurate, succinct and comprehensible representation of the target concept. To overcome this problem, researchers have proposed algorithms for automatic construction of features. The majority of these algorithms use a limited predefined set of operators for building new features. In this paper we propose a generalized and flexible framework that is capable of generating features from any given set of constructor functions. These can be domain-independent functions such as arithmetic and logic operators, or domain-dependent operators that rely on partial knowledge on the part of the user. The paper describes an algorithm which receives as input a set of classified objects, a set of attributes, and a specification for a set of constructor functions that contains their domains, ranges and properties. The algorithm produces as output a set of generated features that can be used by standard concept learners to create improved classifiers. The algorithm maintains a set of its best generated features and improves this set iteratively. During each iteration, the algorithm performs a beam search over its defined feature space and constructs new features by applying constructor functions to the members of its current feature set. The search is guided by general heuristic measures that are not confined to a specific feature representation. The algorithm was applied to a variety of classification problems and was able to generate features that were strongly related to the underlying target concepts. These features also significantly improved the accuracy achieved by standard concept learners, for a ...
Automatic Feature Construction and a Simple Rule Induction Algorithm for Skin Detection
- In Proc. of the ICML Workshop on Machine Learning in Computer Vision
, 2002
"... Many vision systems use skin detection as a principal component. Skin detection algorithms, normally evaluate a single and thus limited color model, such as HSV, Y C r C b , YUV, RGB, normalized RGB, etc. Their limited performance, however, suggests that they are looking at the incorrect color model ..."
Abstract
-
Cited by 33 (1 self)
- Add to MetaCart
(Show Context)
Many vision systems use skin detection as a principal component. Skin detection algorithms, normally evaluate a single and thus limited color model, such as HSV, Y C r C b , YUV, RGB, normalized RGB, etc. Their limited performance, however, suggests that they are looking at the incorrect color models.
Constructing X-of-N Attributes for Decision Tree Learning
- Machine Learning
, 1998
"... . While many constructive induction algorithms focus on generating new binary attributes, this paper explores novel methods of constructing nominal and numeric attributes. We propose a new constructive operator, X-of-N. An X-of-N representation is a set containing one or more attribute-value pairs. ..."
Abstract
-
Cited by 20 (0 self)
- Add to MetaCart
. While many constructive induction algorithms focus on generating new binary attributes, this paper explores novel methods of constructing nominal and numeric attributes. We propose a new constructive operator, X-of-N. An X-of-N representation is a set containing one or more attribute-value pairs. For a given instance, the value of an X-of-N representation corresponds to the number of its attribute-value pairs that are true of the instance. A single X-of-N representation can directly and simply represent any concept that can be represented by a single conjunctive, a single disjunctive, or a single M-of-N representation commonly used for constructive induction, and the reverse is not true. In this paper, we describe a constructive decision tree learning algorithm, called XofN. When building decision trees, this algorithm creates one X-of-N representation, either as a nominal attribute or as a numeric attribute, at each decision node. The construction of X-of-N representations is carrie...
Constructing New Attributes for Decision Tree Learning
, 1996
"... A well-known fundamental limitation of selective induction algorithms is that when tasksupplied attributes are not adequate for, or directly relevant to, describing hypotheses, their performance in terms of prediction accuracy and/or theory complexity is poor. One solution to this problem is constru ..."
Abstract
-
Cited by 8 (3 self)
- Add to MetaCart
A well-known fundamental limitation of selective induction algorithms is that when tasksupplied attributes are not adequate for, or directly relevant to, describing hypotheses, their performance in terms of prediction accuracy and/or theory complexity is poor. One solution to this problem is constructive induction. It constructs, by using task-supplied attributes, new attributes that are expected to be more appropriate than the task-supplied attributes for describing the target concepts. This thesis focuses on constructive induction with decision trees as the theory description language. It explores: (1) novel approaches to constructing new binary attributes using existing constructive operators, and (2) novel methods of constructing new nominal and new continuous-valued attributes based on a newly proposed constructive operator. The thesis investigates a fixed rule-based approach to constructing new binary attributes for decision tree learning. It generates conjunctions from producti...
Building Intelligent Learning Database Systems
- AI Magazine
, 2000
"... Induction and deduction are two opposite operations in data mining applications. Induction extracts knowledge in the form of, say, rules or decision trees from existing data, and deduction applies induction results to interpret new data. An intelligent learning database (ILDB) system integrates m ..."
Abstract
-
Cited by 7 (2 self)
- Add to MetaCart
Induction and deduction are two opposite operations in data mining applications. Induction extracts knowledge in the form of, say, rules or decision trees from existing data, and deduction applies induction results to interpret new data. An intelligent learning database (ILDB) system integrates machine learning techniques with database and knowledge base technology. It starts with existing database technology and performs both induction and deduction. The integration of database technology, induction (from machine learning), and deduction (from knowledgebased systems) plays a key role in the construction of ILDB systems, as does the design of efficient induction and deduction algorithms. This paper presents a system structure for ILDB systems, and discusses practical issues for ILDB applications, such as instance selection and structured induction. 1 Introduction Over the past thirty years database research has evolved technologies that are now widely used in almost every co...
Continuous-valued X-of-N Attributes Versus Nominal X-of-N Attributes for Constructive Induction: A Case Study
- for Young Computer Scientists, Peking University
, 1995
"... : An X-of-N is a set containing one or more attribute-value pairs. For a given instance, its value corresponds to the number of its attribute-value pairs that are true. In this paper, we explore the characteristics and performance of continuousvalued X-of-N attributes versus nominal X-of-N attribute ..."
Abstract
-
Cited by 5 (4 self)
- Add to MetaCart
: An X-of-N is a set containing one or more attribute-value pairs. For a given instance, its value corresponds to the number of its attribute-value pairs that are true. In this paper, we explore the characteristics and performance of continuousvalued X-of-N attributes versus nominal X-of-N attributes for constructive induction. Nominal X-of-Ns are more representationally powerful than continuous-valued X-of-Ns, but the former suffer the "fragmentation" problem, although some mechanisms such as subsetting can help to solve the problem. Two approaches to constructive induction using continuous-valued X-of-Ns are described. Continuous-valued X-of-Ns perform better than nominal ones on domains that need X-of-Ns with only one cut point. On domains that need X-of-N representations with more than one cut point, nominal X-of-Ns perform better than continuous-valued ones. Experimental results on a set of artificial and real-world domains support these statements. 1. Introduction A wide variet...
Constructing X-of-N attributes with a genetic algorithm
- In Proc. of the Genetic and Evolutionary Computation Conference
, 2002
"... The predictive accuracy obtained by a classification algorithm is strongly dependent on the quality of the attributes of the data being mined. When the attributes are little relevant for predicting the class of a record, the predictive accuracy will tend to be low. To combat this problem, a natural ..."
Abstract
-
Cited by 3 (0 self)
- Add to MetaCart
The predictive accuracy obtained by a classification algorithm is strongly dependent on the quality of the attributes of the data being mined. When the attributes are little relevant for predicting the class of a record, the predictive accuracy will tend to be low. To combat this problem, a natural approach consists of constructing new attributes out of the original attributes. Many attribute construction algorithms work by simply constructing conjunctions and/or disjunctions of attribute-value pairs. This kind of representation has a limited expressiveness power to represent attribute interactions. A more expressive representation is X-of-N [Zheng 1995]. An Xof-N condition consists of a set of N attribute-value pairs. The value of an X-of-N condition for a given example
A Comparison of Constructive Induction with Different Types of New Attribute
, 1996
"... : This paper studies the effects on decision tree learning of constructing four types of attribute (conjunctive, disjunctive, M-of-N, and X-of-N representations). To reduce effects of other factors such as tree learning methods, new attribute search strategies, search starting points, evaluation fun ..."
Abstract
-
Cited by 2 (1 self)
- Add to MetaCart
: This paper studies the effects on decision tree learning of constructing four types of attribute (conjunctive, disjunctive, M-of-N, and X-of-N representations). To reduce effects of other factors such as tree learning methods, new attribute search strategies, search starting points, evaluation functions, and stopping criteria, a single tree learning algorithm is developed. With different option settings, it can construct four different types of new attribute, but all other factors are fixed. The study reveals that conjunctive and disjunctive representations have very similar performance in terms of prediction accuracy and theory complexity on a variety of concepts, even on DNF and CNF concepts that are usually thought to be suited only to one of the two kinds of representation. In addition, the study demonstrates that the stronger representation power of M-of-N than conjunction and disjunction and the stronger representation power of X-of-N than these three types of new attribute can...