Ara
Toplam kayıt 17, listelenen: 1-10
Tree Ensembles on the induced discrete space
(Institute of Electrical and Electronics Engineers Inc., 2016-05)
Decision trees are widely used predictive models in machine learning. Recently, K-tree is proposed, where the original discrete feature space is expanded by generating all orderings of values of k discrete attributes and ...
Searching for the optimal ordering of classes in rule induction
(IEEE, 2012-11-15)
Rule induction algorithms such as Ripper, solve a K > 2 class problem by converting it into a sequence of K - 1 two-class problems. As a usual heuristic, the classes are fed into the algorithm in the order of increasing ...
Regularizing soft decision trees
(Springer, 2013)
Recently, we have proposed a new decision tree family called soft decision trees where a node chooses both its left and right children with different probabilities as given by a gating function, different from a hard ...
Budding trees
(IEEE Computer Soc, 2014-08-24)
We propose a new decision tree model, named the budding tree, where a node can be both a leaf and an internal decision node. Each bud node starts as a leaf node, can then grow children, but then later on, if necessary, its ...
On the feature extraction in discrete space
(Elsevier Sci Ltd, 2014-05)
In many pattern recognition applications, feature space expansion is a key step for improving the performance of the classifier. In this paper, we (i) expand the discrete feature space by generating all orderings of values ...
VC-dimension of univariate decision trees
(IEEE-INST Electrical Electronics Engineers Inc, 2015-02-25)
In this paper, we give and prove the lower bounds of the Vapnik-Chervonenkis (VC)-dimension of the univariate decision tree hypothesis class. The VC-dimension of the univariate decision tree depends on the VC-dimension ...
Univariate decision tree induction using maximum margin classification
(Oxford Univ Press, 2012-03)
In many pattern recognition applications, first decision trees are used due to their simplicity and easily interpretable nature. In this paper, we propose a new decision tree learning algorithm called univariate margin ...
VC-dimension of rule sets
(IEEE Computer Soc, 2014-12-04)
In this paper, we give and prove lower bounds of the VC-dimension of the rule set hypothesis class where the input features are binary or continuous. The VC-dimension of the rule set depends on the VC-dimension values of ...
Omnivariate rule induction using a novel pairwise statistical test
(IEEE Computer Soc, 2013-09)
Rule learning algorithms, for example, RIPPER, induces univariate rules, that is, a propositional condition in a rule uses only one feature. In this paper, we propose an omnivariate induction of rules where under each ...
Mapping classifiers and datasets
(Pergamon-Elsevier Science Ltd, 2011-04)
Given the posterior probability estimates of 14 classifiers on 38 datasets, we plot two-dimensional maps of classifiers and datasets using principal component analysis (PCA) and Isomap. The similarity between classifiers ...