TY - GEN
T1 - Naive bayes classification given probability estimation trees
AU - Qin, Zengchang
PY - 2006
Y1 - 2006
N2 - Tree induction is one of the most effective and widely used models in classification. Unfortunately, decision trees such as C4.5 [9] have been found to provide poor probability estimates. By the empirical studies, Provost and Domingos [6] found that Probability Estimation Trees (PETs) give a fairly good probability estimation. However, different from normal decision trees, pruning reduces the performances of PETs. In order to get a good probability estimation, we usually need large trees which are not good in terms of the model transparency. In this paper, two hybrid models by combining the Naive Bayes classifier and PETs are proposed in order to build a model with good performance without losing too much transparency. The first model use Naive Bayes estimation given a PET and the second model use a group of small-sized PETs as Naive Bayes estimators. Empirical studies show that the first model outperforms the PET model at shallow depth and the second model is equivalent to Naive Bayes and PET.
AB - Tree induction is one of the most effective and widely used models in classification. Unfortunately, decision trees such as C4.5 [9] have been found to provide poor probability estimates. By the empirical studies, Provost and Domingos [6] found that Probability Estimation Trees (PETs) give a fairly good probability estimation. However, different from normal decision trees, pruning reduces the performances of PETs. In order to get a good probability estimation, we usually need large trees which are not good in terms of the model transparency. In this paper, two hybrid models by combining the Naive Bayes classifier and PETs are proposed in order to build a model with good performance without losing too much transparency. The first model use Naive Bayes estimation given a PET and the second model use a group of small-sized PETs as Naive Bayes estimators. Empirical studies show that the first model outperforms the PET model at shallow depth and the second model is equivalent to Naive Bayes and PET.
KW - Classification
KW - Decision tree
KW - Hybrid classification model
KW - Naive bayes
KW - Probability estimation tree
UR - https://www.scopus.com/pages/publications/40349086052
U2 - 10.1109/ICMLA.2006.36
DO - 10.1109/ICMLA.2006.36
M3 - 会议稿件
AN - SCOPUS:40349086052
SN - 0769527353
SN - 9780769527352
T3 - Proceedings - 5th International Conference on Machine Learning and Applications, ICMLA 2006
SP - 34
EP - 39
BT - Proceedings - 5th International Conference on Machine Learning and Applications, ICMLA 2006
T2 - 5th International Conference on Machine Learning and Applications, ICMLA 2006
Y2 - 14 December 2006 through 16 December 2006
ER -