C4.5: Programs for Machine Learning (書評)
Citations
8,658 citations
5,246 citations
Cites methods from "C4.5: Programs for Machine Learning..."
...…s, this measure is given by: ScoreC (s, S) = 2I s c (S) Hs(S) + Hc(S) , where Hc(S) is the (log) entropy of the classification in S, Hs(S) is the split entropy (also called split information by Quinlan (1986)), and I sc (S) is the mutual information of the split outcome and the classification....
[...]
4,775 citations
Cites background or methods from "C4.5: Programs for Machine Learning..."
...5: the decision-tree induction method developed by Quinlan (1993)...
[...]
...Table 2 displays the accuracies of the main classification approaches we have discussed throughout the paper using the abbreviations: NB: the naive Bayesian classifier BN: unrestricted Bayesian networks learned with the MDL score TANs: TAN networks learned according to Theorem 2, with smoothed parameters CLs: CL multinet classifier—Bayesian multinets learned according to Theorem 1—with smoothed parameters C4.5: the decision-tree induction method developed by Quinlan (1993) SNB: theselective naive Bayesian classifier, a wrapper-based feature selection applied to naive Bayes, using the implementation of John and Kohavi (1997) In the previous sections we discussed these results in some detail....
[...]
...5 (Quinlan, 1993), a state-of-the-art decision tree learner, we may infer that TAN should perform rather well in comparison to C4....
[...]
...To confirm this prediction, we performed experiments comparing TAN to C4.5, and also to theselective naive Bayesianclassifier (Langley & Sage, 1994; John & Kohavi, 1997)....
[...]
...These results also show that both TAN and the CL multinet classifier are roughly equivalent in terms of accuracy, dominate the naive Bayesian classifier, and compare favorably with both C4.5 and the selective naive Bayesian classifier....
[...]
3,212 citations
2,672 citations
References
8,658 citations
5,246 citations
4,775 citations
3,212 citations
2,672 citations