Minimizing structural risk on decision tree classification

Research output: Chapter in Book/Report/Conference proceedingChapter

3 Citations (Scopus)

Abstract

Tree induction algorithms use heuristic information to obtain decision tree classification. However, there has been little research on how many rules are appropriate for a given set of data, that is, how we can find the best structure leading to desirable generalization performance. In this chapter, an evolutionary multi-objective optimization approach with genetic programming will be applied to the data classification problem in order to find the minimum error rate or the best pattern classifier for each size of decision trees. As a result, we can evaluate the classification performance under various structural complexity of decision trees. Following structural risk minimization suggested by Vapnik, we can determine a desirable number of rules with the best generalization performance. The suggested method is compared with C4.5 application for machine learning data.

Original languageEnglish
Title of host publicationMulti-Objective Machine Learning
EditorsYaochu Jin
Pages241-260
Number of pages20
DOIs
Publication statusPublished - 2006 Dec 12

Publication series

NameStudies in Computational Intelligence
Volume16
ISSN (Print)1860-949X

Fingerprint

Decision trees
Genetic programming
Trees (mathematics)
Multiobjective optimization
Learning systems
Classifiers

All Science Journal Classification (ASJC) codes

  • Artificial Intelligence

Cite this

Kim, D. (2006). Minimizing structural risk on decision tree classification. In Y. Jin (Ed.), Multi-Objective Machine Learning (pp. 241-260). (Studies in Computational Intelligence; Vol. 16). https://doi.org/10.1007/11399346_11
Kim, DaeEun. / Minimizing structural risk on decision tree classification. Multi-Objective Machine Learning. editor / Yaochu Jin. 2006. pp. 241-260 (Studies in Computational Intelligence).
@inbook{4f074461cafc4592b96569a6345b7004,
title = "Minimizing structural risk on decision tree classification",
abstract = "Tree induction algorithms use heuristic information to obtain decision tree classification. However, there has been little research on how many rules are appropriate for a given set of data, that is, how we can find the best structure leading to desirable generalization performance. In this chapter, an evolutionary multi-objective optimization approach with genetic programming will be applied to the data classification problem in order to find the minimum error rate or the best pattern classifier for each size of decision trees. As a result, we can evaluate the classification performance under various structural complexity of decision trees. Following structural risk minimization suggested by Vapnik, we can determine a desirable number of rules with the best generalization performance. The suggested method is compared with C4.5 application for machine learning data.",
author = "DaeEun Kim",
year = "2006",
month = "12",
day = "12",
doi = "10.1007/11399346_11",
language = "English",
isbn = "3540306765",
series = "Studies in Computational Intelligence",
pages = "241--260",
editor = "Yaochu Jin",
booktitle = "Multi-Objective Machine Learning",

}

Kim, D 2006, Minimizing structural risk on decision tree classification. in Y Jin (ed.), Multi-Objective Machine Learning. Studies in Computational Intelligence, vol. 16, pp. 241-260. https://doi.org/10.1007/11399346_11

Minimizing structural risk on decision tree classification. / Kim, DaeEun.

Multi-Objective Machine Learning. ed. / Yaochu Jin. 2006. p. 241-260 (Studies in Computational Intelligence; Vol. 16).

Research output: Chapter in Book/Report/Conference proceedingChapter

TY - CHAP

T1 - Minimizing structural risk on decision tree classification

AU - Kim, DaeEun

PY - 2006/12/12

Y1 - 2006/12/12

N2 - Tree induction algorithms use heuristic information to obtain decision tree classification. However, there has been little research on how many rules are appropriate for a given set of data, that is, how we can find the best structure leading to desirable generalization performance. In this chapter, an evolutionary multi-objective optimization approach with genetic programming will be applied to the data classification problem in order to find the minimum error rate or the best pattern classifier for each size of decision trees. As a result, we can evaluate the classification performance under various structural complexity of decision trees. Following structural risk minimization suggested by Vapnik, we can determine a desirable number of rules with the best generalization performance. The suggested method is compared with C4.5 application for machine learning data.

AB - Tree induction algorithms use heuristic information to obtain decision tree classification. However, there has been little research on how many rules are appropriate for a given set of data, that is, how we can find the best structure leading to desirable generalization performance. In this chapter, an evolutionary multi-objective optimization approach with genetic programming will be applied to the data classification problem in order to find the minimum error rate or the best pattern classifier for each size of decision trees. As a result, we can evaluate the classification performance under various structural complexity of decision trees. Following structural risk minimization suggested by Vapnik, we can determine a desirable number of rules with the best generalization performance. The suggested method is compared with C4.5 application for machine learning data.

UR - http://www.scopus.com/inward/record.url?scp=33845285575&partnerID=8YFLogxK

UR - http://www.scopus.com/inward/citedby.url?scp=33845285575&partnerID=8YFLogxK

U2 - 10.1007/11399346_11

DO - 10.1007/11399346_11

M3 - Chapter

AN - SCOPUS:33845285575

SN - 3540306765

SN - 9783540306764

T3 - Studies in Computational Intelligence

SP - 241

EP - 260

BT - Multi-Objective Machine Learning

A2 - Jin, Yaochu

ER -

Kim D. Minimizing structural risk on decision tree classification. In Jin Y, editor, Multi-Objective Machine Learning. 2006. p. 241-260. (Studies in Computational Intelligence). https://doi.org/10.1007/11399346_11