Multi-objective classification based on bloomy decision tree

Masafumi Gotoh, Yuta Choki, Einoshin Suzuki

Research output: Contribution to journalArticle

Abstract

This paper presents a novel decision-tree induction for a multi-objective data set, i.e. a data set with a multi-dimensional class. Inductive decision-tree learning is one of the frequently-used methods for a single-objective data set, i.e. a data set with a single-dimensional class. However, in a real data analysis, we usually have multiple objectives, and a classifier which explains them simultaneously would be useful. A conventional decision-tree inducer requires transformation of a multi-dimensional class into a singledimensional class, but such a transformation can considerably worsen both accuracy and readability. In order to circumvent this problem we propose a bloomy decision tree which deals with a multi-dimensional class without such transformations. A bloomy decision tree consists of a set of decision nodes each of which splits examples according to their attribute values, and a set of ower nodes each of which decidesa dimension of the class for examples. A flower node appears not only at the fringe of a tree but also inside a tree. Our pruning is executed during tree construction, and evaluates each dimension of the class based on Cramér's V. The proposed method has been implemented as D3-B (Decision tree in Bloom), and tested with eleven benchmark data sets in the machine learning community. The experiments showed that D3-B has higher accuracies in nine data sets than C4.5 and tied with it in the other two data sets. In terms of readability, D3-B has a smaller number of decision nodes in all data sets, and thus outperforms C4.5. Moreover, experts in agriculture evaluated bloomy decision trees, each of which is induced from an agricultural data set, and found them appropriate and interesting.

Original languageEnglish
Pages (from-to)193-201
Number of pages9
JournalTransactions of the Japanese Society for Artificial Intelligence
Volume16
Issue number2
DOIs
Publication statusPublished - Dec 1 2001
Externally publishedYes

Fingerprint

Decision trees
Agriculture
Learning systems
Classifiers
Experiments

All Science Journal Classification (ASJC) codes

  • Artificial Intelligence

Cite this

Multi-objective classification based on bloomy decision tree. / Gotoh, Masafumi; Choki, Yuta; Suzuki, Einoshin.

In: Transactions of the Japanese Society for Artificial Intelligence, Vol. 16, No. 2, 01.12.2001, p. 193-201.

Research output: Contribution to journalArticle

@article{5578b5ac5c274a1b92af67a25648699a,
title = "Multi-objective classification based on bloomy decision tree",
abstract = "This paper presents a novel decision-tree induction for a multi-objective data set, i.e. a data set with a multi-dimensional class. Inductive decision-tree learning is one of the frequently-used methods for a single-objective data set, i.e. a data set with a single-dimensional class. However, in a real data analysis, we usually have multiple objectives, and a classifier which explains them simultaneously would be useful. A conventional decision-tree inducer requires transformation of a multi-dimensional class into a singledimensional class, but such a transformation can considerably worsen both accuracy and readability. In order to circumvent this problem we propose a bloomy decision tree which deals with a multi-dimensional class without such transformations. A bloomy decision tree consists of a set of decision nodes each of which splits examples according to their attribute values, and a set of ower nodes each of which decidesa dimension of the class for examples. A flower node appears not only at the fringe of a tree but also inside a tree. Our pruning is executed during tree construction, and evaluates each dimension of the class based on Cram{\'e}r's V. The proposed method has been implemented as D3-B (Decision tree in Bloom), and tested with eleven benchmark data sets in the machine learning community. The experiments showed that D3-B has higher accuracies in nine data sets than C4.5 and tied with it in the other two data sets. In terms of readability, D3-B has a smaller number of decision nodes in all data sets, and thus outperforms C4.5. Moreover, experts in agriculture evaluated bloomy decision trees, each of which is induced from an agricultural data set, and found them appropriate and interesting.",
author = "Masafumi Gotoh and Yuta Choki and Einoshin Suzuki",
year = "2001",
month = "12",
day = "1",
doi = "10.1527/tjsai.16.193",
language = "English",
volume = "16",
pages = "193--201",
journal = "Transactions of the Japanese Society for Artificial Intelligence",
issn = "1346-0714",
publisher = "Japanese Society for Artificial Intelligence",
number = "2",

}

TY - JOUR

T1 - Multi-objective classification based on bloomy decision tree

AU - Gotoh, Masafumi

AU - Choki, Yuta

AU - Suzuki, Einoshin

PY - 2001/12/1

Y1 - 2001/12/1

N2 - This paper presents a novel decision-tree induction for a multi-objective data set, i.e. a data set with a multi-dimensional class. Inductive decision-tree learning is one of the frequently-used methods for a single-objective data set, i.e. a data set with a single-dimensional class. However, in a real data analysis, we usually have multiple objectives, and a classifier which explains them simultaneously would be useful. A conventional decision-tree inducer requires transformation of a multi-dimensional class into a singledimensional class, but such a transformation can considerably worsen both accuracy and readability. In order to circumvent this problem we propose a bloomy decision tree which deals with a multi-dimensional class without such transformations. A bloomy decision tree consists of a set of decision nodes each of which splits examples according to their attribute values, and a set of ower nodes each of which decidesa dimension of the class for examples. A flower node appears not only at the fringe of a tree but also inside a tree. Our pruning is executed during tree construction, and evaluates each dimension of the class based on Cramér's V. The proposed method has been implemented as D3-B (Decision tree in Bloom), and tested with eleven benchmark data sets in the machine learning community. The experiments showed that D3-B has higher accuracies in nine data sets than C4.5 and tied with it in the other two data sets. In terms of readability, D3-B has a smaller number of decision nodes in all data sets, and thus outperforms C4.5. Moreover, experts in agriculture evaluated bloomy decision trees, each of which is induced from an agricultural data set, and found them appropriate and interesting.

AB - This paper presents a novel decision-tree induction for a multi-objective data set, i.e. a data set with a multi-dimensional class. Inductive decision-tree learning is one of the frequently-used methods for a single-objective data set, i.e. a data set with a single-dimensional class. However, in a real data analysis, we usually have multiple objectives, and a classifier which explains them simultaneously would be useful. A conventional decision-tree inducer requires transformation of a multi-dimensional class into a singledimensional class, but such a transformation can considerably worsen both accuracy and readability. In order to circumvent this problem we propose a bloomy decision tree which deals with a multi-dimensional class without such transformations. A bloomy decision tree consists of a set of decision nodes each of which splits examples according to their attribute values, and a set of ower nodes each of which decidesa dimension of the class for examples. A flower node appears not only at the fringe of a tree but also inside a tree. Our pruning is executed during tree construction, and evaluates each dimension of the class based on Cramér's V. The proposed method has been implemented as D3-B (Decision tree in Bloom), and tested with eleven benchmark data sets in the machine learning community. The experiments showed that D3-B has higher accuracies in nine data sets than C4.5 and tied with it in the other two data sets. In terms of readability, D3-B has a smaller number of decision nodes in all data sets, and thus outperforms C4.5. Moreover, experts in agriculture evaluated bloomy decision trees, each of which is induced from an agricultural data set, and found them appropriate and interesting.

UR - http://www.scopus.com/inward/record.url?scp=18544366981&partnerID=8YFLogxK

UR - http://www.scopus.com/inward/citedby.url?scp=18544366981&partnerID=8YFLogxK

U2 - 10.1527/tjsai.16.193

DO - 10.1527/tjsai.16.193

M3 - Article

VL - 16

SP - 193

EP - 201

JO - Transactions of the Japanese Society for Artificial Intelligence

JF - Transactions of the Japanese Society for Artificial Intelligence

SN - 1346-0714

IS - 2

ER -