CFP last date
20 May 2024
Call for Paper
June Edition
IJCA solicits high quality original research papers for the upcoming June edition of the journal. The last date of research paper submission is 20 May 2024

Submit your paper
Know more
Reseach Article

Effective Decision Tree Learning

by C. Sudarsana Reddy, V. Vasu, B. Kumara Swamy Achari
International Journal of Computer Applications
Foundation of Computer Science (FCS), NY, USA
Volume 82 - Number 9
Year of Publication: 2013
Authors: C. Sudarsana Reddy, V. Vasu, B. Kumara Swamy Achari
10.5120/14141-7690

C. Sudarsana Reddy, V. Vasu, B. Kumara Swamy Achari . Effective Decision Tree Learning. International Journal of Computer Applications. 82, 9 ( November 2013), 1-6. DOI=10.5120/14141-7690

@article{ 10.5120/14141-7690,
author = { C. Sudarsana Reddy, V. Vasu, B. Kumara Swamy Achari },
title = { Effective Decision Tree Learning },
journal = { International Journal of Computer Applications },
issue_date = { November 2013 },
volume = { 82 },
number = { 9 },
month = { November },
year = { 2013 },
issn = { 0975-8887 },
pages = { 1-6 },
numpages = {9},
url = { https://ijcaonline.org/archives/volume82/number9/14141-7690/ },
doi = { 10.5120/14141-7690 },
publisher = {Foundation of Computer Science (FCS), NY, USA},
address = {New York, USA}
}
%0 Journal Article
%1 2024-02-06T21:57:16.826894+05:30
%A C. Sudarsana Reddy
%A V. Vasu
%A B. Kumara Swamy Achari
%T Effective Decision Tree Learning
%J International Journal of Computer Applications
%@ 0975-8887
%V 82
%N 9
%P 1-6
%D 2013
%I Foundation of Computer Science (FCS), NY, USA
Abstract

Classification is a data analysis technique. The decision tree is one of the most popular classification algorithms in current use for data mining because it is more interpretable. Training data sets are not error free due to measurement errors in the data collection process. Traditional decision tree classifiers are constructed without considering any errors in the values of attributes of the training data sets. We extend such classifiers to construct effective decision trees with error corrected training data sets. It is possible to build decision tree classifiers with higher accuracies especially when the measurement errors in the values of the attributes of the training data sets are corrected appropriately before using those training data sets in decision tree learning. Error corrected data sets can be used not only in decision tree learning but also in many data mining techniques. In general, values of attributes in training datasets are always inherently associated with errors. Data errors can be properly handled by using appropriate error models or error correction techniques. Also, sometimes for preserving data privacy, attribute values in the original training data sets are modified so that modified data sets contain data values with some errors. Later on, these modified data sets are reconstructed before applying those tuples to data mining technique. This paper introduces an effective decision tree (EDT) construction algorithm that uses a new error adjusting technique (NEAT) in constructing more accurate decision tree classifiers. The idea behind this new error adjusting technique is that 'many data sets with numerical attributes containing point data values have been collected via repeated measurements' and the process of repeated measurements is the common source of data errors in the training data sets. EDT describes an approach to correct the errors in the values of attributes of the training data sets and then error corrected attribute values of the data sets are used in decision tree learning.

References
  1. Jiawei Han, Micheline Kamber , Data Mining:Concepts and Techniques, Morgan Kaufmann, 2006.
  2. Introduction to Machine Learning EthemAlpaydin
  3. U. M. Fayyad and K. B. Irani, "On the Handling of Continuous –Valued Attributes in Decision tree Generation", Machine Learning, vol. 8, pp. 87-102, 1996.
  4. R. E. Walpole and R. H. Myers, Probability and Statistics for Engineers and Scientists. Macmillan Publishing Company, 1993.
  5. A. Asuncion and D. Newman, UCI Machine Learning Repository, http://www. ics. uci. edu/mlearn/MLRepository. html, 2007.
  6. T. Elomaa and J. Rousu, "General and Efficient Multisplitting of Numerical Attributes," Machine Learning, vol. 36, no. 3, pp. 201- 244, 1999.
  7. J. R. Quinlan, "Improved Use of Continuous attributes in C4. 5" , Journal of Artificial Intelligence Research, 4, pp. 77-90, 1996.
Index Terms

Computer Science
Information Sciences

Keywords

Decision tree Classification Data mining