Back to Search Start Over

An improved decision tree algorithm based on variable precision neighborhood similarity.

Authors :
Liu, Caihui
Lin, Bowen
Lai, Jianying
Miao, Duoqian
Source :
Information Sciences. Nov2022, Vol. 615, p152-166. 15p.
Publication Year :
2022

Abstract

• The existing methods may lead to a contradictory in the transitivity. • A novel neighborhood geometric similarity was firstly defined. • Four new kinds of neighborhood similarities were proposed. • A new VPNRS model is constructed, and a novel decision tree algorithm is proposed. The decision tree algorithm has been widely used in data mining and machine learning due to its high accuracy, low computational cost and high interpretability. However, when dealing with the continuous data, the classical decision tree algorithm needs to replace continuous attributes with discretized attributes by the strategy of discretization. Discretization may cause a loss of information structure, which will affect the performance of classification. To tackle this problem, many researchers have proposed different decision tree methods based on variable precision neighborhood rough sets. However, these methods do not consider the geometric structure of neighborhood systems, which may lead to a contradiction in the transitivity of the equivalence relation. In this paper, we first define a novel neighborhood geometric similarity in a neighborhood system from the perspective of geometry. Second, by combining the neighborhood geometric similarity and the neighborhood algebraic similarity, we propose four new kinds of neighborhood similarities, which can solve the contradictory transitivity of the equivalence relation. Third, a variable precision neighborhood rough set model is constructed using the new similarities, and a novel decision tree algorithm is proposed based on this model, where the degree of attribute dependence is used as the partition measure. Experimental results on 14 selected datasets from the UCI Machine Learning Repository show that our algorithm is effective. The average accuracy of our algorithm is over 90%, which is 10% higher than the classical decision tree algorithms, and the number of leaf nodes increases slightly. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
00200255
Volume :
615
Database :
Academic Search Index
Journal :
Information Sciences
Publication Type :
Periodical
Accession number :
160251259
Full Text :
https://doi.org/10.1016/j.ins.2022.10.043