Back to Search Start Over

Regression tree-based active learning.

Authors :
Jose, Ashna
de Mendonça, João Paulo Almeida
Devijver, Emilie
Jakse, Noël
Monbet, Valérie
Poloni, Roberta
Source :
Data Mining & Knowledge Discovery; Mar2024, Vol. 38 Issue 2, p420-460, 41p
Publication Year :
2024

Abstract

Machine learning algorithms often require large training sets to perform well, but labeling such large amounts of data is not always feasible, as in many applications, substantial human effort and material cost is needed. Finding effective ways to reduce the size of training sets while maintaining the same performance is then crucial: one wants to choose the best sample of fixed size to be labeled among a given population, aiming at an accurate prediction of the response. This challenge has been studied in detail in classification, but not deeply enough in regression, which is known to be a more difficult task for active learning despite its need in practice. Few model-free active learning methods have been proposed that detect the new samples to be labeled using unlabeled data, but they lack the information of the conditional distribution between the response and the features. In this paper, we propose a standard regression tree-based active learning method for regression that improves significantly upon existing active learning approaches. It provides impressive results for small and large training sets and an appreciably low variance within several runs. We also exploit model-free approaches, and adapt them to our algorithm to utilize maximum information. Through experiments on numerous benchmark datasets, we demonstrate that our framework improves existing methods and is effective in learning a regression model from a very limited labeled dataset, reducing the sample size for a fixed level of performance, even with many features. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
13845810
Volume :
38
Issue :
2
Database :
Complementary Index
Journal :
Data Mining & Knowledge Discovery
Publication Type :
Academic Journal
Accession number :
175543362
Full Text :
https://doi.org/10.1007/s10618-023-00951-7