Detailed Information

Cited 42 time in webofscience Cited 64 time in scopus
Metadata Downloads

Double random forest

Full metadata record
DC Field Value Language
dc.contributor.authorHan, Sunwoo-
dc.contributor.authorKim, Hyunjoong-
dc.contributor.authorLee, Yung-Seop-
dc.date.accessioned2023-04-27T22:40:38Z-
dc.date.available2023-04-27T22:40:38Z-
dc.date.issued2020-08-
dc.identifier.issn0885-6125-
dc.identifier.issn1573-0565-
dc.identifier.urihttps://scholarworks.dongguk.edu/handle/sw.dongguk/6373-
dc.description.abstractRandom forest (RF) is one of the most popular parallel ensemble methods, using decision trees as classifiers. One of the hyper-parameters to choose from for RF fitting is the nodesize, which determines the individual tree size. In this paper, we begin with the observation that for many data sets (34 out of 58), the best RF prediction accuracy is achieved when the trees are grown fully by minimizing the nodesize parameter. This observation leads to the idea that prediction accuracy could be further improved if we find a way to generate even bigger trees than the ones with a minimum nodesize. In other words, the largest tree created with the minimum nodesize parameter may not be sufficiently large for the best performance of RF. To produce bigger trees than those by RF, we propose a new classification ensemble method called double random forest (DRF). The new method uses bootstrap on each node during the tree creation process, instead of just bootstrapping once on the root node as in RF. This method, in turn, provides an ensemble of more diverse trees, allowing for more accurate predictions. Finally, for data where RF does not produce trees of sufficient size, we have successfully demonstrated that DRF provides more accurate predictions than RF.-
dc.format.extent18-
dc.language영어-
dc.language.isoENG-
dc.publisherSPRINGER-
dc.titleDouble random forest-
dc.typeArticle-
dc.publisher.location네델란드-
dc.identifier.doi10.1007/s10994-020-05889-1-
dc.identifier.scopusid2-s2.0-85087483727-
dc.identifier.wosid000545057500001-
dc.identifier.bibliographicCitationMACHINE LEARNING, v.109, no.8, pp 1569 - 1586-
dc.citation.titleMACHINE LEARNING-
dc.citation.volume109-
dc.citation.number8-
dc.citation.startPage1569-
dc.citation.endPage1586-
dc.type.docTypeArticle-
dc.description.isOpenAccessY-
dc.description.journalRegisteredClassscie-
dc.description.journalRegisteredClassscopus-
dc.relation.journalResearchAreaComputer Science-
dc.relation.journalWebOfScienceCategoryComputer Science, Artificial Intelligence-
dc.subject.keywordPlusCLASSIFICATION TREES-
dc.subject.keywordPlusALGORITHMS-
dc.subject.keywordPlusENSEMBLES-
dc.subject.keywordAuthorClassification-
dc.subject.keywordAuthorEnsemble-
dc.subject.keywordAuthorRandom forest-
dc.subject.keywordAuthorBootstrap-
dc.subject.keywordAuthorDecision tree-
Files in This Item
There are no files associated with this item.
Appears in
Collections
College of Natural Science > Department of Statistics > 1. Journal Articles

qrcode

Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.

Related Researcher

Researcher Lee, Yung Seop photo

Lee, Yung Seop
College of Natural Science (Department of Statistics)
Read more

Altmetrics

Total Views & Downloads

BROWSE