JMLR

Randomization Can Reduce Both Bias and Variance: A Case Study in Random Forests

Authors
Brian Liu Rahul Mazumder
Research Topics
Experimental Design
Paper Information
  • Journal:
    Journal of Machine Learning Research
  • Added to Tracker:
    Sep 08, 2025
Abstract

We study the often overlooked phenomenon, first noted in Breiman (2001), that random forests appear to reduce bias compared to bagging. Motivated by an interesting paper by Mentch and Zhou (2020), where the authors explain the success of random forests in low signal-to-noise ratio (SNR) settings through regularization, we explore how random forests can capture patterns in the data that bagging ensembles fail to capture. We empirically demonstrate that in the presence of such patterns, random forests reduce bias along with variance and can increasingly outperform bagging ensembles when SNR is high. Our observations offer insights into the real-world success of random forests across a range of SNRs and enhance our understanding of the difference between random forests and bagging ensembles. Our investigations also yield practical insights into the importance of tuning $mtry$ in random forests.

Author Details
Brian Liu
Author
Rahul Mazumder
Author
Research Topics & Keywords
Experimental Design
Research Area
Citation Information
APA Format
Brian Liu & Rahul Mazumder . Randomization Can Reduce Both Bias and Variance: A Case Study in Random Forests. Journal of Machine Learning Research .
BibTeX Format
@article{paper505,
  title = { Randomization Can Reduce Both Bias and Variance: A Case Study in Random Forests },
  author = { Brian Liu and Rahul Mazumder },
  journal = { Journal of Machine Learning Research },
  url = { https://www.jmlr.org/papers/v26/24-0255.html }
}