Examples: visualization, C++, networks, data cleaning, html widgets, ropensci.

Found 1641 packages in 0.26 seconds

randomForest — by Andy Liaw, 2 months ago

Breiman and Cutlers Random Forests for Classification and Regression

Classification and regression based on a forest of trees using random inputs, based on Breiman (2001) .

missForest — by Daniel J. Stekhoven, 3 years ago

Nonparametric Missing Value Imputation using Random Forest

The function 'missForest' in this package is used to impute missing values particularly in the case of mixed-type data. It uses a random forest trained on the observed values of a data matrix to predict the missing values. It can be used to impute continuous and/or categorical data including complex interactions and non-linear relations. It yields an out-of-bag (OOB) imputation error estimate without the need of a test set or elaborate cross-validation. It can be run in parallel to save computation time.

randomForestSRC — by Udaya B. Kogalur, 4 months ago

Fast Unified Random Forests for Survival, Regression, and Classification (RF-SRC)

Fast OpenMP parallel computing of Breiman's random forests for univariate, multivariate, unsupervised, survival, competing risks, class imbalanced classification and quantile regression. New Mahalanobis splitting for correlated outcomes. Extreme random forests and randomized splitting. Suite of imputation methods for missing data. Fast random forests using subsampling. Confidence regions and standard errors for variable importance. New improved holdout importance. Case-specific importance. Minimal depth variable importance. Visualize trees on your Safari or Google Chrome browser. Anonymous random forests for data privacy.

aorsf — by Byron Jaeger, 6 months ago

Accelerated Oblique Random Forests

Fit, interpret, and compute predictions with oblique random forests. Includes support for partial dependence, variable importance, passing customized functions for variable importance and identification of linear combinations of features. Methods for the oblique random survival forest are described in Jaeger et al., (2023) .

MultivariateRandomForest — by Raziur Rahman, 8 years ago

Models Multivariate Cases Using Random Forests

Models and predicts multiple output features in single random forest considering the linear relation among the output features, see details in Rahman et al (2017).

RRF — by Houtao Deng, 17 days ago

Regularized Random Forest

Feature Selection with Regularized Random Forest. This package is based on the 'randomForest' package by Andy Liaw. The key difference is the RRF() function that builds a regularized random forest. Fortran original by Leo Breiman and Adele Cutler, R port by Andy Liaw and Matthew Wiener, Regularized random forest for classification by Houtao Deng, Regularized random forest for regression by Xin Guan. Reference: Houtao Deng (2013) .

VSURF — by Robin Genuer, 2 years ago

Variable Selection Using Random Forests

Three steps variable selection procedure based on random forests. Initially developed to handle high dimensional data (for which number of variables largely exceeds number of observations), the package is very versatile and can treat most dimensions of data, for regression and supervised classification problems. First step is dedicated to eliminate irrelevant variables from the dataset. Second step aims to select all variables related to the response for interpretation purpose. Third step refines the selection by eliminating redundancy in the set of variables selected by the second step, for prediction purpose. Genuer, R. Poggi, J.-M. and Tuleau-Malot, C. (2015) < https://journal.r-project.org/archive/2015-2/genuer-poggi-tuleaumalot.pdf>.

Rborist — by Mark Seligman, 10 months ago

Extensible, Parallelizable Implementation of the Random Forest Algorithm

Scalable implementation of classification and regression forests, as described by Breiman (2001), .

drf — by Loris Michel, 4 years ago

Distributional Random Forests

An implementation of distributional random forests as introduced in Cevid & Michel & Meinshausen & Buhlmann (2020) .

arf — by Marvin N. Wright, 10 months ago

Adversarial Random Forests

Adversarial random forests (ARFs) recursively partition data into fully factorized leaves, where features are jointly independent. The procedure is iterative, with alternating rounds of generation and discrimination. Data becomes increasingly realistic at each round, until original and synthetic samples can no longer be reliably distinguished. This is useful for several unsupervised learning tasks, such as density estimation and data synthesis. Methods for both are implemented in this package. ARFs naturally handle unstructured data with mixed continuous and categorical covariates. They inherit many of the benefits of random forests, including speed, flexibility, and solid performance with default parameters. For details, see Watson et al. (2022) .