Found 1678 packages in 0.03 seconds
Distributional Random Forests
An implementation of distributional random forests as introduced in Cevid & Michel & Meinshausen & Buhlmann (2020)
Adversarial Random Forests
Adversarial random forests (ARFs) recursively partition data into fully factorized leaves, where features are jointly independent. The procedure is iterative, with alternating rounds of generation and discrimination. Data becomes increasingly realistic at each round, until original and synthetic samples can no longer be reliably distinguished. This is useful for several unsupervised learning tasks, such as density estimation and data synthesis. Methods for both are implemented in this package. ARFs naturally handle unstructured data with mixed continuous and categorical covariates. They inherit many of the benefits of random forests, including speed, flexibility, and solid performance with default parameters. For details, see Watson et al. (2023) < https://proceedings.mlr.press/v206/watson23a.html>.
Ordered Random Forests
An implementation of the Ordered Forest estimator as developed
in Lechner & Okasa (2019)
Modified Ordered Random Forest
Nonparametric estimator of the ordered choice model using random forests. The estimator modifies a standard random forest splitting criterion to build a collection of forests, each estimating the conditional probability of a single class. The package also implements a nonparametric estimator of the covariates’ marginal effects.
Random Forests for Longitudinal Data
Random forests are a statistical learning method widely used in many areas of scientific research essentially for its ability to learn complex relationships between input and output variables and also its capacity to handle high-dimensional data. However, current random forests approaches are not flexible enough to handle longitudinal data. In this package, we propose a general approach of random forests for high-dimensional longitudinal data. It includes a flexible stochastic model which allows the covariance structure to vary over time. Furthermore, we introduce a new method which takes intra-individual covariance into consideration to build random forests. The method is fully detailled in Capitaine et.al. (2020)
Visually Exploring Random Forests
Graphic elements for exploring Random Forests using the 'randomForest' or 'randomForestSRC' package for survival, regression and classification forests and 'ggplot2' package plotting.
Permutation Significance for Random Forests
Estimate False Discovery Rates (FDRs) for importance metrics from random forest runs.
Covariance Regression with Random Forests
Covariance Regression with Random Forests (CovRegRF) is a
random forest method for estimating the covariance matrix of a
multivariate response given a set of covariates. Random forest trees
are built with a new splitting rule which is designed to maximize the
distance between the sample covariance matrix estimates of the child
nodes. The method is described in Alakus et al. (2023)
Predictive Inference for Random Forests
An integrated package for constructing random forest prediction intervals using a fast implementation package 'ranger'. This package can apply the following three methods described in Haozhe Zhang, Joshua Zimmerman, Dan Nettleton, and Daniel J. Nordman (2019)
Random Forests for Dependent Data
Fits non-linear regression models on dependant data with Generalised Least Square (GLS) based Random Forest (RF-GLS) detailed in Saha, Basu and Datta (2021)