Found 1812 packages in 0.10 seconds
Variable Importance Measures for Multivariate Random Forests
Calculates two sets of post-hoc variable importance measures for multivariate random forests. The first set of variable importance measures are given by the sum of mean split improvements for splits defined by feature j measured on user-defined examples (i.e., training or testing samples). The second set of importance measures are calculated on a per-outcome variable basis as the sum of mean absolute difference of node values for each split defined by feature j measured on user-defined examples (i.e., training or testing samples). The user can optionally threshold both sets of importance measures to include only splits that are statistically significant as measured using an F-test.
A Unified Framework for Random Forest Prediction Error Estimation
Estimates the conditional error distributions of random forest predictions and common parameters of those distributions, including conditional misclassification rates, conditional mean squared prediction errors, conditional biases, and conditional quantiles, by out-of-bag weighting of out-of-bag prediction errors as proposed by Lu and Hardin (2021). This package is compatible with several existing packages that implement random forests in R.
Significance Level for Random Forest Impurity Importance Scores
Sets a significance level for Random Forest MDI (Mean Decrease in Impurity, Gini or
sum of squares) variable importance scores, using an empirical Bayes approach.
See Dunne et al. (2022)
ROSE Random Forests for Robust Semiparametric Efficient Estimation
ROSE (RObust Semiparametric Efficient) random forests for robust
semiparametric efficient estimation in partially parametric models (containing
generalised partially linear models).
Details can be found in the paper by Young and Shah (2024)
Bootstrap Stacking of Random Forest Models for Heterogeneous Data
Generates and predicts a set of linearly stacked Random Forest models using bootstrap sampling. Individual datasets may be heterogeneous (not all samples have full sets of features). Contains support for parallelization but the user should register their cores before running. This is an extension of the method found in Matlock (2018)
Explaining and Visualizing Random Forests in Terms of Variable Importance
A set of tools to help explain which variables are most important in a random forests. Various variable importance measures are calculated and visualized in different settings in order to get an idea on how their importance changes depending on our criteria (Hemant Ishwaran and Udaya B. Kogalur and Eiran Z. Gorodeski and Andy J. Minn and Michael S. Lauer (2010)
Mixed Effect Random Forests for Small Area Estimation
Mixed Effects Random Forests (MERFs) are a data-driven,
nonparametric alternative to current methods of Small Area Estimation
(SAE). 'SAEforest' provides functions for the estimation of regionally
disaggregated linear and nonlinear indicators using survey sample
data. Included procedures facilitate the estimation of domain-level
economic and inequality metrics and assess associated uncertainty.
Emphasis lies on straightforward interpretation and visualization of results.
From a methodological perspective, the package builds on approaches discussed in
Krennmair and Schmid (2022)
Exploring Heterogeneity in Meta-Analysis using Random Forests
Conduct random forests-based meta-analysis, obtain partial dependence plots for metaforest and classic meta-analyses, and cross-validate and tune metaforest- and classic meta-analyses in conjunction with the caret package. A requirement of classic meta-analysis is that the studies being aggregated are conceptually similar, and ideally, close replications. However, in many fields, there is substantial heterogeneity between studies on the same topic. Classic meta-analysis lacks the power to assess more than a handful of univariate moderators. MetaForest, by contrast, has substantial power to explore heterogeneity in meta-analysis. It can identify important moderators from a larger set of potential candidates (Van Lissa, 2020). This is an appealing quality, because many meta-analyses have small sample sizes. Moreover, MetaForest yields a measure of variable importance which can be used to identify important moderators, and offers partial prediction plots to explore the shape of the marginal relationship between moderators and effect size.
Random Forest Prediction Decomposition and Feature Importance Measure
An R re-implementation of the 'treeinterpreter' package on PyPI
< https://pypi.org/project/treeinterpreter/>. Each prediction can be
decomposed as 'prediction = bias + feature_1_contribution + ... +
feature_n_contribution'. This decomposition is then used to calculate
the Mean Decrease Impurity (MDI) and Mean Decrease Impurity using
out-of-bag samples (MDI-oob) feature importance measures based on the
work of Li et al. (2019)
Integrative Random Forest for Gene Regulatory Network Inference
Provides a flexible integrative algorithm that allows information from prior data, such as protein protein interactions and gene knock-down, to be jointly considered for gene regulatory network inference.