Found 1771 packages in 0.05 seconds
Meta-Analysis Package for R
A comprehensive collection of functions for conducting meta-analyses in R. The package includes functions to calculate various effect sizes or outcome measures, fit equal-, fixed-, random-, and mixed-effects models to such data, carry out moderator and meta-regression analyses, and create various types of meta-analytical plots (e.g., forest, funnel, radial, L'Abbe, Baujat, bubble, and GOSH plots). For meta-analyses of binomial and person-time data, the package also provides functions that implement specialized methods, including the Mantel-Haenszel method, Peto's method, and a variety of suitable generalized linear (mixed-effects) models (i.e., mixed-effects logistic and Poisson regression models). Finally, the package provides functionality for fitting meta-analytic multivariate/multilevel models that account for non-independent sampling errors and/or true effects (e.g., due to the inclusion of multiple treatment studies, multiple endpoints, or other forms of clustering). Network meta-analyses and meta-analyses accounting for known correlation structures (e.g., due to phylogenetic relatedness) can also be conducted. An introduction to the package can be found in Viechtbauer (2010)
Fuzzy Forests
Fuzzy forests, a new algorithm based on random forests,
is designed to reduce the bias seen in random forest feature selection
caused by the presence of correlated features. Fuzzy forests uses
recursive feature elimination random forests to select
features from separate blocks of correlated features where the
correlation within each block of features is high
and the correlation between blocks of features is low.
One final random forest is fit using the surviving features.
This package fits random forests using the 'randomForest' package and
allows for easy use of 'WGCNA' to split features into distinct blocks.
See D. Conn, Ngun, T., C. Ramirez, and G. Li (2019)
Quantile Regression Forests for 'ranger'
This is the implementation of quantile regression forests for the fast random forest package 'ranger'.
Recursive Partitioning for Modeling Survey Data
Functions to allow users to build and analyze design consistent tree and random forest models using survey data from a complex sample design. The tree model algorithm can fit a linear model to survey data in each node obtained by recursively partitioning the data. The splitting variables and selected splits are obtained using a randomized permutation test procedure which adjusted for complex sample design features used to obtain the data. Likewise the model fitting algorithm produces design-consistent coefficients to any specified least squares linear model between the dependent and independent variables used in the end nodes. The main functions return the resulting binary tree or random forest as an object of "rpms" or "rpms_forest" type. The package also provides methods modeling a "boosted" tree or forest model and a tree model for zero-inflated data as well as a number of functions and methods available for use with these object types.
Kernel Factory: An Ensemble of Kernel Machines
Binary classification based on an ensemble of kernel machines ("Ballings, M. and Van den Poel, D. (2013), Kernel Factory: An Ensemble of Kernel Machines. Expert Systems With Applications, 40(8), 2904-2913"). Kernel factory is an ensemble method where each base classifier (random forest) is fit on the kernel matrix of a subset of the training data.
Spatial Machine Learning
Implements a spatial extension of the random forest algorithm
(Georganos et al. (2019)
Double Machine Learning
Yang(2020,
Clustering Analysis Using Survival Tree and Forest Algorithms
An outcome-guided algorithm is developed to identify clusters of samples with similar characteristics and survival rate. The algorithm first builds a random forest and then defines distances between samples based on the fitted random forest. Given the distances, we can apply hierarchical clustering algorithms to define clusters. Details about this method is described in < https://github.com/luyouepiusf/SurvivalClusteringTree>.
Implements Under/Oversampling for Probability Estimation
Implements under/oversampling for probability estimation. To be used with machine learning methods such as AdaBoost, random forests, etc.
Multivariate Outlier Detection and Replacement
Provides a random forest based implementation of the method
described in Chapter 7.1.2 (Regression model based anomaly detection)
of Chandola et al. (2009)