Examples: visualization, C++, networks, data cleaning, html widgets, ropensci.

Found 1771 packages in 0.05 seconds

metafor — by Wolfgang Viechtbauer, 9 months ago

Meta-Analysis Package for R

A comprehensive collection of functions for conducting meta-analyses in R. The package includes functions to calculate various effect sizes or outcome measures, fit equal-, fixed-, random-, and mixed-effects models to such data, carry out moderator and meta-regression analyses, and create various types of meta-analytical plots (e.g., forest, funnel, radial, L'Abbe, Baujat, bubble, and GOSH plots). For meta-analyses of binomial and person-time data, the package also provides functions that implement specialized methods, including the Mantel-Haenszel method, Peto's method, and a variety of suitable generalized linear (mixed-effects) models (i.e., mixed-effects logistic and Poisson regression models). Finally, the package provides functionality for fitting meta-analytic multivariate/multilevel models that account for non-independent sampling errors and/or true effects (e.g., due to the inclusion of multiple treatment studies, multiple endpoints, or other forms of clustering). Network meta-analyses and meta-analyses accounting for known correlation structures (e.g., due to phylogenetic relatedness) can also be conducted. An introduction to the package can be found in Viechtbauer (2010) .

fuzzyforest — by Daniel Conn, 6 years ago

Fuzzy Forests

Fuzzy forests, a new algorithm based on random forests, is designed to reduce the bias seen in random forest feature selection caused by the presence of correlated features. Fuzzy forests uses recursive feature elimination random forests to select features from separate blocks of correlated features where the correlation within each block of features is high and the correlation between blocks of features is low. One final random forest is fit using the surviving features. This package fits random forests using the 'randomForest' package and allows for easy use of 'WGCNA' to split features into distinct blocks. See D. Conn, Ngun, T., C. Ramirez, and G. Li (2019) for further details.

quantregRanger — by Philipp Probst, 8 years ago

Quantile Regression Forests for 'ranger'

This is the implementation of quantile regression forests for the fast random forest package 'ranger'.

rpms — by Daniell Toth, 4 years ago

Recursive Partitioning for Modeling Survey Data

Functions to allow users to build and analyze design consistent tree and random forest models using survey data from a complex sample design. The tree model algorithm can fit a linear model to survey data in each node obtained by recursively partitioning the data. The splitting variables and selected splits are obtained using a randomized permutation test procedure which adjusted for complex sample design features used to obtain the data. Likewise the model fitting algorithm produces design-consistent coefficients to any specified least squares linear model between the dependent and independent variables used in the end nodes. The main functions return the resulting binary tree or random forest as an object of "rpms" or "rpms_forest" type. The package also provides methods modeling a "boosted" tree or forest model and a tree model for zero-inflated data as well as a number of functions and methods available for use with these object types.

kernelFactory — by Michel Ballings, 10 years ago

Kernel Factory: An Ensemble of Kernel Machines

Binary classification based on an ensemble of kernel machines ("Ballings, M. and Van den Poel, D. (2013), Kernel Factory: An Ensemble of Kernel Machines. Expert Systems With Applications, 40(8), 2904-2913"). Kernel factory is an ensemble method where each base classifier (random forest) is fit on the kernel matrix of a subset of the training data.

SpatialML — by Stamatis Kalogirou, 2 years ago

Spatial Machine Learning

Implements a spatial extension of the random forest algorithm (Georganos et al. (2019) ). Allows for a geographically weighted random forest regression including a function to find the optical bandwidth. (Georganos and Kalogirou (2022) < https://www.mdpi.com/2220-9964/11/9/471>).

DMLLZU — by Junchang Zhao, 4 years ago

Double Machine Learning

Yang(2020,) come up with Double Machine Learning model ,it is based on this model, using four machine learning methods-- bagging, Boosting, random forest and neural network, and then based on the four models for two different combinations of the integrated model -- linear model combination and random forest .

SurvivalClusteringTree — by Lu You, a year ago

Clustering Analysis Using Survival Tree and Forest Algorithms

An outcome-guided algorithm is developed to identify clusters of samples with similar characteristics and survival rate. The algorithm first builds a random forest and then defines distances between samples based on the fitted random forest. Given the distances, we can apply hierarchical clustering algorithms to define clusters. Details about this method is described in < https://github.com/luyouepiusf/SurvivalClusteringTree>.

JOUSBoost — by Matthew Olson, 8 years ago

Implements Under/Oversampling for Probability Estimation

Implements under/oversampling for probability estimation. To be used with machine learning methods such as AdaBoost, random forests, etc.

outForest — by Michael Mayer, 2 years ago

Multivariate Outlier Detection and Replacement

Provides a random forest based implementation of the method described in Chapter 7.1.2 (Regression model based anomaly detection) of Chandola et al. (2009) . It works as follows: Each numeric variable is regressed onto all other variables by a random forest. If the scaled absolute difference between observed value and out-of-bag prediction of the corresponding random forest is suspiciously large, then a value is considered an outlier. The package offers different options to replace such outliers, e.g. by realistic values found via predictive mean matching. Once the method is trained on a reference data, it can be applied to new data.