Found 1824 packages in 0.01 seconds
Tune Random Forests Based on Variable Importance & Plot Results
Functions for assessing variable relations and associations prior to modeling with a Random Forest algorithm (although these are relevant for any predictive model). Metrics such as partial correlations and variance inflation factors are tabulated as well as plotted for the user. A function is available for tuning the main Random Forest hyper-parameter based on model performance and variable importance metrics. This grid-search technique provides tables and plots showing the effect of the main hyper-parameter on each of the assessment metrics. It also returns each of the evaluated models to the user. The package also provides superior variable importance plots for individual models. All of the plots are developed so that the user has the ability to edit and improve further upon the plots. Derivations and methodology are described in Bladen (2022) < https://digitalcommons.usu.edu/etd/8587/>.
QSAR Modeling with Multiple Algorithms: MLR, PLS, and Random Forest
Quantitative Structure-Activity Relationship (QSAR) modeling is a valuable tool in computational chemistry and drug design, where it aims to predict the activity or property of chemical compounds based on their molecular structure. In this vignette, we present the 'rQSAR' package, which provides functions for variable selection and QSAR modeling using Multiple Linear Regression (MLR), Partial Least Squares (PLS), and Random Forest algorithms.
Merging of Satellite Datasets with Ground Observations using Random Forests
S3 implementation of the Random Forest MErging Procedure (RF-MEP), which combines two or more satellite-based datasets (e.g., precipitation products, topography) with ground observations to produce a new dataset with improved spatio-temporal distribution of the target field. In particular, this package was developed to merge different Satellite-based Rainfall Estimates (SREs) with measurements from rain gauges, in order to obtain a new precipitation dataset where the time series in the rain gauges are used to correct different types of errors present in the SREs. However, this package might be used to merge other hydrological/environmental satellite fields with point observations. For details, see Baez-Villanueva et al. (2020)
Missing Value Imputation using Random Forest for Prediction Settings
Missing data imputation based on the 'missForest' algorithm (Stekhoven, Daniel J (2012)
Approximate False Positive Rate Control in Selection Frequency for Random Forest
Approximate false positive rate control in selection frequency for
random forest using the methods described by Ender Konukoglu and Melanie Ganz (2014)
Joint Random Forest (JRF) for the Simultaneous Estimation of Multiple Related Networks
Simultaneous estimation of multiple related networks.
Modeling and Map Production using Random Forest and Related Stochastic Models
Creates sophisticated models of training data and validates the models with an independent test set, cross validation, or Out Of Bag (OOB) predictions on the training data. Create graphs and tables of the model validation results. Applies these models to GIS .img files of predictors to create detailed prediction surfaces. Handles large predictor files for map making, by reading in the .img files in chunks, and output to the .txt file the prediction for each data chunk, before reading the next chunk of data.
Performing Monte Carlo Expectation Maximization Random Forest Imputation for Biological Data
Perform missing value imputation for biological data using the random forest algorithm, the imputation aim to keep the original mean and standard deviation consistent after imputation.
Innovative Complex Split Procedures in Random Forests Through Candidate Split Sampling
Implementation of three methods based on the diversity forest (DF) algorithm
(Hornung, 2022,
A Laboratory for Recursive Partytioning
A computational toolbox for recursive partitioning.
The core of the package is ctree(), an implementation of
conditional inference trees which embed tree-structured
regression models into a well defined theory of conditional
inference procedures. This non-parametric class of regression
trees is applicable to all kinds of regression problems, including
nominal, ordinal, numeric, censored as well as multivariate response
variables and arbitrary measurement scales of the covariates.
Based on conditional inference trees, cforest() provides an
implementation of Breiman's random forests. The function mob()
implements an algorithm for recursive partitioning based on
parametric models (e.g. linear models, GLMs or survival
regression) employing parameter instability tests for split
selection. Extensible functionality for visualizing tree-structured
regression models is available. The methods are described in
Hothorn et al. (2006)