Found 531 packages in 0.01 seconds
Risk Regression Models and Prediction Scores for Survival Analysis with Competing Risks
Implementation of the following methods for event history analysis. Risk regression models for survival endpoints also in the presence of competing risks are fitted using binomial regression based on a time sequence of binary event status variables. A formula interface for the Fine-Gray regression model and an interface for the combination of cause-specific Cox regression models. A toolbox for assessing and comparing performance of risk predictions (risk markers and risk prediction models). Prediction performance is measured by the Brier score and the area under the ROC curve for binary possibly time-dependent outcome. Inverse probability of censoring weighting and pseudo values are used to deal with right censored data. Lists of risk markers and lists of risk models are assessed simultaneously. Cross-validation repeatedly splits the data, trains the risk prediction models on one part of each split and then summarizes and compares the performance across splits.
Provides R-Language Code to Examine Quantitative Risk Management Concepts
Provides functions/methods to accompany the book Quantitative Risk Management: Concepts, Techniques and Tools by Alexander J. McNeil, Ruediger Frey, and Paul Embrechts.
Nonlinear Regression for Agricultural Applications
Additional nonlinear regression functions using self-start (SS) algorithms. One of the functions is the Beta growth function proposed by Yin et al. (2003)
Normalizing Transformation Functions
Estimate a suite of normalizing transformations, including a new adaptation of a technique based on ranks which can guarantee normally distributed transformed data if there are no ties: ordered quantile normalization (ORQ). ORQ normalization combines a rank-mapping approach with a shifted logit approximation that allows the transformation to work on data outside the original domain. It is also able to handle new data within the original domain via linear interpolation. The package is built to estimate the best normalizing transformation for a vector consistently and accurately. It implements the Box-Cox transformation, the Yeo-Johnson transformation, three types of Lambert WxF transformations, and the ordered quantile normalization transformation. It estimates the normalization efficacy of other commonly used transformations, and it allows users to specify custom transformations or normalization statistics. Finally, functionality can be integrated into a machine learning workflow via recipes.
Analysis of Quaternary Science Data
Constrained clustering, transfer functions, and other methods for analysing Quaternary science data.
Density, Probability, Quantile ('DPQ') Computations
Computations for approximations and alternatives for the 'DPQ'
(Density (pdf), Probability (cdf) and Quantile) functions for probability
distributions in R.
Primary focus is on (central and non-central) beta, gamma and related
distributions such as the chi-squared, F, and t.
--
For several distribution functions, provide functions implementing formulas from
Johnson, Kotz, and Kemp (1992)
Simple Interactive Controls for R using the 'tcltk' Package
A set of functions to build simple GUI controls for R functions. These are built on the 'tcltk' package. Uses could include changing a parameter on a graph by animating it with a slider or a "doublebutton", up to more sophisticated control panels. Some functions for specific graphical tasks, referred to as 'cartoons', are provided.
Multivariate Functional Principal Component Analysis for Data Observed on Different Dimensional Domains
Calculate a multivariate functional principal component analysis
for data observed on different dimensional domains. The estimation algorithm
relies on univariate basis expansions for each element of the multivariate
functional data (Happ & Greven, 2018)
Tools for Post-Processing Predicted Values
Models can be improved by post-processing class probabilities, by: recalibration, conversion to hard probabilities, assessment of equivocal zones, and other activities. 'probably' contains tools for conducting these operations as well as calibration tools and conformal inference techniques for regression models.
A Multi-Process 'dplyr' Backend
Partition a data frame across multiple worker processes to provide simple multicore parallelism.