Examples: visualization, C++, networks, data cleaning, html widgets, ropensci.

Found 111 packages in 0.24 seconds

QQreflimits — by Jessica J. Kraker, 8 months ago

Reference Limits using QQ Methodology

A collection of routines for finding reference limits using, where appropriate, QQ methodology. All use a data vector X of cases from the reference population. The default is to get the central 95% reference range of the population, namely the 2.5 and 97.5 percentile, with optional adjustment of the range. Along with the reference limits, we want confidence intervals which, for historical reasons, are typically at 90% confidence. A full analysis provides six numbers: – the upper and the lower reference limits, and - each of their confidence intervals. For application details, see Hawkins and Esquivel (2024) .

CohensdpLibrary — by Denis Cousineau, 2 years ago

Cohen's D_p Computation with Confidence Intervals

Computing Cohen's d_p in any experimental designs (between-subject, within-subject, and single-group design). Cousineau (2022) < https://github.com/dcousin3/CohensdpLibrary>; Cohen (1969, ISBN: 0-8058-0283-5).

activPAL — by Craig Speirs, 8 months ago

Advanced Processing and Chart Generation from activPAL Events Files

Contains functions to generate pre-defined summary statistics from activPAL events files < https://www.palt.com/>. The package also contains functions to produce informative graphics that visualise physical activity behaviour and trends. This includes generating graphs that align physical activity behaviour with additional time based observations described by other data sets, such as sleep diaries and continuous glucose monitoring data.

ppwdeming — by Jessica J. Kraker, 2 months ago

Precision Profile Weighted Deming Regression

Weighted Deming regression, also known as 'errors-in-variable' regression, is applied with suitable weights. Weights are modeled via a precision profile; thus the methods implemented here are referred to as precision profile weighted Deming (PWD) regression. The package covers two settings – one where the precision profiles are known either from external studies or from adequate replication of the X and Y readings, and one in which there is a plausible functional form for the precision profiles but the exact (unknown) function must be estimated from the (generally singlicate) readings. The function set includes tools for: estimated standard errors (via jackknifing); standardized-residual analysis function with regression diagnostic tools for normality, linearity and constant variance; and an outlier analysis identifying significant outliers for closer investigation. The following reference provides further information on mathematical derivations and applications. Hawkins, D.M., and J.J. Kraker (2026). 'Precision Profile Weighted Deming Regression for Methods Comparison'. The Journal of Applied Laboratory Medicine 11, 379-392 .

EngrExpt — by Douglas Bates, 14 years ago

Data sets from "Introductory Statistics for Engineering Experimentation"

Datasets from Nelson, Coffin and Copeland "Introductory Statistics for Engineering Experimentation" (Elsevier, 2003) with sample code.

DEGRE — by Douglas Terra Machado, 4 years ago

Inferring Differentially Expressed Genes using Generalized Linear Mixed Models

Genes that are differentially expressed between two or more experimental conditions can be detected in RNA-Seq. A high biological variability may impact the discovery of these genes once it may be divergent between the fixed effects. However, this variability can be covered by the random effects. 'DEGRE' was designed to identify the differentially expressed genes considering fixed and random effects on individuals. These effects are identified earlier in the experimental design matrix. 'DEGRE' has the implementation of preprocessing procedures to clean the near zero gene reads in the count matrix, normalize by 'RLE' published in the 'DESeq2' package, 'Love et al. (2014)' and it fits a regression for each gene using the Generalized Linear Mixed Model with the negative binomial distribution, followed by a Wald test to assess the regression coefficients.

NPCD — by Yi Zheng, 6 years ago

Nonparametric Methods for Cognitive Diagnosis

An array of nonparametric and parametric estimation methods for cognitive diagnostic models, including nonparametric classification of examinee attribute profiles, joint maximum likelihood estimation (JMLE) of examinee attribute profiles and item parameters, and nonparametric refinement of the Q-matrix, as well as conditional maximum likelihood estimation (CMLE) of examinee attribute profiles given item parameters and CMLE of item parameters given examinee attribute profiles. Currently the nonparametric methods in the package support both conjunctive and disjunctive models, and the parametric methods in the package support the DINA model, the DINO model, the NIDA model, the G-NIDA model, and the R-RUM model.

POFIBGE — by Gabriel Assuncao, 4 years ago

Downloading, Reading and Analyzing POF Microdata - Package in Development

Provides tools for downloading, reading and analyzing the POF, a household survey from Brazilian Institute of Geography and Statistics - IBGE. The data must be downloaded from the official website < https://www.ibge.gov.br/>. Further analysis must be made using package 'survey'.

robustGarch — by Echo Liu, a year ago

Robust Garch(1,1) Model

A method for modeling robust generalized autoregressive conditional heteroskedasticity (Garch) (1,1) processes, providing robustness toward additive outliers instead of innovation outliers. This work is based on the methodology described by Muler and Yohai (2008) .

orderly — by Rich FitzJohn, 4 months ago

Lightweight Reproducible Reporting

Distributed reproducible computing framework, adopting ideas from git, docker and other software. By defining a lightweight interface around the inputs and outputs of an analysis, a lot of the repetitive work for reproducible research can be automated. We define a simple format for organising and describing work that facilitates collaborative reproducible research and acknowledges that all analyses are run multiple times over their lifespans.