Found 10000 packages in 0.01 seconds
Obtaining Stars from Flat Tables
Data in multidimensional systems is obtained from operational systems and is transformed to adapt it to the new structure. Frequently, the operations to be performed aim to transform a flat table into a star schema. Transformations can be carried out using professional extract, transform and load tools or tools intended for data transformation for end users. With the tools mentioned, this transformation can be carried out, but it requires a lot of work. The main objective of this package is to define transformations that allow obtaining stars from flat tables easily. In addition, it includes basic data cleaning, dimension enrichment, incremental data refresh and query operations, adapted to this context.
Processing 'Gen5' 2.06 Exported Data
A collection of functions for processing 'Gen5' 2.06 exported data. 'Gen5' is an essential data analysis software for BioTek plate readers < https://www.biotek.com/products/software-robotics-software/gen5-microplate-reader-and-imager-software/>. This package contains functions for data cleaning, modeling and plotting using exported data from 'Gen5' version 2.06. It exports technically correct data defined in (Edwin de Jonge and Mark van der Loo (2013) < https://cran.r-project.org/doc/contrib/de_Jonge+van_der_Loo-Introduction_to_data_cleaning_with_R.pdf>) for customized analysis. It contains Boltzmann fitting for general kinetic analysis. See < https://www.github.com/yanxianUCSB/gen5helper> for more information, documentation and examples.
Interface for Multilevel Regression and Poststratification
Dual interfaces, graphical and programmatic, designed for
intuitive applications of Multilevel Regression and Poststratification (MRP).
Users can apply the method to a variety of datasets, from electronic health records
to sample survey data, through an end-to-end Bayesian data analysis workflow.
The package provides robust tools for data cleaning, exploratory analysis,
flexible model building, and insightful result visualization. For more details, see
Si et al. (2020) < https://www150.statcan.gc.ca/n1/en/pub/12-001-x/2020002/article/00003-eng.pdf?st=iF1_Fbrh>
and Si (2025)
Develop Text Prediction Models Based on N-Grams
A framework for developing n-gram models for text prediction. It provides data cleaning, data sampling, extracting tokens from text, model generation, model evaluation and word prediction. For information on how n-gram models work we referred to: "Speech and Language Processing" < https://web.archive.org/web/20240919222934/https%3A%2F%2Fweb.stanford.edu%2F~jurafsky%2Fslp3%2F3.pdf>. For optimizing R code and using R6 classes we referred to "Advanced R" < https://adv-r.hadley.nz/r6.html>. For writing R extensions we referred to "R Packages", < https://r-pkgs.org/index.html>.
Language Model Agents in R for AI Workflows and Research
Provides modular, graph-based agents powered by large language models (LLMs) for intelligent task execution in R. Supports structured workflows for tasks such as forecasting, data visualization, feature engineering, data wrangling, data cleaning, 'SQL', code generation, weather reporting, and research-driven question answering. Each agent performs iterative reasoning: recommending steps, generating R code, executing, debugging, and explaining results. Includes built-in support for packages such as 'tidymodels', 'modeltime', 'plotly', 'ggplot2', and 'prophet'. Designed for analysts, developers, and teams building intelligent, reproducible AI workflows in R. Compatible with LLM providers such as 'OpenAI', 'Anthropic', 'Groq', and 'Ollama'. Inspired by the Python package 'langagent'.
Make Dealing with Dates a Little Easier
Functions to work with date-times and time-spans: fast and user friendly parsing of date-time data, extraction and updating of components of a date-time (years, months, days, hours, minutes, and seconds), algebraic manipulation on date-time and time-span objects. The 'lubridate' package has a consistent and memorable syntax that makes working with dates easy and fun.
Read, Iteratively Filter, and Analyze Multiple ECG Datasets
Allows users to quickly load multiple patients' electrocardiographic
(ECG) data at once and conduct relevant time analysis of heart rate variability
(HRV) without manual edits from a physician or data cleaning specialist.
The package provides the unique ability to iteratively filter, plot,
and store time analysis results in a data frame while writing plots to a
predefined folder. This streamlines the workflow for HRV analysis across
multiple datasets. Methods are based on RodrÃguez-Liñares et al. (2011)
Creating Contact and Social Networks
Process spatially- and temporally-discrete data into contact and social networks, and facilitate network analysis by randomizing individuals' movement paths and/or related categorical variables. To use this package, users need only have a dataset containing spatial data (i.e., latitude/longitude, or planar x & y coordinates), individual IDs relating spatial data to specific individuals, and date/time information relating spatial locations to temporal locations. The functionality of this package ranges from data "cleaning" via multiple filtration functions, to spatial and temporal data interpolation, and network creation and analysis. Functions within this package are not limited to describing interpersonal contacts. Package functions can also identify and quantify "contacts" between individuals and fixed areas (e.g., home ranges, water bodies, buildings, etc.). As such, this package is an incredibly useful resource for facilitating epidemiological, ecological, ethological and sociological research.
Performance Loss Rate Analysis Pipeline
The pipeline contained in this package provides tools used in the
Solar Durability and Lifetime Extension Center (SDLE) for the analysis of
Performance Loss Rates (PLR) in real world photovoltaic systems. Functions
included allow for data cleaning, feature correction, power predictive modeling,
PLR determination, and uncertainty bootstrapping through various methods
Furniture for Quantitative Scientists
Contains four main functions (i.e., four pieces of furniture): table1() which produces a well-formatted table of descriptive statistics common as Table 1 in research articles, tableC() which produces a well-formatted table of correlations, tableF() which provides frequency counts, and washer() which is helpful in cleaning up the data. These furniture-themed functions are designed to simplify common tasks in quantitative analysis. Other data summary and cleaning tools are also available.