Examples: visualization, C++, networks, data cleaning, html widgets, ropensci.

Found 1769 packages in 0.02 seconds

qpdf — by Jeroen Ooms, 6 months ago

Split, Combine and Compress PDF Files

Content-preserving transformations transformations of PDF files such as split, combine, and compress. This package interfaces directly to the 'qpdf' C++ library < https://qpdf.sourceforge.io/> and does not require any command line utilities. Note that 'qpdf' does not read actual content from PDF files: to extract text and data you need the 'pdftools' package.

NMproject — by Tarj Sahota, 3 years ago

Script Based 'NONMEM' Model Development

Industrialisation of 'NONMEM' < https://www.iconplc.com/innovation/nonmem/> via fully and rapidly reusable model development 'workflows' entirely within 'RStudio'. Quickly get started with new models by importing 'NONMEM' templates from the built-in code library. Manipulate 'NONMEM' code from within R either via the tracked 'manual edit' interface or 'programmatically' via convenience functions. Script 'workflows' by piping sequences of model building steps from control file creation, to execution, to post-processing and evaluation. Run caching makes 'workflows' R markdown friendly for easy documentation of thoughts and modelling decisions alongside executable code. Share, reuse and recycle 'workflows' for new problems.

roxygen2 — by Hadley Wickham, 3 months ago

In-Line Documentation for R

Generate your Rd documentation, 'NAMESPACE' file, and collation field using specially formatted comments. Writing documentation in-line with code makes it easier to keep your documentation up-to-date as your requirements change. 'roxygen2' is inspired by the 'Doxygen' system for C++.

archivist — by Przemyslaw Biecek, a year ago

Tools for Storing, Restoring and Searching for R Objects

Data exploration and modelling is a process in which a lot of data artifacts are produced. Artifacts like: subsets, data aggregates, plots, statistical models, different versions of data sets and different versions of results. The more projects we work with the more artifacts are produced and the harder it is to manage these artifacts. Archivist helps to store and manage artifacts created in R. Archivist allows you to store selected artifacts as a binary files together with their metadata and relations. Archivist allows to share artifacts with others, either through shared folder or github. Archivist allows to look for already created artifacts by using it's class, name, date of the creation or other properties. Makes it easy to restore such artifacts. Archivist allows to check if new artifact is the exact copy that was produced some time ago. That might be useful either for testing or caching.

shinyFiles — by Thomas Lin Pedersen, 3 years ago

A Server-Side File System Viewer for Shiny

Provides functionality for client-side navigation of the server side file system in shiny apps. In case the app is running locally this gives the user direct access to the file system without the need to "download" files to a temporary location. Both file and folder selection as well as file saving is available.

HCUPtools — by Vikrant Dev Rathore, 7 days ago

Access and Work with HCUP Resources and Datasets

A comprehensive R package for accessing and working with publicly available and free resources from the Agency for Healthcare Research and Quality (AHRQ) Healthcare Cost and Utilization Project (HCUP). The package provides streamlined access to HCUP's Clinical Classifications Software Refined (CCSR) mapping files and Summary Trend Tables, enabling researchers and analysts to efficiently map ICD-10-CM diagnosis codes and ICD-10-PCS procedure codes to CCSR categories and access HCUP statistical reports. Key features include: direct download from HCUP website, multiple output formats (long/wide/default), cross-classification support, version management, citation generation, and intelligent caching. The package does not redistribute HCUP data files but facilitates direct download from the official HCUP website, ensuring users always have access to the latest versions and maintain compliance with HCUP data use policies. This package only accesses free public tools and reports; it does NOT access HCUP databases (NIS, KID, SID, NEDS, etc.) that require purchase. For more information, see < https://hcup-us.ahrq.gov/>.

png — by Simon Urbanek, 3 years ago

Read and write PNG images

This package provides an easy and simple way to read, write and display bitmap images stored in the PNG format. It can read and write both files and in-memory raw vectors.

cache — by Olivier Binette, 4 years ago

Cache and Retrieve Computation Results

Easily cache and retrieve computation results. The package works seamlessly across interactive R sessions, R scripts and Rmarkdown documents.

haven — by Hadley Wickham, 7 months ago

Import and Export 'SPSS', 'Stata' and 'SAS' Files

Import foreign statistical formats into R via the embedded 'ReadStat' C library, < https://github.com/WizardMac/ReadStat>.

sooty — by Michael D. Sumner, 7 months ago

Data Source Catalogues Online for Southern Ocean Ecosystem Research

Obtains lists of files of remote sensing collections for Southern Ocean surface properties. Commonly used data sources of sea surface temperature, sea ice concentration, and altimetry products such as sea surface height and sea surface currents are cached in object storage on the Pawsey Supercomputing Research Centre facility. Patterns of working to retrieve data from these object storage catalogues are described. The catalogues include complete collections of datasets Reynolds et al. (2008) "NOAA Optimum Interpolation Sea Surface Temperature (OISST) Analysis, Version 2.1" , Spreen et al. (2008) "Artist Advanced Microwave Scanning Radiometer for Earth Observing System (AMSR-E) sea ice concentration" . In future releases helpers will be added to identify particular data collections and target specific dates for earth observation data for reading, as well as helpers to retrieve data set citation and provenance details. This work was supported by resources provided by the Pawsey Supercomputing Research Centre with funding from the Australian Government and the Government of Western Australia. This software was developed by the Integrated Digital East Antarctica program of the Australian Antarctic Division.