A set of tools for R that enhance reproducibility for data analytics and forecasting. This package aims at making high-level, robust, machine and OS independent tools for making deeply reproducible and reusable content in R.
See updates from latest CRAN and development versions.
Note that versions 1.0.0 and later are not compatible with previous versions.
The current version can be much faster and creates smaller repository files (each with specific options set using Suggests
packages) and allows for different (e.g., RPostgres
backends for the database1 -- not the saved files, however; these are still saved locally).
A reproducible workflow is a series of code steps (e.g., in a script) that, when run, produce the same output from the same inputs every time. The big challenge with such a workflow is that many steps are so time consuming that a scientist tends to not re-run each step every time. After many months of work, it is often unclear if the code will actually function from the start. Is the original dataset still there? Have the packages that were used been updated? Are some of the steps missing because there was some "point and clicking"?
The best way to maintain reproducibility is to have all the code re-run all the time. That way, errors are detected early and can be fixed. The challenge is how to make all the steps fast enough that it becomes convenient to re-run everything from scratch each time.
Caching is the principle tool to achieve this reproducible work-flow.
There are many existing tools that support some notion of caching.
The main tool here, Cache
, can be nested hierarchically, becoming very powerful for the data science developer who is regularly working at many levels of an analysis.
rnorm(1) # give a random number
Cache(rnorm, 1) # generates a random number
Cache(rnorm, 1) # recovers the previous random number because call is identical
A common data problem is starting from a raw (spatial) dataset and getting it into shape for an analysis. Often, copies of a dataset are haphazardly placed in ad hoc local file systems. This makes it particularly difficult to share the workflow. The solution to this is use a canonical location (e.g., cloud storage, permalink to original data provider, etc.) and use tools that are smart enough to download only once.
Get a geospatial dataset. It will be checksummed (locally), meaning if the file is already in place locally, it will not download it again.
# Using dlFun -- a custom download function -- passed to preProcess
test1 <- prepInputs(targetFile = "GADM_2.8_LUX_adm0.rds", # must specify currently
dlFun = "raster::getData", name = "GADM", country = "LUX", level = 0,
path = dPath)
Putting these tools together allows for very rich data flows.
For example, with prepInputs
and using the fun
argument or passing a studyArea
, a raw dataset can be downloaded, loaded into R, and post processed -- all potentially very time consuming steps resulting in a clean, often much smaller dataset.
Wrapping all these with a Cache
can make it very quick.
test1 <- Cache(prepInputs, targetFile = "GADM_2.8_LUX_adm0.rds", # must specify currently
dlFun = "raster::getData", name = "GADM", country = "LUX", level = 0,
path = dPath)
See vignettes and help files for many more real-world examples.
Install from CRAN:
install.packages("reproducible")
Install from GitHub:
#install.packages("devtools")
library("devtools")
install_github("PredictiveEcology/reproducible", dependencies = TRUE)
Install from GitHub:
#install.packages("devtools")
library("devtools")
install_github("PredictiveEcology/reproducible", ref = "development", dependencies = TRUE)
Please see CONTRIBUTING.md
for information on how to contribute to this project.