For the most recent run of the pipeline with tar_make()
where a crew
controller was started, get summary-level information
of the workers.
Usage
tar_crew(store = targets::tar_config_get("store"))
Arguments
- store
Character of length 1, path to the
targets
data store. Defaults totar_config_get("store")
, which in turn defaults to_targets/
. When you set this argument, the value oftar_config_get("store")
is temporarily changed for the current function call. Seetar_config_get()
andtar_config_set()
for details about how to set the data store path persistently for a project.
Value
A data frame one row per crew
worker and the following columns:
controller
: name of thecrew
controller.launches
: number of times the worker was launched.seconds
: number of seconds the worker spent running tasks.targets
: number of targets the worker completed and delivered.
Storage access
Several functions like tar_make()
, tar_read()
, tar_load()
,
tar_meta()
, and tar_progress()
read or modify
the local data store of the pipeline.
The local data store is in flux while a pipeline is running,
and depending on how distributed computing or cloud computing is set up,
not all targets can even reach it. So please do not call these
functions from inside a target as part of a running
pipeline. The only exception is literate programming
target factories in the tarchetypes
package such as tar_render()
and tar_quarto()
.
See also
Other data:
tar_pid()
,
tar_process()
Examples
if (identical(Sys.getenv("TAR_EXAMPLES"), "true")) { # for CRAN
tar_dir({ # tar_dir() runs code from a temp dir for CRAN.
if (requireNamespace("crew", quietly = TRUE)) {
tar_script({
library(targets)
library(tarchetypes)
tar_option_set(controller = crew::crew_controller_local())
list(
tar_target(x, seq_len(2)),
tar_target(y, 2 * x, pattern = map(x))
)
}, ask = FALSE)
tar_make()
tar_process()
tar_process(pid)
}
})
}