Provided by: datalad_0.9.3-1_all
SYNOPSIS
datalad [-h] [-l LEVEL] [--pbs-runner {condor}] [-C PATH] [--version] [--dbg] [--idbg] [-c KEY=VALUE] [--output-format {default,json,json_pp,tailored,'<template>'] [--report- status {success,failure,ok,notneeded,impossible,error}] [--report-type {dataset,file}] [--on-failure {ignore,continue,stop}] [--run-before PLUGINSPEC [PLUGINSPEC ...]] [--run-after PLUGINSPEC [PLUGINSPEC ...]] [--cmd] COMMAND ...
DESCRIPTION
DataLad provides a unified data distribution with the convenience of git-annex repositories as a backend. DataLad command line tools allow to manipulate (obtain, create, update, publish, etc.) datasets and their collections. Commands for dataset operations create Create a new dataset from scratch install Install a dataset from a (remote) source get Get any dataset content (files/directories/subdatasets) add Add files/directories to an existing dataset publish Publish a dataset to a known sibling uninstall Uninstall subdatasets drop Drop file content from datasets remove Remove components from datasets update Update a dataset from a sibling create-sibling Create a dataset sibling on a UNIX-like SSH-accessible machine create-sibling-github Create dataset sibling on Github unlock Unlock file(s) of a dataset save Save the current state of a dataset plugin Generic plugin interface Commands for meta data handling search Search within available in datasets' meta data metadata Metadata manipulation for files and whole datasets aggregate-metadata Aggregate meta data of a dataset for later query Miscellaneous commands test Run internal DataLad (unit)tests crawl Crawl online resource to create or update a dataset crawl-init Initialize crawling configuration ls List summary information about URLs and dataset(s) clean Clean up after DataLad (possible temporary files etc.) add-archive-content Add content of an archive under git annex control download-url Download content run Run an arbitrary command and record its impact on a dataset rerun Re-execute previous `datalad run` commands Plumbing commands annotate-paths Analyze and act upon input paths clone Obtain a dataset copy from a URL or local source (path) create-test-dataset Create test (meta-)dataset diff Report changes of dataset components siblings Manage sibling configuration sshrun Run command on remote machines via SSH subdatasets Report subdatasets and their properties General information Detailed usage information for individual commands is available via command-specific --help, i.e.: datalad <command> --help
OPTIONS
{create, install, get, add, publish, uninstall, drop, remove, update, create-sibling, create-sibling-github, unlock, save, plugin, search, metadata, aggregate-metadata, test, crawl, crawl-init, ls, clean, add-archive-content, download-url, run, rerun, annotate- paths, clone, create-test-dataset, diff, siblings, sshrun, subdatasets} -h, --help, --help-np show this help message. --help-np forcefully disables the use of a pager for displaying the help message -l LEVEL, --log-level LEVEL set logging verbosity level. Choose among critical, error, warning, info, debug. Also you can specify an integer <10 to provide even more debugging information --pbs-runner {condor} execute command by scheduling it via available PBS. For settings, config file will be consulted -C PATH run as if datalad was started in <path> instead of the current working directory. When multiple -C options are given, each subsequent non-absolute -C <path> is interpreted relative to the preceding -C <path>. This option affects the interpretations of the path names in that they are made relative to the working directory caused by the -C option --version show the program's version and license information --dbg enter Python debugger when uncaught exception happens --idbg enter IPython debugger when uncaught exception happens -c KEY=VALUE configuration variable setting. Overrides any configuration read from a file, but is potentially overridden itself by configuration variables in the process environment. --output-format {default, json, json_pp, tailored,'<template>' select format for returned command results. 'default' give one line per result reporting action, status, path and an optional message; 'json' renders a JSON object with all properties for each result (one per line); 'json_pp' pretty-prints JSON spanning multiple lines; 'tailored' enables a command-specific rendering style that is typically tailored to human consumption (no result output otherwise), '<template>' reports any value(s) of any result properties in any format indicated by the template (e.g. '{path}', compare with JSON output for all key-value choices). --report-status {success, failure, ok, notneeded, impossible, error} constrain command result report to records matching the given status. 'success' is a synonym for 'ok' OR 'notneeded', 'failure' stands for 'impossible' OR 'error'. --report-type {dataset, file} constrain command result report to records matching the given type. Can be given more than once to match multiple types. --on-failure {ignore, continue, stop} when an operation fails: 'ignore' and continue with remaining operations, the error is logged but does not lead to a non-zero exit code of the command; 'continue' works like 'ignore', but an error causes a non-zero exit code; 'stop' halts on first failure and yields non-zero exit code. A failure is any result with status 'impossible' or 'error'. --run-before PLUGINSPEC [PLUGINSPEC ...] DataLad plugin to run after the command. PLUGINSPEC is a list comprised of a plugin name plus optional `key=value` pairs with arguments for the plugin call (see `plugin` command documentation for details). This option can be given more than once to run multiple plugins in the order in which they were given. For running plugins that require a --dataset argument it is important to provide the respective dataset as the --dataset argument of the main command, if it is not in the list of plugin arguments. --run-after PLUGINSPEC [PLUGINSPEC ...] Like --run-before, but plugins are executed after the main command has finished. --cmd syntactical helper that can be used to end the list of global command line options before the subcommand label. Options like --run-before can take an arbitray number of arguments and may require to be followed by a single --cmd in order to enable identification of the subcommand. "Control Your Data"
AUTHORS
datalad is developed by The DataLad Team and Contributors <team@datalad.org>.