bionic (1) datalad.1.gz

Provided by: datalad_0.9.3-1_all bug

SYNOPSIS

       datalad  [-h]  [-l  LEVEL]  [--pbs-runner {condor}] [-C PATH] [--version] [--dbg] [--idbg] [-c KEY=VALUE]
              [--output-format           {default,json,json_pp,tailored,'<template>']           [--report-status
              {success,failure,ok,notneeded,impossible,error}]   [--report-type   {dataset,file}]  [--on-failure
              {ignore,continue,stop}]  [--run-before  PLUGINSPEC  [PLUGINSPEC  ...]]   [--run-after   PLUGINSPEC
              [PLUGINSPEC ...]] [--cmd]  COMMAND  ...

DESCRIPTION

       DataLad provides a unified data distribution with the convenience of git-annex repositories as a backend.
       DataLad command line tools allow to manipulate (obtain, create, update, publish, etc.) datasets and their
       collections.

   Commands for dataset operations
       create Create a new dataset from scratch

       install
              Install a dataset from a (remote) source

       get    Get any dataset content (files/directories/subdatasets)

       add    Add files/directories to an existing dataset

       publish
              Publish a dataset to a known sibling

       uninstall
              Uninstall subdatasets

       drop   Drop file content from datasets

       remove Remove components from datasets

       update Update a dataset from a sibling

       create-sibling
              Create a dataset sibling on a UNIX-like SSH-accessible machine

       create-sibling-github
              Create dataset sibling on Github

       unlock Unlock file(s) of a dataset

       save   Save the current state of a dataset

       plugin Generic plugin interface

   Commands for meta data handling
       search Search within available in datasets' meta data

       metadata
              Metadata manipulation for files and whole datasets

       aggregate-metadata
              Aggregate meta data of a dataset for later query

   Miscellaneous commands
       test   Run internal DataLad (unit)tests

       crawl  Crawl online resource to create or update a dataset

       crawl-init
              Initialize crawling configuration

       ls     List summary information about URLs and dataset(s)

       clean  Clean up after DataLad (possible temporary files etc.)

       add-archive-content
              Add content of an archive under git annex control

       download-url
              Download content

       run    Run an arbitrary command and record its impact on a dataset

       rerun  Re-execute previous `datalad run` commands

   Plumbing commands
       annotate-paths
              Analyze and act upon input paths

       clone  Obtain a dataset copy from a URL or local source (path)

       create-test-dataset
              Create test (meta-)dataset

       diff   Report changes of dataset components

       siblings
              Manage sibling configuration

       sshrun Run command on remote machines via SSH

       subdatasets
              Report subdatasets and their properties

   General information
       Detailed  usage  information  for  individual  commands  is  available via command-specific --help, i.e.:
       datalad <command> --help

OPTIONS

       {create, install, get, add, publish, uninstall, drop,  remove,  update,  create-sibling,  create-sibling-
       github,  unlock,  save, plugin, search, metadata, aggregate-metadata, test, crawl, crawl-init, ls, clean,
       add-archive-content,  download-url,  run,  rerun,  annotate-paths,  clone,   create-test-dataset,   diff,
       siblings, sshrun, subdatasets}

       -h, --help, --help-np
              show  this  help message. --help-np forcefully disables the use of a pager for displaying the help
              message

       -l LEVEL, --log-level LEVEL
              set logging verbosity level. Choose among critical, error, warning,  info,  debug.  Also  you  can
              specify an integer <10 to provide even more debugging information

       --pbs-runner {condor}
              execute command by scheduling it via available PBS. For settings, config file will be consulted

       -C PATH
              run as if datalad was started in <path> instead of the current working directory. When multiple -C
              options are given, each subsequent non-absolute -C <path> is interpreted relative to the preceding
              -C  <path>.  This  option  affects  the  interpretations  of  the path names in that they are made
              relative to the working directory caused by the -C option

       --version
              show the program's version and license information

       --dbg  enter Python debugger when uncaught exception happens

       --idbg enter IPython debugger when uncaught exception happens

       -c KEY=VALUE
              configuration variable setting. Overrides any configuration read from a file, but  is  potentially
              overridden itself by configuration variables in the process environment.

       --output-format {default, json, json_pp, tailored,'<template>'
              select  format  for returned command results. 'default' give one line per result reporting action,
              status, path and an optional message; 'json' renders a JSON object with all  properties  for  each
              result  (one per line); 'json_pp' pretty-prints JSON spanning multiple lines; 'tailored' enables a
              command-specific rendering style that is typically tailored to human consumption (no result output
              otherwise),  '<template>' reports any value(s) of any result properties in any format indicated by
              the template (e.g. '{path}', compare with JSON output for all key-value choices).

       --report-status {success, failure, ok, notneeded, impossible, error}
              constrain command result report to records matching the given status. 'success' is a  synonym  for
              'ok' OR 'notneeded', 'failure' stands for 'impossible' OR 'error'.

       --report-type {dataset, file}
              constrain command result report to records matching the given type. Can be given more than once to
              match multiple types.

       --on-failure {ignore, continue, stop}
              when an operation fails: 'ignore' and continue with remaining operations, the error is logged  but
              does not lead to a non-zero exit code of the command; 'continue' works like 'ignore', but an error
              causes a non-zero exit code; 'stop' halts on first  failure  and  yields  non-zero  exit  code.  A
              failure is any result with status 'impossible' or 'error'.

       --run-before PLUGINSPEC [PLUGINSPEC ...]
              DataLad  plugin  to  run  after  the command. PLUGINSPEC is a list comprised of a plugin name plus
              optional `key=value` pairs with arguments for the plugin call (see `plugin` command  documentation
              for  details).  This  option  can  be given more than once to run multiple plugins in the order in
              which they were given. For running plugins that require a --dataset argument it  is  important  to
              provide  the respective dataset as the --dataset argument of the main command, if it is not in the
              list of plugin arguments.

       --run-after PLUGINSPEC [PLUGINSPEC ...]
              Like --run-before, but plugins are executed after the main command has finished.

       --cmd  syntactical helper that can be used to end the list of global  command  line  options  before  the
              subcommand  label.  Options  like  --run-before  can  take an arbitray number of arguments and may
              require to be followed by a single --cmd in order to enable identification of the subcommand.

              "Control Your Data"

AUTHORS

        datalad is developed by The DataLad Team and Contributors <team@datalad.org>.