Provided by: pcs_0.9.149-1_amd64 bug

NAME

       pcs - pacemaker/corosync configuration system

SYNOPSIS

       pcs [-f file] [-h] [commands]...

DESCRIPTION

       Control and configure pacemaker and corosync.

OPTIONS

       -h, --help
              Display usage and exit

       -f file
              Perform actions on file instead of active CIB

       --debug
              Print all network traffic and external commands run

       --version
              Print pcs version information

   Commands:
       cluster
              Configure cluster options and nodes

       resource
              Manage cluster resources

       stonith
              Configure fence devices

       constraint
              Set resource constraints

       property
              Set pacemaker properties

       acl    Set pacemaker access control lists

       status View cluster status

       config View and manage cluster configuration

       pcsd   Manage pcs daemon

       node   Manage cluster nodes

   resource
       show [resource id] [--full] [--groups]
              Show  all  currently  configured  resources  or if a resource is specified show the
              options for the  configured  resource.   If  --full  is  specified  all  configured
              resource  options  will  be  displayed.  If --groups is specified, only show groups
              (and their resources).

       list [<standard|provider|type>] [--nodesc]
              Show list of all  available  resources,  optionally  filtered  by  specified  type,
              standard  or  provider.  If --nodesc is used then descriptions of resources are not
              printed.

       describe <standard:provider:type|type>
              Show options for the specified resource

       create <resource  id>  <standard:provider:type|type>  [resource  options]  [op  <operation
       action>  <operation  options>  [<operation  action>  <operation  options>]...] [meta <meta
       options>...] [--clone <clone options> | --master <master options>  |  --group  <group  id>
       [--before <resource id> | --after <resource id>]] [--disabled] [--wait[=n]]
              Create  specified  resource.   If  --clone is used a clone resource is created.  If
              --master is specified a master/slave resource is created.  If --group is  specified
              the  resource  is  added  to  the  group named.  You can use --before or --after to
              specify the position of the added resource  relatively  to  some  resource  already
              existing  in  the  group.   If  --disabled is specified the resource is not started
              automatically.  If --wait is specified, pcs will wait up to  'n'  seconds  for  the
              resource  to  start  and  then  return  0  if  the resource is started, or 1 if the
              resource has not yet started.  If 'n' is not specified it defaults to 60 minutes.

              Example: Create a new resource called 'VirtualIP'  with  IP  address  192.168.0.99,
              netmask  of  32,  monitored  everything  30  seconds,  on eth2: pcs resource create
              VirtualIP ocf:heartbeat:IPaddr2 ip=192.168.0.99 cidr_netmask=32 nic=eth2 op monitor
              interval=30s

       delete <resource id|group id|master id|clone id>
              Deletes  the  resource,  group,  master  or  clone  (and  all  resources within the
              group/master/clone).

       enable <resource id> [--wait[=n]]
              Allow the cluster to start the resource. Depending on the rest of the configuration
              (constraints,  options, failures, etc), the resource may remain stopped.  If --wait
              is specified, pcs will wait up to 'n' seconds for the resource to  start  and  then
              return  0 if the resource is started, or 1 if the resource has not yet started.  If
              'n' is not specified it defaults to 60 minutes.

       disable <resource id> [--wait[=n]]
              Attempt to stop the resource if it is running and forbid the cluster from  starting
              it  again.   Depending  on  the  rest  of  the configuration (constraints, options,
              failures, etc), the resource may remain started.  If --wait is specified, pcs  will
              wait  up  to 'n' seconds for the resource to stop and then return 0 if the resource
              is stopped or 1 if the resource has not  stopped.   If  'n'  is  not  specified  it
              defaults to 60 minutes.

       restart <resource id> [node] [--wait=n]
              Restart  the  resource  specified.  If a node is specified and if the resource is a
              clone or master/slave it will be restarted only on the node specified.   If  --wait
              is  specified, then we will wait up to 'n' seconds for the resource to be restarted
              and return 0 if the restart was successful or 1 if it was not.

       debug-start <resource id> [--full]
              This command will force the specified resource to start on this node  ignoring  the
              cluster  recommendations  and  print  the output from starting the resource.  Using
              --full will give more detailed output.  This is mainly used for debugging resources
              that fail to start.

       debug-stop <resource id> [--full]
              This  command  will  force the specified resource to stop on this node ignoring the
              cluster recommendations and print the output from  stopping  the  resource.   Using
              --full will give more detailed output.  This is mainly used for debugging resources
              that fail to stop.

       debug-promote <resource id> [--full]
              This command will force the specified resource to be promoted on this node ignoring
              the  cluster  recommendations  and  print  the  output from promoting the resource.
              Using --full will give more detailed output.  This is  mainly  used  for  debugging
              resources that fail to promote.

       debug-demote <resource id> [--full]
              This  command will force the specified resource to be demoted on this node ignoring
              the cluster recommendations and print the output from demoting the resource.  Using
              --full will give more detailed output.  This is mainly used for debugging resources
              that fail to demote.

       debug-monitor <resource id> [--full]
              This command will force the specified resource to be moniored on this node ignoring
              the  cluster  recommendations  and  print  the output from monitoring the resource.
              Using --full will give more detailed output.  This is  mainly  used  for  debugging
              resources that fail to be monitored.

       move <resource id> [destination node] [--master] [lifetime=<lifetime>] [--wait[=n]]
              Move  the  resource off the node it is currently running on by creating a -INFINITY
              location constraint to ban the node.  If destination node is specified the resource
              will  be  moved  to that node by creating an INFINITY location constraint to prefer
              the destination node.  If --master is used the scope of the command is  limited  to
              the  master  role  and you must use the master id (instead of the resource id).  If
              lifetime is specified then the constraint will expire after that time, otherwise it
              defaults  to infinity and the constraint can be cleared manually with 'pcs resource
              clear' or 'pcs constraint delete'.  If --wait is specified, pcs will wait up to 'n'
              seconds  for  the  resource to move and then return 0 on success or 1 on error.  If
              'n' is not specified it defaults to 60  minutes.   If  you  want  the  resource  to
              preferably  avoid  running  on  some nodes but be able to failover to them use 'pcs
              location avoids'.

       ban <resource id> [node] [--master] [lifetime=<lifetime>] [--wait[=n]]
              Prevent the resource id specified from running on the node (or on the current  node
              it  is  running  on  if  no  node  is  specified)  by creating a -INFINITY location
              constraint.  If --master is used the scope of the command is limited to the  master
              role  and  you must use the master id (instead of the resource id).  If lifetime is
              specified then the constraint will expire after that time, otherwise it defaults to
              infinity  and  the  constraint can be cleared manually with 'pcs resource clear' or
              'pcs constraint delete'.  If --wait is specified, pcs will wait up to  'n'  seconds
              for the resource to move and then return 0 on success or 1 on error.  If 'n' is not
              specified it defaults to 60 minutes.  If you want the resource to preferably  avoid
              running on some nodes but be able to failover to them use 'pcs location avoids'.

       clear <resource id> [node] [--master] [--wait[=n]]
              Remove  constraints  created by move and/or ban on the specified resource (and node
              if specified). If --master is used the scope of  the  command  is  limited  to  the
              master role and you must use the master id (instead of the resource id).  If --wait
              is specified, pcs will  wait  up  to  'n'  seconds  for  the  operation  to  finish
              (including  starting  and/or  moving resources if appropriate) and then return 0 on
              success or 1 on error.  If 'n' is not specified it defaults to 60 minutes.

       standards
              List available resource agent standards supported by this installation. (OCF,  LSB,
              etc.)

       providers
              List available OCF resource agent providers

       agents [standard[:provider]]
              List available agents optionally filtered by standard and provider

       update  <resource  id> [resource options] [op [<operation action> <operation options>]...]
       [meta <meta operations>...] [--wait[=n]]
              Add/Change options to specified resource, clone or  multi-state  resource.   If  an
              operation  (op) is specified it will update the first found operation with the same
              action on the specified resource, if no operation with that action  exists  then  a
              new  operation  will  be  created.   (WARNING:  all existing options on the updated
              operation will be reset if not specified.)  If you want to create multiple  monitor
              operations  you  should  use  the  'op  add'  & 'op remove' commands.  If --wait is
              specified, pcs will wait up to 'n' seconds for the changes to take effect and  then
              return  0  if  the  changes  have  been  processed  or  1 otherwise.  If 'n' is not
              specified it defaults to 60 minutes.

       op add <resource id> <operation action> [operation properties]
              Add operation for specified resource

       op remove <resource id> <operation action> [<operation properties>...]
              Remove specified operation (note: you must specify the exact  operation  properties
              to properly remove an existing operation).

       op remove <operation id>
              Remove the specified operation id

       op defaults [options]
              Set  default  values  for  operations,  if  no  options are passed, lists currently
              configured defaults

       meta <resource id | group id | master id | clone id> <meta options> [--wait[=n]]
              Add specified options to the specified  resource,  group,  master/slave  or  clone.
              Meta  options  should  be  in  the  format of name=value, options may be removed by
              setting an option without a value.  If --wait is specified, pcs will wait up to 'n'
              seconds  for  the changes to take effect and then return 0 if the changes have been
              processed or 1 otherwise.  If 'n' is not  specified  it  defaults  to  60  minutes.
              Example: pcs resource meta TestResource failure-timeout=50 stickiness=

       group add <group id> <resource id> [resource id] ... [resource id] [--before <resource id>
       | --after <resource id>] [--wait[=n]]
              Add the specified resource to the group, creating the group if it does  not  exist.
              If  the resource is present in another group it is moved to the new group.  You can
              use --before or --after to specify the position of the added  resources  relatively
              to  some  resource already existing in the group.  If --wait is specified, pcs will
              wait up to 'n' seconds for the operation to finish (including moving  resources  if
              appropriate)  and  then return 0 on success or 1 on error.  If 'n' is not specified
              it defaults to 60 minutes.

       group remove <group id> <resource id> [resource id] ... [resource id] [--wait[=n]]
              Remove the specified resource(s) from the  group,  removing  the  group  if  it  no
              resources  remain.  If --wait is specified, pcs will wait up to 'n' seconds for the
              operation to finish (including moving resources if appropriate) and then  return  0
              on success or 1 on error.  If 'n' is not specified it defaults to 60 minutes.

       ungroup <group id> [resource id] ... [resource id] [--wait[=n]]
              Remove  the group (Note: this does not remove any resources from the cluster) or if
              resources are specified, remove the specified resources from the group.  If  --wait
              is  specified,  pcs  will  wait  up  to  'n'  seconds  for  the operation to finish
              (including moving resources if appropriate) and the return 0 on  success  or  1  on
              error.  If 'n' is not specified it defaults to 60 minutes.

       clone <resource id | group id> [clone options]... [--wait[=n]]
              Setup  up  the specified resource or group as a clone.  If --wait is specified, pcs
              will wait up to 'n' seconds for the operation to finish (including  starting  clone
              instances  if  appropriate)  and then return 0 on success or 1 on error.  If 'n' is
              not specified it defaults to 60 minutes.

       unclone <resource id | group id> [--wait[=n]]
              Remove the clone which contains the specified group or resource  (the  resource  or
              group  will  not  be  removed).   If  --wait  is specified, pcs will wait up to 'n'
              seconds for  the  operation  to  finish  (including  stopping  clone  instances  if
              appropriate)  and  then return 0 on success or 1 on error.  If 'n' is not specified
              it defaults to 60 minutes.

       master [<master/slave id>] <resource id | group id> [options] [--wait[=n]]
              Configure a resource or group as a multi-state (master/slave) resource.  If  --wait
              is  specified,  pcs  will  wait  up  to  'n'  seconds  for  the operation to finish
              (including starting and promoting  resource  instances  if  appropriate)  and  then
              return  0  on  success  or  1  on error.  If 'n' is not specified it defaults to 60
              minutes.  Note: to remove a master you must remove the resource/group it contains.

       manage <resource id> ... [resource n]
              Set resources listed to managed mode (default)

       unmanage <resource id> ... [resource n]
              Set resources listed to unmanaged mode

       defaults [options]
              Set default values for  resources,  if  no  options  are  passed,  lists  currently
              configured defaults

       cleanup [<resource id>]
              Cleans  up  the  resource  in  the  lrmd  (useful  to reset the resource status and
              failcount). This tells the cluster to forget the operation history  of  a  resource
              and  re-detect  its  current  state.  This can be useful to purge knowledge of past
              failures that have since been resolved. If a resource id is not specified then  all
              resources/stonith devices will be cleaned up.

       failcount show <resource id> [node]
              Show  current  failcount for specified resource from all nodes or only on specified
              node

       failcount reset <resource id> [node]
              Reset failcount for specified resource on all nodes or only on specified node. This
              tells the cluster to forget how many times a resource has failed in the past.  This
              may allow the resource to be started or moved to a more preferred location.

       relocate dry-run [resource1] [resource2] ...
              The same as 'relocate run' but has no effect on the cluster.

       relocate run [resource1] [resource2] ...
              Relocate specified resources  to  their  preferred  nodes.   If  no  resources  are
              specified,  relocate all resources.  This command calculates the preferred node for
              each resource  while  ignoring  resource  stickiness.   Then  it  creates  location
              constraints  which will cause the resources to move to their preferred nodes.  Once
              the resources have been moved the constraints are deleted automatically.  Note that
              the  preferred  node  is  calculated  based on current cluster status, constraints,
              location of resources and other settings and thus it might change over time.

       relocate show
              Display current status of  resources  and  their  optimal  node  ignoring  resource
              stickiness.

       relocate clear
              Remove all constraints created by the 'relocate run' command.

       utilization [<resource id> [<name>=<value> ...]]
              Add  specified  utilization  options  to  specified  resource.  If  resource is not
              specified, shows utilization of all  resources.  If  utilization  options  are  not
              specified, shows utilization of specified resource. Utilization option should be in
              format name=value, value has to be integer. Options may be removed  by  setting  an
              option without a value. Example: pcs resource utilization TestResource cpu= ram=20

   cluster
       auth [node] [...] [-u username] [-p password] [--force] [--local]
              Authenticate  pcs  to  pcsd  on  nodes  specified,  or  on  all nodes configured in
              corosync.conf if no  nodes  are  specified  (authorization  tokens  are  stored  in
              ~/.pcs/tokens  or  /var/lib/pcsd/tokens  for  root).  By default all nodes are also
              authenticated to each other, using --local only authenticates the local  node  (and
              does  not authenticate the remote nodes with each other).  Using --force forces re-
              authentication to occur.

       setup  [--start]  [--local]  [--enable]  --name  <cluster  name>   <node1[,node1-altaddr]>
       [node2[,node2-altaddr]]  [..] [--transport <udpu|udp>] [--rrpmode active|passive] [--addr0
       <addr/net> [[[--mcast0 <address>] [--mcastport0 <port>] [--ttl0 <ttl>]] |  [--broadcast0]]
       [--addr1  <addr/net>  [[[--mcast1  <address>]  [--mcastport1  <port>]  [--ttl1  <ttl>]]  |
       [--broadcast1]]]]            [--wait_for_all=<0|1>]             [--auto_tie_breaker=<0|1>]
       [--last_man_standing=<0|1>  [--last_man_standing_window=<time  in  ms>]] [--ipv6] [--token
       <timeout>] [--token_coefficient  <timeout>]  [--join  <timeout>]  [--consensus  <timeout>]
       [--miss_count_const <count>] [--fail_recv_const <failures>]
              Configure  corosync  and  sync configuration out to listed nodes. --local will only
              perform changes on the local node, --start will  also  start  the  cluster  on  the
              specified  nodes,  --enable  will  enable  corosync  and pacemaker on node startup,
              --transport allows specification of corosync transport (default: udpu; udp for CMAN
              clusters),  --rrpmode  allows you to set the RRP mode of the system. Currently only
              'passive'  is  supported  or  tested  (using  'active'  is  not  recommended).  The
              --wait_for_all, --auto_tie_breaker, --last_man_standing, --last_man_standing_window
              options are all documented in corosync's votequorum(5) man page.

              --ipv6 will configure corosync to use ipv6 (instead of ipv4)

              --token <timeout> sets time in milliseconds until a token loss  is  declared  after
              not receiving a token (default 1000 ms)

              --token_coefficient  <timeout>  sets time in milliseconds used for clusters with at
              least 3 nodes as  a  coefficient  for  real  token  timeout  calculation  (token  +
              (number_of_nodes - 2) * token_coefficient) (default 650 ms)

              --join  <timeout>  sets  time in milliseconds to wait for join messages (default 50
              ms)

              --consensus <timeout> sets time  in  milliseconds  to  wait  for  consensus  to  be
              achieved before starting a new round of membership configuration (default 1200 ms)

              --miss_count_const <count> sets the maximum number of times on receipt of a token a
              message is checked for retransmission before a  retransmission  occurs  (default  5
              messages)

              --fail_recv_const  <failures>  specifies  how  many  rotations of the token without
              receiving any messages when messages should be received  may  occur  before  a  new
              configuration is formed (default 2500 failures)

              Configuring Redundant Ring Protocol (RRP)

              When  using  udpu  specifying nodes, specify the ring 0 address first followed by a
              ',' and then the ring 1 address.

              Example: pcs cluster setup --name cname nodeA-0,nodeA-1 nodeB-0,nodeB-1

              When using udp, using --addr0 and --addr1 will allow you to configure rrp mode  for
              corosync.   It's  recommended  to use a network (instead of IP address) for --addr0
              and --addr1 so the  same  corosync.conf  file  can  be  used  around  the  cluster.
              --mcast0   defaults   to   239.255.1.1   and   --mcast1  defaults  to  239.255.2.1,
              --mcastport0/1 default to 5405 and ttl defaults to 1. If --broadcast is  specified,
              --mcast0/1, --mcastport0/1 & --ttl0/1 are ignored.

       start [--all] [node] [...]
              Start  corosync  &  pacemaker on specified node(s), if a node is not specified then
              corosync & pacemaker are started on the local node.  If  --all  is  specified  then
              corosync & pacemaker are started on all nodes.

       stop [--all] [node] [...]
              Stop  corosync  &  pacemaker  on specified node(s), if a node is not specified then
              corosync & pacemaker are stopped on the local node.  If  --all  is  specified  then
              corosync & pacemaker are stopped on all nodes.

       kill   Force corosync and pacemaker daemons to stop on the local node (performs kill -9).

       enable [--all] [node] [...]
              Configure corosync & pacemaker to run on node boot on specified node(s), if node is
              not specified then corosync & pacemaker are enabled on the local node. If --all  is
              specified then corosync & pacemaker are enabled on all nodes.

       disable [--all] [node] [...]
              Configure  corosync  &  pacemaker  to not run on node boot on specified node(s), if
              node is not specified then corosync & pacemaker are disabled on the local node.  If
              --all is specified then corosync & pacemaker are disabled on all nodes. (Note: this
              is the default after installation)

       standby [<node>] | --all
              Put specified node into standby mode (the node specified will no longer be able  to
              host  resources),  if no node or options are specified the current node will be put
              into standby mode, if --all is specified all nodes will be put into standby mode.

       unstandby [<node>] | --all
              Remove node from standby mode  (the  node  specified  will  now  be  able  to  host
              resources),  if  no  node or options are specified the current node will be removed
              from standby mode, if --all is specified all nodes will  be  removed  from  standby
              mode.

       remote-node add <hostname> <resource id> [options]
              Enables  the specified resource as a remote-node resource on the specified hostname
              (hostname should be the same as 'uname -n')

       remote-node remove <hostname>
              Disables any resources configured to  be  remote-node  resource  on  the  specified
              hostname (hostname should be the same as 'uname -n')

       status View current cluster status (an alias of 'pcs status cluster')

       pcsd-status [node] [...]
              Get  current  status  of  pcsd  on  nodes  specified, or on all nodes configured in
              corosync.conf if no nodes are specified

       sync   Sync corosync configuration to all nodes  found  from  current  corosync.conf  file
              (cluster.conf on systems running Corosync 1.x)

       quorum unblock
              Cancel  waiting for all nodes when establishing quorum.  Useful in situations where
              you know the cluster is inquorate, but you are confident that  the  cluster  should
              proceed with resource management regardless.

       cib [filename] [scope=<scope> | --config]
              Get  the  raw  xml  from  the  CIB  (Cluster  Information  Base).  If a filename is
              provided, we save the cib to that file, otherwise  the  cib  is  printed.   Specify
              scope  to  get  a  specific  section  of  the  CIB.  Valid values of the scope are:
              configuration,   nodes,   resources,   constraints,    crm_config,    rsc_defaults,
              op_defaults, status.  --config is the same as scope=configuration.  Use of --config
              is recommended.  Do not specify a scope if you need to get  the  whole  CIB  or  be
              warned in the case of outdated CIB on cib-push.

       cib-push <filename> [scope=<scope> | --config]
              Push  the  raw  xml from <filename> to the CIB (Cluster Information Base).  You can
              obtain the CIB by running the 'pcs cluster cib' command, which is recommended first
              step when you want to perform desired modifications (pcs -f <command>) for the one-
              off push.  Specify scope to push a specific section of the CIB.   Valid  values  of
              the   scope   are:   configuration,   nodes,  resources,  constraints,  crm_config,
              rsc_defaults, op_defaults.  --config is the same as  scope=configuration.   Use  of
              --config  is recommended.  Do not specify a scope if you need to push the whole CIB
              or be warned in the case of outdated CIB.  WARNING: the selected scope of  the  CIB
              will be overwritten by the current content of the specified file.

       cib-upgrade
              Upgrade the CIB to conform to the latest version of the document schema

       edit [scope=<scope> | --config]
              Edit  the  cib in the editor specified by the $EDITOR environment variable and push
              out any changes upon saving.  Specify scope to edit a specific section of the  CIB.
              Valid  values  of  the  scope  are:  configuration,  nodes, resources, constraints,
              crm_config,   rsc_defaults,    op_defaults.     --config    is    the    same    as
              scope=configuration.   Use  of  --config is recommended.  Do not specify a scope if
              you need to edit the whole CIB or be warned in the case of outdated CIB.

       node add <node[,node-altaddr]> [--start] [--enable]
              Add the node to corosync.conf and corosync on all nodes in the cluster and sync the
              new   corosync.conf   to  the  new  node.   If  --start  is  specified  also  start
              corosync/pacemaker  on  the   new   node,   if   --enable   is   specified   enable
              corosync/pacemaker on new node.  When using Redundant Ring Protocol (RRP) with udpu
              transport, specify the ring 0 address first followed by a ',' and then the  ring  1
              address.

       node remove <node>
              Shutdown  specified  node  and  remove  it from pacemaker and corosync on all other
              nodes in the cluster

       uidgid List the current configured uids and gids of users allowed to connect to corosync

       uidgid add [uid=<uid>] [gid=<gid>]
              Add the specified uid and/or gid to the list of users/groups allowed to connect  to
              corosync

       uidgid rm [uid=<uid>] [gid=<gid>]
              Remove  the  specified  uid  and/or  gid  from  the list of users/groups allowed to
              connect to corosync

       corosync [node]
              Get the corosync.conf from the specified node or from the current node if node  not
              specified

       reload corosync
              Reload the corosync configuration on the current node

       destroy [--all]
              Permanently destroy the cluster on the current node, killing all corosync/pacemaker
              processes removing all cib files and the  corosync.conf  file.   Using  --all  will
              attempt  to  destroy  the cluster on all nodes configure in the corosync.conf file.
              WARNING: This command permantly removes any cluster  configuration  that  has  been
              created. It is recommended to run 'pcs cluster stop' before destroying the cluster.

       verify [-V] [filename]
              Checks  the  pacemaker configuration (cib) for syntax and common conceptual errors.
              If no filename is specified  the  check  is  performed  on  the  currently  running
              cluster.  If -V is used more verbose output will be printed

       report [--from "YYYY-M-D H:M:S" [--to "YYYY-M-D" H:M:S"]] dest
              Create  a tarball containing everything needed when reporting cluster problems.  If
              --from and --to are not used, the report will include the past 24 hours.

   stonith
       show [stonith id] [--full]
              Show all currently configured stonith devices or if a stonith id is specified  show
              the  options  for  the  configured  stonith  device.   If  --full  is specified all
              configured stonith options will be displayed

       list [filter] [--nodesc]
              Show list of all available stonith agents (if filter is provided then only  stonith
              agents matching the filter will be shown). If --nodesc is used then descriptions of
              stonith agents are not printed.

       describe <stonith agent>
              Show options for specified stonith agent

       create <stonith id> <stonith device type> [stonith device options] [op <operation  action>
       <operation options> [<operation action> <operation options>]...] [meta <meta options>...]
              Create stonith device with specified type and options

       update <stonith id> [stonith device options]
              Add/Change options to specified stonith id

       delete <stonith id>
              Remove stonith id from configuration

       cleanup [<stonith id>]
              Cleans  up  the  stonith  device  in  the  lrmd  (useful  to  reset  the status and
              failcount).  This tells the cluster to forget the operation history  of  a  stonith
              device  and  re-detect its current state.  This can be useful to purge knowledge of
              past failures that have since been resolved. If a stonith id is not specified  then
              all resources/stonith devices will be cleaned up.

       level  Lists all of the fencing levels currently configured

       level add <level> <node> <devices>
              Add the fencing level for the specified node with a comma separated list of devices
              (stonith ids) to attempt for that node at that level. Fence levels are attempted in
              numerical  order  (starting  with  1)  if a level succeeds (meaning all devices are
              successfully fenced in that level) then no other levels are tried, and the node  is
              considered fenced.

       level remove <level> [node id] [stonith id] ... [stonith id]
              Removes the fence level for the level, node and/or devices specified If no nodes or
              devices are specified then the fence level is removed

       level clear [node|stonith id(s)]
              Clears the fence levels on the node (or stonith id) specified or clears  all  fence
              levels  if  a  node/stonith  id  is  not specified.  If more than one stonith id is
              specified they must be separated by a comma and no spaces.   Example:  pcs  stonith
              level clear dev_a,dev_b

       level verify
              Verifies all fence devices and nodes specified in fence levels exist

       fence <node> [--off]
              Fence  the node specified (if --off is specified, use the 'off' API call to stonith
              which will turn the node off instead of rebooting it)

       confirm <node>
              Confirm that the host specified is currently down.  This  command  should  ONLY  be
              used when the node specified has already been confirmed to be down.

              WARNING:  if  this  node  is  not actually down data corruption/cluster failure can
              occur.

   acl
       [show] List all current access control lists

       enable Enable access control lists

       disable
              Disable access control lists

       role create <role id> [description=<description>] [((read | write | deny) (xpath <query> |
       id <id>))...]
              Create a role with the id and (optional) description specified.  Each role can also
              have an unlimited number of permissions  (read/write/deny)  applied  to  either  an
              xpath query or the id of a specific element in the cib

       role delete <role id>
              Delete the role specified and remove it from any users/groups it was assigned to

       role assign <role id> [to] <username/group>
              Assign a role to a user or group already created with 'pcs acl user/group create'

       role unassign <role id> [from] <username/group>
              Remove a role from the specified user

       user create <username> <role id> [<role id>]...
              Create an ACL for the user specified and assign roles to the user

       user delete <username>
              Remove  the user specified (and roles assigned will be unassigned for the specified
              user)

       group create <group> <role id> [<role id>]...
              Create an ACL for the group specified and assign roles to the group

       group delete <group>
              Remove the group specified (and roles assigned will be unassigned for the specified
              group)

       permission add <role id> ((read | write | deny) (xpath <query> | id <id>))...
              Add the listed permissions to the role specified

       permission delete <permission id>
              Remove the permission id specified (permission id's are listed in parenthesis after
              permissions in 'pcs acl' output)

   property
       list|show [<property> | --all | --defaults]
              List property settings (default: lists configured properties).   If  --defaults  is
              specified  will  show  all  property  defaults,  if  --all  is  specified,  current
              configured properties will be shown with unset properties and their defaults.   Run
              'man pengine' and 'man crmd' to get a description of the properties.

       set [--force | --node <nodename>] <property>=[<value>] [<property>=[<value>] ...]
              Set  specific  pacemaker  properties  (if  the  value is blank then the property is
              removed from the configuration).  If a  property  is  not  recognized  by  pcs  the
              property  will  not be created unless the --force is used. If --node is used a node
              attribute is set on the specified node.  Run 'man pengine' and 'man crmd' to get  a
              description of the properties.

       unset [--node <nodename>] <property>
              Remove  property  from  configuration  (or  remove attribute from specified node if
              --node is used).  Run 'man pengine' and 'man crmd' to  get  a  description  of  the
              properties.

   constraint
       [list|show] --full
              List all current location, order and colocation constraints, if --full is specified
              also list the constraint ids.

       location <resource id> prefers <node[=score]>...
              Create a location constraint on a resource to prefer the specified node  and  score
              (default score: INFINITY)

       location <resource id> avoids <node[=score]>...
              Create  a  location  constraint on a resource to avoid the specified node and score
              (default score: INFINITY)

       location    <resource    id>    rule    [id=<rule    id>]    [resource-discovery=<option>]
       [role=master|slave]    [constraint-id=<id>]    [score=<score>|score-attribute=<attribute>]
       <expression>
              Creates a location rule on the specified resource where the expression  looks  like
              one of the following:
                defined|not_defined <attribute>
                <attribute> lt|gt|lte|gte|eq|ne [string|integer|version] <value>
                date gt|lt <date>
                date in_range <date> to <date>
                date in_range <date> to duration <duration options>...
                date-spec <date spec options>...
                <expression> and|or <expression>
                ( <expression> )
              where  duration  options  and  date  spec  options are: hours, monthdays, weekdays,
              yeardays, months, weeks, years, weekyears, moon. If score is omitted it defaults to
              INFINITY.  If  id  is  omitted  one is generated from the resource id. If resource-
              discovery is omitted it defaults to 'always'.

       location show [resources|nodes [node id|resource id]...] [--full]
              List all the current location constraints, if  'resources'  is  specified  location
              constraints  are displayed per resource (default), if 'nodes' is specified location
              constraints are displayed per node.  If specific nodes or resources  are  specified
              then we only show information about them.  If --full is specified show the internal
              constraint id's as well.

       location add <id> <resource id> <node> <score> [resource-discovery=<option>]
              Add a location constraint with the appropriate  id,  resource  id,  node  name  and
              score. (For more advanced pacemaker usage)

       location remove <id> [<resource id> <node> <score>]
              Remove  a  location  constraint with the appropriate id, resource id, node name and
              score. (For more advanced pacemaker usage)

       order show [--full]
              List all current ordering constraints (if --full is  specified  show  the  internal
              constraint id's as well).

       order [action] <resource id> then [action] <resource id> [options]
              Add an ordering constraint specifying actions (start, stop, promote, demote) and if
              no action is specified the default action will be  start.   Available  options  are
              kind=Optional/Mandatory/Serialize,  symmetrical=true/false,  require-all=true/false
              and id=<constraint-id>.

       order set <resource1> <resource2> [resourceN]... [options]  [set  <resourceX>  <resourceY>
       ... [options]] [setoptions [constraint_options]]
              Create  an  ordered  set of resources. Available options are sequential=true/false,
              require-all=true/false,            action=start/promote/demote/stop             and
              role=Stopped/Started/Master/Slave.       Available      constraint_options      are
              id=<constraint-id>, kind=Optional/Mandatory/Serialize and symmetrical=true/false.

       order remove <resource1> [resourceN]...
              Remove resource from any ordering constraint

       colocation show [--full]
              List all current colocation constraints (if --full is specified show  the  internal
              constraint id's as well).

       colocation  add  [master|slave]  <source resource id> with [master|slave] <target resource
       id> [score] [options] [id=constraint-id]
              Request <source resource> to run on the same node where  pacemaker  has  determined
              <target  resource>  should run.  Positive values of score mean the resources should
              be run on the same node, negative values mean the resources should not  be  run  on
              the same node.  Specifying 'INFINITY' (or '-INFINITY') for the score forces <source
              resource>  to  run  (or  not  run)  with  <target  resource>  (score  defaults   to
              "INFINITY").   A  role can be master or slave (if no role is specified, it defaults
              to 'started').

       colocation  set  <resource1>  <resource2>  [resourceN]...   [options]   [set   <resourceX>
       <resourceY> ... [options]] [setoptions [constraint_options]]
              Create  a  colocation  constraint  with  a  resource  set.  Available  options  are
              sequential=true/false, require-all=true/false, action=start/promote/demote/stop and
              role=Stopped/Started/Master/Slave.  Available  constraint_options  are  id,  score,
              score-attribute and score-attribute-mangle.

       colocation remove <source resource id> <target resource id>
              Remove colocation constraints with <source resource>

       remove [constraint id]...
              Remove constraint(s) or constraint rules with the specified id(s)

       ref <resource>...
              List constraints referencing specified resource

       rule  add  <constraint  id>  [id=<rule  id>]   [role=master|slave]   [score=<score>|score-
       attribute=<attribute>] <expression>
              Add a rule to a constraint where the expression looks like one of the following:
                defined|not_defined <attribute>
                <attribute> lt|gt|lte|gte|eq|ne [string|integer|version] <value>
                date gt|lt <date>
                date in_range <date> to <date>
                date in_range <date> to duration <duration options>...
                date-spec <date spec options>...
                <expression> and|or <expression>
                ( <expression> )
              where  duration  options  and  date  spec  options are: hours, monthdays, weekdays,
              yeardays, months, weeks, years, weekyears, moon If score is ommited it defaults  to
              INFINITY. If id is ommited one is generated from the constraint id.

       rule remove <rule id>
              Remove  a  rule  if a rule id is specified, if rule is last rule in its constraint,
              the constraint will be removed

   status
       [status] [--full | --hide-inactive]
              View all information about the cluster and resources (--full provides more details,
              --hide-inactive hides inactive resources)

       resources
              View current status of cluster resources

       groups View currently configured groups and their resources

       cluster
              View current cluster status

       corosync
              View current membership information as seen by corosync

       nodes [corosync|both|config]
              View  current  status  of  nodes  from pacemaker. If 'corosync' is specified, print
              nodes currently configured in corosync, if 'both' is specified,  print  nodes  from
              both  corosync  & pacemaker.  If 'config' is specified, print nodes from corosync &
              pacemaker configuration.

       pcsd [<node>] ...
              Show the current status  of  pcsd  on  the  specified  nodes.  When  no  nodes  are
              specified, status of all nodes is displayed.

       xml    View xml version of status (output from crm_mon -r -1 -X)

   config
       [show] View full cluster configuration

       backup [filename]
              Creates the tarball containing the cluster configuration files.  If filename is not
              specified the standard output will be used.

       restore [--local] [filename]
              Restores the cluster configuration files on all nodes from the backup.  If filename
              is not specified the standard input will be used.  If --local is specified only the
              files on the current node will be restored.

       checkpoint
              List all available configuration checkpoints.

       checkpoint view <checkpoint_number>
              Show specified configuration checkpoint.

       checkpoint restore <checkpoint_number>
              Restore cluster configuration to specified checkpoint.

       import-cman         output=<filename>          [input=<filename>]          [--interactive]
       [output-format=corosync.conf|cluster.conf]
              Converts  CMAN cluster configuration to Pacemaker cluster configuration.  Converted
              configuration will be saved to 'output' file.  To send  the  configuration  to  the
              cluster  nodes  the  'pcs config restore' command can be used.  If --interactive is
              specified you will be prompted to solve incompatibilities manually.  If no input is
              specified  /etc/cluster/cluster.conf  will be used.  You can force to create output
              containing either cluster.conf or corosync.conf using the output-format option.

       import-cman  output=<filename>   [input=<filename>]   [--interactive]   output-format=pcs-
       commands|pcs-commands-verbose
              Converts  CMAN  cluster configuration to a list of pcs commands which recreates the
              same cluster as Pacemaker  cluster  when  executed.   Commands  will  be  saved  to
              'output' file.  For other options see above.

       export pcs-commands|pcs-commands-verbose output=<filename>
              Creates  a  list of pcs commands which upon execution recreates the current cluster
              running on this node.  Commands will be saved to 'output' file.   Use  pcs-commands
              to  get  a  simple  list  of  commands, whereas pcs-commands-verbose creates a list
              including comments and debug messages.

   pcsd
       certkey <certificate file> <key file>
              Load custom certificate and key files for use in pcsd.

       sync-certificates
              Sync  pcsd  certificates  to  all  nodes  found  from  current  corosync.conf  file
              (cluster.conf  on  systems  running Corosync 1.x).  WARNING: This will restart pcsd
              daemon on the nodes.

       clear-auth [--local] [--remote]
              Removes  all  system  tokens  which  allow  pcs/pcsd  on  the  current  system   to
              authenticate  with remote pcs/pcsd instances and vice-versa.  After this command is
              run this node will need to be re-authenticated with other nodes (using 'pcs cluster
              auth').   Using --local only removes tokens used by local pcs (and pcsd if root) to
              connect to other pcsd instances, using --remote clears authentication  tokens  used
              by remote systems to connect to the local pcsd instance.

   node
       maintenance [--all] | [node]...
              Put  specified  node(s)  into maintenance mode, if no node or options are specified
              the current node will be put into maintenance mode, if --all is specified all nodes
              will be put into maintenace mode.

       unmaintenance [--all] | [node]...
              Remove  node(s)  from  maintenance  mode,  if  no node or options are specified the
              current node will be removed from maintenance mode, if --all is specified all nodes
              will be removed from maintenance mode.

       utilization [<node> [<name>=<value> ...]]
              Add  specified  utilization  options  to  specified node. If node is not specified,
              shows utilization of all nodes. If utilization options  are  not  specified,  shows
              utilization  of  specified node. Utilization option should be in format name=value,
              value has to be integer. Options may be removed by  setting  an  option  without  a
              value. Example: pcs node utilization node1 cpu=4 ram=

EXAMPLES

       Show all resources
              # pcs resource show

       Show options specific to the 'VirtualIP' resource
              # pcs resource show VirtualIP

       Create a new resource called 'VirtualIP' with options
              #    pcs    resource   create   VirtualIP   ocf:heartbeat:IPaddr2   ip=192.168.0.99
              cidr_netmask=32 nic=eth2 op monitor interval=30s

       Create a new resource called 'VirtualIP' with options
              # pcs resource create VirtualIP IPaddr2 ip=192.168.0.99 cidr_netmask=32 nic=eth2 op
              monitor interval=30s

       Change the ip address of VirtualIP and remove the nic option
              # pcs resource update VirtualIP ip=192.168.0.98 nic=

       Delete the VirtualIP resource
              # pcs resource delete VirtualIP

       Create the MyStonith stonith fence_virt device which can fence host 'f1'
              # pcs stonith create MyStonith fence_virt pcmk_host_list=f1

       Set the stonith-enabled property to false on the cluster (which disables stonith)
              # pcs property set stonith-enabled=false