Provided by: glusterfs-server_3.7.6-1ubuntu1_i386 bug

NAME

       gluster - Gluster Console Manager (command line utility)

SYNOPSIS

       gluster

       To run the program and display gluster prompt:

       gluster

       (or)

       To specify a command directly:

       gluster [commands] [options]

DESCRIPTION

       The  Gluster  Console  Manager  is  a  command line utility for elastic
       volume management. You can  run  the  gluster  command  on  any  export
       server. The command enables administrators to perform cloud operations,
       such as creating,  expanding,  shrinking,  rebalancing,  and  migrating
       volumes without needing to schedule server downtime.

COMMANDS

   Volume Commands
        volume info [all|<VOLNAME>]
              Display information about all volumes, or the specified volume.

        volume   create   <NEW-VOLNAME>  [stripe  <COUNT>]  [replica  <COUNT>]
       [disperse     [<COUNT>]]      [redundancy      <COUNT>]      [transport
       <tcp|rdma|tcp,rdma>] <NEW-BRICK> ...
              Create  a  new  volume of the specified type using the specified
              bricks and transport type (the default transport type  is  tcp).
              To  create  a  volume  with both transports (tcp and rdma), give
              'transport tcp,rdma' as an option.

        volume delete <VOLNAME>
              Delete the specified volume.

        volume start <VOLNAME>
              Start the specified volume.

        volume stop <VOLNAME> [force]
              Stop the specified volume.

        volume set <VOLNAME> <OPTION> <PARAMETER> [<OPTION> <PARAMETER>] ...
              Set the volume options.

        volume get <VOLNAME> <OPTION/all>
              Get the volume options.

        volume help
              Display help for the volume command.

   Brick Commands
        volume add-brick <VOLNAME> <NEW-BRICK> ...
              Add the specified brick to the specified volume.

        volume remove-brick <VOLNAME> <BRICK> ...
              Remove the specified brick from the specified volume.

              Note: If you remove the brick, the data  stored  in  that  brick
              will  not  be  available. You can migrate data from one brick to
              another using replace-brick option.

        volume     replace-brick     <VOLNAME>      (<BRICK>      <NEW-BRICK>)
       start|pause|abort|status|commit
              Replace the specified brick.

        volume rebalance <VOLNAME> start
              Start rebalancing the specified volume.

        volume rebalance <VOLNAME> stop
              Stop rebalancing the specified volume.

        volume rebalance <VOLNAME> status
              Display the rebalance status of the specified volume.

   Log Commands
        volume log filename <VOLNAME> [BRICK] <DIRECTORY>
              Set the log directory for the corresponding volume/brick.

        volume log locate <VOLNAME> [BRICK]
              Locate the log file for corresponding volume/brick.

        volume log rotate <VOLNAME> [BRICK]
              Rotate the log file for corresponding volume/brick.

   Peer Commands
        peer probe <HOSTNAME>
              Probe  the  specified peer. In case the <HOSTNAME> given belongs
              to an already probed peer, the peer probe command will  add  the
              hostname to the peer if required.

        peer detach <HOSTNAME>
              Detach the specified peer.

        peer status
              Display the status of peers.

        peer help
              Display help for the peer command.

   Tier Commands
        volume tier <VOLNAME> attach [<replica COUNT>] <NEW-BRICK>...
              Attach  to an existing volume a tier of specified type using the
              specified bricks.

        volume tier <VOLNAME> status
              Display statistics on data migration between the  hot  and  cold
              tiers.

        volume tier <VOLNAME> detach start
              Begin detaching the hot tier from the volume. Data will be moved
              from the hot tier to the cold tier.

        volume tier <VOLNAME> detach commit [force]
              Commit detaching the hot tier from the volume. The  volume  will
              revert to its original state before the hot tier was attached.

        volume tier <VOLNAME> detach status
              Check status of data movement from the hot to cold tier.

        volume tier <VOLNAME> detach stop
              Stop detaching the hot tier from the volume.

   Bitrot Commands
        volume bitrot <VOLNAME> {enable|disable}
              Enable/disable bitrot for volume <VOLNAME>

        volume bitrot <VOLNAME> scrub-throttle {lazy|normal|aggressive}
              Scrub-throttle  value  is  a  measure  of  how  fast or slow the
              scrubber scrubs the filesystem for volume <VOLNAME>

        volume           bitrot           <VOLNAME>            scrub-frequency
       {daily|weekly|biweekly|monthly}
              Scrub frequency for volume <VOLNAME>

        volume bitrot <VOLNAME> scrub {pause|resume}
              Pause/Resume  scrub.  Upon  resume,  scrubber continues where it
              left off.

   Snapshot Commands
        snapshot  create  <snapname>  <volname>  [no-timestamp]   [description
       <description>] [force]
              Creates  a  snapshot  of  a GlusterFS volume. User can provide a
              snap-name and a description to identify the snap. Snap  will  be
              created  by  appending  timestamp in GMT. User can override this
              behaviour using "no-timestamp" option. The description cannot be
              more than 1024 characters. To be able to take a snapshot, volume
              should be present and it should be in started state.

        snapshot restore <snapname>
              Restores an  already  taken  snapshot  of  a  GlusterFS  volume.
              Snapshot  restore is an offline activity therefore if the volume
              is online (in started state) then  the  restore  operation  will
              fail.  Once the snapshot is restored it will not be available in
              the list of snapshots.

        snapshot clone <clonename> <snapname>
              Create a clone of a snapshot volume, the resulting  volume  will
              be  GlusterFS  volume. User can provide a clone-name. To be able
              to take a clone, snapshot should be present and it should be  in
              activated state.

        snapshot delete ( all | <snapname> | volume <volname> )
              If  snapname is specified then mentioned snapshot is deleted. If
              volname is  specified  then  all  snapshots  belonging  to  that
              particular  volume is deleted. If keyword *all* is used then all
              snapshots belonging to the system is deleted.

        snapshot list [volname]
              Lists all snapshots taken. If volname is provided, then only the
              snapshots belonging to that particular volume is listed.

        snapshot info [snapname | (volume <volname>)]
              This  command  gives information such as snapshot name, snapshot
              UUID, time at which snapshot was created, and it lists down  the
              snap-volume-name,  number  of snapshots already taken and number
              of snapshots still available for that particular volume, and the
              state of the snapshot. If snapname is specified then info of the
              mentioned  snapshot is  displayed.  If volname is specified then
              info of all snapshots belonging to that volume is displayed.  If
              both  snapname and  volname  is  not specified then info of  all
              the snapshots present in the system are displayed.

        snapshot status [snapname | (volume <volname>)]
              This  command gives status of the snapshot. The details included
              are snapshot brick path, volume group(LVM  details),  status  of
              the  snapshot  bricks, PID of the bricks, data percentage filled
              for that particular volume group to which the  snapshots  belong
              to, and total size of the logical volume.

              If  snapname  is specified then status of the mentioned snapshot
              is displayed.  If  volname  is  specified  then  status  of  all
              snapshots  belonging  to  that  volume  is  displayed.  If  both
              snapname and volname is not specified then  status  of  all  the
              snapshots present in the system are displayed.

        snapshot  config  [volname]  ([snap-max-hard-limit <count>] [snap-max-
       soft-limit <percent>]) | ([auto-delete <enable|disable>]) | ([activate-
       on-create <enable|disable>])
              Displays and sets the snapshot config values.

              snapshot  config  without  any  keywords  displays  the snapshot
              config values of all  volumes  in  the  system.  If  volname  is
              provided,  then  the  snapshot  config  values of that volume is
              displayed.

              Snapshot config command along  with  keywords  can  be  used  to
              change  the  existing config values. If volname is provided then
              config value of that volume is changed, else it will  set/change
              the system limit.

              snap-max-soft-limit  and  auto-delete  are  global options, that
              will be inherited by all volumes in the system and cannot be set
              to individual volumes.

              snap-max-hard-limit  can be set globally, as well as per volume.
              The lowest limit between the global system limit and the  volume
              specific  limit, becomes the "Effective snap-max-hard-limit" for
              a volume.

              snap-max-soft-limit is a percentage value, which is  applied  on
              the  "Effective snap-max-hard-limit" to get the "Effective snap-
              max-soft-limit".

              When auto-delete feature is  enabled,  then  upon  reaching  the
              "Effective  snap-max-soft-limit", with every successful snapshot
              creation, the oldest snapshot will be deleted.

              When auto-delete feature is disabled,  then  upon  reaching  the
              "Effective  snap-max-soft-limit",  the  user gets a warning with
              every successful snapshot creation.

              When auto-delete feature is disabled,  then  upon  reaching  the
              "Effective  snap-max-hard-limit",  further   snapshot  creations
              will not be allowed.

              activate-on-create  is  disabled  by  default.  If  you   enable
              activate-on-create,  then  further  snapshot  will  be activated
              during the time of snapshot creation.

        snapshot activate <snapname>
              Activates the mentioned snapshot.

              Note : By default the  snapshot  is  activated  during  snapshot
              creation.

        snapshot deactivate <snapname>
              Deactivates the mentioned snapshot.

        snapshot help
              Display help for the snapshot commands.

   Self-heal Commands
        volume heal <VOLNAME>
              Triggers index self heal for the files that need healing.

        volume heal  <VOLNAME> [enable | disable]
              Enable/disable self-heal-daemon for volume <VOLNAME>.

        volume heal <VOLNAME> full
              Triggers self heal on all the files.

        volume heal <VOLNAME> info
              Lists the files that need healing.

        volume heal <VOLNAME> info split-brain
              Lists the files which are in split-brain state.

        volume heal <VOLNAME> statistics
              Lists the crawl statistics.

        volume heal <VOLNAME> statistics heal-count
              Displays the count of files to be healed.

        volume     heal     <VOLNAME>     statistics     heal-count    replica
       <HOSTNAME:BRICKNAME>
              Displays the number of files to  be  healed  from  a  particular
              replica   subvolume  to  which  the  brick  <HOSTNAME:BRICKNAME>
              belongs.

        volume heal <VOLNAME> split-brain bigger-file <FILE>
              Performs healing of <FILE> which is in split-brain  by  choosing
              the bigger file in the replica as source.

        volume heal <VOLNAME> split-brain source-brick <HOSTNAME:BRICKNAME>
              Selects  <HOSTNAME:BRICKNAME>  as  the  source for all the files
              that are in split-brain in that replica and heals them.

        volume heal <VOLNAME>  split-brain  source-brick  <HOSTNAME:BRICKNAME>
       <FILE>
              Selects the split-brained <FILE> present in <HOSTNAME:BRICKNAME>
              as source and completes heal.

   Other Commands
        help  Display the command options.

        quit  Exit the gluster command line interface.

FILES

       /var/lib/glusterd/*

SEE ALSO

       fusermount(1), mount.glusterfs(8), glusterfs(8), glusterd(8)

COPYRIGHT

       Copyright(c) 2006-2011  Gluster, Inc.  <http://www.gluster.com>