Provided by: slurm-llnl_2.6.5-1_amd64 bug

NAME

       sinfo - view information about SLURM nodes and partitions.

SYNOPSIS

       sinfo [OPTIONS...]

DESCRIPTION

       sinfo is used to view partition and node information for a system running SLURM.

OPTIONS

       -a, --all
              Display information about all partitions. This causes information to be displayed about partitions
              that are configured as hidden and partitions that are unavailable to user's group.

       -b, --bgl
              Display information about bglblocks (on Blue Gene systems only).

       -d, --dead
              If set only report state information for non-responding (dead) nodes.

       -e, --exact
              If set, do not group node information on multiple nodes unless their configurations to be reported
              are  identical. Otherwise cpu count, memory size, and disk space for nodes will be listed with the
              minimum value followed by a "+" for nodes with the same partition and state (e.g., "250+").

       -h, --noheader
              Do not print a header on the output.

       --help Print a message describing all sinfo options.

       --hide Do not display information about hidden partitions. By default, partitions that are configured  as
              hidden  or  are  not available to the user's group will not be displayed (i.e. this is the default
              behavior).

       -i <seconds>, --iterate=<seconds>
              Print the state on a periodic basis.  Sleep for the indicated number of seconds  between  reports.
              By default, prints a time stamp with the header.

       -l, --long
              Print more detailed information.  This is ignored if the --format option is specified.

       -M, --clusters=<string>
              Clusters  to  issue  commands  to.   Multiple cluster names may be comma separated.  A value of of
              'all' will query to run on all clusters.

       -n <nodes>, --nodes=<nodes>
              Print information only about the specified node(s).  Multiple nodes  may  be  comma  separated  or
              expressed  using  a  node range expression. For example "linux[00-07]" would indicate eight nodes,
              "linux00" through "linux07."  Performance of the command can be measurably  improved  for  systems
              with large numbers of nodes when a single node name is specified.

       -N, --Node
              Print  information  in  a  node-oriented  format.   The  default  is  to  print  information  in a
              partition-oriented format.  This is ignored if the --format option is specified.

       -o <output_format>, --format=<output_format>
              Specify the information to be displayed using an sinfo format string. Format strings transparently
              used by sinfo when running with various options are

              default        "%P %5a %.10l %.5D %6t %N"

              --summarize    "%P %5a %.10l %16F %N"

              --long         "%P %5a %.10l %.8s %4r %5h %10g %.5D %11T %N"

              --Node         "%N %.5D %9P %6t"

              --long --Node  "%N %.5D %9P %11T %.4c %.8z %.6m %.8d %.6w %8f %20E"

              --list-reasons "%20E %9u %19H %N"

              --long --list-reasons
                             "%20E %12U %19H %6t %N"

              In the above format strings the use of "#" represents the maximum length of an  node  list  to  be
              printed.

              The field specifications available include:

              %a  State/availability of a partition

              %A  Number  of  nodes  by state in the format "allocated/idle".  Do not use this with a node state
                  option ("%t" or "%T") or the different node states will be placed on separate lines.

              %B  The max number of CPUs per node available to jobs in the partition.

              %c  Number of CPUs per node

              %C  Number of CPUs by state in the format "allocated/idle/other/total". Do not  use  this  with  a
                  node  state  option  ("%t"  or  "%T")  or the different node states will be placed on separate
                  lines.

              %d  Size of temporary disk space per node in megabytes

              %D  Number of nodes

              %E  The reason a node is unavailable (down, drained, or draining states).

              %f  Features associated with the nodes

              %F  Number of nodes by state in the format "allocated/idle/other/total".  Do not use this  with  a
                  node  state  option  ("%t"  or  "%T")  or the different node states will be placed on separate
                  lines.

              %g  Groups which may use the nodes

              %G  Generic resources (gres) associated with the nodes

              %h  Jobs may share nodes, "yes", "no", or "force"

              %H  Print the timestamp of the reason a node is unavailable.

              %l  Maximum time for any job in the format "days-hours:minutes:seconds"

              %L  Default time for any job in the format "days-hours:minutes:seconds"

              %m  Size of memory per node in megabytes

              %M  PreemptionMode

              %n  List of node hostnames

              %N  List of node names

              %o  List of node communication addresses

              %O  CPU load of a node

              %p  Partition scheduling priority

              %P  Partition name followed by "*" for the default partition, also see %R

              %r  Only user root may initiate jobs, "yes" or "no"

              %R  Partition name, also see %P

              %s  Maximum job size in nodes

              %S  Allowed allocating nodes

              %t  State of nodes, compact form

              %T  State of nodes, extended form

              %u  Print the user name of who set the reason a node is unavailable.

              %U  Print the user name and uid of who set the reason a node is unavailable.

              %w  Scheduling weight of the nodes

              %X  Number of sockets per node

              %Y  Number of cores per socket

              %Z  Number of threads per core

              %z  Extended processor information: number of sockets, cores, threads (S:C:T) per node

              %.<*>
                  right justification of the field

              %<Number><*>
                  size of field

       -p <partition>, --partition=<partition>
              Print information only about the specified partition.

       -r, --responding
              If set only report state information for responding nodes.

       -R, --list-reasons
              List reasons nodes are in the down, drained, fail or failing  state.   When  nodes  are  in  these
              states  SLURM  supports  optional inclusion of a "reason" string by an administrator.  This option
              will display the first 35 characters of the reason field and list of nodes with  that  reason  for
              all  nodes that are, by default, down, drained, draining or failing.  This option may be used with
              other node filtering options (e.g. -r, -d, -t, -n), however, combinations of  these  options  that
              result  in  a  list  of nodes that are not down or drained or failing will not produce any output.
              When used with -l the output additionally includes the current node state.

       -s, --summarize
              List only a partition state summary with no node state details.  This is ignored if  the  --format
              option is specified.

       -S <sort_list>, --sort=<sort_list>
              Specification  of  the  order  in  which  records  should  be  reported.  This uses the same field
              specification as the <output_format>.  Multiple sorts may be performed by  listing  multiple  sort
              fields  separated by commas.  The field specifications may be preceded by "+" or "-" for ascending
              (default) and descending order respectively.  The  partition  field  specification,  "P",  may  be
              preceded  by  a  "#"  to  report  partitions  in  the  same  order  that  they  appear  in SLURM's
              configuration file, slurm.conf.  For example, a sort value of "+P,-m"  requests  that  records  be
              printed  in  order  of increasing partition name and within a partition by decreasing memory size.
              The default value of sort is "#P,-t"  (partitions  ordered  as  configured  then  decreasing  node
              state).  If the --Node option is selected, the default sort value is "N" (increasing node name).

       -t <states> , --states=<states>
              List  nodes  only  having  the  given  state(s).   Multiple  states may be comma separated and the
              comparison is case insensitive.  Possible values include  (case  insensitive):  ALLOC,  ALLOCATED,
              COMP,  COMPLETING,  DOWN, DRAIN (for node in DRAINING or DRAINED states), DRAINED, DRAINING, FAIL,
              FAILING, IDLE, MAINT, NO_RESPOND, POWER_SAVE, UNK, and UNKNOWN.  By default nodes in the specified
              state are reported whether they are responding or not.  The --dead and --responding options may be
              used to filtering nodes by the responding flag.

       -T, --reservation
              Only display information about SLURM reservations.

       --usage
              Print a brief message listing the sinfo options.

       -v, --verbose
              Provide detailed event logging through program execution.

       -V, --version
              Print version information and exit.

OUTPUT FIELD DESCRIPTIONS

       AVAIL  Partition state: up or down.

       CPUS   Count of CPUs (processors) on each node.

       S:C:T  Count of sockets (S), cores (C), and threads (T) on these nodes.

       SOCKETS
              Count of sockets on these nodes.

       CORES  Count of cores on these nodes.

       THREADS
              Count of threads on these nodes.

       GROUPS Resource allocations in this partition are restricted to the named groups.  all indicates that all
              groups may use this partition.

       JOB_SIZE
              Minimum and maximum node count that can be allocated to any user job.  A single  number  indicates
              the  minimum and maximum node count are the same.  infinite is used to identify partitions without
              a maximum node count.

       TIMELIMIT
              Maximum time limit for any user job in days-hours:minutes:seconds.  infinite is used  to  identify
              partitions without a job time limit.

       MEMORY Size of real memory in megabytes on these nodes.

       NODELIST or BP_LIST (BlueGene systems only)
              Names of nodes associated with this configuration/partition.

       NODES  Count of nodes with this particular configuration.

       NODES(A/I)
              Count of nodes with this particular configuration by node state in the form "available/idle".

       NODES(A/I/O/T)
              Count   of   nodes   with   this   particular   configuration   by   node   state   in   the  form
              "available/idle/other/total".

       PARTITION
              Name of a partition.  Note that the suffix "*" identifies the default partition.

       ROOT   Is the ability to allocate resources in this partition restricted to user root, yes or no.

       SHARE  Will jobs allocated resources in this partition share those resources.  no indicates resources are
              never shared.  exclusive  indicates  whole  nodes  are  dedicated  to  jobs  (equivalent  to  srun
              --exclusive  option, may be used even with shared/cons_res managing individual processors).  force
              indicates resources are always available to be shared.  yes indicates resource may  be  shared  or
              not per job's resource allocation.

       STATE  State  of  the  nodes.   Possible  states include: allocated, completing, down, drained, draining,
              fail, failing, idle, and unknown plus their abbreviated forms: alloc,  comp,  donw,  drain,  drng,
              fail,  failg,  idle,  and  unk  respectively.   Note that the suffix "*" identifies nodes that are
              presently not responding.

       TMP_DISK
              Size of temporary disk space in megabytes on these nodes.

NODE STATE CODES

       Node state codes are shortened as required for the field size.  If the node state  code  is  followed  by
       "*",  this indicates the node is presently not responding and will not be allocated any new work.  If the
       node remains non-responsive, it will be placed in the DOWN state  (except  in  the  case  of  COMPLETING,
       DRAINED, DRAINING, FAIL, FAILING nodes).

       If  the  node  state code is followed by "~", this indicates the node is presently in a power saving mode
       (typically running at reduced frequency).  If the node state code is followed by "#", this indicates  the
       node is presently being powered up or configured.

       ALLOCATED   The node has been allocated to one or more jobs.

       ALLOCATED+  The  node is allocated to one or more active jobs plus one or more jobs are in the process of
                   COMPLETING.

       COMPLETING  All jobs associated with this node are in the process of COMPLETING.  This node state will be
                   removed when all of the job's processes have terminated and the SLURM epilog program (if any)
                   has terminated. See the Epilog parameter description in the  slurm.conf  man  page  for  more
                   information.

       DOWN        The  node  is  unavailable for use. SLURM can automatically place nodes in this state if some
                   failure occurs. System administrators may also explicitly place nodes in  this  state.  If  a
                   node  resumes  normal  operation,  SLURM  can  automatically  return  it  to service. See the
                   ReturnToService and SlurmdTimeout parameter descriptions in the slurm.conf(5)  man  page  for
                   more information.

       DRAINED     The  node  is  unavailable  for  use  per  system administrator request.  See the update node
                   command in the scontrol(1) man page or the slurm.conf(5) man page for more information.

       DRAINING    The node is currently executing a job, but will not be allocated to additional jobs. The node
                   state will be changed to state DRAINED when the last job on it completes.  Nodes  enter  this
                   state  per  system  administrator request. See the update node command in the scontrol(1) man
                   page or the slurm.conf(5) man page for more information.

       FAIL        The node is expected to fail soon  and  is  unavailable  for  use  per  system  administrator
                   request.   See  the  update node command in the scontrol(1) man page or the slurm.conf(5) man
                   page for more information.

       FAILING     The node is currently executing a job, but is expected to fail soon and  is  unavailable  for
                   use  per  system  administrator  request.  See the update node command in the scontrol(1) man
                   page or the slurm.conf(5) man page for more information.

       IDLE        The node is not allocated to any jobs and is available for use.

       MAINT       The node is currently in a reservation with a flag value of "maintenance" or is scheduled  to
                   be rebooted.

       UNKNOWN     The SLURM controller has just started and the node's state has not yet been determined.

ENVIRONMENT VARIABLES

       Some  sinfo  options  may be set via environment variables. These environment variables, along with their
       corresponding options, are listed below. (Note: Commandline options will always override these settings.)

       SINFO_ALL           -a, --all

       SINFO_FORMAT        -o <output_format>, --format=<output_format>

       SINFO_PARTITION     -p <partition>, --partition=<partition>

       SINFO_SORT          -S <sort>, --sort=<sort>

       SLURM_CLUSTERS      Same as --clusters

       SLURM_CONF          The location of the SLURM configuration file.

       SLURM_TIME_FORMAT   Specify the format used to report time stamps.  A  value  of  standard,  the  default
                           value, generates output in the form "year-month-dateThour:minute:second".  A value of
                           relative  returns  only  "hour:minute:second" if the current day.  For other dates in
                           the current year  it  prints  the  "hour:minute"  preceded  by  "Tomorr"  (tomorrow),
                           "Ystday"  (yesterday),  the  name  of the day for the coming week (e.g. "Mon", "Tue",
                           etc.), otherwise the date (e.g. "25 Apr").  For other years it returns a  date  month
                           and year without a time (e.g.  "6 Jun 2012").  Another suggested value is "%a %T" for
                           a  day of week and time stamp (e.g.  "Mon 12:34:56"). All of the time stamps use a 24
                           hour format.

EXAMPLES

       Report basic node and partition configurations:

       > sinfo
       PARTITION AVAIL TIMELIMIT NODES STATE  NODELIST
       batch     up     infinite     2 alloc  adev[8-9]
       batch     up     infinite     6 idle   adev[10-15]
       debug*    up        30:00     8 idle   adev[0-7]

       Report partition summary information:

       > sinfo -s
       PARTITION AVAIL TIMELIMIT NODES(A/I/O/T) NODELIST
       batch     up     infinite 2/6/0/8        adev[8-15]
       debug*    up        30:00 0/8/0/8        adev[0-7]

       Report more complete information about the partition debug:

       > sinfo --long --partition=debug
       PARTITION AVAIL TIMELIMIT JOB_SIZE ROOT SHARE GROUPS NODES STATE NODELIST
       debug*    up        30:00        8 no   no    all        8 idle  dev[0-7]

       Report only those nodes that are in state DRAINED:

       > sinfo --states=drained
       PARTITION AVAIL NODES TIMELIMIT STATE  NODELIST
       debug*    up        2     30:00 drain  adev[6-7]

       Report node-oriented information with details and exact matches:

       > sinfo -Nel
       NODELIST    NODES PARTITION STATE  CPUS MEMORY TMP_DISK WEIGHT FEATURES REASON
       adev[0-1]       2 debug*    idle      2   3448    38536     16 (null)   (null)
       adev[2,4-7]     5 debug*    idle      2   3384    38536     16 (null)   (null)
       adev3           1 debug*    idle      2   3394    38536     16 (null)   (null)
       adev[8-9]       2 batch     allocated 2    246    82306     16 (null)   (null)
       adev[10-15]     6 batch     idle      2    246    82306     16 (null)   (null)

       Report only down, drained and draining nodes and their reason field:

       > sinfo -R
       REASON                              NODELIST
       Memory errors                       dev[0,5]
       Not Responding                      dev8

COPYING

       Copyright (C) 2002-2007 The Regents of the University of  California.   Produced  at  Lawrence  Livermore
       National Laboratory (cf, DISCLAIMER).
       Copyright (C) 2008-2009 Lawrence Livermore National Security.
       Copyright (C) 2010-2013 SchedMD LLC.

       This file is part of SLURM, a resource management program.  For details, see <http://slurm.schedmd.com/>.

       SLURM  is  free  software;  you  can  redistribute it and/or modify it under the terms of the GNU General
       Public License as published by the Free Software Foundation; either version 2 of the License, or (at your
       option) any later version.

       SLURM is distributed in the hope that it will be useful, but  WITHOUT  ANY  WARRANTY;  without  even  the
       implied  warranty  of  MERCHANTABILITY  or  FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public
       License for more details.

SEE ALSO

       scontrol(1), smap(1), squeue(1),  slurm_load_ctl_conf  (3),  slurm_load_jobs  (3),  slurm_load_node  (3),
       slurm_load_partitions   (3),   slurm_reconfigure   (3),   slurm_shutdown   (3),   slurm_update_job   (3),
       slurm_update_node (3), slurm_update_partition (3), slurm.conf(5)

sinfo 2.6                                         December 2012                                         SINFO(1)