Provided by: slurm-llnl_2.3.2-1ubuntu1_amd64 bug

NAME

       sinfo - view information about SLURM nodes and partitions.

SYNOPSIS

       sinfo [OPTIONS...]

DESCRIPTION

       sinfo is used to view partition and node information for a system running SLURM.

OPTIONS

       -a, --all
              Display  information  about  all  partions. This causes information to be displayed
              about partitions that are configured as hidden and partitions that are  unavailable
              to user's group.

       -b, --bgl
              Display information about bglblocks (on Blue Gene systems only).

       -d, --dead
              If set only report state information for non-responding (dead) nodes.

       -e, --exact
              If set, do not group node information on multiple nodes unless their configurations
              to be reported are identical. Otherwise cpu count, memory size, and disk space  for
              nodes  will  be  listed with the minimum value followed by a "+" for nodes with the
              same partition and state (e.g., "250+").

       -h, --noheader
              Do not print a header on the output.

       --help Print a message describing all sinfo options.

       --hide Do not display information about hidden partitions. By default, partitions that are
              configured as hidden or are not available to the user's group will not be displayed
              (i.e. this is the default behavior).

       -i <seconds>, --iterate=<seconds>
              Print the state on a periodic basis.  Sleep for the  indicated  number  of  seconds
              between reports.  By default, prints a time stamp with the header.

       -l, --long
              Print  more  detailed  information.   This  is  ignored  if  the --format option is
              specified.

       -M, --clusters=<string>
              Clusters to issue commands to.  Multiple cluster names may be comma  separated.   A
              value of of 'all' will query to run on all clusters.

       -n <nodes>, --nodes=<nodes>
              Print  information  only  about the specified node(s).  Multiple nodes may be comma
              separated or expressed using a node range expression.  For  example  "linux[00-07]"
              would indicate eight nodes, "linux00" through "linux07."

       -N, --Node
              Print  information  in a node-oriented format.  The default is to print information
              in a partition-oriented  format.   This  is  ignored  if  the  --format  option  is
              specified.

       -o <output_format>, --format=<output_format>
              Specify  the  information  to  be  displayed  using  an sinfo format string. Format
              strings transparently used by sinfo when running with various options are

              default        "%9P %5a %.10l %.5D %6t %N"

              --summarize    "%9P %5a %.10l %16F %N"

              --long         "%9P %5a %.10l %.8s %4r %5h %10g %.5D %11T %N"

              --Node         "%N %.5D %9P %6t"

              --long --Node  "%N %.5D %9P %11T %.4c %.8z %.6m %.8d %.6w %8f %20R"

              --list-reasons "%20R %9u %19H %N"

              --long --list-reasons
                             "%20R %12u %19H %6t %N"

              In the above format strings the use of "#" represents the maximum length of an node
              list to be printed.

              The field specifications available include:

              %a  State/availability of a partition

              %A  Number  of nodes by state in the format "allocated/idle".  Do not use this with
                  a node state option ("%t" or "%T") or the different node states will be  placed
                  on separate lines.

              %c  Number of CPUs per node

              %C  Number  of CPUs by state in the format "allocated/idle/other/total". Do not use
                  this with a node state option ("%t" or "%T") or the different node states  will
                  be placed on separate lines.

              %d  Size of temporary disk space per node in megabytes

              %D  Number of nodes

              %E  The  reason a node is unavailable (down, drained, or draining states).  This is
                  the same as %R except the entries will be sorted by time rather than the reason
                  string.

              %f  Features associated with the nodes

              %F  Number  of  nodes  by state in the format "allocated/idle/other/total".  Do not
                  use this with a node state option ("%t" or "%T") or the different  node  states
                  will be placed on separate lines.

              %g  Groups which may use the nodes

              %G  Generic resources (gres) associated with the nodes

              %h  Jobs may share nodes, "yes", "no", or "force"

              %H  Print the timestamp of the reason a node is unavailable.

              %l  Maximum time for any job in the format "days-hours:minutes:seconds"

              %L  Default time for any job in the format "days-hours:minutes:seconds"

              %m  Size of memory per node in megabytes %M PreemptionMode

              %n  List of node hostnames

              %N  List of node names

              %o  List of node communication addresses

              %P  Partition name

              %r  Only user root may initiate jobs, "yes" or "no"

              %R  The  reason  a  node  is  unavailable (down, drained, draining, fail or failing
                  states)

              %s  Maximum job size in nodes

              %S  Allowed allocating nodes

              %t  State of nodes, compact form

              %T  State of nodes, extended form

              %u  Print the user name of who set the reason a node is unavailable.

              %U  Print the user name and uid of who set the reason a node is unavailable.

              %w  Scheduling weight of the nodes

              %X  Number of sockets per node

              %Y  Number of cores per socket

              %Z  Number of threads per core

              %z  Extended processor information: number of sockets, cores, threads  (S:C:T)  per
                  node

              %.<*>
                  right justification of the field

              %<Number><*>
                  size of field

       -p <partition>, --partition=<partition>
              Print information only about the specified partition.

       -r, --responding
              If set only report state information for responding nodes.

       -R, --list-reasons
              List reasons nodes are in the down, drained, fail or failing state.  When nodes are
              in these states SLURM supports optional  inclusion  of  a  "reason"  string  by  an
              administrator.   This  option  will  display  the first 35 characters of the reason
              field and list of nodes with that reason for all nodes that are, by default,  down,
              drained,  draining  or  failing.  This option may be used with other node filtering
              options (e.g. -r, -d, -t, -n), however, combinations of these options  that  result
              in  a  list  of  nodes that are not down or drained or failing will not produce any
              output.  When used with -l the output additionally includes the current node state.

       -s, --summarize
              List only a partition state summary with no node state details.  This is ignored if
              the --format option is specified.

       -S <sort_list>, --sort=<sort_list>
              Specification of the order in which records should be reported.  This uses the same
              field specifciation as the <output_format>.  Multiple sorts  may  be  performed  by
              listing  multiple sort fields separated by commas.  The field specifications may be
              preceded by "+" or "-" for assending (default) and  desending  order  respectively.
              The  partition  field  specification,  "P",  may  be  preceded  by  a "#" to report
              partitions in the same order that  they  appear  in  SLURM's   configuration  file,
              slurm.conf.   For example, a sort value of "+P,-m" requests that records be printed
              in order of increasing partition name and within a partition by  decreasing  memory
              size.   The default value of sort is "#P,-t" (partitions ordered as configured then
              decreasing node state).  If the --Node option is selected, the default  sort  value
              is "N" (increasing node name).

       -t <states> , --states=<states>
              List  nodes only having the given state(s).  Multiple states may be comma separated
              and  the  comparison  is  case  insensitive.    Possible   values   include   (case
              insensitive): ALLOC, ALLOCATED, COMP, COMPLETING, DOWN, DRAIN (for node in DRAINING
              or DRAINED states), DRAINED, DRAINING,  FAIL,  FAILING,  IDLE,  MAINT,  NO_RESPOND,
              POWER_SAVE, UNK, and UNKNOWN.  By default nodes in the specified state are reported
              whether they are responding or not.  The --dead and  --responding  options  may  be
              used to filtering nodes by the responding flag.

       --usage
              Print a brief message listing the sinfo options.

       -v, --verbose
              Provide detailed event logging through program execution.

       -V, --version
              Print version information and exit.

OUTPUT FIELD DESCRIPTIONS

       AVAIL  Partition state: up or down.

       CPUS   Count of CPUs (processors) on each node.

       S:C:T  Count of sockets (S), cores (C), and threads (T) on these nodes.

       SOCKETS
              Count of sockets on these nodes.

       CORES  Count of cores on these nodes.

       THREADS
              Count of threads on these nodes.

       GROUPS Resource  allocations  in  this  partition are restricted to the named groups.  all
              indicates that all groups may use this partition.

       JOB_SIZE
              Minimum and maximum node count that can be allocated to any  user  job.   A  single
              number indicates the minimum and maximum node count are the same.  infinite is used
              to identify partitions without a maximum node count.

       TIMELIMIT
              Maximum time limit for any user job  in  days-hours:minutes:seconds.   infinite  is
              used to identify partitions without a job time limit.

       MEMORY Size of real memory in megabytes on these nodes.

       NODELIST or BP_LIST (BlueGene systems only)
              Names of nodes associated with this configuration/partition.

       NODES  Count of nodes with this particular configuration.

       NODES(A/I)
              Count  of  nodes  with  this  particular  configuration  by  node state in the form
              "available/idle".

       NODES(A/I/O/T)
              Count of nodes with this  particular  configuration  by  node  state  in  the  form
              "available/idle/other/total".

       PARTITION
              Name of a partition.  Note that the suffix "*" identifies the default partition.

       ROOT   Is the ability to allocate resources in this partition restricted to user root, yes
              or no.

       SHARE  Will jobs  allocated  resources  in  this  partition  share  those  resources.   no
              indicates  resources  are  never  shared.   exclusive  indicates  whole  nodes  are
              dedicated to jobs (equivalent to srun --exclusive option, may  be  used  even  with
              shared/cons_res  managing  individual  processors).   force indicates resources are
              always available to be shared.  yes indicates resource may be  shared  or  not  per
              job's resource allocation.

       STATE  State of the nodes.  Possible states include: allocated, completing, down, drained,
              draining, fail, failing, idle, and unknown plus  their  abbreviated  forms:  alloc,
              comp,  donw,  drain,  drng, fail, failg, idle, and unk respectively.  Note that the
              suffix "*" identifies nodes that are presently not responding.

       TMP_DISK
              Size of temporary disk space in megabytes on these nodes.

NODE STATE CODES

       Node state codes are shortened as required for the field size.  If the node state code  is
       followed  by  "*",  this  indicates  the  node is presently not responding and will not be
       allocated any new work.  If the node remains non-responsive, it will be placed in the DOWN
       state (except in the case of COMPLETING, DRAINED, DRAINING, FAIL, FAILING nodes).

       If the node state code is followed by "~", this indicates the node is presently in a power
       saving mode (typically running at reduced frequency).  If the node state code is  followed
       by "#", this indicates the node is presently being powered up or configured.

       ALLOCATED   The node has been allocated to one or more jobs.

       ALLOCATED+  The  node is allocated to one or more active jobs plus one or more jobs are in
                   the process of COMPLETING.

       COMPLETING  All jobs associated with this node are in the  process  of  COMPLETING.   This
                   node state will be removed when all of the job's processes have terminated and
                   the SLURM epilog program (if any) has terminated.  See  the  Epilog  parameter
                   description in the slurm.conf man page for more information.

       DOWN        The  node  is unavailable for use. SLURM can automatically place nodes in this
                   state if some failure occurs. System administrators may also explicitly  place
                   nodes   in  this  state.  If  a  node  resumes  normal  operation,  SLURM  can
                   automatically return it to service. See the ReturnToService and  SlurmdTimeout
                   parameter descriptions in the slurm.conf(5) man page for more information.

       DRAINED     The  node  is  unavailable  for use per system administrator request.  See the
                   update node command in the scontrol(1) man page or the slurm.conf(5) man  page
                   for more information.

       DRAINING    The node is currently executing a job, but will not be allocated to additional
                   jobs. The node state will be changed to state DRAINED when the last job on  it
                   completes.  Nodes  enter  this state per system administrator request. See the
                   update node command in the scontrol(1) man page or the slurm.conf(5) man  page
                   for more information.

       FAIL        The  node  is  expected  to  fail  soon  and is unavailable for use per system
                   administrator request.  See the update node command  in  the  scontrol(1)  man
                   page or the slurm.conf(5) man page for more information.

       FAILING     The  node  is  currently  executing a job, but is expected to fail soon and is
                   unavailable for use per system administrator request.   See  the  update  node
                   command  in  the  scontrol(1)  man page or the slurm.conf(5) man page for more
                   information.

       IDLE        The node is not allocated to any jobs and is available for use.

       MAINT       The node is currently in a reservation with a flag value of "maintainence".

       UNKNOWN     The SLURM controller has just started and the node's state has  not  yet  been
                   determined.

ENVIRONMENT VARIABLES

       Some  sinfo  options  may  be  set via environment variables. These environment variables,
       along with their corresponding options, are listed below. (Note: Commandline options  will
       always override these settings.)

       SINFO_ALL           -a, --all

       SINFO_FORMAT        -o <output_format>, --format=<output_format>

       SINFO_PARTITION     -p <partition>, --partition=<partition>

       SINFO_SORT          -S <sort>, --sort=<sort>

       SLURM_CLUSTERS      Same as --clusters

       SLURM_CONF          The location of the SLURM configuration file.

       SLURM_TIME_FORMAT   Specify  the  format  used to report time stamps. A value of standard,
                           the  default  value,  generates  output  in  the   form   "year-month-
                           dateThour:minute:second".    A   value   of   relative   returns  only
                           "hour:minute:second" if the current  day.   For  other  dates  in  the
                           current   year  it  prints  the  "hour:minute"  preceded  by  "Tomorr"
                           (tomorrow), "Ystday" (yesterday), the name of the day for  the  coming
                           week  (e.g.  "Mon",  "Tue", etc.), otherwise the date (e.g. "25 Apr").
                           For other years it returns a date month and year without a time  (e.g.
                           "6  Jun  2012").  Another suggested value is "%a %T" for a day of week
                           and time stamp (e.g.  "Mon 12:34:56"). All of the time stamps use a 24
                           hour format.

EXAMPLES

       Report basic node and partition configurations:

       > sinfo
       PARTITION AVAIL TIMELIMIT NODES STATE  NODELIST
       batch     up     infinite     2 alloc  adev[8-9]
       batch     up     infinite     6 idle   adev[10-15]
       debug*    up        30:00     8 idle   adev[0-7]

       Report partition summary information:

       > sinfo -s
       PARTITION AVAIL TIMELIMIT NODES(A/I/O/T) NODELIST
       batch     up     infinite 2/6/0/8        adev[8-15]
       debug*    up        30:00 0/8/0/8        adev[0-7]

       Report more complete information about the partition debug:

       > sinfo --long --partition=debug
       PARTITION AVAIL TIMELIMIT JOB_SIZE ROOT SHARE GROUPS NODES STATE NODELIST
       debug*    up        30:00        8 no   no    all        8 idle  dev[0-7]

       Report only those nodes that are in state DRAINED:

       > sinfo --states=drained
       PARTITION AVAIL NODES TIMELIMIT STATE  NODELIST
       debug*    up        2     30:00 drain  adev[6-7]

       Report node-oriented information with details and exact matches:

       > sinfo -Nel
       NODELIST    NODES PARTITION STATE  CPUS MEMORY TMP_DISK WEIGHT FEATURES REASON
       adev[0-1]       2 debug*    idle      2   3448    38536     16 (null)   (null)
       adev[2,4-7]     5 debug*    idle      2   3384    38536     16 (null)   (null)
       adev3           1 debug*    idle      2   3394    38536     16 (null)   (null)
       adev[8-9]       2 batch     allocated 2    246    82306     16 (null)   (null)
       adev[10-15]     6 batch     idle      2    246    82306     16 (null)   (null)

       Report only down, drained and draining nodes and their reason field:

       > sinfo -R
       REASON                              NODELIST
       Memory errors                       dev[0,5]
       Not Responding                      dev8

COPYING

       Copyright  (C)  2002-2007  The  Regents  of  the  University of California.  Copyright (C)
       2008-2009 Lawrence Livermore National  Security.   Portions  Copyright  (C)  2010  SchedMD
       <http://www.schedmd.com>.    Produced  at  Lawrence  Livermore  National  Laboratory  (cf,
       DISCLAIMER).  CODE-OCEC-09-009. All rights reserved.

       This  file  is  part  of  SLURM,  a  resource  management  program.   For   details,   see
       <http://www.schedmd.com/slurmdocs/>.

       SLURM  is  free  software; you can redistribute it and/or modify it under the terms of the
       GNU General Public License as published by the Free Software Foundation; either version  2
       of the License, or (at your option) any later version.

       SLURM is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without
       even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
       GNU General Public License for more details.

SEE ALSO

       scontrol(1),   smap(1),   squeue(1),   slurm_load_ctl_conf   (3),   slurm_load_jobs   (3),
       slurm_load_node (3), slurm_load_partitions (3), slurm_reconfigure (3), slurm_shutdown (3),
       slurm_update_job (3), slurm_update_node (3), slurm_update_partition (3), slurm.conf(5)