focal (1) sinfo.1.gz

Provided by: slurm-client_19.05.5-1_amd64 bug

NAME

       sinfo - view information about Slurm nodes and partitions.

SYNOPSIS

       sinfo [OPTIONS...]

DESCRIPTION

       sinfo is used to view partition and node information for a system running Slurm.

OPTIONS

       -a, --all
              Display information about all partitions. This causes information to be displayed about partitions
              that are configured as hidden and partitions that are unavailable to user's group.

       -d, --dead
              If set only report state information for non-responding (dead) nodes.

       -e, --exact
              If set, do not group node information on multiple nodes unless their configurations to be reported
              are  identical. Otherwise cpu count, memory size, and disk space for nodes will be listed with the
              minimum value followed by a "+" for nodes with the same partition and state (e.g., "250+").

       --federation
              Show all partitions from the federation if a member of one.

       -h, --noheader
              Do not print a header on the output.

       --help Print a message describing all sinfo options.

       --hide Do not display information about hidden partitions. By default, partitions that are configured  as
              hidden  or  are  not available to the user's group will not be displayed (i.e. this is the default
              behavior).

       -i <seconds>, --iterate=<seconds>
              Print the state on a periodic basis.  Sleep for the indicated number of seconds  between  reports.
              By default, prints a time stamp with the header.

       --local
              Show only jobs local to this cluster. Ignore other clusters in this federation (if any). Overrides
              --federation.

       -l, --long
              Print more detailed information.  This is ignored if the --format option is specified.

       -M, --clusters=<string>
              Clusters to issue commands to.  Multiple cluster names may be comma  separated.   A  value  of  of
              'all'  will  query  to  run on all clusters.  Note that the SlurmDBD must be up for this option to
              work properly.  This option implicitly sets the --local option.

       -n <nodes>, --nodes=<nodes>
              Print information only about the specified node(s).  Multiple nodes  may  be  comma  separated  or
              expressed  using  a  node range expression. For example "linux[00-07]" would indicate eight nodes,
              "linux00" through "linux07."  Performance of the command can be measurably  improved  for  systems
              with large numbers of nodes when a single node name is specified.

       --noconvert
              Don't convert units from their original type (e.g. 2048M won't be converted to 2G).

       -N, --Node
              Print  information  in  a node-oriented format with one line per node and partition. That is, if a
              node belongs to more than one partition, then one line for each node-partition pair will be shown.
              If  --partition  is  also  specified, then only one line per node in this partition is shown.  The
              default is to print information in a partition-oriented format.  This is ignored if  the  --format
              option is specified.

       -o <output_format>, --format=<output_format>
              Specify the information to be displayed using an sinfo format string. Format strings transparently
              used by sinfo when running with various options are

              default        "%#P %.5a %.10l %.6D %.6t %N"

              --summarize    "%#P %.5a %.10l %.16F  %N"

              --long         "%#P %.5a %.10l %.10s %.4r %.8h %.10g %.6D %.11T %N"

              --Node         "%#N %.6D %#P %6t"

              --long --Node  "%#N %.6D %#P %.11T %.4c %.8z %.6m %.8d %.6w %.8f %20E"

              --list-reasons "%20E %9u %19H %N"

              --long --list-reasons
                             "%20E %12U %19H %6t %N"

              In the above format strings, the use of "#" represents the maximum length of any partition name or
              node  list  to be printed.  A pass is made over the records to be printed to establish the size in
              order to align the sinfo output, then a second pass is made over the records to print them.   Note
              that  the literal character "#" itself is not a valid field length specification, but is only used
              to document this behaviour.

              The field specifications available include:

              %all  Print all fields available for this data type with a vertical bar separating each field.

              %a    State/availability of a partition

              %A    Number of nodes by state in the format "allocated/idle".  Do not use this with a node  state
                    option ("%t" or "%T") or the different node states will be placed on separate lines.

              %b    Features currently active on the nodes, also see %f

              %B    The max number of CPUs per node available to jobs in the partition.

              %c    Number of CPUs per node

              %C    Number  of  CPUs by state in the format "allocated/idle/other/total". Do not use this with a
                    node state option ("%t" or "%T") or the different node states will  be  placed  on  separate
                    lines.

              %d    Size of temporary disk space per node in megabytes

              %D    Number of nodes

              %e    Free memory of a node

              %E    The reason a node is unavailable (down, drained, or draining states).

              %f    Features available the nodes, also see %b

              %F    Number  of  nodes by state in the format "allocated/idle/other/total".  Note the use of this
                    format option with a node state format option ("%t" or "%T") will result  in  the  different
                    node states being be reported on separate lines.

              %g    Groups which may use the nodes

              %G    Generic resources (gres) associated with the nodes

              %h    Jobs may oversubscribe compute resources (i.e. CPUs), "yes", "no", "exclusive" or "force"

              %H    Print the timestamp of the reason a node is unavailable.

              %I    Partition job priority weighting factor.

              %l    Maximum time for any job in the format "days-hours:minutes:seconds"

              %L    Default time for any job in the format "days-hours:minutes:seconds"

              %m    Size of memory per node in megabytes

              %M    PreemptionMode

              %n    List of node hostnames

              %N    List of node names

              %o    List of node communication addresses

              %O    CPU load of a node

              %p    Partition scheduling tier priority.

              %P    Partition name followed by "*" for the default partition, also see %R

              %r    Only user root may initiate jobs, "yes" or "no"

              %R    Partition name, also see %P

              %s    Maximum job size in nodes

              %S    Allowed allocating nodes

              %t    State of nodes, compact form

              %T    State of nodes, extended form

              %u    Print the user name of who set the reason a node is unavailable.

              %U    Print the user name and uid of who set the reason a node is unavailable.

              %v    Print the version of the running slurmd daemon.

              %V    Print the cluster name if running in a federation

              %w    Scheduling weight of the nodes

              %X    Number of sockets per node

              %Y    Number of cores per socket

              %Z    Number of threads per core

              %z    Extended processor information: number of sockets, cores, threads (S:C:T) per node

              %.<*> right justification of the field

              %<Number><*>
                    size of field

       -O <output_format>, --Format=<output_format>
              Specify    the    information    to    be   displayed.    Also   see   the   -o   <output_format>,
              --format=<output_format> option described below (which supports greater flexibility in formatting,
              but  does  not  support  access  to  all  fields because we ran out of letters).  Requests a comma
              separated list of job information to be displayed.

              The format of each field is "type[:[.]size]"

              size    is the minimum field size.  If no size is specified, 20 characters will  be  allocated  to
                      print the information.

               .      indicates  the  output  should be right justified and size must be specified.  By default,
                      output is left justified.

              Valid type specifications include:

              all   Print all fields available in the  -o  format  for  this  data  type  with  a  vertical  bar
                    separating each field.

              allocmem
                    Prints the amount of allocated memory on a node.

              allocnodes
                    Allowed allocating nodes.

              available
                    State/availability of a partition.

              cluster
                    Print the cluster name if running in a federation

              cpus  Number of CPUs per node.

              cpusload
                    CPU load of a node.

              freemem
                    Free memory of a node.

              cpusstate
                    Number  of  CPUs by state in the format "allocated/idle/other/total". Do not use this with a
                    node state option ("%t" or "%T") or the different node states will  be  placed  on  separate
                    lines.

              cores Number of cores per socket.

              defaulttime
                    Default time for any job in the format "days-hours:minutes:seconds".

              disk  Size of temporary disk space per node in megabytes.

              features
                    Features available on the nodes. Also see features_act.

              features_act
                    Features currently active on the nodes. Also see features.

              groups
                    Groups which may use the nodes.

              gres  Generic resources (gres) associated with the nodes.

              gresused
                    Generic resources (gres) currently in use on the nodes.

              maxcpuspernode
                    The max number of CPUs per node available to jobs in the partition.

              memory
                    Size of memory per node in megabytes.

              nodes Number of nodes.

              nodeaddr
                    List of node communication addresses.

              nodeai
                    Number  of nodes by state in the format "allocated/idle".  Do not use this with a node state
                    option ("%t" or "%T") or the different node states will be placed on separate lines.

              nodeaiot
                    Number of nodes by state in the format "allocated/idle/other/total".  Do not use this with a
                    node  state  option  ("%t"  or "%T") or the different node states will be placed on separate
                    lines.

              nodehost
                    List of node hostnames.

              nodelist
                    List of node names.

              oversubscribe
                    Jobs may oversubscribe compute resources (i.e. CPUs), "yes", "no", "exclusive" or "force".

              partition
                    Partition name followed by "*" for the default partition, also see %R.

              partitionname
                    Partition name, also see %P.

              port  Node TCP port.

              preemptmode
                    PreemptionMode.

              priorityjobfactor
                    Partition factor used by priority/multifactor plugin in calculating job priority.

              prioritytier or priority
                    Partition scheduling tier priority.

              reason
                    The reason a node is unavailable (down, drained, or draining states).

              root  Only user root may initiate jobs, "yes" or "no".

              size  Maximum job size in nodes.

              statecompact
                    State of nodes, compact form.

              statelong
                    State of nodes, extended form.

              sockets
                    Number of sockets per node.

              socketcorethread
                    Extended processor information: number of sockets, cores, threads (S:C:T) per node.

              time  Maximum time for any job in the format "days-hours:minutes:seconds".

              timestamp
                    Print the timestamp of the reason a node is unavailable.

              threads
                    Number of threads per core.

              user  Print the user name of who set the reason a node is unavailable.

              userlong
                    Print the user name and uid of who set the reason a node is unavailable.

              version
                    Print the version of the running slurmd daemon.

              weight
                    Scheduling weight of the nodes.

       -p <partition>, --partition=<partition>
              Print information only about the specified partition(s).  Multiple  partitions  are  separated  by
              commas.

       -r, --responding
              If set only report state information for responding nodes.

       -R, --list-reasons
              List  reasons  nodes  are  in  the  down, drained, fail or failing state.  When nodes are in these
              states Slurm supports optional inclusion of a "reason" string by an  administrator.   This  option
              will  display  the  first 20 characters of the reason field and list of nodes with that reason for
              all nodes that are, by default, down, drained, draining or failing.  This option may be used  with
              other  node  filtering  options (e.g. -r, -d, -t, -n), however, combinations of these options that
              result in a list of nodes that are not down or drained or failing will  not  produce  any  output.
              When used with -l the output additionally includes the current node state.

       -s, --summarize
              List  only  a partition state summary with no node state details.  This is ignored if the --format
              option is specified.

       -S <sort_list>, --sort=<sort_list>
              Specification of the order in which  records  should  be  reported.   This  uses  the  same  field
              specification  as  the  <output_format>.  Multiple sorts may be performed by listing multiple sort
              fields separated by commas.  The field specifications may be preceded by "+" or "-" for  ascending
              (default)  and  descending  order  respectively.   The  partition field specification, "P", may be
              preceded by  a  "#"  to  report  partitions  in  the  same  order  that  they  appear  in  Slurm's
              configuration  file,  slurm.conf.   For  example, a sort value of "+P,-m" requests that records be
              printed in order of increasing partition name and within a partition by  decreasing  memory  size.
              The  default  value  of  sort  is  "#P,-t"  (partitions ordered as configured then decreasing node
              state).  If the --Node option is selected, the default sort value is "N" (increasing node name).

       -t <states> , --states=<states>
              List nodes only having the given state(s).   Multiple  states  may  be  comma  separated  and  the
              comparison  is  case  insensitive.   Possible values include (case insensitive): ALLOC, ALLOCATED,
              COMP, COMPLETING, DOWN, DRAIN (for node in DRAINING or DRAINED states), DRAINED,  DRAINING,  FAIL,
              FUTURE,  FUTR,  IDLE,  MAINT,  MIX,  MIXED,  NO_RESPOND, NPC, PERFCTRS, POWER_DOWN, POWERING_DOWN,
              POWER_UP, RESV, RESERVED, UNK, and UNKNOWN.  By default nodes in the specified state are  reported
              whether  they are responding or not.  The --dead and --responding options may be used to filtering
              nodes by the responding flag.

       -T, --reservation
              Only display information about Slurm reservations.

       --usage
              Print a brief message listing the sinfo options.

       -v, --verbose
              Provide detailed event logging through program execution.

       -V, --version
              Print version information and exit.

OUTPUT FIELD DESCRIPTIONS

       AVAIL  Partition state: up or down.

       CPUS   Count of CPUs (processors) on each node.

       S:C:T  Count of sockets (S), cores (C), and threads (T) on these nodes.

       SOCKETS
              Count of sockets on these nodes.

       CORES  Count of cores on these nodes.

       THREADS
              Count of threads on these nodes.

       GROUPS Resource allocations in this partition are restricted to the named groups.  all indicates that all
              groups may use this partition.

       JOB_SIZE
              Minimum  and  maximum node count that can be allocated to any user job.  A single number indicates
              the minimum and maximum node count are the same.  infinite is used to identify partitions  without
              a maximum node count.

       TIMELIMIT
              Maximum  time  limit for any user job in days-hours:minutes:seconds.  infinite is used to identify
              partitions without a job time limit.

       MEMORY Size of real memory in megabytes on these nodes.

       NODELIST
              Names of nodes associated with this configuration/partition.

       NODES  Count of nodes with this particular configuration.

       NODES(A/I)
              Count of nodes with this particular configuration by node state in the form "available/idle".

       NODES(A/I/O/T)
              Count  of   nodes   with   this   particular   configuration   by   node   state   in   the   form
              "available/idle/other/total".

       PARTITION
              Name of a partition.  Note that the suffix "*" identifies the default partition.

       PORT   Local TCP port used by slurmd on the node.

       ROOT   Is the ability to allocate resources in this partition restricted to user root, yes or no.

       OVERSUBSCRIBE
              Will jobs allocated resources in this partition oversubscribe those compute resources (i.e. CPUs).
              no indicates resources are never oversubscribed.  exclusive indicates whole nodes are dedicated to
              jobs  (equivalent  to  srun  --exclusive  option,  may  be used even with select/cons_res managing
              individual processors).  force indicates resources are always available to be oversubscribed.  yes
              indicates resource may be oversubscribed or not per job's resource allocation.

       STATE  State  of  the  nodes.   Possible  states include: allocated, completing, down, drained, draining,
              fail, failing, future, idle, maint, mixed, perfctrs, power_down, power_up, reserved,  and  unknown
              plus Their abbreviated forms: alloc, comp, down, drain, drng, fail, failg, futr, idle, maint, mix,
              npc, pow_dn, pow_up, resv, and unk respectively.  Note that the suffix "*" identifies  nodes  that
              are presently not responding.

       TMP_DISK
              Size of temporary disk space in megabytes on these nodes.

NODE STATE CODES

       Node  state  codes  are shortened as required for the field size.  These node states may be followed by a
       special character to identify state flags associated with the node.  The  following  node  sufficies  and
       states are used:

       *   The  node  is  presently  not responding and will not be allocated any new work.  If the node remains
           non-responsive, it will be placed in the DOWN state (except  in  the  case  of  COMPLETING,  DRAINED,
           DRAINING, FAIL, FAILING nodes).

       ~   The node is presently in a power saving mode (typically running at reduced frequency).

       #   The node is presently being powered up or configured.

       %   The node is presently being powered down.

       $   The node is currently in a reservation with a flag value of "maintenance".

       @   The node is pending reboot.

       ALLOCATED   The node has been allocated to one or more jobs.

       ALLOCATED+  The  node is allocated to one or more active jobs plus one or more jobs are in the process of
                   COMPLETING.

       COMPLETING  All jobs associated with this node are in the process of COMPLETING.  This node state will be
                   removed when all of the job's processes have terminated and the Slurm epilog program (if any)
                   has terminated. See the Epilog parameter description in the  slurm.conf  man  page  for  more
                   information.

       DOWN        The  node  is  unavailable for use. Slurm can automatically place nodes in this state if some
                   failure occurs. System administrators may also explicitly place nodes in  this  state.  If  a
                   node  resumes  normal  operation,  Slurm  can  automatically  return  it  to service. See the
                   ReturnToService and SlurmdTimeout parameter descriptions in the slurm.conf(5)  man  page  for
                   more information.

       DRAINED     The  node  is  unavailable  for  use  per  system administrator request.  See the update node
                   command in the scontrol(1) man page or the slurm.conf(5) man page for more information.

       DRAINING    The node is currently executing a job, but will not be allocated to additional jobs. The node
                   state  will  be  changed to state DRAINED when the last job on it completes. Nodes enter this
                   state per system administrator request. See the update node command in  the  scontrol(1)  man
                   page or the slurm.conf(5) man page for more information.

       FAIL        The  node  is  expected  to  fail  soon  and  is unavailable for use per system administrator
                   request.  See the update node command in the scontrol(1) man page or  the  slurm.conf(5)  man
                   page for more information.

       FAILING     The  node  is  currently executing a job, but is expected to fail soon and is unavailable for
                   use per system administrator request.  See the update node command  in  the  scontrol(1)  man
                   page or the slurm.conf(5) man page for more information.

       FUTURE      The node is currently not fully configured, but expected to be available at some point in the
                   indefinite future for use.

       IDLE        The node is not allocated to any jobs and is available for use.

       MAINT       The node is currently in a reservation with a flag value of "maintainence".

       REBOOT      The node is currently scheduled to be rebooted.

       MIXED       The node has some of its CPUs ALLOCATED while others are IDLE.

       PERFCTRS (NPC)
                   Network Performance Counters associated with this node are in use, rendering this node as not
                   usable for any other jobs

       POWER_DOWN  The node is currently powered down and not capable of running any jobs.

       POWERING_DOWN
                   The node is currently powering down and not capable of running any jobs.

       POWER_UP    The node is currently in the process of being powered up.

       RESERVED    The node is in an advanced reservation and not generally available.

       UNKNOWN     The Slurm controller has just started and the node's state has not yet been determined.

ENVIRONMENT VARIABLES

       Some  sinfo  options  may be set via environment variables. These environment variables, along with their
       corresponding options, are listed below. (Note: Commandline options will always override these settings.)

       SINFO_ALL           -a, --all

       SINFO_FEDERATION    Same as --federation

       SINFO_FORMAT        -o <output_format>, --format=<output_format>

       SINFO_LOCAL         Same as --local

       SINFO_PARTITION     -p <partition>, --partition=<partition>

       SINFO_SORT          -S <sort>, --sort=<sort>

       SLURM_CLUSTERS      Same as --clusters

       SLURM_CONF          The location of the Slurm configuration file.

       SLURM_TIME_FORMAT   Specify the format used to report time stamps.  A  value  of  standard,  the  default
                           value, generates output in the form "year-month-dateThour:minute:second".  A value of
                           relative returns only "hour:minute:second" if the current day.  For  other  dates  in
                           the  current  year  it  prints  the  "hour:minute"  preceded  by "Tomorr" (tomorrow),
                           "Ystday" (yesterday), the name of the day for the coming  week  (e.g.  "Mon",  "Tue",
                           etc.),  otherwise  the date (e.g. "25 Apr").  For other years it returns a date month
                           and year without a time (e.g.  "6 Jun 2012"). All of the time stamps use  a  24  hour
                           format.

                           A valid strftime() format can also be specified. For example, a value of "%a %T" will
                           report the day of the week and a time stamp (e.g. "Mon 12:34:56").

EXAMPLES

       Report basic node and partition configurations:

       > sinfo
       PARTITION AVAIL TIMELIMIT NODES STATE  NODELIST
       batch     up     infinite     2 alloc  adev[8-9]
       batch     up     infinite     6 idle   adev[10-15]
       debug*    up        30:00     8 idle   adev[0-7]

       Report partition summary information:

       > sinfo -s
       PARTITION AVAIL TIMELIMIT NODES(A/I/O/T) NODELIST
       batch     up     infinite 2/6/0/8        adev[8-15]
       debug*    up        30:00 0/8/0/8        adev[0-7]

       Report more complete information about the partition debug:

       > sinfo --long --partition=debug
       PARTITION AVAIL TIMELIMIT JOB_SIZE ROOT OVERSUBS GROUPS NODES STATE NODELIST
       debug*    up        30:00        8 no   no       all        8 idle  dev[0-7]

       Report only those nodes that are in state DRAINED:

       > sinfo --states=drained
       PARTITION AVAIL NODES TIMELIMIT STATE  NODELIST
       debug*    up        2     30:00 drain  adev[6-7]

       Report node-oriented information with details and exact matches:

       > sinfo -Nel
       NODELIST    NODES PARTITION STATE  CPUS MEMORY TMP_DISK WEIGHT FEATURES REASON
       adev[0-1]       2 debug*    idle      2   3448    38536     16 (null)   (null)
       adev[2,4-7]     5 debug*    idle      2   3384    38536     16 (null)   (null)
       adev3           1 debug*    idle      2   3394    38536     16 (null)   (null)
       adev[8-9]       2 batch     allocated 2    246    82306     16 (null)   (null)
       adev[10-15]     6 batch     idle      2    246    82306     16 (null)   (null)

       Report only down, drained and draining nodes and their reason field:

       > sinfo -R
       REASON                              NODELIST
       Memory errors                       dev[0,5]
       Not Responding                      dev8

COPYING

       Copyright (C) 2002-2007 The Regents of the University of  California.   Produced  at  Lawrence  Livermore
       National Laboratory (cf, DISCLAIMER).
       Copyright (C) 2008-2009 Lawrence Livermore National Security.
       Copyright (C) 2010-2017 SchedMD LLC.

       This    file    is    part    of    Slurm,   a   resource   management   program.    For   details,   see
       <https://slurm.schedmd.com/>.

       Slurm is free software; you can redistribute it and/or modify it under  the  terms  of  the  GNU  General
       Public License as published by the Free Software Foundation; either version 2 of the License, or (at your
       option) any later version.

       Slurm is distributed in the hope that it will be useful, but  WITHOUT  ANY  WARRANTY;  without  even  the
       implied  warranty  of  MERCHANTABILITY  or  FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public
       License for more details.

SEE ALSO

       scontrol(1), smap(1), squeue(1),  slurm_load_ctl_conf  (3),  slurm_load_jobs  (3),  slurm_load_node  (3),
       slurm_load_partitions   (3),   slurm_reconfigure   (3),   slurm_shutdown   (3),   slurm_update_job   (3),
       slurm_update_node (3), slurm_update_partition (3), slurm.conf(5)