Provided by: slurm-client_17.11.2-1build1_amd64 

NAME
sinfo - view information about Slurm nodes and partitions.
SYNOPSIS
sinfo [OPTIONS...]
DESCRIPTION
sinfo is used to view partition and node information for a system running Slurm.
OPTIONS
-a, --all
Display information about all partitions. This causes information to be displayed about partitions
that are configured as hidden and partitions that are unavailable to user's group.
-b, --bgl
Display information about bglblocks (on Blue Gene systems only).
-d, --dead
If set only report state information for non-responding (dead) nodes.
-e, --exact
If set, do not group node information on multiple nodes unless their configurations to be reported
are identical. Otherwise cpu count, memory size, and disk space for nodes will be listed with the
minimum value followed by a "+" for nodes with the same partition and state (e.g., "250+").
--federation
Show all partitions from the federation if a member of one.
-h, --noheader
Do not print a header on the output.
--help Print a message describing all sinfo options.
--hide Do not display information about hidden partitions. By default, partitions that are configured as
hidden or are not available to the user's group will not be displayed (i.e. this is the default
behavior).
-i <seconds>, --iterate=<seconds>
Print the state on a periodic basis. Sleep for the indicated number of seconds between reports.
By default, prints a time stamp with the header.
--local
Show only jobs local to this cluster. Ignore other clusters in this federation (if any). Overrides
--federation.
-l, --long
Print more detailed information. This is ignored if the --format option is specified.
-M, --clusters=<string>
Clusters to issue commands to. Multiple cluster names may be comma separated. A value of of
'all' will query to run on all clusters. Note that the SlurmDBD must be up for this option to
work properly. This option implicitly sets the --local option.
-n <nodes>, --nodes=<nodes>
Print information only about the specified node(s). Multiple nodes may be comma separated or
expressed using a node range expression. For example "linux[00-07]" would indicate eight nodes,
"linux00" through "linux07." Performance of the command can be measurably improved for systems
with large numbers of nodes when a single node name is specified.
--noconvert
Don't convert units from their original type (e.g. 2048M won't be converted to 2G).
-N, --Node
Print information in a node-oriented format with one line per node. The default is to print
information in a partition-oriented format. This is ignored if the --format option is specified.
-o <output_format>, --format=<output_format>
Specify the information to be displayed using an sinfo format string. Format strings transparently
used by sinfo when running with various options are
default "%#P %.5a %.10l %.6D %.6t %N"
--summarize "%#P %.5a %.10l %.16F %N"
--long "%#P %.5a %.10l %.10s %.4r %.8h %.10g %.6D %.11T %N"
--Node "%#N %.6D %#P %6t"
--long --Node "%#N %.6D %#P %.11T %.4c %.8z %.6m %.8d %.6w %.8f %20E"
--list-reasons "%20E %9u %19H %N"
--long --list-reasons
"%20E %12U %19H %6t %N"
In the above format strings, the use of "#" represents the maximum length of any partition name or
node list to be printed. A pass is made over the records to be printed to establish the size in
order to align the sinfo output, then a second pass is made over the records to print them. Note
that the literal character "#" itself is not a valid field length specification, but is only used
to document this behaviour.
The field specifications available include:
%all Print all fields available for this data type with a vertical bar separating each field.
%a State/availability of a partition
%A Number of nodes by state in the format "allocated/idle". Do not use this with a node state
option ("%t" or "%T") or the different node states will be placed on separate lines.
%b Features currently active on the nodes, also see %f
%B The max number of CPUs per node available to jobs in the partition.
%c Number of CPUs per node
%C Number of CPUs by state in the format "allocated/idle/other/total". Do not use this with a
node state option ("%t" or "%T") or the different node states will be placed on separate
lines.
%d Size of temporary disk space per node in megabytes
%D Number of nodes
%e Free memory of a node
%E The reason a node is unavailable (down, drained, or draining states).
%f Features available the nodes, also see %b
%F Number of nodes by state in the format "allocated/idle/other/total". Note the use of this
format option with a node state format option ("%t" or "%T") will result in the different
node states being be reported on separate lines.
%g Groups which may use the nodes
%G Generic resources (gres) associated with the nodes
%h Jobs may oversubscribe compute resources (i.e. CPUs), "yes", "no", "exclusive" or "force"
%H Print the timestamp of the reason a node is unavailable.
%I Partition job priority weighting factor.
%l Maximum time for any job in the format "days-hours:minutes:seconds"
%L Default time for any job in the format "days-hours:minutes:seconds"
%m Size of memory per node in megabytes
%M PreemptionMode
%n List of node hostnames
%N List of node names
%o List of node communication addresses
%O CPU load of a node
%p Partition scheduling tier priority.
%P Partition name followed by "*" for the default partition, also see %R
%r Only user root may initiate jobs, "yes" or "no"
%R Partition name, also see %P
%s Maximum job size in nodes
%S Allowed allocating nodes
%t State of nodes, compact form
%T State of nodes, extended form
%u Print the user name of who set the reason a node is unavailable.
%U Print the user name and uid of who set the reason a node is unavailable.
%v Print the version of the running slurmd daemon.
%V Print the cluster name if running in a federation
%w Scheduling weight of the nodes
%X Number of sockets per node
%Y Number of cores per socket
%Z Number of threads per core
%z Extended processor information: number of sockets, cores, threads (S:C:T) per node
%.<*> right justification of the field
%<Number><*>
size of field
-O <output_format>, --Format=<output_format>
Specify the information to be displayed. Also see the -o <output_format>,
--format=<output_format> option described below (which supports greater flexibility in formatting,
but does not support access to all fields because we ran out of letters). Requests a comma
separated list of job information to be displayed.
The format of each field is "type[:[.]size]"
size is the minimum field size. If no size is specified, 20 characters will be allocated to
print the information.
. indicates the output should be right justified and size must be specified. By default,
output is left justified.
Valid type specifications include:
all Print all fields available in the -o format for this data type with a vertical bar
separating each field.
allocmem
Prints the amount of allocated memory on a node.
allocnodes
Allowed allocating nodes.
available
State/availability of a partition.
cluster
Print the cluster name if running in a federation
cpus Number of CPUs per node.
cpusload
CPU load of a node.
freemem
Free memory of a node.
cpusstate
Number of CPUs by state in the format "allocated/idle/other/total". Do not use this with a
node state option ("%t" or "%T") or the different node states will be placed on separate
lines.
cores Number of cores per socket.
defaulttime
Default time for any job in the format "days-hours:minutes:seconds".
disk Size of temporary disk space per node in megabytes.
features
Features available on the nodes. Also see features_act.
features_act
Features currently active on the nodes. Also see features.
groups
Groups which may use the nodes.
gres Generic resources (gres) associated with the nodes.
maxcpuspernode
The max number of CPUs per node available to jobs in the partition.
memory
Size of memory per node in megabytes.
nodes Number of nodes.
nodeaddr
List of node communication addresses.
nodeai
Number of nodes by state in the format "allocated/idle". Do not use this with a node state
option ("%t" or "%T") or the different node states will be placed on separate lines.
nodeaiot
Number of nodes by state in the format "allocated/idle/other/total". Do not use this with a
node state option ("%t" or "%T") or the different node states will be placed on separate
lines.
nodehost
List of node hostnames.
nodelist
List of node names.
oversubscribe
Jobs may oversubscribe compute resources (i.e. CPUs), "yes", "no", "exclusive" or "force".
partition
Partition name followed by "*" for the default partition, also see %R.
partitionname
Partition name, also see %P.
port Node TCP port.
preemptmode
PreemptionMode.
priorityjobfactor
Partition factor used by priority/multifactor plugin in calculating job priority.
prioritytier or priority
Partition scheduling tier priority.
reason
The reason a node is unavailable (down, drained, or draining states).
root Only user root may initiate jobs, "yes" or "no".
size Maximum job size in nodes.
statecompact
State of nodes, compact form.
statelong
State of nodes, extended form.
sockets
Number of sockets per node.
socketcorethread
Extended processor information: number of sockets, cores, threads (S:C:T) per node.
time Maximum time for any job in the format "days-hours:minutes:seconds".
timestamp
Print the timestamp of the reason a node is unavailable.
threads
Number of threads per core.
user Print the user name of who set the reason a node is unavailable.
userlong
Print the user name and uid of who set the reason a node is unavailable.
version
Print the version of the running slurmd daemon.
weight
Scheduling weight of the nodes.
-p <partition>, --partition=<partition>
Print information only about the specified partition(s). Multiple partitions are separated by
commas.
-r, --responding
If set only report state information for responding nodes.
-R, --list-reasons
List reasons nodes are in the down, drained, fail or failing state. When nodes are in these
states Slurm supports optional inclusion of a "reason" string by an administrator. This option
will display the first 35 characters of the reason field and list of nodes with that reason for
all nodes that are, by default, down, drained, draining or failing. This option may be used with
other node filtering options (e.g. -r, -d, -t, -n), however, combinations of these options that
result in a list of nodes that are not down or drained or failing will not produce any output.
When used with -l the output additionally includes the current node state.
-s, --summarize
List only a partition state summary with no node state details. This is ignored if the --format
option is specified.
-S <sort_list>, --sort=<sort_list>
Specification of the order in which records should be reported. This uses the same field
specification as the <output_format>. Multiple sorts may be performed by listing multiple sort
fields separated by commas. The field specifications may be preceded by "+" or "-" for ascending
(default) and descending order respectively. The partition field specification, "P", may be
preceded by a "#" to report partitions in the same order that they appear in Slurm's
configuration file, slurm.conf. For example, a sort value of "+P,-m" requests that records be
printed in order of increasing partition name and within a partition by decreasing memory size.
The default value of sort is "#P,-t" (partitions ordered as configured then decreasing node
state). If the --Node option is selected, the default sort value is "N" (increasing node name).
-t <states> , --states=<states>
List nodes only having the given state(s). Multiple states may be comma separated and the
comparison is case insensitive. Possible values include (case insensitive): ALLOC, ALLOCATED,
COMP, COMPLETING, DOWN, DRAIN (for node in DRAINING or DRAINED states), DRAINED, DRAINING, ERR,
ERROR, FAIL, FUTURE, FUTR, IDLE, MAINT, MIX, MIXED, NO_RESPOND, NPC, PERFCTRS, POWER_DOWN,
POWER_UP, RESV, RESERVED, UNK, and UNKNOWN. By default nodes in the specified state are reported
whether they are responding or not. The --dead and --responding options may be used to filtering
nodes by the responding flag.
-T, --reservation
Only display information about Slurm reservations.
--usage
Print a brief message listing the sinfo options.
-v, --verbose
Provide detailed event logging through program execution.
-V, --version
Print version information and exit.
OUTPUT FIELD DESCRIPTIONS
AVAIL Partition state: up or down.
CPUS Count of CPUs (processors) on each node.
S:C:T Count of sockets (S), cores (C), and threads (T) on these nodes.
SOCKETS
Count of sockets on these nodes.
CORES Count of cores on these nodes.
THREADS
Count of threads on these nodes.
GROUPS Resource allocations in this partition are restricted to the named groups. all indicates that all
groups may use this partition.
JOB_SIZE
Minimum and maximum node count that can be allocated to any user job. A single number indicates
the minimum and maximum node count are the same. infinite is used to identify partitions without
a maximum node count.
TIMELIMIT
Maximum time limit for any user job in days-hours:minutes:seconds. infinite is used to identify
partitions without a job time limit.
MEMORY Size of real memory in megabytes on these nodes.
NODELIST or BP_LIST (BlueGene systems only)
Names of nodes associated with this configuration/partition.
NODES Count of nodes with this particular configuration.
NODES(A/I)
Count of nodes with this particular configuration by node state in the form "available/idle".
NODES(A/I/O/T)
Count of nodes with this particular configuration by node state in the form
"available/idle/other/total".
PARTITION
Name of a partition. Note that the suffix "*" identifies the default partition.
PORT Local TCP port used by slurmd on the node.
ROOT Is the ability to allocate resources in this partition restricted to user root, yes or no.
OVERSUBSCRIBE
Will jobs allocated resources in this partition oversubscribe those compute resources (i.e. CPUs).
no indicates resources are never oversubscribed. exclusive indicates whole nodes are dedicated to
jobs (equivalent to srun --exclusive option, may be used even with select/cons_res managing
individual processors). force indicates resources are always available to be oversubscribed. yes
indicates resource may be oversubscribed or not per job's resource allocation.
STATE State of the nodes. Possible states include: allocated, completing, down, drained, draining,
fail, failing, future, idle, maint, mixed, perfctrs, power_down, power_up, reserved, and unknown
plus Their abbreviated forms: alloc, comp, down, drain, drng, fail, failg, futr, idle, maint, mix,
npc, pow_dn, pow_up, resv, and unk respectively. Note that the suffix "*" identifies nodes that
are presently not responding.
TMP_DISK
Size of temporary disk space in megabytes on these nodes.
NODE STATE CODES
Node state codes are shortened as required for the field size. These node states may be followed by a
special character to identify state flags associated with the node. The following node sufficies and
states are used:
* The node is presently not responding and will not be allocated any new work. If the node remains
non-responsive, it will be placed in the DOWN state (except in the case of COMPLETING, DRAINED,
DRAINING, FAIL, FAILING nodes).
~ The node is presently in a power saving mode (typically running at reduced frequency).
# The node is presently being powered up or configured.
$ The node is currently in a reservation with a flag value of "maintenance".
@ The node is pending reboot.
ALLOCATED The node has been allocated to one or more jobs.
ALLOCATED+ The node is allocated to one or more active jobs plus one or more jobs are in the process of
COMPLETING.
COMPLETING All jobs associated with this node are in the process of COMPLETING. This node state will be
removed when all of the job's processes have terminated and the Slurm epilog program (if any)
has terminated. See the Epilog parameter description in the slurm.conf man page for more
information.
DOWN The node is unavailable for use. Slurm can automatically place nodes in this state if some
failure occurs. System administrators may also explicitly place nodes in this state. If a
node resumes normal operation, Slurm can automatically return it to service. See the
ReturnToService and SlurmdTimeout parameter descriptions in the slurm.conf(5) man page for
more information.
DRAINED The node is unavailable for use per system administrator request. See the update node
command in the scontrol(1) man page or the slurm.conf(5) man page for more information.
DRAINING The node is currently executing a job, but will not be allocated to additional jobs. The node
state will be changed to state DRAINED when the last job on it completes. Nodes enter this
state per system administrator request. See the update node command in the scontrol(1) man
page or the slurm.conf(5) man page for more information.
ERROR The node is currently in an error state and not capable of running any jobs. Slurm can
automatically place nodes in this state if some failure occurs. System administrators may
also explicitly place nodes in this state. If a node resumes normal operation, Slurm can
automatically return it to service. See the ReturnToService and SlurmdTimeout parameter
descriptions in the slurm.conf(5) man page for more information.
FAIL The node is expected to fail soon and is unavailable for use per system administrator
request. See the update node command in the scontrol(1) man page or the slurm.conf(5) man
page for more information.
FAILING The node is currently executing a job, but is expected to fail soon and is unavailable for
use per system administrator request. See the update node command in the scontrol(1) man
page or the slurm.conf(5) man page for more information.
FUTURE The node is currently not fully configured, but expected to be available at some point in the
indefinite future for use.
IDLE The node is not allocated to any jobs and is available for use.
MAINT The node is currently in a reservation with a flag value of "maintainence".
REBOOT The node is currently scheduled to be rebooted.
MIXED The node has some of its CPUs ALLOCATED while others are IDLE.
PERFCTRS (NPC)
Network Performance Counters associated with this node are in use, rendering this node as not
usable for any other jobs
POWER_DOWN The node is currently powered down and not capable of running any jobs.
POWER_UP The node is currently in the process of being powered up.
RESERVED The node is in an advanced reservation and not generally available.
UNKNOWN The Slurm controller has just started and the node's state has not yet been determined.
ENVIRONMENT VARIABLES
Some sinfo options may be set via environment variables. These environment variables, along with their
corresponding options, are listed below. (Note: Commandline options will always override these settings.)
SINFO_ALL -a, --all
SINFO_FEDERATION Same as --federation
SINFO_FORMAT -o <output_format>, --format=<output_format>
SINFO_LOCAL Same as --local
SINFO_PARTITION -p <partition>, --partition=<partition>
SINFO_SORT -S <sort>, --sort=<sort>
SLURM_CLUSTERS Same as --clusters
SLURM_CONF The location of the Slurm configuration file.
SLURM_TIME_FORMAT Specify the format used to report time stamps. A value of standard, the default
value, generates output in the form "year-month-dateThour:minute:second". A value of
relative returns only "hour:minute:second" if the current day. For other dates in
the current year it prints the "hour:minute" preceded by "Tomorr" (tomorrow),
"Ystday" (yesterday), the name of the day for the coming week (e.g. "Mon", "Tue",
etc.), otherwise the date (e.g. "25 Apr"). For other years it returns a date month
and year without a time (e.g. "6 Jun 2012"). All of the time stamps use a 24 hour
format.
A valid strftime() format can also be specified. For example, a value of "%a %T" will
report the day of the week and a time stamp (e.g. "Mon 12:34:56").
EXAMPLES
Report basic node and partition configurations:
> sinfo
PARTITION AVAIL TIMELIMIT NODES STATE NODELIST
batch up infinite 2 alloc adev[8-9]
batch up infinite 6 idle adev[10-15]
debug* up 30:00 8 idle adev[0-7]
Report partition summary information:
> sinfo -s
PARTITION AVAIL TIMELIMIT NODES(A/I/O/T) NODELIST
batch up infinite 2/6/0/8 adev[8-15]
debug* up 30:00 0/8/0/8 adev[0-7]
Report more complete information about the partition debug:
> sinfo --long --partition=debug
PARTITION AVAIL TIMELIMIT JOB_SIZE ROOT OVERSUBS GROUPS NODES STATE NODELIST
debug* up 30:00 8 no no all 8 idle dev[0-7]
Report only those nodes that are in state DRAINED:
> sinfo --states=drained
PARTITION AVAIL NODES TIMELIMIT STATE NODELIST
debug* up 2 30:00 drain adev[6-7]
Report node-oriented information with details and exact matches:
> sinfo -Nel
NODELIST NODES PARTITION STATE CPUS MEMORY TMP_DISK WEIGHT FEATURES REASON
adev[0-1] 2 debug* idle 2 3448 38536 16 (null) (null)
adev[2,4-7] 5 debug* idle 2 3384 38536 16 (null) (null)
adev3 1 debug* idle 2 3394 38536 16 (null) (null)
adev[8-9] 2 batch allocated 2 246 82306 16 (null) (null)
adev[10-15] 6 batch idle 2 246 82306 16 (null) (null)
Report only down, drained and draining nodes and their reason field:
> sinfo -R
REASON NODELIST
Memory errors dev[0,5]
Not Responding dev8
COPYING
Copyright (C) 2002-2007 The Regents of the University of California. Produced at Lawrence Livermore
National Laboratory (cf, DISCLAIMER).
Copyright (C) 2008-2009 Lawrence Livermore National Security.
Copyright (C) 2010-2017 SchedMD LLC.
This file is part of Slurm, a resource management program. For details, see
<https://slurm.schedmd.com/>.
Slurm is free software; you can redistribute it and/or modify it under the terms of the GNU General
Public License as published by the Free Software Foundation; either version 2 of the License, or (at your
option) any later version.
Slurm is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the
implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public
License for more details.
SEE ALSO
scontrol(1), smap(1), squeue(1), slurm_load_ctl_conf (3), slurm_load_jobs (3), slurm_load_node (3),
slurm_load_partitions (3), slurm_reconfigure (3), slurm_shutdown (3), slurm_update_job (3),
slurm_update_node (3), slurm_update_partition (3), slurm.conf(5)
November 2016 Slurm Commands sinfo(1)