Provided by: collectd-core_5.4.0-3ubuntu2.2_amd64 bug

NAME

       collectd.conf - Configuration for the system statistics collection daemon collectd

SYNOPSIS

         BaseDir "/path/to/data/"
         PIDFile "/path/to/pidfile/collectd.pid"
         Server  "123.123.123.123" 12345

         LoadPlugin cpu
         LoadPlugin load

         <LoadPlugin df>
           Interval 3600
         </LoadPlugin>

         LoadPlugin ping
         <Plugin ping>
           Host "example.org"
           Host "provider.net"
         </Plugin>

DESCRIPTION

       This config file controls how the system statistics collection daemon collectd behaves. The most
       significant option is LoadPlugin, which controls which plugins to load. These plugins ultimately define
       collectd's behavior.

       The syntax of this config file is similar to the config file of the famous Apache webserver. Each line
       contains either an option (a key and a list of one or more values) or a section-start or -end. Empty
       lines and everything after a non-quoted hash-symbol ("#") is ignored. Keys are unquoted strings,
       consisting only of alphanumeric characters and the underscore ("_") character. Keys are handled case
       insensitive by collectd itself and all plugins included with it. Values can either be an unquoted string,
       a quoted string (enclosed in double-quotes) a number or a boolean expression. Unquoted strings consist of
       only alphanumeric characters and underscores ("_") and do not need to be quoted. Quoted strings are
       enclosed in double quotes ("""). You can use the backslash character ("\") to include double quotes as
       part of the string. Numbers can be specified in decimal and floating point format (using a dot "." as
       decimal separator), hexadecimal when using the "0x" prefix and octal with a leading zero (0).  Boolean
       values are either true or false.

       Lines may be wrapped by using "\" as the last character before the newline.  This allows long lines to be
       split into multiple lines. Quoted strings may be wrapped as well. However, those are treated special in
       that whitespace at the beginning of the following lines will be ignored, which allows for nicely
       indenting the wrapped lines.

       The configuration is read and processed in order, i.e. from top to bottom. So the plugins are loaded in
       the order listed in this config file. It is a good idea to load any logging plugins first in order to
       catch messages from plugins during configuration. Also, the "LoadPlugin" option must occur before the
       appropriate "<Plugin ...>" block.

GLOBAL OPTIONS

       BaseDir Directory
           Sets  the  base  directory.  This  is  the directory beneath all RRD-files are created. Possibly more
           subdirectories are created. This is also the working directory for the daemon.

       LoadPlugin Plugin
           Loads the plugin Plugin. This is required to  load  plugins,  unless  the  AutoLoadPlugin  option  is
           enabled (see below). Without any loaded plugins, collectd will be mostly useless.

           Only  the  first LoadPlugin statement or block for a given plugin name has any effect. This is useful
           when you want to split up the configuration into smaller  files  and  want  each  file  to  be  "self
           contained",  i.e.  it contains a Plugin block and then appropriate LoadPlugin statement. The downside
           is that if you have  multiple  conflicting  LoadPlugin  blocks,  e.g.  when  they  specify  different
           intervals,  only  one  of  them  (the  first one encountered) will take effect and all others will be
           silently ignored.

           LoadPlugin may either be a simple  configuration  statement  or  a  block  with  additional  options,
           affecting the behavior of LoadPlugin. A simple statement looks like this:

            LoadPlugin "cpu"

           Options  inside  a  LoadPlugin  block can override default settings and influence the way plugins are
           loaded, e.g.:

            <LoadPlugin perl>
              Globals true
              Interval 60
            </LoadPlugin>

           The following options are valid inside LoadPlugin blocks:

           Globals true|false
               If enabled, collectd will export all global symbols of the plugin (and of all libraries loaded as
               dependencies of the plugin) and, thus, makes those symbols  available  for  resolving  unresolved
               symbols in subsequently loaded plugins if that is supported by your system.

               This  is  useful  (or  possibly  even  required),  e.g.,  when  loading a plugin that embeds some
               scripting language into the daemon (e.g.  the  Perl  and  Python  plugins).  Scripting  languages
               usually  provide means to load extensions written in C. Those extensions require symbols provided
               by the interpreter, which is loaded as a dependency of the respective collectd plugin.   See  the
               documentation of those plugins (e.g., collectd-perl(5) or collectd-python(5)) for details.

               By  default,  this  is  disabled.  As a special exception, if the plugin name is either "perl" or
               "python", the default is changed to enabled in order to keep the average user from ever having to
               deal with this low level linking stuff.

           Interval Seconds
               Sets a plugin-specific interval for  collecting  metrics.  This  overrides  the  global  Interval
               setting.  If  a  plugin  provides  own support for specifying an interval, that setting will take
               precedence.

       AutoLoadPlugin false|true
           When set to false (the default), each plugin needs to be  loaded  explicitly,  using  the  LoadPlugin
           statement  documented  above.  If  a  <Plugin ...> block is encountered and no configuration handling
           callback for this plugin has been registered, a warning is logged and the block is ignored.

           When set to true, explicit LoadPlugin statements are not required. Each <Plugin ...> block acts as if
           it was immediately preceded by a LoadPlugin statement. LoadPlugin statements are still  required  for
           plugins that don't provide any configuration, e.g. the Load plugin.

       Include Path [pattern]
           If Path points to a file, includes that file. If Path points to a directory, recursively includes all
           files  within  that  directory and its subdirectories. If the "wordexp" function is available on your
           system, shell-like wildcards are  expanded  before  files  are  included.  This  means  you  can  use
           statements like the following:

             Include "/etc/collectd.d/*.conf"

           Starting  with  version 5.3, this may also be a block in which further options affecting the behavior
           of Include may be specified. The following option is currently allowed:

             <Include "/etc/collectd.d">
               Filter "*.conf"
             </Include>

           Filter pattern
               If the "fnmatch" function is available on your system,  a  shell-like  wildcard  pattern  may  be
               specified  to  filter  which  files  to include. This may be used in combination with recursively
               including a directory to easily  be  able  to  arbitrarily  mix  configuration  files  and  other
               documents  (e.g.  README  files).   The  given  example is similar to the first example above but
               includes all files matching "*.conf" in any subdirectory of "/etc/collectd.d":

                 Include "/etc/collectd.d" "*.conf"

           If more than one files are included by a single  Include  option,  the  files  will  be  included  in
           lexicographical  order  (as  defined  by  the  "strcmp"  function).  Thus, you can e. g. use numbered
           prefixes to specify the order in which the files are loaded.

           To prevent loops and shooting yourself in the foot in interesting ways the nesting is  limited  to  a
           depth  of 8 levels, which should be sufficient for most uses. Since symlinks are followed it is still
           possible to crash the daemon by looping symlinks. In our opinion significant stupidity should  result
           in an appropriate amount of pain.

           It  is  no  problem to have a block like "<Plugin foo>" in more than one file, but you cannot include
           files from within blocks.

       PIDFile File
           Sets where to write the PID file to. This file is overwritten when it exists  and  deleted  when  the
           program is stopped. Some init-scripts might override this setting using the -P command-line option.

       PluginDir Directory
           Path to the plugins (shared objects) of collectd.

       TypesDB File [File ...]
           Set  one  or  more files that contain the data-set descriptions. See types.db(5) for a description of
           the format of this file.

       Interval Seconds
           Configures the interval in which to query the read plugins. Obviously smaller values lead to a higher
           system load produced by collectd, while higher values lead to more coarse statistics.

           Warning: You should set this once and then never touch it again. If you do, you will have  to  delete
           all  your  RRD  files  or  know  some  serious  RRDtool  magic! (Assuming you're using the RRDtool or
           RRDCacheD plugin.)

       Timeout Iterations
           Consider a value list "missing" when no update has been read or received for  Iterations  iterations.
           By  default,  collectd  considers a value list missing when no update has been received for twice the
           update interval. Since this setting uses iterations, the maximum allowed time without update  depends
           on the Interval information contained in each value list. This is used in the Threshold configuration
           to dispatch notifications about missing values, see collectd-threshold(5) for details.

       ReadThreads Num
           Number  of threads to start for reading plugins. The default value is 5, but you may want to increase
           this if you have more than five plugins that take a long time to read. Mostly those are plugins  that
           do  network-IO.  Setting  this  to a value higher than the number of registered read callbacks is not
           recommended.

       WriteThreads Num
           Number of threads to start for dispatching value lists to write plugins. The default value is 5,  but
           you  may  want  to  increase this if you have more than five plugins that may take relatively long to
           write to.

       WriteQueueLimitHigh HighNum
       WriteQueueLimitLow LowNum
           Metrics are read by the read threads and then put into a queue to be handled by the write threads. If
           one of the write plugins is slow (e.g. network timeouts, I/O saturation of the disk) this queue  will
           grow.  In  order  to  avoid running into memory issues in such a case, you can limit the size of this
           queue.

           By default, there is no limit and memory may grow indefinitely. This is most likely not an issue  for
           clients, i.e. instances that only handle the local metrics. For servers it is recommended to set this
           to a non-zero value, though.

           You  can  set  the  limits  using  WriteQueueLimitHigh  and WriteQueueLimitLow.  Each of them takes a
           numerical argument which is the number of metrics in the queue. If there are HighNum metrics  in  the
           queue,  any  new metrics will be dropped. If there are less than LowNum metrics in the queue, all new
           metrics will be enqueued. If the number of metrics currently in  the  queue  is  between  LowNum  and
           HighNum,  the  metric  is dropped with a probability that is proportional to the number of metrics in
           the queue (i.e. it increases linearly until it reaches 100%.)

           If WriteQueueLimitHigh is set to non-zero and WriteQueueLimitLow is unset, the latter will default to
           half of WriteQueueLimitHigh.

           If you do not want to randomly drop values when the queue size is between LowNum and HighNum, set  If
           WriteQueueLimitHigh and WriteQueueLimitLow to same value.

       Hostname Name
           Sets  the  hostname that identifies a host. If you omit this setting, the hostname will be determined
           using the gethostname(2) system call.

       FQDNLookup true|false
           If Hostname is determined automatically this setting controls whether or not the daemon should try to
           figure out the "fully qualified domain name", FQDN.  This is done using a lookup of the name returned
           by "gethostname". This option is enabled by default.

       PreCacheChain ChainName
       PostCacheChain ChainName
           Configure the name  of  the  "pre-cache  chain"  and  the  "post-cache  chain".  Please  see  "FILTER
           CONFIGURATION" below on information on chains and how these setting change the daemon's behavior.

PLUGIN OPTIONS

       Some  plugins  may  register  own  options.  These  options must be enclosed in a "Plugin"-Section. Which
       options exist depends on the plugin used. Some plugins require external configuration, too.  The  "apache
       plugin",  for  example,  required  "mod_status" to be configured in the webserver you're going to collect
       data from. These plugins are listed below as well, even if they don't require  any  configuration  within
       collectd's configuration file.

       A  list  of  all plugins and a short summary for each plugin can be found in the README file shipped with
       the sourcecode and hopefully binary packets as well.

   Plugin "aggregation"
       The Aggregation plugin makes it possible to aggregate several values into one using aggregation functions
       such as sum, average, min and max.  This can be put to a wide variety of uses, e.g. average and total CPU
       statistics for your entire fleet.

       The grouping is powerful but, as with many powerful tools, may be a  bit  difficult  to  wrap  your  head
       around.  The  grouping  will  therefore  be demonstrated using an example: The average and sum of the CPU
       usage across all CPUs of each host is to be calculated.

       To select all the affected values for our example, set "Plugin cpu" and "Type cpu". The other values  are
       left  unspecified,  meaning "all values". The Host, Plugin, PluginInstance, Type and TypeInstance options
       work as if they were specified in the "WHERE" clause of an "SELECT" SQL statement.

         Plugin "cpu"
         Type "cpu"

       Although the Host, PluginInstance (CPU number, i.e. 0, 1, 2, ...)  and TypeInstance (idle, user,  system,
       ...)  fields  are  left  unspecified in the example, the intention is to have a new value for each host /
       type instance pair. This is achieved by "grouping" the values using the  "GroupBy"  option.   It  can  be
       specified multiple times to group by more than one field.

         GroupBy "Host"
         GroupBy "TypeInstance"

       We  do  neither  specify nor group by plugin instance (the CPU number), so all metrics that differ in the
       CPU number only will be aggregated. Each  aggregation  needs  at  least  one  such  field,  otherwise  no
       aggregation would take place.

       The full example configuration looks like this:

        <Plugin "aggregation">
          <Aggregation>
            Plugin "cpu"
            Type "cpu"

            GroupBy "Host"
            GroupBy "TypeInstance"

            CalculateSum true
            CalculateAverage true
          </Aggregation>
        </Plugin>

       There are a couple of limitations you should be aware of:

       •   The Type cannot be left unspecified, because it is not reasonable to add apples to oranges. Also, the
           internal lookup structure won't work if you try to group by type.

       •   There must be at least one unspecified, ungrouped field. Otherwise nothing will be aggregated.

       As  you  can  see  in  the  example  above,  each aggregation has its own Aggregation block. You can have
       multiple aggregation blocks and aggregation blocks may match the same values, i.e.  one  value  list  can
       update multiple aggregations. The following options are valid inside Aggregation blocks:

       Host Host
       Plugin Plugin
       PluginInstance PluginInstance
       Type Type
       TypeInstance TypeInstance
           Selects  the  value  lists  to  be added to this aggregation. Type must be a valid data set name, see
           types.db(5) for details.

           If the string starts with and ends with a slash  ("/"),  the  string  is  interpreted  as  a  regular
           expression.  The  regex  flavor used are POSIX extended regular expressions as described in regex(7).
           Example usage:

            Host "/^db[0-9]\\.example\\.com$/"

       GroupBy Host|Plugin|PluginInstance|TypeInstance
           Group valued by the specified field. The GroupBy option may be repeated to group by multiple fields.

       SetHost Host
       SetPlugin Plugin
       SetPluginInstance PluginInstance
       SetTypeInstance TypeInstance
           Sets the appropriate part of the identifier to the provided string.

           The PluginInstance should include the placeholder "%{aggregation}" which will be  replaced  with  the
           aggregation  function,  e.g.  "average".  Not  including  the  placeholder will result in duplication
           warnings and/or messed up values if more than one aggregation function are enabled.

           The following example calculates the average usage of all "even" CPUs:

            <Plugin "aggregation">
              <Aggregation>
                Plugin "cpu"
                PluginInstance "/[0,2,4,6,8]$/"
                Type "cpu"

                SetPlugin "cpu"
                SetPluginInstance "even-%{aggregation}"

                GroupBy "Host"
                GroupBy "TypeInstance"

                CalculateAverage true
              </Aggregation>
            </Plugin>

           This will create the files:

           •   foo.example.com/cpu-even-average/cpu-idle

           •   foo.example.com/cpu-even-average/cpu-system

           •   foo.example.com/cpu-even-average/cpu-user

           •   ...

       CalculateNum true|false
       CalculateSum true|false
       CalculateAverage true|false
       CalculateMinimum true|false
       CalculateMaximum true|false
       CalculateStddev true|false
           Boolean options for enabling calculation of the number of value lists, their sum,  average,  minimum,
           maximum and / or standard deviation. All options are disabled by default.

   Plugin "amqp"
       The  AMQMP plugin can be used to communicate with other instances of collectd or third party applications
       using an AMQP message broker. Values are sent to or received from  the  broker,  which  handles  routing,
       queueing and possibly filtering or messages.

        <Plugin "amqp">
          # Send values to an AMQP broker
          <Publish "some_name">
            Host "localhost"
            Port "5672"
            VHost "/"
            User "guest"
            Password "guest"
            Exchange "amq.fanout"
        #   ExchangeType "fanout"
        #   RoutingKey "collectd"
        #   Persistent false
        #   Format "command"
        #   StoreRates false
        #   GraphitePrefix "collectd."
        #   GraphiteEscapeChar "_"
          </Publish>

          # Receive values from an AMQP broker
          <Subscribe "some_name">
            Host "localhost"
            Port "5672"
            VHost "/"
            User "guest"
            Password "guest"
            Exchange "amq.fanout"
        #   ExchangeType "fanout"
        #   Queue "queue_name"
        #   RoutingKey "collectd.#"
          </Subscribe>
        </Plugin>

       The  plugin's configuration consists of a number of Publish and Subscribe blocks, which configure sending
       and receiving of values respectively. The two blocks are very similar,  so  unless  otherwise  noted,  an
       option  can be used in either block. The name given in the blocks starting tag is only used for reporting
       messages, but may be used to support flushing of certain Publish blocks in the future.

       Host Host
           Hostname or IP-address of the AMQP broker.  Defaults  to  the  default  behavior  of  the  underlying
           communications library, rabbitmq-c, which is "localhost".

       Port Port
           Service  name  or  port  number on which the AMQP broker accepts connections. This argument must be a
           string, even if the numeric form is used. Defaults to "5672".

       VHost VHost
           Name of the virtual host on the AMQP broker to use. Defaults to "/".

       User User
       Password Password
           Credentials used to authenticate to the AMQP broker. By default "guest"/"guest" is used.

       Exchange Exchange
           In Publish blocks, this option specifies the exchange to send values to.   By  default,  "amq.fanout"
           will be used.

           In  Subscribe  blocks this option is optional. If given, a binding between the given exchange and the
           queue is created, using the routing key if configured. See the Queue and RoutingKey options below.

       ExchangeType Type
           If given, the plugin will try to create the configured exchange with this type after connecting. When
           in a Subscribe block, the queue will then be bound to this exchange.

       Queue Queue (Subscribe only)
           Configures the queue name to subscribe to. If no queue name was configures explicitly, a unique queue
           name will be created by the broker.

       RoutingKey Key
           In Publish blocks, this configures the routing key to set on all outgoing messages. If not given, the
           routing key will be computed from the identifier of the value. The host, plugin,  type  and  the  two
           instances are concatenated together using dots as the separator and all containing dots replaced with
           slashes.  For  example  "collectd.host/example/com.cpu.0.cpu.user". This makes it possible to receive
           only specific values using a "topic" exchange.

           In Subscribe blocks, configures the routing key used when creating a binding between an exchange  and
           the  queue.  The  usual  wildcards  can  be used to filter messages when using a "topic" exchange. If
           you're only interested in CPU statistics, you  could  use  the  routing  key  "collectd.*.cpu.#"  for
           example.

       Persistent true|false (Publish only)
           Selects  the  delivery method to use. If set to true, the persistent mode will be used, i.e. delivery
           is guaranteed. If set to false (the default), the transient delivery mode will be used, i.e. messages
           may be lost due to high load, overflowing queues or similar issues.

       Format Command|JSON|Graphite (Publish only)
           Selects the format in which messages are sent to the broker. If set to Command (the default),  values
           are  sent  as  "PUTVAL"  commands  which  are  identical  to the syntax used by the Exec and UnixSock
           plugins. In this case, the "Content-Type" header field will be set to "text/collectd".

           If set to JSON, the values are encoded in the  JavaScript  Object  Notation,  an  easy  and  straight
           forward exchange format. The "Content-Type" header field will be set to "application/json".

           If  set  to  Graphite,  values  are  encoded  in  the  Graphite  format,  which  is "<metric> <value>
           <timestamp>\n". The "Content-Type" header field will be set to "text/graphite".

           A subscribing client should use the "Content-Type" header  field  to  determine  how  to  decode  the
           values. Currently, the AMQP plugin itself can only decode the Command format.

       StoreRates true|false (Publish only)
           Determines  whether  or  not  "COUNTER", "DERIVE" and "ABSOLUTE" data sources are converted to a rate
           (i.e. a "GAUGE" value). If set to false (the default), no  conversion  is  performed.  Otherwise  the
           conversion is performed using the internal value cache.

           Please note that currently this option is only used if the Format option has been set to JSON.

       GraphitePrefix (Publish and Format=Graphite only)
           A  prefix  can be added in the metric name when outputting in the Graphite format.  It's added before
           the Host name.  Metric name will be "<prefix><host><postfix><plugin><type><name>"

       GraphitePostfix (Publish and Format=Graphite only)
           A postfix can be added in the metric name when outputting in the Graphite format.  It's  added  after
           the Host name.  Metric name will be "<prefix><host><postfix><plugin><type><name>"

       GraphiteEscapeChar (Publish and Format=Graphite only)
           Specify  a  character  to  replace  dots (.) in the host part of the metric name.  In Graphite metric
           name, dots are used as separators between different metric parts (host, plugin,  type).   Default  is
           "_" (Underscore).

   Plugin "apache"
       To  configure the "apache"-plugin you first need to configure the Apache webserver correctly. The Apache-
       plugin "mod_status" needs to be loaded and  working  and  the  "ExtendedStatus"  directive  needs  to  be
       enabled. You can use the following snipped to base your Apache config upon:

         ExtendedStatus on
         <IfModule mod_status.c>
           <Location /mod_status>
             SetHandler server-status
           </Location>
         </IfModule>

       Since  its  "mod_status"  module is very similar to Apache's, lighttpd is also supported. It introduces a
       new field, called "BusyServers", to count the number of currently connected clients. This field  is  also
       supported.

       The configuration of the Apache plugin consists of one or more "<Instance />" blocks. Each block requires
       one string argument as the instance name. For example:

        <Plugin "apache">
          <Instance "www1">
            URL "http://www1.example.com/mod_status?auto"
          </Instance>
          <Instance "www2">
            URL "http://www2.example.com/mod_status?auto"
          </Instance>
        </Plugin>

       The  instance  name will be used as the plugin instance. To emulate the old (version 4) behavior, you can
       use an empty string (""). In order for the plugin to work correctly, each instance name must  be  unique.
       This is not enforced by the plugin and it is your responsibility to ensure it.

       The following options are accepted within each Instance block:

       URL http://host/mod_status?auto
           Sets the URL of the "mod_status" output. This needs to be the output generated by "ExtendedStatus on"
           and  it  needs  to  be  the machine readable output generated by appending the "?auto" argument. This
           option is mandatory.

       User Username
           Optional user name needed for authentication.

       Password Password
           Optional password needed for authentication.

       VerifyPeer true|false
           Enable or disable peer SSL certificate verification. See <http://curl.haxx.se/docs/sslcerts.html> for
           details. Enabled by default.

       VerifyHost true|false
           Enable or disable peer host name verification. If enabled, the plugin checks if the "Common Name"  or
           a  "Subject  Alternate  Name"  field of the SSL certificate matches the host name provided by the URL
           option. If this identity  check  fails,  the  connection  is  aborted.  Obviously,  only  works  when
           connecting to a SSL enabled server. Enabled by default.

       CACert File
           File  that  holds  one or more SSL certificates. If you want to use HTTPS you will possibly need this
           option. What CA certificates come bundled with "libcurl" and are checked by default  depends  on  the
           distribution you use.

   Plugin "apcups"
       Host Hostname
           Hostname  of  the host running apcupsd. Defaults to localhost. Please note that IPv6 support has been
           disabled unless someone can confirm or decline that apcupsd can handle it.

       Port Port
           TCP-Port to connect to. Defaults to 3551.

       ReportSeconds true|false
           If set to true, the time reported in the "timeleft" metric will be converted to seconds. This is  the
           recommended  setting.  If  set  to  false,  the default for backwards compatibility, the time will be
           reported in minutes.

   Plugin "aquaero"
       This plugin collects the value of the available sensors in an Aquaero 5  board.  Aquaero 5  is  a  water-
       cooling  controller  board, manufactured by Aqua Computer GmbH <http://www.aquacomputer.de/>, with a USB2
       connection for monitoring and configuration. The board can handle  multiple  temperature  sensors,  fans,
       water  pumps  and water level sensors and adjust the output settings such as fan voltage or power used by
       the water pump based on the available inputs using a configurable controller included in the board.  This
       plugin collects all the available inputs as well as some of the output values chosen by this  controller.
       The plugin is based on the libaquaero5 library provided by aquatools-ng.

       Device DevicePath
           Device   path   of   the   Aquaero 5's  USB  HID  (human  interface  device),  usually  in  the  form
           "/dev/usb/hiddevX". If this option is no set the plugin will try to auto-detect  the  Aquaero  5  USB
           device based on vendor-ID and product-ID.

   Plugin "ascent"
       This  plugin collects information about an Ascent server, a free server for the "World of Warcraft" game.
       This plugin gathers the information by fetching the XML status page using "libcurl" and parses  it  using
       "libxml2".

       The configuration options are the same as for the "apache" plugin above:

       URL http://localhost/ascent/status/
           Sets the URL of the XML status output.

       User Username
           Optional user name needed for authentication.

       Password Password
           Optional password needed for authentication.

       VerifyPeer true|false
           Enable or disable peer SSL certificate verification. See <http://curl.haxx.se/docs/sslcerts.html> for
           details. Enabled by default.

       VerifyHost true|false
           Enable  or disable peer host name verification. If enabled, the plugin checks if the "Common Name" or
           a "Subject Alternate Name" field of the SSL certificate matches the host name  provided  by  the  URL
           option.  If  this  identity  check  fails,  the  connection  is  aborted.  Obviously, only works when
           connecting to a SSL enabled server. Enabled by default.

       CACert File
           File that holds one or more SSL certificates. If you want to use HTTPS you will  possibly  need  this
           option.  What  CA  certificates come bundled with "libcurl" and are checked by default depends on the
           distribution you use.

   Plugin "bind"
       Starting with BIND 9.5.0, the most widely used DNS server software provides  extensive  statistics  about
       queries,  responses  and  lots  of  other information.  The bind plugin retrieves this information that's
       encoded in XML and provided via HTTP and submits the values to collectd.

       To use this plugin, you first need to tell BIND to make this information available. This is done with the
       "statistics-channels" configuration option:

        statistics-channels {
          inet localhost port 8053;
        };

       The configuration follows the grouping that can be seen when looking at the data with an XSLT  compatible
       viewer,  such  as  a  modern  web  browser.  It's probably a good idea to make yourself familiar with the
       provided values, so you can understand what the collected statistics actually mean.

       Synopsis:

        <Plugin "bind">
          URL "http://localhost:8053/"
          ParseTime       false
          OpCodes         true
          QTypes          true

          ServerStats     true
          ZoneMaintStats  true
          ResolverStats   false
          MemoryStats     true

          <View "_default">
            QTypes        true
            ResolverStats true
            CacheRRSets   true

            Zone "127.in-addr.arpa/IN"
          </View>
        </Plugin>

       The bind plugin accepts the following configuration options:

       URL URL
           URL from which to retrieve the XML data. If not specified, "http://localhost:8053/" will be used.

       ParseTime true|false
           When set to true, the time provided by BIND will be parsed and used to dispatch the values. When  set
           to false, the local time source is queried.

           This  setting  is  set  to  true  by  default  for  backwards compatibility; setting this to false is
           recommended to avoid problems with timezones and localization.

       OpCodes true|false
           When enabled, statistics about the  "OpCodes",  for  example  the  number  of  "QUERY"  packets,  are
           collected.

           Default: Enabled.

       QTypes true|false
           When  enabled,  the  number  of  incoming  queries  by query types (for example "A", "MX", "AAAA") is
           collected.

           Default: Enabled.

       ServerStats true|false
           Collect global server statistics, such as requests received over IPv4 and IPv6,  successful  queries,
           and failed updates.

           Default: Enabled.

       ZoneMaintStats true|false
           Collect  zone  maintenance statistics, mostly information about notifications (zone updates) and zone
           transfers.

           Default: Enabled.

       ResolverStats true|false
           Collect resolver statistics, i. e. statistics about outgoing requests (e. g. queries over IPv4,  lame
           servers). Since the global resolver counters apparently were removed in BIND 9.5.1 and 9.6.0, this is
           disabled by default. Use the ResolverStats option within a View "_default" block instead for the same
           functionality.

           Default: Disabled.

       MemoryStats
           Collect global memory statistics.

           Default: Enabled.

       View Name
           Collect statistics about a specific "view". BIND can behave different, mostly depending on the source
           IP-address  of  the request. These different configurations are called "views". If you don't use this
           feature, you most likely are only interested in the "_default" view.

           Within a <View name> block, you can specify which information you want to collect about a view. If no
           View block is configured, no detailed view statistics will be collected.

           QTypes true|false
               If enabled, the number of outgoing queries by query type (e. g. "A", "MX") is collected.

               Default: Enabled.

           ResolverStats true|false
               Collect resolver statistics, i. e. statistics about outgoing requests (e. g. queries  over  IPv4,
               lame servers).

               Default: Enabled.

           CacheRRSets true|false
               If  enabled,  the  number  of entries ("RR sets") in the view's cache by query type is collected.
               Negative entries (queries which resulted in an error, for example names that do  not  exist)  are
               reported with a leading exclamation mark, e. g. "!A".

               Default: Enabled.

           Zone Name
               When  given,  collect  detailed  information  about  the  given zone in the view. The information
               collected if very similar to the global ServerStats information (see above).

               You can repeat this option to collect detailed information about multiple zones.

               By default no detailed zone information is collected.

   Plugin "cgroups"
       This plugin collects the CPU user/system time for each cgroup by reading the cpuacct.stat  files  in  the
       first cpuacct-mountpoint (typically /sys/fs/cgroup/cpu.cpuacct on machines using systemd).

       CGroup Directory
           Select  cgroup  based on the name. Whether only matching cgroups are collected or if they are ignored
           is controlled by the IgnoreSelected option; see below.

       IgnoreSelected true|false
           Invert the selection: If set to true, all cgroups except the ones that match any one of the  criteria
           are collected. By default only selected cgroups are collected if a selection is made. If no selection
           is configured at all, all cgroups are selected.

   Plugin "cpufreq"
       This plugin doesn't have any options. It reads /sys/devices/system/cpu/cpu0/cpufreq/scaling_cur_freq (for
       the first CPU installed) to get the current CPU frequency. If this file does not exist make sure cpufreqd
       (<http://cpufreqd.sourceforge.net/>)  or  a  similar  tool  is  installed and an "cpu governor" (that's a
       kernel module) is loaded.

   Plugin "csv"
       DataDir Directory
           Set the directory to store CSV-files under. Per default CSV-files are generated beneath the  daemon's
           working  directory, i. e. the BaseDir.  The special strings stdout and stderr can be used to write to
           the standard output and standard error channels, respectively. This, of course, only makes much sense
           when collectd is running in foreground- or non-daemon-mode.

       StoreRates true|false
           If set to true, convert counter values to rates. If set to false (the  default)  counter  values  are
           stored as is, i. e. as an increasing integer number.

   Plugin "curl"
       The  curl plugin uses the libcurl (<http://curl.haxx.se/>) to read web pages and the match infrastructure
       (the same code used by the tail plugin) to use regular expressions with the received data.

       The following example will read the current value of AMD stock from Google's finance  page  and  dispatch
       the value to collectd.

         <Plugin curl>
           <Page "stock_quotes">
             URL "http://finance.google.com/finance?q=NYSE%3AAMD"
             User "foo"
             Password "bar"
             <Match>
               Regex "<span +class=\"pr\"[^>]*> *([0-9]*\\.[0-9]+) *</span>"
               DSType "GaugeAverage"
               # Note: `stock_value' is not a standard type.
               Type "stock_value"
               Instance "AMD"
             </Match>
           </Page>
         </Plugin>

       In  the  Plugin  block,  there  may  be one or more Page blocks, each defining a web page and one or more
       "matches" to be performed on the returned data. The string argument to the Page block is used  as  plugin
       instance.

       The following options are valid within Page blocks:

       URL URL
           URL  of the web site to retrieve. Since a regular expression will be used to extract information from
           this data, non-binary data is a big plus here ;)

       User Name
           Username to use if authorization is required to read the page.

       Password Password
           Password to use if authorization is required to read the page.

       VerifyPeer true|false
           Enable or disable peer SSL certificate verification. See <http://curl.haxx.se/docs/sslcerts.html> for
           details. Enabled by default.

       VerifyHost true|false
           Enable or disable peer host name verification. If enabled, the plugin checks if the "Common Name"  or
           a  "Subject  Alternate  Name"  field of the SSL certificate matches the host name provided by the URL
           option. If this identity  check  fails,  the  connection  is  aborted.  Obviously,  only  works  when
           connecting to a SSL enabled server. Enabled by default.

       CACert file
           File  that  holds  one or more SSL certificates. If you want to use HTTPS you will possibly need this
           option. What CA certificates come bundled with "libcurl" and are checked by default  depends  on  the
           distribution you use.

       Header Header
           A HTTP header to add to the request. Multiple headers are added if this option is specified more than
           once.

       Post Body
           Specifies  that  the HTTP operation should be a POST instead of a GET. The complete data to be posted
           is given as the argument.  This option will usually need to be accompanied by a Header option to  set
           an appropriate "Content-Type" for the post body (e.g. to "application/x-www-form-urlencoded").

       MeasureResponseTime true|false
           Measure  response  time  for  the  request.  If this setting is enabled, Match blocks (see below) are
           optional. Disabled by default.

       <Match>
           One or more Match blocks that define how to match information in the data returned by "libcurl".  The
           "curl"  plugin  uses  the  same  infrastructure  that's  used by the "tail" plugin, so please see the
           documentation of the "tail" plugin below on how  matches  are  defined.  If  the  MeasureResponseTime
           option is set to true, Match blocks are optional.

   Plugin "curl_json"
       The    curl_json    plugin    collects    values    from    JSON   data   to   be   parsed   by   libyajl
       (<http://www.lloydforge.org/projects/yajl/>) retrieved via  either  libcurl  (<http://curl.haxx.se/>)  or
       read  directly  from  a  unix socket. The former can be used, for example, to collect values from CouchDB
       documents (which are stored JSON notation), and the latter to collect values from a uWSGI stats socket.

       The following example will collect several values from the built-in "_stats" runtime statistics module of
       CouchDB (<http://wiki.apache.org/couchdb/Runtime_Statistics>).

         <Plugin curl_json>
           <URL "http://localhost:5984/_stats">
             Instance "httpd"
             <Key "httpd/requests/count">
               Type "http_requests"
             </Key>

             <Key "httpd_request_methods/*/count">
               Type "http_request_methods"
             </Key>

             <Key "httpd_status_codes/*/count">
               Type "http_response_codes"
             </Key>
           </URL>
         </Plugin>

       This example will collect data directly from a uWSGI "Stats Server" socket.

         <Plugin curl_json>
           <Sock "/var/run/uwsgi.stats.sock">
             Instance "uwsgi"
             <Key "workers/*/requests">
               Type "http_requests"
             </Key>

             <Key "workers/*/apps/*/requests">
               Type "http_requests"
             </Key>
           </Sock>
         </Plugin>

       In the Plugin block, there may be one or more URL blocks, each defining a URL  to  be  fetched  via  HTTP
       (using  libcurl)  or Sock blocks defining a unix socket to read JSON from directly.  Each of these blocks
       may have one or more Key blocks.

       The Key string argument must be in a path format. Each component is used to match the key from a JSON map
       or the index of an JSON array. If a path component of a Key is a * wildcard, the values for all map  keys
       or array indices will be collectd.

       The following options are valid within URL blocks:

       Instance Instance
           Sets the plugin instance to Instance.

       User Name =item Password Password =item VerifyPeer true|false =item VerifyHost true|false =item CACert
       file =item Header Header =item Post Body
           These  options  behave  exactly  equivalent to the appropriate options of the cURL plugin. Please see
           there for a detailed description.

       The following options are valid within Key blocks:

       Type Type
           Sets the type used to dispatch the values to the daemon. Detailed information about types  and  their
           configuration can be found in types.db(5). This option is mandatory.

       Instance Instance
           Type-instance to use. Defaults to the current map key or current string array element value.

   Plugin "curl_xml"
       The curl_xml plugin uses libcurl (<http://curl.haxx.se/>) and libxml2 (<http://xmlsoft.org/>) to retrieve
       XML data via cURL.

        <Plugin "curl_xml">
          <URL "http://localhost/stats.xml">
            Host "my_host"
            Instance "some_instance"
            User "collectd"
            Password "thaiNg0I"
            VerifyPeer true
            VerifyHost true
            CACert "/path/to/ca.crt"

            <XPath "table[@id=\"magic_level\"]/tr">
              Type "magic_level"
              #InstancePrefix "prefix-"
              InstanceFrom "td[1]"
              ValuesFrom "td[2]/span[@class=\"level\"]"
            </XPath>
          </URL>
        </Plugin>

       In  the  Plugin  block,  there  may  be  one  or more URL blocks, each defining a URL to be fetched using
       libcurl. Within each URL block there are options which specify the  connection  parameters,  for  example
       authentication information, and one or more XPath blocks.

       Each  XPath block specifies how to get one type of information. The string argument must be a valid XPath
       expression which returns a list of "base elements". One value is dispatched for each "base element".  The
       type  instance  and  values  are looked up using further XPath expressions that should be relative to the
       base element.

       Within the URL block the following options are accepted:

       Host Name
           Use Name as the host name when submitting values. Defaults to the global host name setting.

       Instance Instance
           Use Instance as the plugin instance when submitting values. Defaults to an empty  string  (no  plugin
           instance).

       Namespace Prefix URL
           If  an XPath expression references namespaces, they must be specified with this option. Prefix is the
           "namespace prefix" used in the XML document.  URL is the "namespace name", an URI reference  uniquely
           identifying the namespace. The option can be repeated to register multiple namespaces.

           Examples:

             Namespace "s" "http://schemas.xmlsoap.org/soap/envelope/"
             Namespace "m" "http://www.w3.org/1998/Math/MathML"

       User User
       Password Password
       VerifyPeer true|false
       VerifyHost true|false
       CACert CA Cert File
       Header Header
       Post Body
           These  options  behave  exactly  equivalent to the appropriate options of the cURL plugin. Please see
           there for a detailed description.

       <XPath XPath-expression>
           Within each URL block, there must be one or more XPath blocks. Each XPath block specifies how to  get
           one type of information. The string argument must be a valid XPath expression which returns a list of
           "base elements". One value is dispatched for each "base element".

           Within the XPath block the following options are accepted:

           Type Type
               Specifies  the  Type  used  for submitting patches. This determines the number of values that are
               required / expected and whether the strings are parsed as signed or unsigned integer or as double
               values. See types.db(5) for details.  This option is required.

           InstancePrefix InstancePrefix
               Prefix the type instance with InstancePrefix. The values are simply concatenated together without
               any separator.  This option is optional.

           InstanceFrom InstanceFrom
               Specifies a XPath expression to use for determining the type instance. The XPath expression  must
               return  exactly one element. The element's value is then used as type instance, possibly prefixed
               with InstancePrefix (see above).

               This value is required. As a special exception, if the "base XPath expression" (the  argument  to
               the XPath block) returns exactly one argument, then this option may be omitted.

           ValuesFrom ValuesFrom [ValuesFrom ...]
               Specifies  one  or  more  XPath  expression  to  use  for reading the values. The number of XPath
               expressions must match the number of data sources in the type specified with  Type  (see  above).
               Each  XPath  expression  must return exactly one element. The element's value is then parsed as a
               number and used as value for the appropriate value in the value list dispatched to the daemon.

   Plugin "dbi"
       This plugin uses the dbi library (<http://libdbi.sourceforge.net/>)  to  connect  to  various  databases,
       execute  SQL statements and read back the results. dbi is an acronym for "database interface" in case you
       were wondering about the name. You can configure how each column is to be interpreted and the plugin will
       generate one or more data sets from each row returned according to these rules.

       Because the plugin is very generic, the configuration is a  little  more  complex  than  those  of  other
       plugins. It usually looks something like this:

         <Plugin dbi>
           <Query "out_of_stock">
             Statement "SELECT category, COUNT(*) AS value FROM products WHERE in_stock = 0 GROUP BY category"
             # Use with MySQL 5.0.0 or later
             MinVersion 50000
             <Result>
               Type "gauge"
               InstancePrefix "out_of_stock"
               InstancesFrom "category"
               ValuesFrom "value"
             </Result>
           </Query>
           <Database "product_information">
             Driver "mysql"
             DriverOption "host" "localhost"
             DriverOption "username" "collectd"
             DriverOption "password" "aZo6daiw"
             DriverOption "dbname" "prod_info"
             SelectDB "prod_info"
             Query "out_of_stock"
           </Database>
         </Plugin>

       The  configuration  above defines one query with one result and one database. The query is then linked to
       the database with the Query option within the <Database> block. You can have any number  of  queries  and
       databases  and  you  can  also  use the Include statement to split up the configuration file in multiple,
       smaller files. However, the <Query> block must  precede  the  <Database>  blocks,  because  the  file  is
       interpreted from top to bottom!

       The following is a complete list of options:

       Query blocks

       Query  blocks  define SQL statements and how the returned data should be interpreted. They are identified
       by the name that is given in the opening line of the block. Thus the name needs to be unique. Other  than
       that, the name is not used in collectd.

       In  each  Query  block, there is one or more Result blocks. Result blocks define which column holds which
       value or instance information. You can use multiple Result blocks to  create  multiple  values  from  one
       returned  row. This is especially useful, when queries take a long time and sending almost the same query
       again and again is not desirable.

       Example:

         <Query "environment">
           Statement "select station, temperature, humidity from environment"
           <Result>
             Type "temperature"
             # InstancePrefix "foo"
             InstancesFrom "station"
             ValuesFrom "temperature"
           </Result>
           <Result>
             Type "humidity"
             InstancesFrom "station"
             ValuesFrom "humidity"
           </Result>
         </Query>

       The following options are accepted:

       Statement SQL
           Sets the statement that should be executed on the server. This is not interpreted  by  collectd,  but
           simply  passed  to  the database server. Therefore, the SQL dialect that's used depends on the server
           collectd is connected to.

           The query has to return at least two columns, one for the instance and one value. You cannot omit the
           instance, even if the statement is guaranteed to always return exactly one line. In  that  case,  you
           can usually specify something like this:

             Statement "SELECT \"instance\", COUNT(*) AS value FROM table"

           (That  works  with  MySQL  but  may  not be valid SQL according to the spec. If you use a more strict
           database server, you may have to select from a dummy table or something.)

           Please note that some databases, for example Oracle, will fail if you include a semicolon at the  end
           of the statement.

       MinVersion Version
       MaxVersion Value
           Only use this query for the specified database version. You can use these options to provide multiple
           queries  with  the  same  name  but  with a slightly different syntax. The plugin will use only those
           queries, where the specified minimum and maximum versions fit the version of the database in use.

           The database version is determined by "dbi_conn_get_engine_version",  see  the  libdbi  documentation
           <http://libdbi.sourceforge.net/docs/programmers-guide/reference-conn.html#DBI-CONN-GET-ENGINE-
           VERSION> for details. Basically, each part of the version is assumed to be in the range from 00 to 99
           and all dots are removed. So version "4.1.2" becomes "40102", version "5.0.42" becomes "50042".

           Warning:  The  plugin will use all matching queries, so if you specify multiple queries with the same
           name and overlapping ranges, weird stuff will happen. Don't to it! A valid example would be something
           along these lines:

             MinVersion 40000
             MaxVersion 49999
             ...
             MinVersion 50000
             MaxVersion 50099
             ...
             MinVersion 50100
             # No maximum

           In the above example, there are three ranges that don't overlap.  The  last  one  goes  from  version
           "5.1.0" to infinity, meaning "all later versions". Versions before "4.0.0" are not specified.

       Type Type
           The  type  that's  used  for  each  line  returned. See types.db(5) for more details on how types are
           defined. In short: A type is a predefined layout of data and the number of values and type of  values
           has to match the type definition.

           If you specify "temperature" here, you need exactly one gauge column. If you specify "if_octets", you
           will need two counter columns. See the ValuesFrom setting below.

           There must be exactly one Type option inside each Result block.

       InstancePrefix prefix
           Prepends prefix to the type instance. If InstancesFrom (see below) is not given, the string is simply
           copied.  If  InstancesFrom  is  given, prefix and all strings returned in the appropriate columns are
           concatenated together, separated by dashes ("-").

       InstancesFrom column0 [column1 ...]
           Specifies the columns whose values will be used to create the "type-instance" for each  row.  If  you
           specify  more  than one column, the value of all columns will be joined together with dashes ("-") as
           separation characters.

           The plugin itself does not check whether  or  not  all  built  instances  are  different.  It's  your
           responsibility  to  assure  that  each  is  unique.  This  is  especially true, if you do not specify
           InstancesFrom: You have to make sure that only one row is returned in this case.

           If neither InstancePrefix nor InstancesFrom is given, the type-instance will be empty.

       ValuesFrom column0 [column1 ...]
           Names the columns whose content is used as the actual data for the data sets that are  dispatched  to
           the  daemon.  How  many such columns you need is determined by the Type setting above. If you specify
           too many or not enough columns, the plugin will complain about that and no data will be submitted  to
           the daemon.

           The  actual  data  type  in the columns is not that important. The plugin will automatically cast the
           values to the right type if it know how to do that. So  it  should  be  able  to  handle  integer  an
           floating point types, as well as strings (if they include a number at the beginning).

           There must be at least one ValuesFrom option inside each Result block.

       Database blocks

       Database  blocks  define  a  connection  to a database and which queries should be sent to that database.
       Since the used "dbi" library can handle a wide variety of databases, the configuration is  very  generic.
       If in doubt, refer to libdbi's documentation - we stick as close to the terminology used there.

       Each  database needs a "name" as string argument in the starting tag of the block. This name will be used
       as "PluginInstance" in the values submitted to the daemon. Other than that, that name is not used.

       Driver Driver
           Specifies the driver to use to connect to the database. In many cases those drivers are  named  after
           the  database they can connect to, but this is not a technical necessity. These drivers are sometimes
           referred to as "DBD", DataBase Driver, and some distributions ship them in separate packages. Drivers
           for    the    "dbi"    library    are    developed    by    the     libdbi-drivers     project     at
           <http://libdbi-drivers.sourceforge.net/>.

           You  need  to  give the driver name as expected by the "dbi" library here. You should be able to find
           that in the documentation for each driver. If you mistype the driver name, the  plugin  will  dump  a
           list of all known driver names to the log.

       DriverOption Key Value
           Sets  driver-specific  options.  What  option a driver supports can be found in the documentation for
           each driver, somewhere at  <http://libdbi-drivers.sourceforge.net/>.  However,  the  options  "host",
           "username", "password", and "dbname" seem to be de facto standards.

           DBDs  can  register two types of options: String options and numeric options. The plugin will use the
           "dbi_conn_set_option"   function   when   the   configuration   provides    a    string    and    the
           "dbi_conn_require_option_numeric"  function  when  the  configuration provides a number. So these two
           lines will actually result in different calls being used:

             DriverOption "Port" 1234      # numeric
             DriverOption "Port" "1234"    # string

           Unfortunately, drivers are not too keen to report errors when an unknown option is passed to them, so
           invalid settings here may go unnoticed. This is not the plugin's fault, it will report errors  if  it
           gets  them from the library / the driver. If a driver complains about an option, the plugin will dump
           a complete list of  all  options  understood  by  that  driver  to  the  log.  There  is  no  way  to
           programatically  find  out  if  an option expects a string or a numeric argument, so you will have to
           refer to the appropriate DBD's documentation to find this out. Sorry.

       SelectDB Database
           In some cases, the database name you connect with is not the  database  name  you  want  to  use  for
           querying  data.  If  this option is set, the plugin will "select" (switch to) that database after the
           connection is established.

       Query QueryName
           Associates the query named QueryName with this database connection. The query  needs  to  be  defined
           before  this statement, i. e. all query blocks you want to refer to must be placed above the database
           block you want to refer to them from.

       Host Hostname
           Sets the host field of value lists to Hostname  when  dispatching  values.  Defaults  to  the  global
           hostname setting.

   Plugin "df"
       Device Device
           Select partitions based on the devicename.

       MountPoint Directory
           Select partitions based on the mountpoint.

       FSType FSType
           Select partitions based on the filesystem type.

       IgnoreSelected true|false
           Invert  the  selection:  If  set  to  true,  all partitions except the ones that match any one of the
           criteria are collected. By default only selected partitions are collected if a selection is made.  If
           no selection is configured at all, all partitions are selected.

       ReportByDevice true|false
           Report using the device name rather than the mountpoint. i.e. with this false, (the default), it will
           report a disk as "root", but with it true, it will be "sda1" (or whichever).

       ReportInodes true|false
           Enables  or  disables reporting of free, reserved and used inodes. Defaults to inode collection being
           disabled.

           Enable this option if inodes are a scarce resource for you, usually  because  many  small  files  are
           stored on the disk. This is a usual scenario for mail transfer agents and web caches.

       ValuesAbsolute true|false
           Enables or disables reporting of free, used and used disk space in 1K-blocks.  Defaults to true.

       ValuesPercentage true|false
           Enables or disables reporting of free, used and used disk space in percentage.  Defaults to false.

           This  is  useful  for  deploying  collectd  on the cloud, where machines with different disk size may
           exist. Then it is more practical to configure thresholds based on relative disk size.

   Plugin "disk"
       The "disk" plugin collects information about the usage of physical disks and logical disks  (partitions).
       Values  collected  are  the  number of octets written to and read from a disk or partition, the number of
       read/write operations issued to the disk and a rather complex "time" it took for  these  commands  to  be
       issued.

       Using  the  following  two options you can ignore some disks or configure the collection only of specific
       disks.

       Disk Name
           Select the disk Name. Whether it is collected or ignored depends on the IgnoreSelected  setting,  see
           below. As with other plugins that use the daemon's ignorelist functionality, a string that starts and
           ends with a slash is interpreted as a regular expression. Examples:

             Disk "sdd"
             Disk "/hda[34]/"

       IgnoreSelected true|false
           Sets  whether selected disks, i. e. the ones matches by any of the Disk statements, are ignored or if
           all other disks are ignored. The behavior (hopefully) is intuitive: If no Disk option is  configured,
           all  disks are collected. If at least one Disk option is given and no IgnoreSelected or set to false,
           only matching disks will be collected. If IgnoreSelected is set to  true,  all  disks  are  collected
           except the ones matched.

   Plugin "dns"
       Interface Interface
           The  dns  plugin  uses libpcap to capture dns traffic and analyzes it. This option sets the interface
           that should be used. If this option is not set, or set to "any", the plugin will try to  get  packets
           from all interfaces. This may not work on certain platforms, such as Mac OS X.

       IgnoreSource IP-address
           Ignore packets that originate from this address.

       SelectNumericQueryTypes true|false
           Enabled by default, collects unknown (and thus presented as numeric only) query types.

   Plugin "email"
       SocketFile Path
           Sets the socket-file which is to be created.

       SocketGroup Group
           If  running  as  root  change  the  group  of  the UNIX-socket after it has been created. Defaults to
           collectd.

       SocketPerms Permissions
           Change the file permissions of the UNIX-socket after it has been created.  The  permissions  must  be
           given as a numeric, octal value as you would pass to chmod(1). Defaults to 0770.

       MaxConns Number
           Sets  the maximum number of connections that can be handled in parallel. Since this many threads will
           be started immediately setting this to a very high value will waste valuable resources. Defaults to 5
           and will be forced to be at most 16384 to prevent typos and dumb mistakes.

   Plugin "ethstat"
       The ethstat plugin collects information about network interface cards (NICs) by talking directly with the
       underlying kernel driver using ioctl(2).

       Synopsis:

        <Plugin "ethstat">
          Interface "eth0"
          Map "rx_csum_offload_errors" "if_rx_errors" "checksum_offload"
          Map "multicast" "if_multicast"
        </Plugin>

       Options:

       Interface Name
           Collect statistical information about interface Name.

       Map Name Type [TypeInstance]
           By default, the plugin will submit values as type "derive" and type instance set to Name, the name of
           the metric as reported by the driver. If an appropriate  Map  option  exists,  the  given  Type  and,
           optionally, TypeInstance will be used.

       MappedOnly true|false
           When  set  to true, only metrics that can be mapped to to a type will be collected, all other metrics
           will be ignored. Defaults to false.

   Plugin "exec"
       Please make sure to read collectd-exec(5) before using this plugin. It contains valuable  information  on
       when the executable is executed and the output that is expected from it.

       Exec User[:[Group]] Executable [<arg> [<arg> ...]]
       NotificationExec User[:[Group]] Executable [<arg> [<arg> ...]]
           Execute  the  executable Executable as user User. If the user name is followed by a colon and a group
           name, the effective group is set to that group.  The real group and saved-set group will  be  set  to
           the  default  group of that user. If no group is given the effective group ID will be the same as the
           real group ID.

           Please note that in order to change the user and/or group the daemon needs superuser  privileges.  If
           the daemon is run as an unprivileged user you must specify the same user/group here. If the daemon is
           run with superuser privileges, you must supply a non-root user here.

           The executable may be followed by optional arguments that are passed to the program. Please note that
           due  to  the  configuration  parsing  numbers  and  boolean  values may be changed. If you want to be
           absolutely sure that something is passed as-is please enclose it in quotes.

           The Exec and NotificationExec statements change the semantics of the  programs  executed,  i. e.  the
           data  passed  to  them  and  the  response  expected from them. This is documented in great detail in
           collectd-exec(5).

   Plugin "filecount"
       The "filecount" plugin counts the number of files in a certain directory  (and  its  subdirectories)  and
       their combined size. The configuration is very straight forward:

         <Plugin "filecount">
           <Directory "/var/qmail/queue/mess">
             Instance "qmail-message"
           </Directory>
           <Directory "/var/qmail/queue/todo">
             Instance "qmail-todo"
           </Directory>
           <Directory "/var/lib/php5">
             Instance "php5-sessions"
             Name "sess_*"
           </Directory>
         </Plugin>

       The  example  above  counts  the  number  of  files  in  QMail's queue directories and the number of PHP5
       sessions. Jfiy: The "todo" queue holds the messages that QMail has not yet looked at, the "message" queue
       holds the messages that were classified into "local" and "remote".

       As you can see, the configuration consists of one or more "Directory" blocks, each of which  specifies  a
       directory in which to count the files. Within those blocks, the following options are recognized:

       Instance Instance
           Sets   the  plugin  instance  to  Instance.  That  instance  name  must  be  unique,  but  it's  your
           responsibility, the plugin doesn't check for that. If not given, the instance is set to the directory
           name with all slashes replaced by underscores and all leading underscores removed.

       Name Pattern
           Only count files that match Pattern,  where  Pattern  is  a  shell-like  wildcard  as  understood  by
           fnmatch(3). Only the filename is checked against the pattern, not the entire path. In case this makes
           it easier for you: This option has been named after the -name parameter to find(1).

       MTime Age
           Count only files of a specific age: If Age is greater than zero, only files that haven't been touched
           in  the  last Age seconds are counted. If Age is a negative number, this is inversed. For example, if
           -60 is specified, only files that have been modified in the last minute will be counted.

           The number can also be followed by a "multiplier" to easily specify a larger timespan. When given  in
           this notation, the argument must in quoted, i. e.  must be passed as string. So the -60 could also be
           written  as  "-1m"  (one  minute).  Valid multipliers are "s" (second), "m" (minute), "h" (hour), "d"
           (day), "w" (week), and "y" (year). There is no "month" multiplier. You can  also  specify  fractional
           numbers, e. g. "0.5d" is identical to "12h".

       Size Size
           Count  only  files  of  a specific size. When Size is a positive number, only files that are at least
           this big are counted. If Size is a negative number, this is inversed, i. e. only files  smaller  than
           the absolute value of Size are counted.

           As  with  the  MTime option, a "multiplier" may be added. For a detailed description see above. Valid
           multipliers here are "b" (byte), "k" (kilobyte), "m" (megabyte), "g" (gigabyte), "t" (terabyte),  and
           "p" (petabyte). Please note that there are 1000 bytes in a kilobyte, not 1024.

       Recursive true|false
           Controls whether or not to recurse into subdirectories. Enabled by default.

       IncludeHidden true|false
           Controls  whether  or not to include "hidden" files and directories in the count.  "Hidden" files and
           directories are those, whose name begins with a dot.  Defaults to false, i.e. by default hidden files
           and directories are ignored.

   Plugin "GenericJMX"
       The GenericJMX plugin is written in Java and therefore documented in collectd-java(5).

   Plugin "gmond"
       The gmond plugin received the multicast traffic sent  by  gmond,  the  statistics  collection  daemon  of
       Ganglia.  Mappings  for  the  standard  "metrics"  are  built-in, custom mappings may be added via Metric
       blocks, see below.

       Synopsis:

        <Plugin "gmond">
          MCReceiveFrom "239.2.11.71" "8649"
          <Metric "swap_total">
            Type "swap"
            TypeInstance "total"
            DataSource "value"
          </Metric>
          <Metric "swap_free">
            Type "swap"
            TypeInstance "free"
            DataSource "value"
          </Metric>
        </Plugin>

       The following metrics are built-in:

       •   load_one, load_five, load_fifteen

       •   cpu_user, cpu_system, cpu_idle, cpu_nice, cpu_wio

       •   mem_free, mem_shared, mem_buffers, mem_cached, mem_total

       •   bytes_in, bytes_out

       •   pkts_in, pkts_out

       Available configuration options:

       MCReceiveFrom MCGroup [Port]
           Sets sets the multicast group and UDP port to which to subscribe.

           Default: 239.2.11.71 / 8649

       <Metric Name>
           These blocks add a new metric conversion to the internal table. Name,  the  string  argument  to  the
           Metric block, is the metric name as used by Ganglia.

           Type Type
               Type to map this metric to. Required.

           TypeInstance Instance
               Type-instance to use. Optional.

           DataSource Name
               Data  source  to  map this metric to. If the configured type has exactly one data source, this is
               optional. Otherwise the option is required.

   Plugin "hddtemp"
       To get values from hddtemp collectd connects to localhost (127.0.0.1), port 7634/tcp. The Host  and  Port
       options  can  be  used  to  change  these  default values, see below. "hddtemp" has to be running to work
       correctly. If "hddtemp" is not running timeouts may appear which may interfere with other statistics..

       The hddtemp homepage can be found at <http://www.guzu.net/linux/hddtemp.php>.

       Host Hostname
           Hostname to connect to. Defaults to 127.0.0.1.

       Port Port
           TCP-Port to connect to. Defaults to 7634.

   Plugin "interface"
       Interface Interface
           Select this interface. By default these interfaces will  then  be  collected.  For  a  more  detailed
           description see IgnoreSelected below.

       IgnoreSelected true|false
           If  no configuration if given, the traffic-plugin will collect data from all interfaces. This may not
           be practical, especially for loopback- and similar interfaces. Thus, you can use the Interface-option
           to pick the interfaces you're interested in. Sometimes, however, it's easier/preferred to collect all
           interfaces except a few ones. This option enables you to do that: By setting IgnoreSelected  to  true
           the effect of Interface is inverted: All selected interfaces are ignored and all other interfaces are
           collected.

   Plugin "ipmi"
       Sensor Sensor
           Selects sensors to collect or to ignore, depending on IgnoreSelected.

       IgnoreSelected true|false
           If  no  configuration  if  given,  the  ipmi  plugin will collect data from all sensors found of type
           "temperature", "voltage", "current" and "fanspeed".  This option enables you to do that:  By  setting
           IgnoreSelected  to  true  the  effect of Sensor is inverted: All selected sensors are ignored and all
           other sensors are collected.

       NotifySensorAdd true|false
           If a sensor appears after initialization time of a minute a notification is sent.

       NotifySensorRemove true|false
           If a sensor disappears a notification is sent.

       NotifySensorNotPresent true|false
           If you have for example dual power supply and one of them is (un)plugged then a notification is sent.

   Plugin "iptables"
       Chain Table Chain [Comment|Number [Name]]
           Select the rules to count. If only Table and Chain are given, this plugin will collect  the  counters
           of all rules which have a comment-match. The comment is then used as type-instance.

           If  Comment  or  Number  is  given,  only  the rule with the matching comment or the nth rule will be
           collected. Again, the comment (or the number) will be used as the type-instance.

           If Name is supplied, it will be used as the type-instance instead of the comment or the number.

   Plugin "irq"
       Irq Irq
           Select this irq. By default these irqs will then be collected. For a more  detailed  description  see
           IgnoreSelected below.

       IgnoreSelected true|false
           If  no  configuration  if  given,  the  irq-plugin  will  collect data from all irqs. This may not be
           practical, especially if no interrupts happen. Thus, you can use the Irq-option to pick the interrupt
           you're interested in.  Sometimes, however, it's easier/preferred to collect all interrupts  except  a
           few  ones. This option enables you to do that: By setting IgnoreSelected to true the effect of Irq is
           inverted: All selected interrupts are ignored and all other interrupts are collected.

   Plugin "java"
       The Java plugin makes it possible to write extensions for collectd in Java.  This section only  discusses
       the  syntax  and semantic of the configuration options. For more in-depth information on the Java plugin,
       please read collectd-java(5).

       Synopsis:

        <Plugin "java">
          JVMArg "-verbose:jni"
          JVMArg "-Djava.class.path=/opt/collectd/lib/collectd/bindings/java"
          LoadPlugin "org.collectd.java.Foobar"
          <Plugin "org.collectd.java.Foobar">
            # To be parsed by the plugin
          </Plugin>
        </Plugin>

       Available configuration options:

       JVMArg Argument
           Argument that is to be passed to the Java Virtual Machine (JVM).  This  works  exactly  the  way  the
           arguments to the java binary on the command line work.  Execute "java --help" for details.

           Please  note  that all these options must appear before (i. e. above) any other options! When another
           option is found, the JVM will be started and later options will have to be ignored!

       LoadPlugin JavaClass
           Instantiates a new JavaClass object. The constructor of this object very likely then registers one or
           more callback methods with the server.

           See collectd-java(5) for details.

           When the first such option is found, the virtual machine (JVM) is created. This means that all JVMArg
           options must appear before (i. e. above) all LoadPlugin options!

       Plugin Name
           The entire block is passed to the Java plugin as an org.collectd.api.OConfigItem object.

           For this to work, the plugin has to register a configuration callback first, see "config callback" in
           collectd-java(5). This means, that the Plugin block must  appear  after  the  appropriate  LoadPlugin
           block.  Also  note, that Name depends on the (Java) plugin registering the callback and is completely
           independent from the JavaClass argument passed to LoadPlugin.

   Plugin "libvirt"
       This plugin allows CPU, disk and network load to be collected for virtualized guests on the machine. This
       means that these characteristics can be collected for guest systems without installing  any  software  on
       them  -  collectd  only  runs  on  the  hosting  system.  The  statistics  are  collected through libvirt
       (<http://libvirt.org/>).

       Only Connection is required.

       Connection uri
           Connect to the hypervisor given by uri. For example if using Xen use:

            Connection "xen:///"

           Details which URIs allowed are given at <http://libvirt.org/uri.html>.

       RefreshInterval seconds
           Refresh the list of domains and devices every seconds. The default is 60 seconds. Setting this to  be
           the  same  or smaller than the Interval will cause the list of domains and devices to be refreshed on
           every iteration.

           Refreshing the devices in particular is quite a costly operation, so if your virtualization setup  is
           static  you  might  consider  increasing  this.  If  this  option is set to 0, refreshing is disabled
           completely.

       Domain name
       BlockDevice name:dev
       InterfaceDevice name:dev
       IgnoreSelected true|false
           Select which domains and devices are collected.

           If IgnoreSelected is not given or false then only the listed domains  and  disk/network  devices  are
           collected.

           If  IgnoreSelected  is true then the test is reversed and the listed domains and disk/network devices
           are ignored, while the rest are collected.

           The domain name and device names may use a regular expression, if the name is surrounded by /.../ and
           collectd was compiled with support for regexps.

           The default is to collect statistics for all domains and all their devices.

           Example:

            BlockDevice "/:hdb/"
            IgnoreSelected "true"

           Ignore all hdb devices on any domain, but other block devices (eg. hda) will be collected.

       HostnameFormat name|uuid|hostname|...
           When the libvirt plugin logs data, it sets the hostname of  the  collected  data  according  to  this
           setting.  The  default  is  to  use  the  guest name as provided by the hypervisor, which is equal to
           setting name.

           uuid means use the guest's UUID. This  is  useful  if  you  want  to  track  the  same  guest  across
           migrations.

           hostname  means  to  use the global Hostname setting, which is probably not useful on its own because
           all guests will appear to have the same name.

           You can also specify combinations of these fields. For example name uuid  means  to  concatenate  the
           guest name and UUID (with a literal colon character between, thus "foo:1234-1234-1234-1234").

       InterfaceFormat name|address
           When the libvirt plugin logs interface data, it sets the name of the collected data according to this
           setting.  The  default  is  to  use the path as provided by the hypervisor (the "dev" property of the
           target node), which is equal to setting name.

           address means use the interface's mac address. This is useful since the interface path  might  change
           between reboots of a guest or across migrations.

   Plugin "logfile"
       LogLevel debug|info|notice|warning|err
           Sets the log-level. If, for example, set to notice, then all events with severity notice, warning, or
           err will be written to the logfile.

           Please note that debug is only available if collectd has been compiled with debugging support.

       File File
           Sets the file to write log messages to. The special strings stdout and stderr can be used to write to
           the standard output and standard error channels, respectively. This, of course, only makes much sense
           when collectd is running in foreground- or non-daemon-mode.

       Timestamp true|false
           Prefix all lines printed by the current time. Defaults to true.

       PrintSeverity true|false
           When  enabled,  all  lines  are  prefixed  by the severity of the log message, for example "warning".
           Defaults to false.

       Note: There is no need to notify the daemon after moving or removing the log file  (e. g.  when  rotating
       the logs). The plugin reopens the file for each line it writes.

   Plugin "lpar"
       The  LPAR  plugin  reads  CPU  statistics of Logical Partitions, a virtualization technique for IBM POWER
       processors. It takes into account CPU time stolen from or donated to a  partition,  in  addition  to  the
       usual user, system, I/O statistics.

       The following configuration options are available:

       CpuPoolStats false|true
           When  enabled,  statistics  about  the processor pool are read, too. The partition needs to have pool
           authority in order to be able to acquire this information.  Defaults to false.

       ReportBySerial false|true
           If enabled, the serial of the physical machine the partition is currently running on is  reported  as
           hostname  and  the logical hostname of the machine is reported in the plugin instance. Otherwise, the
           logical hostname will be used (just like other plugins)  and  the  plugin  instance  will  be  empty.
           Defaults to false.

   Plugin "mbmon"
       The "mbmon plugin" uses mbmon to retrieve temperature, voltage, etc.

       Be  default  collectd  connects  to localhost (127.0.0.1), port 411/tcp. The Host and Port options can be
       used to change these values, see below.  "mbmon" has to be running to work correctly. If "mbmon"  is  not
       running timeouts may appear which may interfere with other statistics..

       "mbmon"  must be run with the -r option ("print TAG and Value format"); Debian's /etc/init.d/mbmon script
       already does this, other people will need to ensure that this is the case.

       Host Hostname
           Hostname to connect to. Defaults to 127.0.0.1.

       Port Port
           TCP-Port to connect to. Defaults to 411.

   Plugin "md"
       The "md plugin" collects information from Linux Software-RAID devices (md).

       All reported values are of the type "md_disks". Reported type instances are active, failed  (present  but
       not operational), spare (hot stand-by) and missing (physically absent) disks.

       Device Device
           Select  md devices based on device name. The device name is the basename of the device, i.e. the name
           of the block device without the leading "/dev/".  See IgnoreSelected for more details.

       IgnoreSelected true|false
           Invert device selection: If set to true,  all  md  devices  except  those  listed  using  Device  are
           collected. If false (the default), only those listed are collected. If no configuration is given, the
           md plugin will collect data from all md devices.

   Plugin "memcachec"
       The  "memcachec  plugin"  connects  to a memcached server, queries one or more given pages and parses the
       returned data according to user specification.  The matches used are the same as the matches used in  the
       "curl" and "tail" plugins.

       In  order  to  talk  to the memcached server, this plugin uses the libmemcached library. Please note that
       there is another library with a very similar name, libmemcache (notice the missing  `d'),  which  is  not
       applicable.

       Synopsis of the configuration:

        <Plugin "memcachec">
          <Page "plugin_instance">
            Server "localhost"
            Key "page_key"
            <Match>
              Regex "(\\d+) bytes sent"
              DSType CounterAdd
              Type "ipt_octets"
              Instance "type_instance"
            </Match>
          </Page>
        </Plugin>

       The configuration options are:

       <Page Name>
           Each  Page  block  defines  one page to be queried from the memcached server.  The block requires one
           string argument which is used as plugin instance.

       Server Address
           Sets the server address to connect to when querying the page. Must be inside a Page block.

       Key Key
           When connected to the memcached server, asks for the page Key.

       <Match>
           Match blocks define which strings to look for and how  matches  substrings  are  interpreted.  For  a
           description of match blocks, please see "Plugin tail".

   Plugin "memcached"
       The  memcached  plugin  connects  to  a  memcached server and queries statistics about cache utilization,
       memory and bandwidth used.  <http://www.danga.com/memcached/>

        <Plugin "memcached">
          <Instance "name">
            Host "memcache.example.com"
            Port 11211
          </Instance>
        </Plugin>

       The plugin configuration consists of one or more Instance blocks which specify one  memcached  connection
       each. Within the Instance blocks, the following options are allowed:

       Host Hostname
           Hostname to connect to. Defaults to 127.0.0.1.

       Port Port
           TCP-Port to connect to. Defaults to 11211.

       Socket Path
           Connect  to  memcached  using  the UNIX domain socket at Path. If this setting is given, the Host and
           Port settings are ignored.

   Plugin "mic"
       The mic plugin gathers CPU statistics, memory usage and temperatures from Intel's  Many  Integrated  Core
       (MIC) systems.

       Synopsis:

        <Plugin mic>
          ShowCPU true
          ShowCPUCores true
          ShowMemory true

          ShowTemperatures true
          Temperature vddg
          Temperature vddq
          IgnoreSelectedTemperature true

          ShowPower true
          Power total0
          Power total1
          IgnoreSelectedPower true
        </Plugin>

       The following options are valid inside the Plugin mic block:

       ShowCPU true|false
           If enabled (the default) a sum of the CPU usage accross all cores is reported.

       ShowCPUCores true|false
           If enabled (the default) per-core CPU usage is reported.

       ShowMemory true|false
           If enabled (the default) the physical memory usage of the MIC system is reported.

       ShowTemperatures true|false
           If enabled (the default) various temperatures of the MIC system are reported.

       Temperature Name
           This  option  controls which temperatures are being reported. Whether matching temperatures are being
           ignored or only matching temperatures are reported depends on the  IgnoreSelectedTemperature  setting
           below. By default all temperatures are reported.

       IgnoreSelectedTemperature false|true
           Controls  the  behavior  of  the  Temperature  setting  above.  If  set  to  false (the default) only
           temperatures matching a Temperature option are reported or, if no Temperature  option  is  specified,
           all  temperatures  are  reported.  If  set  to  true, matching temperatures are ignored and all other
           temperatures are reported.

           Known temperature names are:

           die Die of the CPU

           devmem
               Device Memory

           fin Fan In

           fout
               Fan Out

           vccp
               Voltage ccp

           vddg
               Voltage ddg

           vddq
               Voltage ddq

       ShowPower true|false
           If enabled (the default) various temperatures of the MIC system are reported.

       Power Name
           This option controls which power readings are being reported. Whether  matching  power  readings  are
           being ignored or only matching power readings are reported depends on the IgnoreSelectedPower setting
           below. By default all power readings are reported.

       IgnoreSelectedPower false|true
           Controls  the  behavior of the Power setting above. If set to false (the default) only power readings
           matching a Power option are reported or, if no Power option is  specified,  all  power  readings  are
           reported.  If  set  to  true,  matching  power  readings are ignored and all other power readings are
           reported.

           Known power names are:

           total0
               Total power utilization averaged over Time Window 0 (uWatts).

           total1
               Total power utilization averaged over Time Window 0 (uWatts).

           inst
               Instantaneous power (uWatts).

           imax
               Max instantaneous power (uWatts).

           pcie
               PCI-E connector power (uWatts).

           c2x3
               2x3 connector power (uWatts).

           c2x4
               2x4 connector power (uWatts).

           vccp
               Core rail (uVolts).

           vddg
               Uncore rail (uVolts).

           vddq
               Memory subsystem rail (uVolts).

   Plugin "modbus"
       The modbus plugin connects to a Modbus "slave" via Modbus/TCP and  reads  register  values.  It  supports
       reading  single  registers  (unsigned  16 bit  values), large integer values (unsigned 32 bit values) and
       floating point values (two registers interpreted as IEEE floats in big endian notation).

       Synopsis:

        <Data "voltage-input-1">
          RegisterBase 0
          RegisterType float
          Type voltage
          Instance "input-1"
        </Data>

        <Data "voltage-input-2">
          RegisterBase 2
          RegisterType float
          Type voltage
          Instance "input-2"
        </Data>

        <Host "modbus.example.com">
          Address "192.168.0.42"
          Port    "502"
          Interval 60

          <Slave 1>
            Instance "power-supply"
            Collect  "voltage-input-1"
            Collect  "voltage-input-2"
          </Slave>
        </Host>

       <Data Name> blocks
           Data blocks define a mapping between register numbers and the "types" used by collectd.

           Within <Data /> blocks, the following options are allowed:

           RegisterBase Number
               Configures the base register to read from the device. If the option RegisterType has been set  to
               Uint32  or  Float,  this  and the next register will be read (the register number is increased by
               one).

           RegisterType Int16|Int32|Uint16|Uint32|Float
               Specifies what kind of data is returned by the device. If the type is Int32, Uint32 or Float, two
               16 bit registers will be read and the data is combined into one value. Defaults to Uint16.

           Type Type
               Specifies the "type" (data set) to use when dispatching the value to  collectd.  Currently,  only
               data sets with exactly one data source are supported.

           Instance Instance
               Sets  the  type instance to use when dispatching the value to collectd. If unset, an empty string
               (no type instance) is used.

       <Host Name> blocks
           Host blocks are used to specify to which hosts to connect and what data to read from their  "slaves".
           The string argument Name is used as hostname when dispatching the values to collectd.

           Within <Host /> blocks, the following options are allowed:

           Address Hostname
               Specifies  the node name (the actual network address) used to connect to the host. This may be an
               IP address or a hostname. Please note that the used libmodbus library only supports IPv4  at  the
               moment.

           Port Service
               Specifies  the port used to connect to the host. The port can either be given as a number or as a
               service name. Please note that the Service argument must be a string, even if ports are given  in
               their numerical form. Defaults to "502".

           Interval Interval
               Sets  the  interval (in seconds) in which the values will be collected from this host. By default
               the global Interval setting will be used.

           <Slave ID>
               Over each TCP connection, multiple Modbus devices may be reached. The slave ID is used to specify
               which device should be addressed. For each device you want to query,  one  Slave  block  must  be
               given.

               Within <Slave /> blocks, the following options are allowed:

               Instance Instance
                   Specify  the  plugin  instance  to  use  when dispatching the values to collectd.  By default
                   "slave_ID" is used.

               Collect DataName
                   Specifies which data to retrieve from the device. DataName must be the  same  string  as  the
                   Name  argument  passed to a Data block. You can specify this option multiple times to collect
                   more than one value from a slave. At least one Collect option is mandatory.

   Plugin "mysql"
       The "mysql plugin" requires mysqlclient to be installed. It  connects  to  one  or  more  databases  when
       started  and keeps the connection up as long as possible. When the connection is interrupted for whatever
       reason it will try to re-connect. The plugin will complain loudly in case anything goes wrong.

       This plugin issues the MySQL "SHOW STATUS" / "SHOW GLOBAL STATUS" command and collects information  about
       MySQL  network  traffic,  executed  statements,  requests,  the query cache and threads by evaluating the
       "Bytes_{received,sent}", "Com_*", "Handler_*", "Qcache_*" and "Threads_*" return values. Please refer  to
       the MySQL reference manual, 5.1.6. Server Status Variables for an explanation of these values.

       Optionally,  master  and  slave  statistics  may be collected in a MySQL replication setup. In that case,
       information about the synchronization state of the nodes  are  collected  by  evaluating  the  "Position"
       return  value  of the "SHOW MASTER STATUS" command and the "Seconds_Behind_Master", "Read_Master_Log_Pos"
       and "Exec_Master_Log_Pos" return values of the "SHOW SLAVE  STATUS"  command.  See  the  MySQL  reference
       manual, 12.5.5.21 SHOW MASTER STATUS Syntax and 12.5.5.31 SHOW SLAVE STATUS Syntax for details.

       Synopsis:

         <Plugin mysql>
           <Database foo>
             Host "hostname"
             User "username"
             Password "password"
             Port "3306"
             MasterStats true
           </Database>

           <Database bar>
             Host "localhost"
             Socket "/var/run/mysql/mysqld.sock"
             SlaveStats true
             SlaveNotifications true
           </Database>
         </Plugin>

       A Database block defines one connection to a MySQL database. It accepts a single argument which specifies
       the  name  of  the  database.  None  of  the other options are required. MySQL will use default values as
       documented in the section "mysql_real_connect()" in the MySQL reference manual.

       Host Hostname
           Hostname of the database server. Defaults to localhost.

       User Username
           Username to use when connecting to the database. The user does not have to be granted any  privileges
           (which  is  synonym  to  granting  the  "USAGE"  privilege),  unless you want to collectd replication
           statistics (see MasterStats and SlaveStats below). In this case,  the  user  needs  the  "REPLICATION
           CLIENT" (or "SUPER") privileges. Else, any existing MySQL user will do.

       Password Password
           Password needed to log into the database.

       Database Database
           Select  this  database.  Defaults to no database which is a perfectly reasonable option for what this
           plugin does.

       Port Port
           TCP-port to connect to. The port must be specified in its numeric form, but it must be  passed  as  a
           string nonetheless. For example:

             Port "3306"

           If Host is set to localhost (the default), this setting has no effect.  See the documentation for the
           "mysql_real_connect" function for details.

       Socket Socket
           Specifies the path to the UNIX domain socket of the MySQL server. This option only has any effect, if
           Host  is set to localhost (the default).  Otherwise, use the Port option above. See the documentation
           for the "mysql_real_connect" function for details.

       MasterStats true|false
       SlaveStats true|false
           Enable the collection of master / slave statistics in a replication setup. In order to be able to get
           access to these statistics, the user needs special privileges. See the User documentation above.

       SlaveNotifications true|false
           If enabled, the plugin sends a notification if the replication slave I/O and / or SQL threads are not
           running.

   Plugin "netapp"
       The netapp plugin can collect various performance and capacity information from a NetApp filer using  the
       NetApp API.

       Please  note that NetApp has a wide line of products and a lot of different software versions for each of
       these products. This plugin was developed for a NetApp  FAS3040  running  OnTap  7.2.3P8  and  tested  on
       FAS2050 7.3.1.1L1, FAS3140 7.2.5.1 and FAS3020 7.2.4P9. It should work for most combinations of model and
       software version but it is very hard to test this.  If you have used this plugin with other models and/or
       software version, feel free to send us a mail to tell us about the results, even if it's just a short "It
       works".

       To collect these data collectd will log in to the NetApp via HTTP(S) and HTTP basic authentication.

       Do  not use a regular user for this! Create a special collectd user with just the minimum of capabilities
       needed. The user only needs the "login-http-admin" capability as well as a few more  depending  on  which
       data will be collected.  Required capabilities are documented below.

       Synopsis

        <Plugin "netapp">
          <Host "netapp1.example.com">
           Protocol      "https"
           Address       "10.0.0.1"
           Port          443
           User          "username"
           Password      "aef4Aebe"
           Interval      30

           <WAFL>
             Interval 30
             GetNameCache   true
             GetDirCache    true
             GetBufferCache true
             GetInodeCache  true
           </WAFL>

           <Disks>
             Interval 30
             GetBusy true
           </Disks>

           <VolumePerf>
             Interval 30
             GetIO      "volume0"
             IgnoreSelectedIO      false
             GetOps     "volume0"
             IgnoreSelectedOps     false
             GetLatency "volume0"
             IgnoreSelectedLatency false
           </VolumePerf>

           <VolumeUsage>
             Interval 30
             GetCapacity "vol0"
             GetCapacity "vol1"
             IgnoreSelectedCapacity false
             GetSnapshot "vol1"
             GetSnapshot "vol3"
             IgnoreSelectedSnapshot false
           </VolumeUsage>

           <Quota>
             Interval 60
           </Quota>

           <Snapvault>
             Interval 30
           </Snapvault>

           <System>
             Interval 30
             GetCPULoad     true
             GetInterfaces  true
             GetDiskOps     true
             GetDiskIO      true
           </System>

           <VFiler vfilerA>
             Interval 60

             SnapVault true
             # ...
           </VFiler>
          </Host>
        </Plugin>

       The netapp plugin accepts the following configuration options:

       Host Name
           A  host  block  defines  one  NetApp filer. It will appear in collectd with the name you specify here
           which does not have to be its real name nor its hostname (see the Address option below).

       VFiler Name
           A VFiler block may only be used inside a host block. It accepts all the  same  options  as  the  Host
           block  (except for cascaded VFiler blocks) and will execute all NetApp API commands in the context of
           the specified VFiler(R). It will appear in collectd with the name you specify  here  which  does  not
           have  to  be  its real name. The VFiler name may be specified using the VFilerName option. If this is
           not specified, it will default to the name you specify here.

           The VFiler block inherits all connection related settings from  the  surrounding  Host  block  (which
           appear before the VFiler block) but they may be overwritten inside the VFiler block.

           This  feature  is useful, for example, when using a VFiler as SnapVault target (supported since OnTap
           8.1). In that case, the SnapVault statistics are not available in the host filer (vfiler0)  but  only
           in the respective VFiler context.

       Protocol httpd|http
           The protocol collectd will use to query this host.

           Optional

           Type: string

           Default: https

           Valid options: http, https

       Address Address
           The hostname or IP address of the host.

           Optional

           Type: string

           Default: The "host" block's name.

       Port Port
           The TCP port to connect to on the host.

           Optional

           Type: integer

           Default: 80 for protocol "http", 443 for protocol "https"

       User User
       Password Password
           The username and password to use to login to the NetApp.

           Mandatory

           Type: string

       VFilerName Name
           The  name of the VFiler in which context to execute API commands. If not specified, the name provided
           to the VFiler block will be used instead.

           Optional

           Type: string

           Default: name of the VFiler block

           Note: This option may only be used inside VFiler blocks.

       Interval Interval
           TODO

       The following options decide what kind of data will be collected. You can either use them as a block  and
       fine tune various parameters inside this block, use them as a single statement to just accept all default
       values, or omit it to not collect any data.

       The following options are valid inside all blocks:

       Interval Seconds
           Collect the respective statistics every Seconds seconds. Defaults to the host specific setting.

       The System block

       This will collect various performance data about the whole system.

       Note: To get this data the collectd user needs the "api-perf-object-get-instances" capability.

       Interval Seconds
           Collect disk statistics every Seconds seconds.

       GetCPULoad true|false
           If  you  set  this  option to true the current CPU usage will be read. This will be the average usage
           between all CPUs in your NetApp without any information about individual CPUs.

           Note: These are the same values that the NetApp CLI command "sysstat" returns in the "CPU" field.

           Optional

           Type: boolean

           Default: true

           Result: Two value lists of type "cpu", and type instances "idle" and "system".

       GetInterfaces true|false
           If you set this option to true the current traffic of the network interfaces will be read. This  will
           be  the  total  traffic  over  all interfaces of your NetApp without any information about individual
           interfaces.

           Note: This is the same values that the NetApp CLI command "sysstat" returns in the "Net kB/s" field.

           Or is it?

           Optional

           Type: boolean

           Default: true

           Result: One value list of type "if_octects".

       GetDiskIO true|false
           If you set this option to true the current IO throughput will be read. This will be the total  IO  of
           your NetApp without any information about individual disks, volumes or aggregates.

           Note: This is the same values that the NetApp CLI command "sysstat" returns in the "Disk kB/s" field.

           Optional

           Type: boolean

           Default: true

           Result: One value list of type "disk_octets".

       GetDiskOps true|false
           If  you  set  this  option to true the current number of HTTP, NFS, CIFS, FCP, iSCSI, etc. operations
           will be read. This will be the total number of operations on  your  NetApp  without  any  information
           about individual volumes or aggregates.

           Note:  These  are the same values that the NetApp CLI command "sysstat" returns in the "NFS", "CIFS",
           "HTTP", "FCP" and "iSCSI" fields.

           Optional

           Type: boolean

           Default: true

           Result: A variable number of value lists of type "disk_ops_complex".  Each  type  of  operation  will
           result in one value list with the name of the operation as type instance.

       The WAFL block

       This  will  collect  various  performance  data about the WAFL file system. At the moment this just means
       cache performance.

       Note: To get this data the collectd user needs the "api-perf-object-get-instances" capability.

       Note: The interface to get these values is classified as "Diagnostics" by NetApp. This means that  it  is
       not guaranteed to be stable even between minor releases.

       Interval Seconds
           Collect disk statistics every Seconds seconds.

       GetNameCache true|false
           Optional

           Type: boolean

           Default: true

           Result: One value list of type "cache_ratio" and type instance "name_cache_hit".

       GetDirCache true|false
           Optional

           Type: boolean

           Default: true

           Result: One value list of type "cache_ratio" and type instance "find_dir_hit".

       GetInodeCache true|false
           Optional

           Type: boolean

           Default: true

           Result: One value list of type "cache_ratio" and type instance "inode_cache_hit".

       GetBufferCache true|false
           Note: This is the same value that the NetApp CLI command "sysstat" returns in the "Cache hit" field.

           Optional

           Type: boolean

           Default: true

           Result: One value list of type "cache_ratio" and type instance "buf_hash_hit".

       The Disks block

       This will collect performance data about the individual disks in the NetApp.

       Note: To get this data the collectd user needs the "api-perf-object-get-instances" capability.

       Interval Seconds
           Collect disk statistics every Seconds seconds.

       GetBusy true|false
           If  you  set  this  option to true the busy time of all disks will be calculated and the value of the
           busiest disk in the system will be written.

           Note: This is the same values that the NetApp CLI command "sysstat" returns in the "Disk util" field.
           Probably.

           Optional

           Type: boolean

           Default: true

           Result: One value list of type "percent" and type instance "disk_busy".

       The VolumePerf block

       This will collect various performance data about the individual volumes.

       You can select which data to collect about which volume using the  following  options.  They  follow  the
       standard ignorelist semantic.

       Note: To get this data the collectd user needs the api-perf-object-get-instances capability.

       Interval Seconds
           Collect volume performance data every Seconds seconds.

       GetIO Volume
       GetOps Volume
       GetLatency Volume
           Select  the  given  volume  for IO, operations or latency statistics collection.  The argument is the
           name of the volume without the "/vol/" prefix.

           Since the standard ignorelist functionality is used here, you can use a string  starting  and  ending
           with  a slash to specify regular expression matching: To match the volumes "vol0", "vol2" and "vol7",
           you can use this regular expression:

             GetIO "/^vol[027]$/"

           If no regular expression is specified, an exact match is required. Both, regular and  exact  matching
           are case sensitive.

           If  no  volume  was specified at all for either of the three options, that data will be collected for
           all available volumes.

       IgnoreSelectedIO true|false
       IgnoreSelectedOps true|false
       IgnoreSelectedLatency true|false
           When set to true, the volumes selected for IO, operations or latency statistics  collection  will  be
           ignored and the data will be collected for all other volumes.

           When  set  to false, data will only be collected for the specified volumes and all other volumes will
           be ignored.

           If no volumes have been specified with  the  above  Get*  options,  all  volumes  will  be  collected
           regardless of the IgnoreSelected* option.

           Defaults to false

       The VolumeUsage block

       This will collect capacity data about the individual volumes.

       Note: To get this data the collectd user needs the api-volume-list-info capability.

       Interval Seconds
           Collect volume usage statistics every Seconds seconds.

       GetCapacity VolumeName
           The  current  capacity  of the volume will be collected. This will result in two to four value lists,
           depending on the configuration of the volume. All data sources are of type "df_complex" with the name
           of the volume as plugin_instance.

           There will be type_instances "used" and "free" for the number of used  and  available  bytes  on  the
           volume.  If the volume has some space reserved for snapshots, a type_instance "snap_reserved" will be
           available.  If the volume has SIS enabled, a type_instance "sis_saved" will be available. This is the
           number of bytes saved by the SIS feature.

           Note:  The current NetApp API has a bug that results in this value being reported as a 32 bit number.
           This plugin tries to guess the correct number which works most of  the  time.   If  you  see  strange
           values here, bug NetApp support to fix this.

           Repeat this option to specify multiple volumes.

       IgnoreSelectedCapacity true|false
           Specify  whether  to  collect  only the volumes selected by the GetCapacity option or to ignore those
           volumes. IgnoreSelectedCapacity defaults to false. However, if no GetCapacity option is specified  at
           all, all capacities will be selected anyway.

       GetSnapshot VolumeName
           Select volumes from which to collect snapshot information.

           Usually,  the  space  used  for  snapshots  is  included in the space reported as "used". If snapshot
           information is collected as well, the space used for snapshots is subtracted from the used space.

           To make things even more interesting, it is possible to reserve space to be used  for  snapshots.  If
           the  space  required  for  snapshots  is  less than that reserved space, there is "reserved free" and
           "reserved used" space in addition to "free" and "used". If the space required for  snapshots  exceeds
           the  reserved  space,  that  part  allocated  in the normal space is subtracted from the "used" space
           again.

           Repeat this option to specify multiple volumes.

       IgnoreSelectedSnapshot
           Specify whether to collect only the volumes selected by the GetSnapshot option  or  to  ignore  those
           volumes.  IgnoreSelectedSnapshot defaults to false. However, if no GetSnapshot option is specified at
           all, all capacities will be selected anyway.

       The Quota block

       This will collect (tree) quota statistics (used disk space and number of used files). This  mechanism  is
       useful  to  get  usage  information  for  single  qtrees.   In case the quotas are not used for any other
       purpose, an entry similar to the following in "/etc/quotas" would be sufficient:

         /vol/volA/some_qtree tree - - - - -

       After adding the entry, issue "quota on -w volA" on the NetApp filer.

       Interval Seconds
           Collect SnapVault(R) statistics every Seconds seconds.

       The SnapVault block

       This will collect statistics about the time and traffic of SnapVault(R) transfers.

       Interval Seconds
           Collect SnapVault(R) statistics every Seconds seconds.

   Plugin "netlink"
       The "netlink" plugin uses a netlink socket  to  query  the  Linux  kernel  about  statistics  of  various
       interface and routing aspects.

       Interface Interface
       VerboseInterface Interface
           Instruct  the  plugin  to  collect interface statistics. This is basically the same as the statistics
           provided by the "interface" plugin (see above) but potentially much more detailed.

           When configuring with Interface only the basic statistics will be collected, namely octets,  packets,
           and  errors. These statistics are collected by the "interface" plugin, too, so using both at the same
           time is no benefit.

           When configured with VerboseInterface all counters except the basic ones, so that no data needs to be
           collected twice if you use the "interface" plugin.  This includes dropped packets, received multicast
           packets, collisions and a whole zoo of differentiated RX and TX errors. You  can  try  the  following
           command to get an idea of what awaits you:

             ip -s -s link list

           If Interface is All, all interfaces will be selected.

       QDisc Interface [QDisc]
       Class Interface [Class]
       Filter Interface [Filter]
           Collect the octets and packets that pass a certain qdisc, class or filter.

           QDiscs  and  classes are identified by their type and handle (or classid).  Filters don't necessarily
           have a handle, therefore the parent's handle is used.  The notation used  in  collectd  differs  from
           that  used  in tc(1) in that it doesn't skip the major or minor number if it's zero and doesn't print
           special ids by their name. So, for example, a qdisc may be identified by "pfifo_fast-1:0" even though
           the minor number of all qdiscs is zero and thus not displayed by tc(1).

           If QDisc, Class, or Filter is given without the second argument, i. .e. without  an  identifier,  all
           qdiscs, classes, or filters that are associated with that interface will be collected.

           Since  a  filter itself doesn't necessarily have a handle, the parent's handle is used. This may lead
           to problems when more than one filter is attached to a qdisc or class. This isn't nice, but we  don't
           know how this could be done any better. If you have a idea, please don't hesitate to tell us.

           As with the Interface option you can specify All as the interface, meaning all interfaces.

           Here are some examples to help you understand the above text more easily:

             <Plugin netlink>
               VerboseInterface "All"
               QDisc "eth0" "pfifo_fast-1:0"
               QDisc "ppp0"
               Class "ppp0" "htb-1:10"
               Filter "ppp0" "u32-1:0"
             </Plugin>

       IgnoreSelected
           The behavior is the same as with all other similar plugins: If nothing is selected at all, everything
           is  collected.  If  some things are selected using the options described above, only these statistics
           are collected. If you set IgnoreSelected to true, this behavior  is  inverted,  i. e.  the  specified
           statistics will not be collected.

   Plugin "network"
       The  Network plugin sends data to a remote instance of collectd, receives data from a remote instance, or
       both at the same time. Data which has been received from the network is usually  not  transmitted  again,
       but this can be activated, see the Forward option below.

       The default IPv6 multicast group is "ff18::efc0:4a42". The default IPv4 multicast group is 239.192.74.66.
       The default UDP port is 25826.

       Both,  Server  and Listen can be used as single option or as block. When used as block, given options are
       valid for this socket only. The following example will export the metrics twice: Once  to  an  "internal"
       server (without encryption and signing) and one to an external server (with cryptographic signature):

        <Plugin "network">
          # Export to an internal server
          # (demonstrates usage without additional options)
          Server "collectd.internal.tld"

          # Export to an external server
          # (demonstrates usage with signature options)
          <Server "collectd.external.tld">
            SecurityLevel "sign"
            Username "myhostname"
            Password "ohl0eQue"
          </Server>
        </Plugin>

       <Server Host [Port]>
           The  Server  statement/block  sets  the server to send datagrams to. The statement may occur multiple
           times to send each datagram to multiple destinations.

           The argument Host may be a hostname, an IPv4 address or an IPv6 address. The optional second argument
           specifies a port number or a service name. If not given, the default, 25826, is used.

           The following options are recognized within Server blocks:

           SecurityLevel Encrypt|Sign|None
               Set the security you require for network communication. When the security level has been  set  to
               Encrypt,  data  sent over the network will be encrypted using AES-256. The integrity of encrypted
               packets is ensured using  SHA-1.  When  set  to  Sign,  transmitted  data  is  signed  using  the
               HMAC-SHA-256 message authentication code. When set to None, data is sent without any security.

               This feature is only available if the network plugin was linked with libgcrypt.

           Username Username
               Sets  the  username  to transmit. This is used by the server to lookup the password. See AuthFile
               below. All security levels except None require this setting.

               This feature is only available if the network plugin was linked with libgcrypt.

           Password Password
               Sets a password (shared secret) for this socket. All security levels  except  None  require  this
               setting.

               This feature is only available if the network plugin was linked with libgcrypt.

           Interface Interface name
               Set  the outgoing interface for IP packets. This applies at least to IPv6 packets and if possible
               to IPv4. If this option is  not  applicable,  undefined  or  a  non-existent  interface  name  is
               specified,  the default behavior is to let the kernel choose the appropriate interface. Be warned
               that the manual selection of an interface for unicast traffic is only necessary in rare cases.

       <Listen Host [Port]>
           The Listen statement sets the interfaces to bind to. When multiple statements are  found  the  daemon
           will bind to multiple interfaces.

           The  argument  Host  may  be  a  hostname,  an  IPv4 address or an IPv6 address. If the argument is a
           multicast address the daemon will join that multicast group.  The optional second argument  specifies
           a port number or a service name. If not given, the default, 25826, is used.

           The following options are recognized within "<Listen>" blocks:

           SecurityLevel Encrypt|Sign|None
               Set  the  security you require for network communication. When the security level has been set to
               Encrypt, only encrypted data will be accepted. The integrity  of  encrypted  packets  is  ensured
               using  SHA-1. When set to Sign, only signed and encrypted data is accepted. When set to None, all
               data will be accepted. If an AuthFile option was given (see below), encrypted data  is  decrypted
               if possible.

               This feature is only available if the network plugin was linked with libgcrypt.

           AuthFile Filename
               Sets  a  file  in  which  usernames  are  mapped to passwords. These passwords are used to verify
               signatures and to decrypt encrypted network packets. If SecurityLevel is set  to  None,  this  is
               optional.  If  given,  signed  data  is  verified and encrypted packets are decrypted. Otherwise,
               signed data is accepted without checking the signature and encrypted data  cannot  be  decrypted.
               For the other security levels this option is mandatory.

               The  file  format  is  very  simple: Each line consists of a username followed by a colon and any
               number of spaces followed by the password. To demonstrate, an example file could look like this:

                 user0: foo
                 user1: bar

               Each time a packet is received, the modification time of the file is checked  using  stat(2).  If
               the  file  has  been changed, the contents is re-read. While the file is being read, it is locked
               using fcntl(2).

           Interface Interface name
               Set the incoming interface for IP packets explicitly. This applies at least to IPv6  packets  and
               if possible to IPv4. If this option is not applicable, undefined or a non-existent interface name
               is  specified,  the default behavior is, to let the kernel choose the appropriate interface. Thus
               incoming traffic gets only accepted, if it arrives on the given interface.

       TimeToLive 1-255
           Set the time-to-live of sent packets. This applies to all, unicast and multicast, and IPv4  and  IPv6
           packets.  The  default  is  to not change this value.  That means that multicast packets will be sent
           with a TTL of 1 (one) on most operating systems.

       MaxPacketSize 1024-65535
           Set the maximum size for datagrams received over the  network.  Packets  larger  than  this  will  be
           truncated.  Defaults  to 1452 bytes, which is the maximum payload size that can be transmitted in one
           Ethernet frame using IPv6 / UDP.

           On the server side, this limit should be set to the largest value used on any client.  Likewise,  the
           value on the client must not be larger than the value on the server, or data will be lost.

           Compatibility:  Versions  prior to version 4.8 used a fixed sized buffer of 1024 bytes. Versions 4.8,
           4.9 and 4.10 used a default value of 1024 bytes to avoid problems  when  sending  data  to  an  older
           server.

       Forward true|false
           If  set to true, write packets that were received via the network plugin to the sending sockets. This
           should only be activated when the Listen- and Server-statements differ. Otherwise packets may be send
           multiple times to the same multicast group. While this results in more network traffic than necessary
           it's not a huge problem since the plugin has a duplicate detection, so the values will not loop.

       ReportStats true|false
           The network plugin cannot only receive and send statistics,  it  can  also  create  statistics  about
           itself. Collected data included the number of received and sent octets and packets, the length of the
           receive  queue and the number of values handled. When set to true, the Network plugin will make these
           statistics available. Defaults to false.

   Plugin "nginx"
       This plugin collects the number of connections  and  requests  handled  by  the  "nginx  daemon"  (speak:
       engine X),    a    HTTP    and    mail    server/proxy.   It   queries   the   page   provided   by   the
       "ngx_http_stub_status_module"   module,   which   isn't   compiled   by   default.   Please   refer    to
       <http://wiki.codemongers.com/NginxStubStatusModule>  for more information on how to compile and configure
       nginx and this module.

       The following options are accepted by the "nginx plugin":

       URL http://host/nginx_status
           Sets the URL of the "ngx_http_stub_status_module" output.

       User Username
           Optional user name needed for authentication.

       Password Password
           Optional password needed for authentication.

       VerifyPeer true|false
           Enable or disable peer SSL certificate verification. See <http://curl.haxx.se/docs/sslcerts.html> for
           details. Enabled by default.

       VerifyHost true|false
           Enable or disable peer host name verification. If enabled, the plugin checks if the "Common Name"  or
           a  "Subject  Alternate  Name"  field of the SSL certificate matches the host name provided by the URL
           option. If this identity  check  fails,  the  connection  is  aborted.  Obviously,  only  works  when
           connecting to a SSL enabled server. Enabled by default.

       CACert File
           File  that  holds  one or more SSL certificates. If you want to use HTTPS you will possibly need this
           option. What CA certificates come bundled with "libcurl" and are checked by default  depends  on  the
           distribution you use.

   Plugin "notify_desktop"
       This plugin sends a desktop notification to a notification daemon, as defined in the Desktop Notification
       Specification. To actually display the notifications, notification-daemon is required and collectd has to
       be  able  to  access the X server (i. e., the "DISPLAY" and "XAUTHORITY" environment variables have to be
       set correctly) and the D-Bus message bus.

       The        Desktop        Notification        Specification        can        be         found         at
       <http://www.galago-project.org/specs/notification/>.

       OkayTimeout timeout
       WarningTimeout timeout
       FailureTimeout timeout
           Set  the  timeout,  in milliseconds, after which to expire the notification for "OKAY", "WARNING" and
           "FAILURE" severities respectively. If zero has been specified, the displayed notification will not be
           closed at all - the user has to do so herself. These options default to 5000. If  a  negative  number
           has been specified, the default is used as well.

   Plugin "notify_email"
       The notify_email plugin uses the ESMTP library to send notifications to a configured email address.

       libESMTP is available from <http://www.stafford.uklinux.net/libesmtp/>.

       Available configuration options:

       From Address
           Email address from which the emails should appear to come from.

           Default: "root@localhost"

       Recipient Address
           Configures  the  email  address(es)  to which the notifications should be mailed.  May be repeated to
           send notifications to multiple addresses.

           At least one Recipient must be present for the plugin to work correctly.

       SMTPServer Hostname
           Hostname of the SMTP server to connect to.

           Default: "localhost"

       SMTPPort Port
           TCP port to connect to.

           Default: 25

       SMTPUser Username
           Username for ASMTP authentication. Optional.

       SMTPPassword Password
           Password for ASMTP authentication. Optional.

       Subject Subject
           Subject-template to use when sending emails. There must be exactly  two  string-placeholders  in  the
           subject,  given  in  the  standard  printf(3)  syntax,  i. e. %s. The first will be replaced with the
           severity, the second with the hostname.

           Default: "Collectd notify: %s@%s"

   Plugin "ntpd"
       Host Hostname
           Hostname of the host running ntpd. Defaults to localhost.

       Port Port
           UDP-Port to connect to. Defaults to 123.

       ReverseLookups true|false
           Sets whether or not to perform reverse lookups on peers. Since the name or IP-address may be used  in
           a  filename  it  is  recommended  to disable reverse lookups. The default is to do reverse lookups to
           preserve backwards compatibility, though.

       IncludeUnitID true|false
           When a peer is a refclock, include the unit ID in the type instance.  Defaults to false for  backward
           compatibility.

           If  two  refclock  peers  use  the  same  driver  and  this  is  false,  the plugin will try to write
           simultaneous measurements from both to the same type instance.  This will result in error messages in
           the log and only one set of measurements making it through.

   Plugin "nut"
       UPS upsname@hostname[:port]
           Add a UPS to collect data from. The format is identical to the one accepted by upsc(8).

   Plugin "olsrd"
       The olsrd plugin connects to the TCP port opened by the  txtinfo  plugin  of  the  Optimized  Link  State
       Routing daemon and reads information about the current state of the meshed network.

       The following configuration options are understood:

       Host Host
           Connect to Host. Defaults to "localhost".

       Port Port
           Specifies the port to connect to. This must be a string, even if you give the port as a number rather
           than a service name. Defaults to "2006".

       CollectLinks No|Summary|Detail
           Specifies what information to collect about links, i. e. direct connections of the daemon queried. If
           set to No, no information is collected. If set to Summary, the number of links and the average of all
           link  quality (LQ) and neighbor link quality (NLQ) values is calculated.  If set to Detail LQ and NLQ
           are collected per link.

           Defaults to Detail.

       CollectRoutes No|Summary|Detail
           Specifies what information to collect  about  routes  of  the  daemon  queried.  If  set  to  No,  no
           information  is  collected. If set to Summary, the number of routes and the average metric and ETX is
           calculated. If set to Detail metric and ETX are collected per route.

           Defaults to Summary.

       CollectTopology No|Summary|Detail
           Specifies what information to collect about the global topology. If set  to  No,  no  information  is
           collected. If set to Summary, the number of links in the entire topology and the average link quality
           (LQ) is calculated.  If set to Detail LQ and NLQ are collected for each link in the entire topology.

           Defaults to Summary.

   Plugin "onewire"
       EXPERIMENTAL! See notes below.

       The  "onewire"  plugin  uses  the owcapi library from the owfs project <http://owfs.org/> to read sensors
       connected via the onewire bus.

       Currently only temperature sensors (sensors with the family code 10, e. g. DS1820, DS18S20,  DS1920)  can
       be  read.  If  you  have other sensors you would like to have included, please send a sort request to the
       mailing list.

       Hubs (the DS2409 chips) are working, but read the note, why this plugin is experimental, below.

       Device Device
           Sets the device to read the values from. This can either be a  "real"  hardware  device,  such  as  a
           serial port or an USB port, or the address of the owserver(1) socket, usually localhost:4304.

           Though   the  documentation  claims  to  automatically  recognize  the  given  address  format,  with
           version 2.7p4  we  had  to  specify  the  type  explicitly.  So  with  that  version,  the  following
           configuration worked for us:

             <Plugin onewire>
               Device "-s localhost:4304"
             </Plugin>

           This directive is required and does not have a default value.

       Sensor Sensor
           Selects  sensors  to  collect  or  to  ignore,  depending  on  IgnoreSelected, see below. Sensors are
           specified without the family byte at the beginning, to you'd use "F10FCA000800", and not include  the
           leading 10. family byte and point.

       IgnoreSelected true|false
           If  no  configuration if given, the onewire plugin will collect data from all sensors found. This may
           not be practical, especially if sensors are added and removed  regularly.  Sometimes,  however,  it's
           easier/preferred  to  collect  only specific sensors or all sensors except a few specified ones. This
           option enables you to do that: By setting IgnoreSelected to true the effect of  Sensor  is  inverted:
           All selected interfaces are ignored and all other interfaces are collected.

       Interval Seconds
           Sets  the interval in which all sensors should be read. If not specified, the global Interval setting
           is used.

       EXPERIMENTAL! The "onewire" plugin is experimental, because it doesn't yet work with big setups. It works
       with one sensor being attached to one controller, but as soon as you throw in a couple  more  senors  and
       maybe  a  hub  or two, reading all values will take more than ten seconds (the default interval). We will
       probably add some separate thread for reading the sensors and some cache or something like that, but it's
       not done yet. We will try to maintain backwards compatibility in the future, but we can't promise. So  in
       short:  If  it  works  for  you:  Great!  But  keep  in mind that the config might change, though this is
       unlikely. Oh, and if you want to help improving this plugin, just send a  short  notice  to  the  mailing
       list. Thanks :)

   Plugin "openvpn"
       The  OpenVPN  plugin  reads  a  status  file  maintained  by OpenVPN and gathers traffic statistics about
       connected clients.

       To set up OpenVPN to write to the status file periodically, use the --status  option  of  OpenVPN.  Since
       OpenVPN  can  write  two  different  formats,  you  need to set the required format, too. This is done by
       setting --status-version to 2.

       So, in a nutshell you need:

         openvpn $OTHER_OPTIONS \
           --status "/var/run/openvpn-status" 10 \
           --status-version 2

       Available options:

       StatusFile File
           Specifies the location of the status file.

       ImprovedNamingSchema true|false
           When enabled, the filename of the status file will be  used  as  plugin  instance  and  the  client's
           "common  name"  will  be  used as type instance. This is required when reading multiple status files.
           Enabling this option is recommended, but to maintain backwards compatibility this option is  disabled
           by default.

       CollectCompression true|false
           Sets  whether  or  not  statistics  about  the  compression used by OpenVPN should be collected. This
           information is only available in single mode. Enabled by default.

       CollectIndividualUsers true|false
           Sets whether or not traffic information is collected for each connected client individually.  If  set
           to  false,  currently  no  traffic  data  is collected at all because aggregating this data in a save
           manner is tricky. Defaults to true.

       CollectUserCount true|false
           When enabled, the number of currently connected clients or users is collected.   This  is  especially
           interesting  when  CollectIndividualUsers  is disabled, but can be configured independently from that
           option. Defaults to false.

   Plugin "oracle"
       The "oracle" plugin uses the OracleX Call Interface (OCI) to connect to an OracleX Database and lets  you
       execute  SQL  statements there. It is very similar to the "dbi" plugin, because it was written around the
       same time. See the "dbi" plugin's documentation above for details.

         <Plugin oracle>
           <Query "out_of_stock">
             Statement "SELECT category, COUNT(*) AS value FROM products WHERE in_stock = 0 GROUP BY category"
             <Result>
               Type "gauge"
               # InstancePrefix "foo"
               InstancesFrom "category"
               ValuesFrom "value"
             </Result>
           </Query>
           <Database "product_information">
             ConnectID "db01"
             Username "oracle"
             Password "secret"
             Query "out_of_stock"
           </Database>
         </Plugin>

       Query blocks

       The Query blocks are handled identically to the  Query  blocks  of  the  "dbi"  plugin.  Please  see  its
       documentation above for details on how to specify queries.

       Database blocks

       Database blocks define a connection to a database and which queries should be sent to that database. Each
       database  needs  a  "name" as string argument in the starting tag of the block. This name will be used as
       "PluginInstance" in the values submitted to the daemon. Other than that, that name is not used.

       ConnectID ID
           Defines the "database alias" or "service name" to connect to. Usually, these names are defined in the
           file named "$ORACLE_HOME/network/admin/tnsnames.ora".

       Host Host
           Hostname to use when dispatching values for this database. Defaults to using the global  hostname  of
           the collectd instance.

       Username Username
           Username used for authentication.

       Password Password
           Password used for authentication.

       Query QueryName
           Associates  the  query  named  QueryName with this database connection. The query needs to be defined
           before this statement, i. e. all query blocks you want to refer to must be placed above the  database
           block you want to refer to them from.

   Plugin "perl"
       This  plugin  embeds  a  Perl-interpreter  into  collectd  and provides an interface to collectd's plugin
       system. See collectd-perl(5) for its documentation.

   Plugin "pinba"
       The Pinba plugin receives profiling information from Pinba, an extension for the PHP interpreter. At  the
       end  of  executing a script, i.e. after a PHP-based webpage has been delivered, the extension will send a
       UDP packet containing timing information, peak memory usage and so on. The  plugin  will  wait  for  such
       packets, parse them and account the provided information, which is then dispatched to the daemon once per
       interval.

       Synopsis:

        <Plugin pinba>
          Address "::0"
          Port "30002"
          # Overall statistics for the website.
          <View "www-total">
            Server "www.example.com"
          </View>
          # Statistics for www-a only
          <View "www-a">
            Host "www-a.example.com"
            Server "www.example.com"
          </View>
          # Statistics for www-b only
          <View "www-b">
            Host "www-b.example.com"
            Server "www.example.com"
          </View>
        </Plugin>

       The plugin provides the following configuration options:

       Address Node
           Configures  the  address  used  to  open  a listening socket. By default, plugin will bind to the any
           address "::0".

       Port Service
           Configures the port (service) to bind to. By default the default Pinba port "30002" will be used. The
           option accepts service names in addition to port numbers and thus requires a string argument.

       <View Name> block
           The packets sent by the Pinba extension include the hostname of the server, the server name (the name
           of the virtual host) and the script that was executed.  Using View blocks it is possible to  separate
           the data into multiple groups to get more meaningful statistics. Each packet is added to all matching
           groups, so that a packet may be accounted for more than once.

           Host Host
               Matches  the  hostname  of the system the webserver / script is running on. This will contain the
               result of the gethostname(2) system call. If not configured, all hostnames will be accepted.

           Server Server
               Matches the name of the virtual host, i.e. the contents of the  $_SERVER["SERVER_NAME"]  variable
               when within PHP. If not configured, all server names will be accepted.

           Script Script
               Matches  the  name  of the script name, i.e. the contents of the $_SERVER["SCRIPT_NAME"] variable
               when within PHP. If not configured, all script names will be accepted.

   Plugin "ping"
       The Ping plugin starts a new thread which sends ICMP "ping" packets to the configured hosts  periodically
       and  measures  the  network latency. Whenever the "read" function of the plugin is called, it submits the
       average latency, the standard deviation and the drop rate for each host.

       Available configuration options:

       Host IP-address
           Host to ping periodically. This option may be repeated several times to ping multiple hosts.

       Interval Seconds
           Sets the interval in which to send ICMP echo packets to  the  configured  hosts.   This  is  not  the
           interval  in  which  statistics  are  queries from the plugin but the interval in which the hosts are
           "pinged". Therefore, the setting here should be smaller than or equal to the global Interval setting.
           Fractional times, such as "1.24" are allowed.

           Default: 1.0

       Timeout Seconds
           Time to wait for a response from the host to which an ICMP packet had been sent. If a reply  was  not
           received  after  Seconds  seconds,  the  host is assumed to be down or the packet to be dropped. This
           setting must be smaller than the Interval setting above for the plugin to work correctly.  Fractional
           arguments are accepted.

           Default: 0.9

       TTL 0-255
           Sets the Time-To-Live of generated ICMP packets.

       SourceAddress host
           Sets the source address to use. host may either be a numerical network address or a network hostname.

       Device name
           Sets  the  outgoing  network device to be used. name has to specify an interface name (e. g. "eth0").
           This might not be supported by all operating systems.

       MaxMissed Packets
           Trigger a DNS resolve after the host has not replied to Packets packets.  This  enables  the  use  of
           dynamic DNS services (like dyndns.org) with the ping plugin.

           Default: -1 (disabled)

   Plugin "postgresql"
       The "postgresql" plugin queries statistics from PostgreSQL databases. It keeps a persistent connection to
       all  configured  databases  and  tries to reconnect if the connection has been interrupted. A database is
       configured by specifying a Database block as described below. The default statistics are  collected  from
       PostgreSQL's  statistics  collector  which thus has to be enabled for this plugin to work correctly. This
       should usually be the case by default. See the section  "The  Statistics  Collector"  of  the  PostgreSQL
       Documentation for details.

       By  specifying  custom  database queries using a Query block as described below, you may collect any data
       that is available from some PostgreSQL database. This way, you are able to access statistics of  external
       daemons  which  are  available  in  a PostgreSQL database or use future or special statistics provided by
       PostgreSQL without the need to upgrade your collectd installation.

       Starting with version 5.2, the "postgresql" plugin supports writing data to PostgreSQL databases as well.
       This has been implemented in a generic way. You need to specify an  SQL  statement  which  will  then  be
       executed  by collectd in order to write the data (see below for details). The benefit of that approach is
       that there is no fixed database layout. Rather, the layout may be optimized for the current setup.

       The PostgreSQL Documentation manual can be found at <http://www.postgresql.org/docs/manuals/>.

         <Plugin postgresql>
           <Query magic>
             Statement "SELECT magic FROM wizard WHERE host = $1;"
             Param hostname
             <Result>
               Type gauge
               InstancePrefix "magic"
               ValuesFrom magic
             </Result>
           </Query>

           <Query rt36_tickets>
             Statement "SELECT COUNT(type) AS count, type \
                               FROM (SELECT CASE \
                                            WHEN resolved = 'epoch' THEN 'open' \
                                            ELSE 'resolved' END AS type \
                                            FROM tickets) type \
                               GROUP BY type;"
             <Result>
               Type counter
               InstancePrefix "rt36_tickets"
               InstancesFrom "type"
               ValuesFrom "count"
             </Result>
           </Query>

           <Writer sqlstore>
             Statement "SELECT collectd_insert($1, $2, $3, $4, $5, $6, $7, $8, $9);"
             StoreRates true
           </Writer>

           <Database foo>
             Host "hostname"
             Port "5432"
             User "username"
             Password "secret"
             SSLMode "prefer"
             KRBSrvName "kerberos_service_name"
             Query magic
           </Database>

           <Database bar>
             Interval 300
             Service "service_name"
             Query backend # predefined
             Query rt36_tickets
           </Database>

           <Database qux>
             # ...
             Writer sqlstore
             CommitInterval 10
           </Database>
         </Plugin>

       The Query block defines one database query which may later be used by a database definition. It accepts a
       single mandatory argument which specifies the name of the query. The names of  all  queries  have  to  be
       unique  (see  the  MinVersion  and MaxVersion options below for an exception to this rule). The following
       configuration options are available to define the query:

       In each Query block, there is one or more Result blocks. Result blocks define how to  handle  the  values
       returned from the query. They define which column holds which value and how to dispatch that value to the
       daemon.  Multiple Result blocks may be used to extract multiple values from a single query.

       Statement sql query statement
           Specify  the  sql  query statement which the plugin should execute. The string may contain the tokens
           $1, $2, etc. which are used to reference  the  first,  second,  etc.  parameter.  The  value  of  the
           parameters  is  specified  by  the  Param  configuration option - see below for details. To include a
           literal $ character followed by a number, surround it with single quotes (').

           Any SQL command which may return data (such as "SELECT" or "SHOW") is allowed.  Note,  however,  that
           only  a  single command may be used. Semicolons are allowed as long as a single non-empty command has
           been specified only.

           The returned lines will be handled separately one after another.

       Param hostname|database|username|interval
           Specify the parameters which should be passed to the SQL query. The parameters are referred to in the
           SQL query as $1, $2, etc. in the same order as they appear in the configuration file.  The  value  of
           the parameter is determined depending on the value of the Param option as follows:

           hostname
               The  configured  hostname  of  the  database  connection.  If  a  UNIX domain socket is used, the
               parameter expands to "localhost".

           database
               The name of the database of the current connection.

           instance
               The name of the database plugin instance. See the Instance option of the  database  specification
               below for details.

           username
               The username used to connect to the database.

           interval
               The interval with which this database is queried (as specified by the database specific or global
               Interval options).

           Please note that parameters are only supported by PostgreSQL's protocol version 3 and above which was
           introduced in version 7.4 of PostgreSQL.

       Type type
           The  type  name to be used when dispatching the values. The type describes how to handle the data and
           where to store it. See types.db(5) for more details on types and their configuration. The number  and
           type of values (as selected by the ValuesFrom option) has to match the type of the given name.

           This option is required inside a Result block.

       InstancePrefix prefix
       InstancesFrom column0 [column1 ...]
           Specify  how  to  create the "TypeInstance" for each data set (i. e. line).  InstancePrefix defines a
           static prefix that will be prepended to all type instances. InstancesFrom defines  the  column  names
           whose  values will be used to create the type instance. Multiple values will be joined together using
           the hyphen ("-") as separation character.

           The plugin itself does not check whether or not  all  built  instances  are  different.  It  is  your
           responsibility to assure that each is unique.

           Both options are optional. If none is specified, the type instance will be empty.

       ValuesFrom column0 [column1 ...]
           Names  the  columns whose content is used as the actual data for the data sets that are dispatched to
           the daemon. How many such columns you need is determined by the Type setting as explained  above.  If
           you  specify  too many or not enough columns, the plugin will complain about that and no data will be
           submitted to the daemon.

           The actual data type, as seen by PostgreSQL, is not that important as long as it represents  numbers.
           The  plugin will automatically cast the values to the right type if it know how to do that. For that,
           it uses the strtoll(3) and strtod(3) functions, so anything supported by those functions is supported
           by the plugin as well.

           This option is required inside a Result block and  may  be  specified  multiple  times.  If  multiple
           ValuesFrom options are specified, the columns are read in the given order.

       MinVersion version
       MaxVersion version
           Specify  the  minimum  or  maximum  version  of  PostgreSQL that this query should be used with. Some
           statistics might only be available with certain versions of PostgreSQL. This allows  you  to  specify
           multiple  queries  with the same name but which apply to different versions, thus allowing you to use
           the same configuration in a heterogeneous environment.

           The version has to be specified as the concatenation of the major, minor  and  patch-level  versions,
           each represented as two-decimal-digit numbers. For example, version 8.2.3 will become 80203.

       The   following   predefined   queries   are   available   (the   definitions   can   be   found  in  the
       postgresql_default.conf file which, by default, is available at "prefix/share/collectd/"):

       backends
           This query collects the number of backends, i. e. the number of connected clients.

       transactions
           This query collects the numbers of committed and rolled-back transactions of the user tables.

       queries
           This query  collects  the  numbers  of  various  table  modifications  (i. e.   insertions,  updates,
           deletions) of the user tables.

       query_plans
           This query collects the numbers of various table scans and returned tuples of the user tables.

       table_states
           This query collects the numbers of live and dead rows in the user tables.

       disk_io
           This query collects disk block access counts for user tables.

       disk_usage
           This query collects the on-disk size of the database in bytes.

       In  addition,  the  following  detailed  queries are available by default. Please note that each of those
       queries collects information by table, thus, potentially producing a lot of data.  For  details  see  the
       description of the non-by_table queries above.

       queries_by_table
       query_plans_by_table
       table_states_by_table
       disk_io_by_table

       The  Writer  block defines a PostgreSQL writer backend. It accepts a single mandatory argument specifying
       the name of the writer. This will then be used in the Database specification in  order  to  activate  the
       writer instance. The names of all writers have to be unique. The following options may be specified:

       Statement sql statement
           This  mandatory  option specifies the SQL statement that will be executed for each submitted value. A
           single SQL statement is allowed only. Anything after the first semicolon will be ignored.

           Nine parameters will be passed to the statement and should be specified as tokens $1, $2, through  $9
           in the statement string. The following values are made available through those parameters:

           $1  The timestamp of the queried value as a floating point number.

           $2  The hostname of the queried value.

           $3  The plugin name of the queried value.

           $4  The plugin instance of the queried value. This value may be NULL if there is no plugin instance.

           $5  The type of the queried value (cf. types.db(5)).

           $6  The type instance of the queried value. This value may be NULL if there is no type instance.

           $7  An  array of names for the submitted values (i. e., the name of the data sources of the submitted
               value-list).

           $8  An array of types for the submitted values (i. e., the type of the data sources of the  submitted
               value-list;  "counter", "gauge", ...). Note, that if StoreRates is enabled (which is the default,
               see below), all types will be "gauge".

           $9  An array of the submitted values. The dimensions of the value name and value arrays match.

           In general, it is advisable to create and call a custom function in the PostgreSQL database for  this
           purpose. Any procedural language supported by PostgreSQL will do (see chapter "Server Programming" in
           the PostgreSQL manual for details).

       StoreRates false|true
           If  set  to  true  (the default), convert counter values to rates. If set to false counter values are
           stored as is, i. e. as an increasing integer number.

       The Database block defines one PostgreSQL database for which to collect statistics. It accepts  a  single
       mandatory  argument which specifies the database name. None of the other options are required. PostgreSQL
       will use default values as documented in the section "CONNECTING TO A DATABASE" in the  psql(1)  manpage.
       However,  be  aware  that  those  defaults  may  be influenced by the user collectd is run as and special
       environment variables. See the manpage for details.

       Interval seconds
           Specify the interval with which the database should be queried. The default  is  to  use  the  global
           Interval setting.

       CommitInterval seconds
           This  option  may  be  used  for  database  connections which have "writers" assigned (see above). If
           specified, it causes a writer to put several updates into a single transaction. This transaction will
           last for the specified amount of time. By default,  each  update  will  be  executed  in  a  separate
           transaction.  Each  transaction  generates  a  fair amount of overhead which can, thus, be reduced by
           activating this option. The draw-back is, that data covering the specified amount  of  time  will  be
           lost,  for  example,  if  a  single  statement within the transaction fails or if the database server
           crashes.

       Host hostname
           Specify the hostname or IP of the PostgreSQL server to connect to. If the value begins with a  slash,
           it is interpreted as the directory name in which to look for the UNIX domain socket.

           This  option  is also used to determine the hostname that is associated with a collected data set. If
           it has been omitted or either begins with with a slash or equals localhost it will be  replaced  with
           the global hostname definition of collectd. Any other value will be passed literally to collectd when
           dispatching values. Also see the global Hostname and FQDNLookup options.

       Port port
           Specify the TCP port or the local UNIX domain socket file extension of the server.

       User username
           Specify the username to be used when connecting to the server.

       Password password
           Specify the password to be used when connecting to the server.

       SSLMode disable|allow|prefer|require
           Specify  whether  to  use  an  SSL  connection  when  contacting  the server. The following modes are
           supported:

       Instance name
           Specify the plugin instance name that should be used instead of  the  database  name  (which  is  the
           default,  if this option has not been specified). This allows to query multiple databases of the same
           name on the same host (e.g.  when running multiple database server versions in parallel).

           disable
               Do not use SSL at all.

           allow
               First, try to connect without using SSL. If that fails, try using SSL.

           prefer (default)
               First, try to connect using SSL. If that fails, try without using SSL.

           require
               Use SSL only.

       KRBSrvName kerberos_service_name
           Specify the Kerberos service name to use when authenticating with  Kerberos  5  or  GSSAPI.  See  the
           sections "Kerberos authentication" and "GSSAPI" of the PostgreSQL Documentation for details.

       Service service_name
           Specify  the PostgreSQL service name to use for additional parameters. That service has to be defined
           in pg_service.conf and holds additional  connection  parameters.  See  the  section  "The  Connection
           Service File" in the PostgreSQL Documentation for details.

       Query query
           Specifies a query which should be executed in the context of the database connection. This may be any
           of  the  predefined  or  user-defined queries. If no such option is given, it defaults to "backends",
           "transactions", "queries", "query_plans", "table_states", "disk_io" and "disk_usage" (unless a Writer
           has been specified). Else, the specified queries are used only.

       Writer writer
           Assigns the specified writer backend to the database connection. This causes all collected data to be
           send to the database using the settings defined in the writer configuration (see the section  "FILTER
           CONFIGURATION" below for details on how to selectively send data to certain plugins).

           Each  writer  will  register a flush callback which may be used when having long transactions enabled
           (see the CommitInterval option above). When issuing the FLUSH command (see  collectd-unixsock(5)  for
           details) the current transaction will be committed right away. Two different kinds of flush callbacks
           are available with the "postgresql" plugin:

           postgresql
               Flush all writer backends.

           postgresql-database
               Flush all writers of the specified database only.

   Plugin "powerdns"
       The  "powerdns"  plugin  queries  statistics  from an authoritative PowerDNS nameserver and/or a PowerDNS
       recursor. Since both offer a wide variety of values, many of  which  are  probably  meaningless  to  most
       users,  but  may  be  useful  for  some. So you may chose which values to collect, but if you don't, some
       reasonable defaults will be collected.

         <Plugin "powerdns">
           <Server "server_name">
             Collect "latency"
             Collect "udp-answers" "udp-queries"
             Socket "/var/run/pdns.controlsocket"
           </Server>
           <Recursor "recursor_name">
             Collect "questions"
             Collect "cache-hits" "cache-misses"
             Socket "/var/run/pdns_recursor.controlsocket"
           </Recursor>
           LocalSocket "/opt/collectd/var/run/collectd-powerdns"
         </Plugin>

       Server and Recursor block
           The Server block defines one authoritative server to  query,  the  Recursor  does  the  same  for  an
           recursing  server.  The  possible options in both blocks are the same, though. The argument defines a
           name for the server / recursor and is required.

           Collect Field
               Using the Collect statement you can select which values to collect. Here, you specify the name of
               the values as used by the PowerDNS servers, e. g.  "dlg-only-drops", "answers10-100".

               The method of getting the values differs for Server and Recursor blocks: When querying the server
               a "SHOW *" command is issued in any case, because that's the only way of getting multiple  values
               out  of  the server at once.  collectd then picks out the values you have selected. When querying
               the recursor, a command is generated to query exactly these values. So  if  you  specify  invalid
               fields  when querying the recursor, a syntax error may be returned by the daemon and collectd may
               not collect any values at all.

               If no Collect statement is given, the following Server values will be collected:

               latency
               packetcache-hit
               packetcache-miss
               packetcache-size
               query-cache-hit
               query-cache-miss
               recursing-answers
               recursing-questions
               tcp-answers
               tcp-queries
               udp-answers
               udp-queries

               The following Recursor values will be collected by default:

               noerror-answers
               nxdomain-answers
               servfail-answers
               sys-msec
               user-msec
               qa-latency
               cache-entries
               cache-hits
               cache-misses
               questions

               Please note that up to that point collectd doesn't know what values are available on  the  server
               and  values that are added do not need a change of the mechanism so far. However, the values must
               be mapped to collectd's naming scheme, which is done using a lookup table that  lists  all  known
               values.  If values are added in the future and collectd does not know about them, you will get an
               error much like this:

                 powerdns plugin: submit: Not found in lookup table: foobar = 42

               In this case please file a bug report with the collectd team.

           Socket Path
               Configures the path to the UNIX domain socket to be  used  when  connecting  to  the  daemon.  By
               default  "${localstatedir}/run/pdns.controlsocket"  will  be used for an authoritative server and
               "${localstatedir}/run/pdns_recursor.controlsocket" will be used for the recursor.

       LocalSocket Path
           Querying the recursor is done using UDP. When using UDP over UNIX domain sockets, the  client  socket
           needs  a  name  in  the  file  system, too. You can set this local name to Path using the LocalSocket
           option. The default is "prefix/var/run/collectd-powerdns".

   Plugin "processes"
       Process Name
           Select more detailed statistics of processes matching this name. The statistics collected  for  these
           selected processes are size of the resident segment size (RSS), user- and system-time used, number of
           processes and number of threads, io data (where available) and minor and major pagefaults.

       ProcessMatch name regex
           Similar  to  the  Process option this allows to select more detailed statistics of processes matching
           the specified regex (see regex(7) for details). The statistics of all matching processes  are  summed
           up  and  dispatched  to  the daemon using the specified name as an identifier. This allows to "group"
           several processes together. name must not contain slashes.

   Plugin "protocols"
       Collects a lot of information about various network protocols, such as IP, TCP, UDP, etc.

       Available configuration options:

       Value Selector
           Selects whether or not to select  a  specific  value.  The  string  being  matched  is  of  the  form
           "Protocol:ValueName",  where  Protocol will be used as the plugin instance and ValueName will be used
           as type instance. An example of the string being used would be "Tcp:RetransSegs".

           You can use regular expressions to match a large number of values with just one configuration option.
           To select all "extended" TCP values, you could use the following statement:

             Value "/^TcpExt:/"

           Whether only matched  values  are  selected  or  all  matched  values  are  ignored  depends  on  the
           IgnoreSelected.  By default, only matched values are selected.  If no value is configured at all, all
           values will be selected.

       IgnoreSelected true|false
           If set to true, inverts the selection made by Value, i. e. all matching values will be ignored.

   Plugin "python"
       This plugin embeds a Python-interpreter into collectd and provides  an  interface  to  collectd's  plugin
       system. See collectd-python(5) for its documentation.

   Plugin "routeros"
       The "routeros" plugin connects to a device running RouterOS, the Linux-based operating system for routers
       by  MikroTik.  The  plugin  uses  librouteros  to  connect and reads information about the interfaces and
       wireless connections of the device. The configuration supports querying multiple routers:

         <Plugin "routeros">
           <Router>
             Host "router0.example.com"
             User "collectd"
             Password "secr3t"
             CollectInterface true
             CollectCPULoad true
             CollectMemory true
           </Router>
           <Router>
             Host "router1.example.com"
             User "collectd"
             Password "5ecret"
             CollectInterface true
             CollectRegistrationTable true
             CollectDF true
             CollectDisk true
           </Router>
         </Plugin>

       As you can see above, the configuration of the routeros plugin consists of one or more  <Router>  blocks.
       Within each block, the following options are understood:

       Host Host
           Hostname or IP-address of the router to connect to.

       Port Port
           Port  name  or  port  number used when connecting. If left unspecified, the default will be chosen by
           librouteros, currently "8728". This option expects a string argument, even when a numeric port number
           is given.

       User User
           Use the user name User to authenticate. Defaults to "admin".

       Password Password
           Set the password used to authenticate.

       CollectInterface true|false
           When set to true, interface statistics will be collected for all interfaces present  on  the  device.
           Defaults to false.

       CollectRegistrationTable true|false
           When set to true, information about wireless LAN connections will be collected. Defaults to false.

       CollectCPULoad true|false
           When  set  to  true, information about the CPU usage will be collected. The number is a dimensionless
           value where zero indicates no CPU usage at all.  Defaults to false.

       CollectMemory true|false
           When enabled, the amount of used and free memory will be collected. How used memory is calculated  is
           unknown, for example whether or not caches are counted as used space.  Defaults to false.

       CollectDF true|false
           When enabled, the amount of used and free disk space will be collected.  Defaults to false.

       CollectDisk true|false
           When enabled, the number of sectors written and bad blocks will be collected.  Defaults to false.

   Plugin "redis"
       The Redis plugin connects to one or more Redis servers and gathers information about each server's state.
       For each server there is a Node block which configures the connection parameters for this node.

         <Plugin redis>
           <Node "example">
               Host "localhost"
               Port "6379"
               Timeout 2000
           </Node>
         </Plugin>

       The  information  shown in the synopsis above is the default configuration which is used by the plugin if
       no configuration is present.

       Node Nodename
           The Node block identifies a new Redis node, that is a new Redis instance running in an specified host
           and port. The name for node is a canonical identifier which is used as plugin instance. It is limited
           to 64 characters in length.

       Host Hostname
           The Host option is the hostname or IP-address where the Redis instance is running on.

       Port Port
           The Port option is the TCP port on which the Redis instance accepts  connections.  Either  a  service
           name  of  a  port  number  may  be  given. Please note that numerical port numbers must be given as a
           string, too.

       Password Password
           Use Password to authenticate when connecting to Redis.

       Timeout Timeout in miliseconds
           The Timeout option set the socket timeout for  node  response.  Since  the  Redis  read  function  is
           blocking,  you  should  keep  this value as low as possible. Keep in mind that the sum of all Timeout
           values for all Nodes should be lower than Interval defined globally.

   Plugin "rrdcached"
       The "rrdcached" plugin uses the RRDtool accelerator daemon, rrdcached(1), to store values to RRD files in
       an efficient manner. The combination of the "rrdcached" plugin and the "rrdcached" daemon is very similar
       to the way the "rrdtool" plugin works (see below). The added  abstraction  layer  provides  a  number  of
       benefits, though: Because the cache is not within "collectd" anymore, it does not need to be flushed when
       "collectd"  is  to  be  restarted. This results in much shorter (if any) gaps in graphs, especially under
       heavy load. Also, the "rrdtool" command line utility is aware of the daemon so that it can  flush  values
       to  disk  automatically  when needed. This allows to integrate automated flushing of values into graphing
       solutions much more easily.

       There are disadvantages, though: The daemon may reside on a different host, so it may not be possible for
       "collectd" to create the appropriate RRD files anymore. And even if "rrdcached" runs on the same host, it
       may run in a different base directory, so relative paths may do weird stuff if you're not careful.

       So the recommended configuration is to let "collectd" and "rrdcached" run on the same host, communicating
       via a UNIX domain socket. The DataDir setting should be set to an absolute path, so that a  changed  base
       directory does not result in RRD files being created / expected in the wrong place.

       DaemonAddress Address
           Address  of  the  daemon  as  understood  by  the  "rrdc_connect"  function  of  the RRD library. See
           rrdcached(1) for details. Example:

             <Plugin "rrdcached">
               DaemonAddress "unix:/var/run/rrdcached.sock"
             </Plugin>

       DataDir Directory
           Set the base directory in which the RRD files reside. If this is a relative path, it is  relative  to
           the working base directory of the "rrdcached" daemon!  Use of an absolute path is recommended.

       CreateFiles true|false
           Enables  or  disables  the creation of RRD files. If the daemon is not running locally, or DataDir is
           set to a relative path, this will not work as expected. Default is true.

       CreateFilesAsync false|true
           When enabled, new RRD files are enabled asynchronously, using a separate  thread  that  runs  in  the
           background.  This prevents writes to block, which is a problem especially when many hundreds of files
           need to be created at once. However, since the purpose of creating the files asynchronously is not to
           block until the file is available, values before the file  is  available  will  be  discarded.   When
           disabled (the default) files are created synchronously, blocking for a short while, while the file is
           being written.

       StepSize Seconds
           Force  the  stepsize  of newly created RRD-files. Ideally (and per default) this setting is unset and
           the stepsize is set to the interval in which the data is collected. Do not use this option unless you
           absolutely have to for some reason. Setting this option may cause problems with  the  "snmp  plugin",
           the "exec plugin" or when the daemon is set up to receive data from other hosts.

       HeartBeat Seconds
           Force  the  heartbeat  of  newly  created  RRD-files.  This setting should be unset in which case the
           heartbeat is set to twice the StepSize which should equal the interval in which data is collected. Do
           not set this option unless you have a very good reason to do so.

       RRARows NumRows
           The "rrdtool plugin" calculates the number of PDPs per CDP based on the StepSize, this setting and  a
           timespan. This plugin creates RRD-files with three times five RRAs, i. e. five RRAs with the CFs MIN,
           AVERAGE,  and  MAX.  The five RRAs are optimized for graphs covering one hour, one day, one week, one
           month, and one year.

           So for each timespan, it  calculates  how  many  PDPs  need  to  be  consolidated  into  one  CDP  by
           calculating:
             number of PDPs = timespan / (stepsize * rrarows)

           Bottom line is, set this no smaller than the width of you graphs in pixels. The default is 1200.

       RRATimespan Seconds
           Adds  an RRA-timespan, given in seconds. Use this option multiple times to have more then one RRA. If
           this option is never used, the built-in default of (3600, 86400, 604800, 2678400, 31622400) is used.

           For more information on how RRA-sizes are calculated see RRARows above.

       XFF Factor
           Set the "XFiles Factor". The default is 0.1. If unsure, don't set this option.  Factor must be in the
           range "[0.0-1.0)", i.e. between zero (inclusive) and one (exclusive).

   Plugin "rrdtool"
       You can use the settings StepSize, HeartBeat, RRARows, and XFF to fine-tune your RRD-files.  Please  read
       rrdcreate(1) if you encounter problems using these settings. If you don't want to dive into the depths of
       RRDtool, you can safely ignore these settings.

       DataDir Directory
           Set  the  directory to store RRD files under. By default RRD files are generated beneath the daemon's
           working directory, i.e. the BaseDir.

       CreateFilesAsync false|true
           When enabled, new RRD files are enabled asynchronously, using a separate  thread  that  runs  in  the
           background.  This prevents writes to block, which is a problem especially when many hundreds of files
           need to be created at once. However, since the purpose of creating the files asynchronously is not to
           block until the file is available, values before the file  is  available  will  be  discarded.   When
           disabled (the default) files are created synchronously, blocking for a short while, while the file is
           being written.

       StepSize Seconds
           Force  the  stepsize  of newly created RRD-files. Ideally (and per default) this setting is unset and
           the stepsize is set to the interval in which the data is collected. Do not use this option unless you
           absolutely have to for some reason. Setting this option may cause problems with  the  "snmp  plugin",
           the "exec plugin" or when the daemon is set up to receive data from other hosts.

       HeartBeat Seconds
           Force  the  heartbeat  of  newly  created  RRD-files.  This setting should be unset in which case the
           heartbeat is set to twice the StepSize which should equal the interval in which data is collected. Do
           not set this option unless you have a very good reason to do so.

       RRARows NumRows
           The "rrdtool plugin" calculates the number of PDPs per CDP based on the StepSize, this setting and  a
           timespan.  This plugin creates RRD-files with three times five RRAs, i.e. five RRAs with the CFs MIN,
           AVERAGE, and MAX. The five RRAs are optimized for graphs covering one hour, one day,  one  week,  one
           month, and one year.

           So  for  each  timespan,  it  calculates  how  many  PDPs  need  to  be  consolidated into one CDP by
           calculating:
             number of PDPs = timespan / (stepsize * rrarows)

           Bottom line is, set this no smaller than the width of you graphs in pixels. The default is 1200.

       RRATimespan Seconds
           Adds an RRA-timespan, given in seconds. Use this option multiple times to have more then one RRA.  If
           this option is never used, the built-in default of (3600, 86400, 604800, 2678400, 31622400) is used.

           For more information on how RRA-sizes are calculated see RRARows above.

       XFF Factor
           Set the "XFiles Factor". The default is 0.1. If unsure, don't set this option.  Factor must be in the
           range "[0.0-1.0)", i.e. between zero (inclusive) and one (exclusive).

       CacheFlush Seconds
           When  the "rrdtool" plugin uses a cache (by setting CacheTimeout, see below) it writes all values for
           a certain RRD-file if the oldest value is older than (or equal to) the number of  seconds  specified.
           If  some  RRD-file is not updated anymore for some reason (the computer was shut down, the network is
           broken, etc.) some values may still be in the cache. If CacheFlush is set, then the entire  cache  is
           searched for entries older than CacheTimeout seconds and written to disk every Seconds seconds. Since
           this  is  kind of expensive and does nothing under normal circumstances, this value should not be too
           small.  900 seconds might be a good value, though setting this to 7200 seconds  doesn't  normally  do
           much harm either.

       CacheTimeout Seconds
           If this option is set to a value greater than zero, the "rrdtool plugin" will save values in a cache,
           as  described  above. Writing multiple values at once reduces IO-operations and thus lessens the load
           produced by updating the files.  The trade off is that the graphs kind of "drag behind" and that more
           memory is used.

       WritesPerSecond Updates
           When collecting many statistics with  collectd  and  the  "rrdtool"  plugin,  you  will  run  serious
           performance  problems.  The  CacheFlush  setting  and  the internal update queue assert that collectd
           continues to work just fine even under heavy load, but the system may become  very  unresponsive  and
           slow.  This  is a problem especially if you create graphs from the RRD files on the same machine, for
           example using the "graph.cgi" script included in the "contrib/collection3/" directory.

           This setting is designed for very large setups. Setting this option to a  value  between  25  and  80
           updates  per  second,  depending  on  your  hardware, will leave the server responsive enough to draw
           graphs even while all the cached values are written to disk. Flushed values, i. e.  values  that  are
           forced  to  disk by the FLUSH command, are not effected by this limit. They are still written as fast
           as possible, so that web frontends have up to date data when generating graphs.

           For example: If you have 100,000 RRD files and set WritesPerSecond to 30 updates per second,  writing
           all  values  to  disk  will  take approximately 56 minutes. Together with the flushing ability that's
           integrated into "collection3" you'll end up with a responsive and fast system, up to date graphs  and
           basically a "backup" of your values every hour.

       RandomTimeout Seconds
           When set, the actual timeout for each value is chosen randomly between CacheTimeout-RandomTimeout and
           CacheTimeout+RandomTimeout.  The  intention  is  to  avoid high load situations that appear when many
           values timeout at the same time. This is especially  a  problem  shortly  after  the  daemon  starts,
           because all values were added to the internal cache at roughly the same time.

   Plugin "sensors"
       The Sensors plugin uses lm_sensors to retrieve sensor-values. This means that all the needed modules have
       to  be  loaded  and  lm_sensors  has  to  be  configured  (most likely by editing /etc/sensors.conf. Read
       sensors.conf(5) for details.

       The lm_sensors homepage can be found at <http://secure.netroedge.com/~lm78/>.

       SensorConfigFile File
           Read the lm_sensors configuration from File. When unset (recommended), the library's default will  be
           used.

       Sensor chip-bus-address/type-feature
           Selects  the  name of the sensor which you want to collect or ignore, depending on the IgnoreSelected
           below. For example, the option "Sensor it8712-isa-0290/voltage-in1" will  cause  collectd  to  gather
           data for the voltage sensor in1 of the it8712 on the isa bus at the address 0290.

       IgnoreSelected true|false
           If  no configuration if given, the sensors-plugin will collect data from all sensors. This may not be
           practical, especially for uninteresting sensors.  Thus, you can use the  Sensor-option  to  pick  the
           sensors you're interested in. Sometimes, however, it's easier/preferred to collect all sensors except
           a  few  ones.  This  option  enables  you to do that: By setting IgnoreSelected to true the effect of
           Sensor is inverted: All selected sensors are ignored and all other sensors are collected.

   Plugin "sigrok"
       The sigrok plugin uses libsigrok to retrieve  measurements  from  any  device  supported  by  the  sigrok
       <http://sigrok.org/> project.

       Synopsis

        <Plugin sigrok>
          LogLevel 3
          <Device "AC Voltage">
             Driver "fluke-dmm"
             MinimumInterval 10
             Conn "/dev/ttyUSB2"
          </Device>
          <Device "Sound Level">
             Driver "cem-dt-885x"
             Conn "/dev/ttyUSB1"
          </Device>
        </Plugin>

       LogLevel 0-5
           The  sigrok  logging  level  to pass on to the collectd log, as a number between 0 and 5 (inclusive).
           These levels correspond  to  "None",  "Errors",  "Warnings",  "Informational",  "Debug  "and  "Spew",
           respectively.  The default is 2 ("Warnings"). The sigrok log messages, regardless of their level, are
           always submitted to collectd at its INFO log level.

       <Device Name>
           A  sigrok-supported  device,  uniquely  identified  by  this section's options. The Name is passed to
           collectd as the plugin instance.

       Driver DriverName
           The sigrok driver to use for this device.

       Conn ConnectionSpec
           If the device cannot be auto-discovered, or  more  than  one  might  be  discovered  by  the  driver,
           ConnectionSpec specifies the connection string to the device.  It can be of the form of a device path
           (e.g. "/dev/ttyUSB2"),  or,  in case of a non-serial USB-connected device, the USB VendorID.ProductID
           separated by  a  period  (e.g. 0403.6001).  A  USB  device  can  also  be  specified  as  Bus.Address
           (e.g. 1.41).

       SerialComm SerialSpec
           For serial devices with non-standard port settings, this option can be used to specify them in a form
           understood by sigrok, e.g. "9600/8n1".  This should not be necessary; drivers know how to communicate
           with devices they support.

       MinimumInterval Seconds
           Specifies  the minimum time between measurement dispatches to collectd, in seconds. Since some sigrok
           supported devices can acquire measurements many times per second, it may  be  necessary  to  throttle
           these. For example, the RRD plugin cannot process writes more than once per second.

           The default MinimumInterval is 0, meaning measurements received from the device are always dispatched
           to collectd. When throttled, unused measurements are discarded.

   Plugin "snmp"
       Since  the  configuration  of  the  "snmp  plugin"  is  a little more complicated than other plugins, its
       documentation has been moved to an own manpage, collectd-snmp(5). Please see there for details.

   Plugin "statsd"
       The statsd plugin listens to a UDP socket, reads "events" in the statsd protocol and dispatches rates  or
       other aggregates of these numbers periodically.

       The  plugin implements the Counter, Timer, Gauge and Set types which are dispatched as the collectd types
       "derive", "latency", "gauge" and "objects" respectively.

       The following configuration options are valid:

       Host Host
           Bind to the hostname / address Host. By default, the plugin will bind  to  the  "any"  address,  i.e.
           accept packets sent to any of the hosts addresses.

       Port Port
           UDP port to listen to. This can be either a service name or a port number.  Defaults to 8125.

       DeleteCounters false|true
       DeleteTimers false|true
       DeleteGauges false|true
       DeleteSets false|true
           These  options  control what happens if metrics are not updated in an interval.  If set to False, the
           default, metrics are dispatched unchanged, i.e. the rate of counters and size of sets will  be  zero,
           timers report "NaN" and gauges are unchanged. If set to True, the such metrics are not dispatched and
           removed from the internal cache.

       TimerPercentile Percent
           Calculate  and  dispatch  the configured percentile, i.e. compute the latency, so that Percent of all
           reported timers are smaller than or equal to the computed latency. This is useful for cutting off the
           long tail latency, as it's often done in Service Level Agreements (SLAs).

           If not specified, no percentile is calculated / dispatched.

   Plugin "swap"
       The Swap plugin collects information about used and available swap  space.  On  Linux  and  Solaris,  the
       following options are available:

       ReportByDevice false|true
           Configures  how  to report physical swap devices. If set to false (the default), the summary over all
           swap devices is reported only, i.e. the globally used and available space over all devices.  If  true
           is configured, the used and available space of each device will be reported separately.

           This  option  is  only  available  if the Swap plugin can read "/proc/swaps" (under Linux) or use the
           swapctl(2) mechanism (under Solaris).

       ReportBytes false|true
           When enabled, the swap I/O is reported in bytes. When disabled, the default, swap I/O is reported  in
           pages. This option is available under Linux only.

   Plugin "syslog"
       LogLevel debug|info|notice|warning|err
           Sets the log-level. If, for example, set to notice, then all events with severity notice, warning, or
           err will be submitted to the syslog-daemon.

           Please note that debug is only available if collectd has been compiled with debugging support.

       NotifyLevel OKAY|WARNING|FAILURE
           Controls which notifications should be sent to syslog. The default behaviour is not to send any. Less
           severe  notifications  always imply logging more severe notifications: Setting this to OKAY means all
           notifications will be sent to  syslog,  setting  this  to  WARNING  will  send  WARNING  and  FAILURE
           notifications  but  will  dismiss  OKAY  notifications. Setting this option to FAILURE will only send
           failures to syslog.

   Plugin "table"
       The "table plugin" provides generic means to parse tabular  data  and  dispatch  user  specified  values.
       Values  are selected based on column numbers. For example, this plugin may be used to get values from the
       Linux proc(5) filesystem or CSV (comma separated values) files.

         <Plugin table>
           <Table "/proc/slabinfo">
             Instance "slabinfo"
             Separator " "
             <Result>
               Type gauge
               InstancePrefix "active_objs"
               InstancesFrom 0
               ValuesFrom 1
             </Result>
             <Result>
               Type gauge
               InstancePrefix "objperslab"
               InstancesFrom 0
               ValuesFrom 4
             </Result>
           </Table>
         </Plugin>

       The configuration consists of one or more Table blocks, each of  which  configures  one  file  to  parse.
       Within  each  Table  block, there are one or more Result blocks, which configure which data to select and
       how to interpret it.

       The following options are available inside a Table block:

       Instance instance
           If specified, instance is used as the plugin instance. So, in the  above  example,  the  plugin  name
           "table-slabinfo"  would  be  used.  If  omitted,  the filename of the table is used instead, with all
           special characters replaced with an underscore ("_").

       Separator string
           Any character of string is interpreted as a delimiter between the different columns of the  table.  A
           sequence  of  two  or more contiguous delimiters in the table is considered to be a single delimiter,
           i. e. there cannot be any empty columns. The plugin uses the strtok_r(3) function to parse the  lines
           of a table - see its documentation for more details. This option is mandatory.

           A  horizontal  tab,  newline  and  carriage  return  may  be  specified  by  "\\t",  "\\n"  and "\\r"
           respectively. Please note that the double backslashes  are  required  because  of  collectd's  config
           parsing.

       The following options are available inside a Result block:

       Type type
           Sets  the  type used to dispatch the values to the daemon. Detailed information about types and their
           configuration can be found in types.db(5). This option is mandatory.

       InstancePrefix prefix
           If specified, prepend prefix to the type instance. If  omitted,  only  the  InstancesFrom  option  is
           considered for the type instance.

       InstancesFrom column0 [column1 ...]
           If  specified,  the  content  of the given columns (identified by the column number starting at zero)
           will be used to create the type instance for each row. Multiple values (and the instance prefix) will
           be joined together with dashes (-) as separation  character.  If  omitted,  only  the  InstancePrefix
           option is considered for the type instance.

           The  plugin  itself  does  not  check  whether  or  not  all built instances are different. ItXs your
           responsibility to assure that each is unique.  This  is  especially  true,  if  you  do  not  specify
           InstancesFrom: You have to make sure that the table only contains one row.

           If neither InstancePrefix nor InstancesFrom is given, the type instance will be empty.

       ValuesFrom column0 [column1 ...]
           Specifies  the  columns  (identified by the column numbers starting at zero) whose content is used as
           the actual data for the data sets that are dispatched to the daemon. How many such columns  you  need
           is  determined  by  the Type setting above. If you specify too many or not enough columns, the plugin
           will complain about that and no data will be submitted to the daemon. The plugin uses strtoll(3)  and
           strtod(3) to parse counter and gauge values respectively, so anything supported by those functions is
           supported by the plugin as well. This option is mandatory.

   Plugin "tail"
       The "tail plugin" follows logfiles, just like tail(1) does, parses each line and dispatches found values.
       What  is  matched  can  be  configured  by the user using (extended) regular expressions, as described in
       regex(7).

         <Plugin "tail">
           <File "/var/log/exim4/mainlog">
             Instance "exim"
             <Match>
               Regex "S=([1-9][0-9]*)"
               DSType "CounterAdd"
               Type "ipt_bytes"
               Instance "total"
             </Match>
             <Match>
               Regex "\\<R=local_user\\>"
               ExcludeRegex "\\<R=local_user\\>.*mail_spool defer"
               DSType "CounterInc"
               Type "counter"
               Instance "local_user"
             </Match>
           </File>
         </Plugin>

       The config consists of one or more File blocks, each of which configures one  logfile  to  parse.  Within
       each File block, there are one or more Match blocks, which configure a regular expression to search for.

       The Instance option in the File block may be used to set the plugin instance. So in the above example the
       plugin name "tail-foo" would be used.  This plugin instance is for all Match blocks that follow it, until
       the  next Instance option. This way you can extract several plugin instances from one logfile, handy when
       parsing syslog and the like.

       Each Match block has the following options to describe how the match should be performed:

       Regex regex
           Sets the regular expression to use for matching against a line. The first subexpression has to  match
           something  that  can  be  turned  into a number by strtoll(3) or strtod(3), depending on the value of
           "CounterAdd", see below. Because extended regular expressions are  used,  you  do  not  need  to  use
           backslashes  for  subexpressions!  If  in  doubt,  please  consult regex(7). Due to collectd's config
           parsing you need to escape backslashes, though. So if you want to match literal parentheses you  need
           to do the following:

             Regex "SPAM \\(Score: (-?[0-9]+\\.[0-9]+)\\)"

       ExcludeRegex regex
           Sets  an  optional  regular  expression  to use for excluding lines from the match.  An example which
           excludes all connections from localhost from the match:

             ExcludeRegex "127\\.0\\.0\\.1"

       DSType Type
           Sets how the values are cumulated. Type is one of:

           GaugeAverage
               Calculate the average.

           GaugeMin
               Use the smallest number only.

           GaugeMax
               Use the greatest number only.

           GaugeLast
               Use the last number found.

           CounterSet
           DeriveSet
           AbsoluteSet
               The matched number is a counter. Simply sets the internal counter to this value.  Variants  exist
               for "COUNTER", "DERIVE", and "ABSOLUTE" data sources.

           CounterAdd
           DeriveAdd
               Add  the  matched  value to the internal counter. In case of DeriveAdd, the matched number may be
               negative, which will effectively subtract from the internal counter.

           CounterInc
           DeriveInc
               Increase the internal counter by one. These DSType are the only ones that do not use the  matched
               subexpression,  but  simply  count  the  number  of  matched  lines.  Thus, you may use a regular
               expression without submatch in this case.

           As you'd expect the Gauge* types interpret the submatch as a floating point number, using  strtod(3).
           The  Counter*  and AbsoluteSet types interpret the submatch as an unsigned integer using strtoull(3).
           The Derive* types interpret the submatch  as  a  signed  integer  using  strtoll(3).  CounterInc  and
           DeriveInc do not use the submatch at all and it may be omitted in this case.

       Type Type
           Sets  the  type used to dispatch this value. Detailed information about types and their configuration
           can be found in types.db(5).

       Instance TypeInstance
           This optional setting sets the type instance to use.

   Plugin "tail_csv"
       The tail_csv plugin reads files in the CSV format, e.g. the statistics file written by Snort.

       Synopsis:

        <Plugin "tail_csv">
          <Metric "snort-dropped">
              Type "percent"
              Instance "dropped"
              Index 1
          </Metric>
          <File "/var/log/snort/snort.stats">
              Instance "snort-eth0"
              Interval 600
              Collect "snort-dropped"
          </File>
        </Plugin>

       The configuration consists of one or more Metric blocks that define an index into the  line  of  the  CSV
       file  and  how  this  value is mapped to collectd's internal representation. These are followed by one or
       more Instance blocks which configure which file to read, in which interval and which metrics to extract.

       <Metric Name>
           The Metric block configures a new metric to be extracted from the  statistics  file  and  how  it  is
           mapped  on collectd's data model. The string Name is only used inside the Instance blocks to refer to
           this block, so you can use one Metric block for multiple CSV files.

           Type Type
               Configures which Type to use when dispatching this metric. Types are defined in  the  types.db(5)
               file, see the appropriate manual page for more information on specifying types. Only types with a
               single  data source are supported by the tail_csv plugin. The information whether the value is an
               absolute value (i.e. a "GAUGE") or a rate (i.e. a "DERIVE") is taken from the Type's definition.

           Instance TypeInstance
               If set, TypeInstance is used to populate the type instance field  of  the  created  value  lists.
               Otherwise, no type instance is used.

           ValueFrom Index
               Configure  to  read  the  value  from the field with the zero-based index Index.  If the value is
               parsed as signed integer, unsigned integer or double depends on the Type setting, see above.

       <File Path>
           Each File block represents one CSV file to read. There must be at least one File block but there  can
           be multiple if you have multiple CSV files.

           Instance PluginInstance
               Sets the plugin instance used when dispatching the values.

           Collect Metric
               Specifies  which  Metric to collect. This option must be specified at least once, and you can use
               this option multiple times to specify more than one metric to be extracted  from  this  statistic
               file.

           Interval Seconds
               Configures  the  interval  in  which  to  read values from this instance / file.  Defaults to the
               plugin's default interval.

           TimeFrom Index
               Rather than using the local time when dispatching a value, read the timestamp from the field with
               the zero-based index Index. The value is interpreted as seconds since epoch. The value is  parsed
               as a double and may be factional.

   Plugin "teamspeak2"
       The  "teamspeak2  plugin"  connects to the query port of a teamspeak2 server and polls interesting global
       and virtual server data. The plugin can query only one physical server but unlimited virtual servers. You
       can use the following options to configure it:

       Host hostname/ip
           The hostname or ip which identifies the physical server.  Default: 127.0.0.1

       Port port
           The query port of the physical server. This needs to be a string.  Default: "51234"

       Server port
           This option has to be added once for every virtual server the plugin should query.  If  you  want  to
           query the virtual server on port 8767 this is what the option would look like:

             Server "8767"

           This  option, although numeric, needs to be a string, i. e. you must use quotes around it! If no such
           statement is given only global information will be collected.

   Plugin "ted"
       The TED plugin connects to a device of "The Energy Detective", a device  to  measure  power  consumption.
       These devices are usually connected to a serial (RS232) or USB port. The plugin opens a configured device
       and   tries   to   read   the   current   energy   readings.   For   more   information   on  TED,  visit
       <http://www.theenergydetective.com/>.

       Available configuration options:

       Device Path
           Path to the device on which TED is connected. collectd will need read and write permissions  on  that
           file.

           Default: /dev/ttyUSB0

       Retries Num
           Apparently  reading  from  TED  is not that reliable. You can therefore configure a number of retries
           here. You only configure the retries here, to if you specify zero, one reading will be performed (but
           no retries if that fails); if you specify three, a maximum of four readings are  performed.  Negative
           values are illegal.

           Default: 0

   Plugin "tcpconns"
       The  "tcpconns plugin" counts the number of currently established TCP connections based on the local port
       and/or the remote port. Since there may be a lot of connections the default if to count  all  connections
       with a local port, for which a listening socket is opened. You can use the following options to fine-tune
       the ports you are interested in:

       ListeningPorts true|false
           If this option is set to true, statistics for all local ports for which a listening socket exists are
           collected.  The  default  depends  on  LocalPort  and  RemotePort  (see  below): If no port at all is
           specifically selected, the default is to collect listening ports. If specific  ports  (no  matter  if
           local  or  remote  ports)  are selected, this option defaults to false, i. e. only the selected ports
           will be collected unless this option is set to true specifically.

       LocalPort Port
           Count the connections to a specific local port. This can be used to  see  how  many  connections  are
           handled by a specific daemon, e. g. the mailserver.  You have to specify the port in numeric form, so
           for the mailserver example you'd need to set 25.

       RemotePort Port
           Count  the  connections to a specific remote port. This is useful to see how much a remote service is
           used. This is most useful if you want to know how many connections a  local  service  has  opened  to
           remote  services,  e. g.  how many connections a mail server or news server has to other mail or news
           servers, or how many connections a web proxy holds to web servers. You  have  to  give  the  port  in
           numeric form.

   Plugin "thermal"
       ForceUseProcfs true|false
           By default, the Thermal plugin tries to read the statistics from the Linux "sysfs" interface. If that
           is  not  available,  the plugin falls back to the "procfs" interface. By setting this option to true,
           you can force the plugin to use the latter. This option defaults to false.

       Device Device
           Selects the name of the thermal device that you want to collect or ignore, depending on the value  of
           the IgnoreSelected option. This option may be used multiple times to specify a list of devices.

       IgnoreSelected true|false
           Invert  the  selection:  If  set  to  true,  all  devices except the ones that match the device names
           specified by the Device option are collected. By default only selected devices  are  collected  if  a
           selection is made. If no selection is configured at all, all devices are selected.

   Plugin "threshold"
       The Threshold plugin checks values collected or received by collectd against a configurable threshold and
       issues notifications if values are out of bounds.

       Documentation for this plugin is available in the collectd-threshold(5) manual page.

   Plugin "tokyotyrant"
       The TokyoTyrant plugin connects to a TokyoTyrant server and collects a couple metrics: number of records,
       and database size on disk.

       Host Hostname/IP
           The hostname or ip which identifies the server.  Default: 127.0.0.1

       Port Service/Port
           The  query  port  of  the server. This needs to be a string, even if the port is given in its numeric
           form.  Default: 1978

   Plugin "unixsock"
       SocketFile Path
           Sets the socket-file which is to be created.

       SocketGroup Group
           If running as root change the group of the  UNIX-socket  after  it  has  been  created.  Defaults  to
           collectd.

       SocketPerms Permissions
           Change  the  file  permissions  of the UNIX-socket after it has been created. The permissions must be
           given as a numeric, octal value as you would pass to chmod(1). Defaults to 0770.

       DeleteSocket false|true
           If set to true, delete the socket file before calling bind(2), if a file with the given name  already
           exists.  If collectd crashes a socket file may be left over, preventing the daemon from opening a new
           socket when restarted.  Since this is potentially dangerous, this defaults to false.

   Plugin "uuid"
       This plugin, if loaded, causes the Hostname  to  be  taken  from  the  machine's  UUID.  The  UUID  is  a
       universally  unique  designation  for  the  machine,  usually taken from the machine's BIOS. This is most
       useful if the machine is running in a virtual environment  such  as  Xen,  in  which  case  the  UUID  is
       preserved across shutdowns and migration.

       The following methods are used to find the machine's UUID, in order:

       •   Check /etc/uuid (or UUIDFile).

       •   Check for UUID from HAL (<http://www.freedesktop.org/wiki/Software/hal>) if present.

       •   Check for UUID from "dmidecode" / SMBIOS.

       •   Check for UUID from Xen hypervisor.

       If no UUID can be found then the hostname is not modified.

       UUIDFile Path
           Take the UUID from the given file (default /etc/uuid).

   Plugin "varnish"
       The varnish plugin collects information about Varnish, an HTTP accelerator.

       Synopsis:

        <Plugin "varnish">
          <Instance "example">
            CollectCache       true
            CollectConnections true
            CollectBackend     true
            CollectSHM         true
            CollectESI         false
            CollectFetch       false
            CollectHCB         false
            CollectSMA         false
            CollectSMS         false
            CollectSM          false
            CollectTotals      false
            CollectWorkers     false
          </Instance>
        </Plugin>

       The  configuration  consists  of  one  or  more  <Instance Name>  blocks. Name is the parameter passed to
       "varnishd -n". If left empty, it will collectd statistics from  the  default  "varnishd"  instance  (this
       should work fine in most cases).

       Inside each <Instance> blocks, the following options are recognized:

       CollectCache true|false
           Cache hits and misses. True by default.

       CollectConnections true|false
           Number of client connections received, accepted and dropped. True by default.

       CollectBackend true|false
           Back-end connection statistics, such as successful, reused, and closed connections. True by default.

       CollectSHM true|false
           Statistics  about  the  shared  memory log, a memory region to store log messages which is flushed to
           disk when full. True by default.

       CollectBan true|false
           Statistics about ban operations, such as number of bans added, retired, and number of objects  tested
           against ban operations. Only available with Varnish 3.x. False by default.

       CollectDirectorDNS true|false
           DNS director lookup cache statistics. Only available with Varnish 3.x. False by default.

       CollectESI true|false
           Edge Side Includes (ESI) parse statistics. False by default.

       CollectFetch true|false
           Statistics about fetches (HTTP requests sent to the backend). False by default.

       CollectHCB true|false
           Inserts  and  look-ups in the crit bit tree based hash. Look-ups are divided into locked and unlocked
           look-ups. False by default.

       CollectObjects true|false
           Statistics on cached objects: number of objects expired, nuked (prematurely expired),  saved,  moved,
           etc. False by default.

       CollectPurge true|false
           Statistics  about  purge  operations,  such as number of purges added, retired, and number of objects
           tested against purge operations. Only available with Varnish 2.x. False by default.

       CollectSession true|false
           Client session statistics. Number of past and current sessions, session  herd  and  linger  counters,
           etc. False by default.

       CollectSMA true|false
           malloc  or umem (umem_alloc(3MALLOC) based) storage statistics. The umem storage component is Solaris
           specific. Only available with Varnish 2.x. False by default.

       CollectSMS true|false
           synth (synthetic content) storage statistics. This storage component is used internally  only.  False
           by default.

       CollectSM true|false
           file (memory mapped file) storage statistics. Only available with Varnish 2.x.  False by default.

       CollectStruct true|false
           Current  varnish  internal state statistics. Number of current sessions, objects in cache store, open
           connections to backends (with Varnish 2.x), etc. False by default.

       CollectTotals true|false
           Collects overview counters, such as the number of sessions created, the number of requests and  bytes
           transferred. False by default.

       CollectUptime true|false
           Varnish uptime. False by default.

       CollectVCL true|false
           Number of total (available + discarded) VCL (config files). False by default.

       CollectWorkers true|false
           Collect statistics about worker threads. False by default.

   Plugin "vmem"
       The  "vmem" plugin collects information about the usage of virtual memory.  Since the statistics provided
       by the Linux kernel are very detailed, they are collected very detailed. However, to get all the details,
       you have to switch them on manually. Most people just want an overview over, such as the number of  pages
       read from swap space.

       Verbose true|false
           Enables  verbose  collection  of  information.  This will start collecting page "actions", e. g. page
           allocations, (de)activations, steals and so on.  Part of these statistics are  collected  on  a  "per
           zone" basis.

   Plugin "vserver"
       This plugin doesn't have any options. VServer support is only available for Linux. It cannot yet be found
       in  a vanilla kernel, though. To make use of this plugin you need a kernel that has VServer support built
       in, i. e. you need to apply the patches and  compile  your  own  kernel,  which  will  then  provide  the
       /proc/virtual filesystem that is required by this plugin.

       The VServer homepage can be found at <http://linux-vserver.org/>.

       Note:  The  traffic  collected  by  this plugin accounts for the amount of traffic passing a socket which
       might be a lot less than the actual on-wire traffic (e. g. due to headers  and  retransmission).  If  you
       want  to  collect  on-wire traffic you could, for example, use the logging facilities of iptables to feed
       data for the guest IPs into the iptables plugin.

   Plugin "write_graphite"
       The "write_graphite" plugin writes data to Graphite, an open-source metrics storage and graphing project.
       The plugin connects to Carbon, the data layer of Graphite, via TCP or UDP and sends data  via  the  "line
       based"  protocol  (per default using port 2003). The data will be sent in blocks of at most 1428 bytes to
       minimize the number of network packets.

       Synopsis:

        <Plugin write_graphite>
          <Node "example">
            Host "localhost"
            Port "2003"
            Protocol "udp"
            LogSendErrors true
            Prefix "collectd"
          </Node>
        </Plugin>

       The configuration consists of one or more <Node Name> blocks.  Inside  the  Node  blocks,  the  following
       options are recognized:

       Host Address
           Hostname or address to connect to. Defaults to "localhost".

       Port Service
           Service name or port number to connect to. Defaults to 2003.

       Protocol String
           Protocol to use when connecting to Graphite. Defaults to "tcp".

       LogSendErrors false|true
           If  set  to  true (the default), logs errors when sending data to Graphite.  If set to false, it will
           not log the errors. This is especially useful when using Protocol UDP since many times we want to use
           the "fire-and-forget" approach and logging errors fills syslog with unneeded messages.

       Prefix String
           When set, String is added in front of the host name. Dots and whitespace  are  not  escaped  in  this
           string (see EscapeCharacter below).

       Postfix String
           When  set,  String  is  appended to the host name. Dots and whitespace are not escaped in this string
           (see EscapeCharacter below).

       EscapeCharacter Char
           Carbon uses the dot (".") as escape character and doesn't allow whitespace  in  the  identifier.  The
           EscapeCharacter  option  determines  which  character  dots,  whitespace  and  control characters are
           replaced with. Defaults to underscore ("_").

       StoreRates false|true
           If set to true (the default), convert counter values to rates. If set to  false  counter  values  are
           stored as is, i. e. as an increasing integer number.

       SeparateInstances false|true
           If  set  to  true,  the  plugin  instance  and type instance will be in their own path component, for
           example "host.cpu.0.cpu.idle". If set to false (the default), the plugin  and  plugin  instance  (and
           likewise the type and type instance) are put into one component, for example "host.cpu-0.cpu-idle".

       AlwaysAppendDS false|true
           If  set the true, append the name of the Data Source (DS) to the "metric" identifier. If set to false
           (the default), this is only done when there is more than one DS.

   Plugin "write_mongodb"
       The write_mongodb plugin will send values to MongoDB, a schema-less NoSQL database.

       Synopsis:

        <Plugin "write_mongodb">
          <Node "default">
            Host "localhost"
            Port "27017"
            Timeout 1000
            StoreRates true
          </Node>
        </Plugin>

       The plugin can send values to multiple instances of  MongoDB  by  specifying  one  Node  block  for  each
       instance. Within the Node blocks, the following options are available:

       Host Address
           Hostname or address to connect to. Defaults to "localhost".

       Port Service
           Service name or port number to connect to. Defaults to 27017.

       Timeout Timeout
           Set  the  timeout for each operation on MongoDB to Timeout milliseconds.  Setting this option to zero
           means no timeout, which is the default.

       StoreRates false|true
           If set to true (the default), convert counter values to rates. If set to  false  counter  values  are
           stored as is, i.e. as an increasing integer number.

       Database Database
       User User
       Password Password
           Sets  the  information  used  when  authenticating to a MongoDB database. The fields are optional (in
           which case no authentication is attempted), but if you want to use authentication  all  three  fields
           must be set.

   Plugin "write_http"
       This  output  plugin  submits values to an http server by POST them using the PUTVAL plain-text protocol.
       Each destination you want to post data to needs to have one URL block, within which the  destination  can
       be configured further, for example by specifying authentication data.

       Synopsis:

        <Plugin "write_http">
          <URL "http://example.com/post-collectd">
            User "collectd"
            Password "weCh3ik0"
          </URL>
        </Plugin>

       URL  blocks  need  one  string  argument  which is used as the URL to which data is posted. The following
       options are understood within URL blocks.

       User Username
           Optional user name needed for authentication.

       Password Password
           Optional password needed for authentication.

       VerifyPeer true|false
           Enable or disable peer SSL certificate verification. See <http://curl.haxx.se/docs/sslcerts.html> for
           details. Enabled by default.

       VerifyHost true|false
           Enable or disable peer host name verification. If enabled, the plugin checks if the "Common Name"  or
           a  "Subject  Alternate  Name"  field of the SSL certificate matches the host name provided by the URL
           option. If this identity  check  fails,  the  connection  is  aborted.  Obviously,  only  works  when
           connecting to a SSL enabled server. Enabled by default.

       CACert File
           File  that  holds  one or more SSL certificates. If you want to use HTTPS you will possibly need this
           option. What CA certificates come bundled with "libcurl" and are checked by default  depends  on  the
           distribution you use.

       Format Command|JSON
           Format  of  the  output  to generate. If set to Command, will create output that is understood by the
           Exec and UnixSock plugins. When set to JSON, will create output in  the  JavaScript  Object  Notation
           (JSON).

           Defaults to Command.

       StoreRates true|false
           If  set  to  true,  convert counter values to rates. If set to false (the default) counter values are
           stored as is, i. e. as an increasing integer number.

   Plugin "write_riemann"
       The write_riemann plugin will send values to Riemann,  a  powerfull  stream  aggregation  and  monitoring
       system. The plugin sends Protobuf encoded data to Riemann using UDP packets.

       Synopsis:

        <Plugin "write_riemann">
          <Node "example">
            Host "localhost"
            Port "5555"
            Protocol UDP
            StoreRates true
            AlwaysAppendDS false
            TTLFactor 2.0
          </Node>
          Tag "foobar"
        </Plugin>

       The following options are understood by the write_riemann plugin:

       <Node Name>
           The plugin's configuration consists of one or more Node blocks. Each block is given a unique Name and
           specifies  one  connection  to  an  instance  of  Riemann.  Indise the Node block, the following per-
           connection options are understood:

           Host Address
               Hostname or address to connect to. Defaults to "localhost".

           Port Service
               Service name or port number to connect to. Defaults to 5555.

           Protocol UDP|TCP
               Specify the protocol to use when communicating with Riemann. Defaults to UDP.

           StoreRates true|false
               If set to true (the default), convert counter values to rates. If set to false counter values are
               stored as is, i.e. as an increasing integer number.

               This will be reflected in the "ds_type" tag: If StoreRates is enabled, converted values will have
               "rate" appended to the data source type, e.g.  "ds_type:derive:rate".

           AlwaysAppendDS false|true
               If set the true, append the name of the Data Source (DS) to the "service", i.e. the  field  that,
               together  with  the  "host"  field, uniquely identifies a metric in Riemann. If set to false (the
               default), this is only done when there is more than one DS.

           TTLFactor Factor
               Riemann events have a Time to Live (TTL) which  specifies  how  long  each  event  is  considered
               active.  collectd  populates  this  field  based  on  the  metrics interval setting. This setting
               controls the factor with which the interval is multiplied to set the TTL. The  default  value  is
               2.0.  Unless  you  know exactly what you're doing, you should only increase this setting from its
               default value.

       Tag String
           Add the given string as an additional tag to the metric being sent to Riemann.

THRESHOLD CONFIGURATION

       Starting with version 4.3.0 collectd has support for monitoring. By that we mean that the values are  not
       only  stored or sent somewhere, but that they are judged and, if a problem is recognized, acted upon. The
       only action collectd takes itself is to generate and dispatch a "notification". Plugins can  register  to
       receive notifications and perform appropriate further actions.

       Since  systems  and what you expect them to do differ a lot, you can configure thresholds for your values
       freely. This gives you a lot of flexibility but also a lot of responsibility.

       Every time a value is out of range a notification is dispatched. This means that the idle  percentage  of
       your  CPU  needs  to  be less then the configured threshold only once for a notification to be generated.
       There's no such thing as a moving average or similar - at least not now.

       Also, all values that match a threshold are considered to be relevant or "interesting". As a  consequence
       collectd  will  issue  a  notification  if  they  are  not  received  for Timeout iterations. The Timeout
       configuration option is explained in section "GLOBAL OPTIONS". If, for example, Timeout  is  set  to  "2"
       (the  default)  and  some  hosts sends it's CPU statistics to the server every 60 seconds, a notification
       will be dispatched after about 120 seconds. It may take a little longer because the  timeout  is  checked
       only once each Interval on the server.

       When  a  value  comes  within  range again or is received after it was missing, an "OKAY-notification" is
       dispatched.

       Here is a configuration example to get you started. Read below for more information.

        <Threshold>
          <Type "foo">
            WarningMin    0.00
            WarningMax 1000.00
            FailureMin    0.00
            FailureMax 1200.00
            Invert false
            Instance "bar"
          </Type>

          <Plugin "interface">
            Instance "eth0"
            <Type "if_octets">
              FailureMax 10000000
              DataSource "rx"
            </Type>
          </Plugin>

          <Host "hostname">
            <Type "cpu">
              Instance "idle"
              FailureMin 10
            </Type>

            <Plugin "memory">
              <Type "memory">
                Instance "cached"
                WarningMin 100000000
              </Type>
            </Plugin>
          </Host>
        </Threshold>

       There are basically two types of configuration statements: The "Host", "Plugin", and "Type" blocks select
       the value for which a threshold should be configured. The "Plugin" and "Type"  blocks  may  be  specified
       further using the "Instance" option. You can combine the block by nesting the blocks, though they must be
       nested  in the above order, i. e. "Host" may contain either "Plugin" and "Type" blocks, "Plugin" may only
       contain "Type" blocks and "Type" may not contain other blocks. If multiple blocks apply to the same value
       the most specific block is used.

       The other statements specify the threshold to configure.  They  must  be  included  in  a  "Type"  block.
       Currently the following statements are recognized:

       FailureMax Value
       WarningMax Value
           Sets  the  upper  bound  of  acceptable values. If unset defaults to positive infinity. If a value is
           greater than FailureMax a FAILURE notification  will  be  created.  If  the  value  is  greater  than
           WarningMax but less than (or equal to) FailureMax a WARNING notification will be created.

       FailureMin Value
       WarningMin Value
           Sets the lower bound of acceptable values. If unset defaults to negative infinity. If a value is less
           than  FailureMin  a  FAILURE  notification  will be created. If the value is less than WarningMin but
           greater than (or equal to) FailureMin a WARNING notification will be created.

       DataSource DSName
           Some data sets have more than one "data source". Interesting examples are the "if_octets"  data  set,
           which  has  received ("rx") and sent ("tx") bytes and the "disk_ops" data set, which holds "read" and
           "write" operations. The system load data set, "load",  even  has  three  data  sources:  "shortterm",
           "midterm", and "longterm".

           Normally,  all data sources are checked against a configured threshold. If this is undesirable, or if
           you want to specify different limits for each data source, you can use the DataSource option to  have
           a threshold apply only to one data source.

       Invert true|false
           If  set  to  true  the  range  of acceptable values is inverted, i. e.  values between FailureMin and
           FailureMax (WarningMin and WarningMax) are not okay. Defaults to false.

       Persist true|false
           Sets how often notifications are generated. If set to true one notification  will  be  generated  for
           each  value that is out of the acceptable range. If set to false (the default) then a notification is
           only generated if a value is out of range but the previous value was okay.

           This applies to missing values, too: If set to true a notification about a missing value is generated
           once every Interval seconds. If set to false only one such notification is generated until the  value
           appears again.

       Percentage true|false
           If set to true, the minimum and maximum values given are interpreted as percentage value, relative to
           the  other data sources. This is helpful for example for the "df" type, where you may want to issue a
           warning when less than 5 % of the total space is available. Defaults to false.

       Hits Number
           Delay creating the notification until the threshold has been passed Number times. When a notification
           has been generated, or when a subsequent value is inside the threshold, the counter is reset. If, for
           example, a value is collected once every 10 seconds and Hits is set to  3,  a  notification  will  be
           dispatched at most once every 30 seconds.

           This  is  useful  when  short  bursts  are not a problem. If, for example, 100% CPU usage for up to a
           minute is normal (and data is collected every 10 seconds), you could set Hits to  6  to  account  for
           this.

       Hysteresis Number
           When set to non-zero, a hysteresis value is applied when checking minimum and maximum bounds. This is
           useful  for  values  that  increase slowly and fluctuate a bit while doing so. When these values come
           close to the threshold, they may "flap", i.e. switch between failure / warning  case  and  okay  case
           repeatedly.

           If, for example, the threshold is configures as

             WarningMax 100.0
             Hysteresis 1.0

           then  a  Warning  notification  is  created  when  the  value  exceeds 101 and the corresponding Okay
           notification is only created once the value falls below 99, thus avoiding the "flapping".

FILTER CONFIGURATION

       Starting with collectd 4.6 there is a powerful filtering infrastructure implemented in  the  daemon.  The
       concept  has  mostly  been copied from ip_tables, the packet filter infrastructure for Linux. We'll use a
       similar terminology, so that users that are familiar with iptables feel right at home.

   Terminology
       The following are the terms used in the remainder of  the  filter  configuration  documentation.  For  an
       ASCII-art schema of the mechanism, see "General structure" below.

       Match
           A  match  is  a criteria to select specific values. Examples are, of course, the name of the value or
           it's current value.

           Matches are implemented in plugins which you have to load prior to using the match. The name of  such
           plugins starts with the "match_" prefix.

       Target
           A  target  is  some  action that is to be performed with data. Such actions could, for example, be to
           change part of the value's identifier or to ignore the value completely.

           Some of these targets are built into the daemon, see "Built-in  targets"  below.  Other  targets  are
           implemented  in  plugins  which  you have to load prior to using the target. The name of such plugins
           starts with the "target_" prefix.

       Rule
           The combination of any number of matches and at least one target is called a rule. The target actions
           will be performed for all values for which all matches apply. If the rule does not have  any  matches
           associated with it, the target action will be performed for all values.

       Chain
           A  chain  is  a  list  of rules and possibly default targets. The rules are tried in order and if one
           matches, the associated target will be called. If a value is handled by a rule,  it  depends  on  the
           target  whether  or  not any subsequent rules are considered or if traversal of the chain is aborted,
           see "Flow control" below. After all rules have been checked, the default targets will be executed.

   General structure
       The following shows the resulting structure:

        +---------+
        ! Chain   !
        +---------+
             !
             V
        +---------+  +---------+  +---------+  +---------+
        ! Rule    !->! Match   !->! Match   !->! Target  !
        +---------+  +---------+  +---------+  +---------+
             !
             V
        +---------+  +---------+  +---------+
        ! Rule    !->! Target  !->! Target  !
        +---------+  +---------+  +---------+
             !
             V
             :
             :
             !
             V
        +---------+  +---------+  +---------+
        ! Rule    !->! Match   !->! Target  !
        +---------+  +---------+  +---------+
             !
             V
        +---------+
        ! Default !
        ! Target  !
        +---------+

   Flow control
       There are four ways to control which way a value takes through the filter mechanism:

       jump
           The built-in jump target can be used to "call" another chain, i. e.  process the value  with  another
           chain. When the called chain finishes, usually the next target or rule after the jump is executed.

       stop
           The  stop  condition,  signaled for example by the built-in target stop, causes all processing of the
           value to be stopped immediately.

       return
           Causes processing in the current chain to be aborted, but processing  of  the  value  generally  will
           continue.  This  means  that if the chain was called via Jump, the next target or rule after the jump
           will be executed. If the chain was not called by another chain,  control  will  be  returned  to  the
           daemon and it may pass the value to another chain.

       continue
           Most  targets  will  signal the continue condition, meaning that processing should continue normally.
           There is no special built-in target for this condition.

   Synopsis
       The configuration reflects this structure directly:

        PostCacheChain "PostCache"
        <Chain "PostCache">
          <Rule "ignore_mysql_show">
            <Match "regex">
              Plugin "^mysql$"
              Type "^mysql_command$"
              TypeInstance "^show_"
            </Match>
            <Target "stop">
            </Target>
          </Rule>
          <Target "write">
            Plugin "rrdtool"
          </Target>
        </Chain>

       The above configuration example will ignore all values where the plugin field is  "mysql",  the  type  is
       "mysql_command" and the type instance begins with "show_". All other values will be sent to the "rrdtool"
       write  plugin via the default target of the chain. Since this chain is run after the value has been added
       to the cache, the MySQL "show_*" command statistics will be available via the "unixsock" plugin.

   List of configuration options
       PreCacheChain ChainName
       PostCacheChain ChainName
           Configure the name of the "pre-cache chain" and the "post-cache chain". The argument is the name of a
           chain that should be executed before and/or after the values have been added to the cache.

           To understand the implications, it's important you  know  what  is  going  on  inside  collectd.  The
           following diagram shows how values are passed from the read-plugins to the write-plugins:

              +---------------+
              !  Read-Plugin  !
              +-------+-------+
                      !
            + - - - - V - - - - +
            : +---------------+ :
            : !   Pre-Cache   ! :
            : !     Chain     ! :
            : +-------+-------+ :
            :         !         :
            :         V         :
            : +-------+-------+ :  +---------------+
            : !     Cache     !--->!  Value Cache  !
            : !     insert    ! :  +---+---+-------+
            : +-------+-------+ :      !   !
            :         !   ,------------'   !
            :         V   V     :          V
            : +-------+---+---+ :  +-------+-------+
            : !  Post-Cache   +--->! Write-Plugins !
            : !     Chain     ! :  +---------------+
            : +---------------+ :
            :                   :
            :  dispatch values  :
            + - - - - - - - - - +

           After the values are passed from the "read" plugins to the dispatch functions, the pre-cache chain is
           run  first.  The values are added to the internal cache afterwards. The post-cache chain is run after
           the values have been added to the cache. So why is it such a huge deal if chains are  run  before  or
           after the values have been added to this cache?

           Targets  that change the identifier of a value list should be executed before the values are added to
           the cache, so that the name in the cache matches the name that is used in the  "write"  plugins.  The
           "unixsock"  plugin, too, uses this cache to receive a list of all available values. If you change the
           identifier after the value list has been added to the cache, this may easily lead to  confusion,  but
           it's not forbidden of course.

           The  cache is also used to convert counter values to rates. These rates are, for example, used by the
           "value" match (see below). If you use the rate stored in the cache before the new value is added, you
           will use the old, previous rate. Write plugins may use this rate, too,  see  the  "csv"  plugin,  for
           example.  The "unixsock" plugin uses these rates too, to implement the "GETVAL" command.

           Last  but  not  last,  the  stop  target  makes a difference: If the pre-cache chain returns the stop
           condition, the value will not be added to the cache and the post-cache chain will not be run.

       Chain Name
           Adds a new chain with a certain name. This name can be used to refer to a specific chain, for example
           to jump to it.

           Within the Chain block, there can be Rule blocks and Target blocks.

       Rule [Name]
           Adds a new rule to the current chain. The name of the rule is optional and currently has  no  meaning
           for the daemon.

           Within  the Rule block, there may be any number of Match blocks and there must be at least one Target
           block.

       Match Name
           Adds a match to a Rule block. The name specifies what kind of match should  be  performed.  Available
           matches depend on the plugins that have been loaded.

           The  arguments  inside  the  Match  block  are  passed to the plugin implementing the match, so which
           arguments are valid here depends on the plugin being used.  If you  do  not  need  any  to  pass  any
           arguments to a match, you can use the shorter syntax:

            Match "foobar"

           Which is equivalent to:

            <Match "foobar">
            </Match>

       Target Name
           Add  a  target to a rule or a default target to a chain. The name specifies what kind of target is to
           be added. Which targets are available depends on the plugins being loaded.

           The arguments inside the Target block are passed to the plugin  implementing  the  target,  so  which
           arguments  are  valid  here  depends  on  the  plugin being used.  If you do not need any to pass any
           arguments to a target, you can use the shorter syntax:

            Target "stop"

           This is the same as writing:

            <Target "stop">
            </Target>

   Built-in targets
       The following targets are built into the core daemon and therefore need no plugins to be loaded:

       return
           Signals the "return" condition, see the "Flow control" section above. This causes the  current  chain
           to  stop  processing  the  value  and  returns  control  to the calling chain. The calling chain will
           continue processing targets and rules just after the jump target (see below). This is very similar to
           the RETURN target of iptables, see iptables(8).

           This target does not have any options.

           Example:

            Target "return"

       stop
           Signals the "stop" condition, see the "Flow control" section above. This  causes  processing  of  the
           value to be aborted immediately. This is similar to the DROP target of iptables, see iptables(8).

           This target does not have any options.

           Example:

            Target "stop"

       write
           Sends the value to "write" plugins.

           Available options:

           Plugin Name
               Name  of  the  write  plugin  to which the data should be sent. This option may be given multiple
               times to send the data to more than one write plugin.

           If no plugin is explicitly specified, the values will be sent to all available write plugins.

           Example:

            <Target "write">
              Plugin "rrdtool"
            </Target>

       jump
           Starts processing the rules of another chain, see "Flow control" above. If the end of that  chain  is
           reached,  or  a  stop condition is encountered, processing will continue right after the jump target,
           i. e. with the next target or the next rule. This is  similar  to  the  -j  command  line  option  of
           iptables, see iptables(8).

           Available options:

           Chain Name
               Jumps to the chain Name. This argument is required and may appear only once.

           Example:

            <Target "jump">
              Chain "foobar"
            </Target>

   Available matches
       regex
           Matches a value using regular expressions.

           Available options:

           Host Regex
           Plugin Regex
           PluginInstance Regex
           Type Regex
           TypeInstance Regex
               Match  values where the given regular expressions match the various fields of the identifier of a
               value. If multiple regular expressions are given, all regexen must match for a value to match.

           Invert false|true
               When set to true, the result of the match is inverted, i.e. all value  lists  where  all  regular
               expressions apply are not matched, all other value lists are matched. Defaults to false.

           Example:

            <Match "regex">
              Host "customer[0-9]+"
              Plugin "^foobar$"
            </Match>

       timediff
           Matches values that have a time which differs from the time on the server.

           This  match  is  mainly  intended for servers that receive values over the "network" plugin and write
           them to disk using the "rrdtool" plugin. RRDtool  is  very  sensitive  to  the  timestamp  used  when
           updating  the  RRD  files.  In  particular, the time must be ever increasing. If a misbehaving client
           sends one packet with a timestamp far in the future, all further packets with a correct time will  be
           ignored because of that one packet. What's worse, such corrupted RRD files are hard to fix.

           This  match lets one match all values outside a specified time range (relative to the server's time),
           so you can use the stop target (see below) to ignore the value, for example.

           Available options:

           Future Seconds
               Matches all values that are ahead of the server's time by Seconds or more seconds.  Set  to  zero
               for no limit. Either Future or Past must be non-zero.

           Past Seconds
               Matches  all  values that are behind of the server's time by Seconds or more seconds. Set to zero
               for no limit. Either Future or Past must be non-zero.

           Example:

            <Match "timediff">
              Future  300
              Past   3600
            </Match>

           This example matches all values that are five minutes or more ahead of the server  or  one  hour  (or
           more) lagging behind.

       value
           Matches  the  actual  value  of  data  sources  against given minimum / maximum values. If a data-set
           consists of more than one data-source, all  data-sources  must  match  the  specified  ranges  for  a
           positive match.

           Available options:

           Min Value
               Sets the smallest value which still results in a match. If unset, behaves like negative infinity.

           Max Value
               Sets the largest value which still results in a match. If unset, behaves like positive infinity.

           Invert true|false
               Inverts  the  selection.  If the Min and Max settings result in a match, no-match is returned and
               vice versa. Please note that the Invert setting only effects how Min and Max  are  applied  to  a
               specific value. Especially the DataSource and Satisfy settings (see below) are not inverted.

           DataSource DSName [DSName ...]
               Select one or more of the data sources. If no data source is configured, all data sources will be
               checked.  If  the type handled by the match does not have a data source of the specified name(s),
               this will always result in no match (independent of the Invert setting).

           Satisfy Any|All
               Specifies how checking with several data sources is performed. If set to Any, the match  succeeds
               if  one  of the data sources is in the configured range. If set to All the match only succeeds if
               all data sources are within the configured range. Default is All.

               Usually All is used for positive matches, Any is used for negative matches. This means that  with
               All  you  usually  check  that  all values are in a "good" range, while with Any you check if any
               value is within a "bad" range (or outside the "good" range).

           Either Min or Max, but not both, may be unset.

           Example:

            # Match all values smaller than or equal to 100. Matches only if all data
            # sources are below 100.
            <Match "value">
              Max 100
              Satisfy "All"
            </Match>

            # Match if the value of any data source is outside the range of 0 - 100.
            <Match "value">
              Min   0
              Max 100
              Invert true
              Satisfy "Any"
            </Match>

       empty_counter
           Matches all values with one or more data sources of type COUNTER and where  all  counter  values  are
           zero.  These  counters  usually  never  increased  since  they  started  existing  (and are therefore
           uninteresting), or got reset recently or overflowed and you had really, really bad luck.

           Please keep in mind that ignoring such counters can result  in  confusing  behavior:  Counters  which
           hardly  ever  increase will be zero for long periods of time. If the counter is reset for some reason
           (machine or service restarted, usually), the graph will be empty (NAN) for a long  time.  People  may
           not understand why.

       hashed
           Calculates  a hash value of the host name and matches values according to that hash value. This makes
           it possible to divide all hosts into groups and match only values that are in a specific  group.  The
           intended  use is in load balancing, where you want to handle only part of all data and leave the rest
           for other servers.

           The hashing function used tries to distribute the hosts evenly. First, it calculates  a  32 bit  hash
           value using the characters of the hostname:

             hash_value = 0;
             for (i = 0; host[i] != 0; i++)
               hash_value = (hash_value * 251) + host[i];

           The  constant  251  is a prime number which is supposed to make this hash value more random. The code
           then checks the group for this host according to the Total and Match arguments:

             if ((hash_value % Total) == Match)
               matches;
             else
               does not match;

           Please note that when you set Total to  two  (i. e.  you  have  only  two  groups),  then  the  least
           significant bit of the hash value will be the XOR of all least significant bits in the host name. One
           consequence  is  that when you have two hosts, "server0.example.com" and "server1.example.com", where
           the host name differs in one digit only and the digits differ by one, those hosts will never  end  up
           in the same group.

           Available options:

           Match Match Total
               Divide  the  data  into  Total  groups and match all hosts in group Match as described above. The
               groups are numbered from zero, i. e. Match must be smaller than Total. Total  must  be  at  least
               one, although only values greater than one really do make any sense.

               You can repeat this option to match multiple groups, for example:

                 Match 3 7
                 Match 5 7

               The  above config will divide the data into seven groups and match groups three and five. One use
               would be to keep every value on two hosts so that if one fails the  missing  data  can  later  be
               reconstructed from the second host.

           Example:

            # Operate on the pre-cache chain, so that ignored values are not even in the
            # global cache.
            <Chain "PreCache">
              <Rule>
                <Match "hashed">
                  # Divide all received hosts in seven groups and accept all hosts in
                  # group three.
                  Match 3 7
                </Match>
                # If matched: Return and continue.
                Target "return"
              </Rule>
              # If not matched: Return and stop.
              Target "stop"
            </Chain>

   Available targets
       notification
           Creates and dispatches a notification.

           Available options:

           Message String
               This  required  option  sets  the message of the notification. The following placeholders will be
               replaced by an appropriate value:

               %{host}
               %{plugin}
               %{plugin_instance}
               %{type}
               %{type_instance}
                   These placeholders are replaced by the identifier field of the same name.

               %{ds:name}
                   These placeholders are replaced by a (hopefully) human readable representation of the current
                   rate of this data source. If you changed the instance name (using the set or replace targets,
                   see below), it may not be possible to convert counter values to rates.

               Please note that these placeholders are case sensitive!

           Severity "FAILURE"|"WARNING"|"OKAY"
               Sets the severity of the message. If omitted, the severity "WARNING" is used.

           Example:

             <Target "notification">
               Message "Oops, the %{type_instance} temperature is currently %{ds:value}!"
               Severity "WARNING"
             </Target>

       replace
           Replaces parts of the identifier using regular expressions.

           Available options:

           Host Regex Replacement
           Plugin Regex Replacement
           PluginInstance Regex Replacement
           TypeInstance Regex Replacement
               Match the appropriate field with the given regular expression Regex. If  the  regular  expression
               matches,  that  part  that  matches is replaced with Replacement. If multiple places of the input
               buffer match a given regular expression, only the first occurrence will be replaced.

               You can specify each option multiple times to use multiple regular expressions one after another.

           Example:

            <Target "replace">
              # Replace "example.net" with "example.com"
              Host "\\<example.net\\>" "example.com"

              # Strip "www." from hostnames
              Host "\\<www\\." ""
            </Target>

       set Sets part of the identifier of a value to a given string.

           Available options:

           Host String
           Plugin String
           PluginInstance String
           TypeInstance String
               Set the appropriate field to the given string. The strings for plugin instance and type  instance
               may  be  empty,  the strings for host and plugin may not be empty. It's currently not possible to
               set the type of a value this way.

           Example:

            <Target "set">
              PluginInstance "coretemp"
              TypeInstance "core3"
            </Target>

   Backwards compatibility
       If you use collectd with an old configuration, i. e. one without a Chain block, it will behave as it used
       to. This is equivalent to the following configuration:

        <Chain "PostCache">
          Target "write"
        </Chain>

       If you specify a PostCacheChain, the write target will not be added anywhere and you will  have  to  make
       sure  that  it is called where appropriate. We suggest to add the above snippet as default target to your
       "PostCache" chain.

   Examples
       Ignore all values, where the hostname does not contain a dot, i. e. can't be an FQDN.

        <Chain "PreCache">
          <Rule "no_fqdn">
            <Match "regex">
              Host "^[^\.]*$"
            </Match>
            Target "stop"
          </Rule>
          Target "write"
        </Chain>

SEE ALSO

       collectd(1),   collectd-exec(5),   collectd-perl(5),   collectd-unixsock(5),   types.db(5),   hddtemp(8),
       iptables(8), kstat(3KSTAT), mbmon(1), psql(1), regex(7), rrdtool(1), sensors(1)

AUTHOR

       Florian Forster <octo@verplant.org>

5.4.0.git                                          2013-08-18                                   COLLECTD.CONF(5)