Provided by: ovn-central_2.5.9-0ubuntu0.16.04.3_amd64 bug

NAME

       ovn-sb - OVN_Southbound database schema

       This  database  holds  logical  and  physical  configuration and state for the Open Virtual Network (OVN)
       system  to  support  virtual  network  abstraction.    For   an   introduction   to   OVN,   please   see
       ovn-architecture(7).

       The  OVN  Southbound  database  sits at the center of the OVN architecture.  It is the one component that
       speaks    both    southbound    directly    to    all    the    hypervisors     and     gateways,     via
       ovn-controller/ovn-controller-vtep, and northbound to the Cloud Management System, via ovn-northd:

   Database Structure
       The OVN Southbound database contains three classes of data with different properties, as described in the
       sections below.

     Physical Network (PN) data

       PN  tables  contain information about the chassis nodes in the system.  This contains all the information
       necessary to wire the overlay, such as IP addresses, supported tunnel types, and security keys.

       The amount of PN data is small (O(n) in the number of chassis) and it changes infrequently, so it can  be
       replicated to every chassis.

       The Chassis table comprises the PN tables.

     Logical Network (LN) data

       LN  tables  contain  the  topology  of logical switches and routers, ACLs, firewall rules, and everything
       needed to describe how packets traverse a logical network, represented as  logical  datapath  flows  (see
       Logical Datapath Flows, below).

       LN  data  may be large (O(n) in the number of logical ports, ACL rules, etc.).  Thus, to improve scaling,
       each chassis should receive only data related to logical networks in  which  that  chassis  participates.
       Past  experience shows that in the presence of large logical networks, even finer-grained partitioning of
       data, e.g. designing logical flows so that only the chassis hosting a logical port needs  related  flows,
       pays off scale-wise.  (This is not necessary initially but it is worth bearing in mind in the design.)

       The  LN  is  a  slave  of the cloud management system running northbound of OVN.  That CMS determines the
       entire OVN logical configuration and therefore the LN’s content at any  given  time  is  a  deterministic
       function of the CMS’s configuration, although that happens indirectly via the OVN_Northbound database and
       ovn-northd.

       LN  data  is  likely  to  change  more  quickly  than  PN  data.   This is especially true in a container
       environment where VMs are created and  destroyed  (and  therefore  added  to  and  deleted  from  logical
       switches) quickly.

       Logical_Flow and Multicast_Group contain LN data.

     Bindings data

       Bindings  data  link  logical  and  physical  components.   They  show  the  current placement of logical
       components (such as VMs and VIFs) onto chassis, and map logical entities to  the  values  that  represent
       them in tunnel encapsulations.

       Bindings  change  frequently,  at  least  every  time  a VM powers up or down or migrates, and especially
       quickly in a container environment.  The amount of data per VM (or VIF) is small.

       Each chassis is authoritative about the VMs and VIFs that it hosts at any given time and can  efficiently
       flood that state to a central location, so the consistency needs are minimal.

       The Port_Binding and Datapath_Binding tables contain binding data.

   Common Columns
       Some  tables contain a special column named external_ids.  This column has the same form and purpose each
       place that it appears, so we describe it here to save space later.

              external_ids: map of string-string pairs
                     Key-value pairs for use by the software that manages the  OVN  Southbound  database  rather
                     than  by  ovn-controller/ovn-controller-vtep.   In particular, ovn-northd can use key-value
                     pairs in this column to relate entities in the southbound database to higher-level entities
                     (such as entities in the OVN Northbound database).   Individual  key-value  pairs  in  this
                     column may be documented in some cases to aid in understanding and troubleshooting, but the
                     reader should not mistake such documentation as comprehensive.

TABLE SUMMARY

       The  following  list  summarizes  the purpose of each of the tables in the OVN_Southbound database.  Each
       table is described in more detail on a later page.

       Table     Purpose
       Chassis   Physical Network Hypervisor and Gateway Information
       Encap     Encapsulation Types
       Logical_Flow
                 Logical Network Flows
       Multicast_Group
                 Logical Port Multicast Groups
       Datapath_Binding
                 Physical-Logical Datapath Bindings
       Port_Binding
                 Physical-Logical Port Bindings

Chassis TABLE

       Each row in this table represents a hypervisor or gateway (a chassis) in the physical network (PN).  Each
       chassis, via ovn-controller/ovn-controller-vtep, adds and updates its own row, and keeps a  copy  of  the
       remaining rows to determine how to reach other hypervisors.

       When  a  chassis  shuts  down  gracefully,  it  should remove its own row.  (This is not critical because
       resources hosted on the chassis are equally unreachable regardless of whether the row is present.)  If  a
       chassis  shuts  down  permanently  without  removing its row, some kind of manual or automatic cleanup is
       eventually needed; we can devise a process for that as necessary.

   Summary:
       name                          string (must be unique within table)
       Encapsulation Configuration:
         encaps                      set of 1 or more Encaps
       Gateway Configuration:
         vtep_logical_switches       set of strings

   Details:
       name: string (must be unique within table)
              A chassis name, taken from external_ids:system-id  in  the  Open_vSwitch  database’s  Open_vSwitch
              table.  OVN does not prescribe a particular format for chassis names.

     Encapsulation Configuration:
       OVN uses encapsulation to transmit logical dataplane packets between chassis.

       encaps: set of 1 or more Encaps
              Points  to  supported  encapsulation  configurations to transmit logical dataplane packets to this
              chassis.  Each entry is a Encap record that describes the configuration.

     Gateway Configuration:
       A gateway is a chassis that forwards traffic between the OVN-managed part of  a  logical  network  and  a
       physical  VLAN, extending a tunnel-based logical network into a physical network.  Gateways are typically
       dedicated nodes that do not host VMs and will be controlled by ovn-controller-vtep.

       vtep_logical_switches: set of strings
              Stores all VTEP logical switch names connected by this gateway chassis.   The  Port_Binding  table
              entry  with options:vtep-physical-switch equal Chassis name, and options:vtep-logical-switch value
              in Chassis vtep_logical_switches, will be associated with this Chassis.

Encap TABLE

       The encaps column in the Chassis table refers to rows in this table to  identify  how  OVN  may  transmit
       logical    dataplane    packets    to   this   chassis.    Each   chassis,   via   ovn-controller(8)   or
       ovn-controller-vtep(8), adds and updates its own rows and keeps a copy of the remaining rows to determine
       how to reach other chassis.

   Summary:
       type                          string, one of stt, geneve, or vxlan
       options                       map of string-string pairs
       ip                            string

   Details:
       type: string, one of stt, geneve, or vxlan
              The encapsulation to use to transmit packets to this chassis.  Hypervisors must use either  geneve
              or stt.  Gateways may use vxlan, geneve, or stt.

       options: map of string-string pairs
              Options for configuring the encapsulation, e.g. IPsec parameters when IPsec support is introduced.
              No options are currently defined.

       ip: string
              The IPv4 address of the encapsulation tunnel endpoint.

Logical_Flow TABLE

       Each  row  in this table represents one logical flow.  ovn-northd populates this table with logical flows
       that implement the L2 and L3 topologies specified in the OVN_Northbound database.  Each  hypervisor,  via
       ovn-controller,  translates the logical flows into OpenFlow flows specific to its hypervisor and installs
       them into Open vSwitch.

       Logical flows are expressed in an OVN-specific format, described here.  A logical datapath flow  is  much
       like  an OpenFlow flow, except that the flows are written in terms of logical ports and logical datapaths
       instead of physical ports and physical datapaths.  Translation between logical and physical  flows  helps
       to  ensure  isolation  between  logical  datapaths.   (The  logical  flow abstraction also allows the OVN
       centralized components to do less work, since they do  not  have  to  separately  compute  and  push  out
       physical flows to each chassis.)

       The default action when no flow matches is to drop packets.

       Architectural Logical Life Cycle of a Packet

       This  following  description  focuses  on the life cycle of a packet through a logical datapath, ignoring
       physical details of the implementation.  Please refer to Architectural Physical Life Cycle of a Packet in
       ovn-architecture(7) for the physical information.

       The description here is written as if OVN itself  executes  these  steps,  but  in  fact  OVN  (that  is,
       ovn-controller) programs Open vSwitch, via OpenFlow and OVSDB, to execute them on its behalf.

       At  a  high  level, OVN passes each packet through the logical datapath’s logical ingress pipeline, which
       may output the packet to one or more logical port or logical multicast groups.   For  each  such  logical
       output  port, OVN passes the packet through the datapath’s logical egress pipeline, which may either drop
       the packet or deliver it to the destination.  Between the two pipelines,  outputs  to  logical  multicast
       groups  are  expanded  into  logical  ports,  so that the egress pipeline only processes a single logical
       output port at a time.  Between the two pipelines is also  where,  when  necessary,  OVN  encapsulates  a
       packet in a tunnel (or tunnels) to transmit to remote hypervisors.

       In  more detail, to start, OVN searches the Logical_Flow table for a row with correct logical_datapath, a
       pipeline of ingress, a table_id of 0, and a match that is true for the packet.  If  none  is  found,  OVN
       drops  the packet.  If OVN finds more than one, it chooses the match with the highest priority.  Then OVN
       executes each of the actions specified in the  row’s  actions  column,  in  the  order  specified.   Some
       actions, such as those to modify packet headers, require no further details.  The next and output actions
       are special.

       The  next  action  causes  the  above  process  to  be repeated recursively, except that OVN searches for
       table_id of 1 instead of 0.  Similarly, any next action in a row  found  in  that  table  would  cause  a
       further search for a table_id of 2, and so on.  When recursive processing completes, flow control returns
       to the action following next.

       The  output  action  also  introduces  recursion.  Its effect depends on the current value of the outport
       field.  Suppose outport designates a logical port.  First, OVN compares inport to outport;  if  they  are
       equal,  it treats the output as a no-op.  In the common case, where they are different, the packet enters
       the egress pipeline.  This transition to the egress pipeline discards register data, e.g. reg0 ...   reg4
       and  connection  tracking state, to achieve uniform behavior regardless of whether the egress pipeline is
       on a different hypervisor (because registers aren’t preserve across tunnel encapsulation).

       To execute the egress pipeline, OVN again  searches  the  Logical_Flow  table  for  a  row  with  correct
       logical_datapath, a table_id of 0, a match that is true for the packet, but now looking for a pipeline of
       egress.   If  no  matching row is found, the output becomes a no-op.  Otherwise, OVN executes the actions
       for the matching flow (which is chosen from multiple, if necessary, as already described).

       In the egress pipeline, the next action acts as already described, except that it,  of  course,  searches
       for  egress flows.  The output action, however, now directly outputs the packet to the output port (which
       is now fixed, because outport is read-only within the egress pipeline).

       The description earlier assumed that outport referred to a logical port.   If  it  instead  designates  a
       logical  multicast group, then the description above still applies, with the addition of fan-out from the
       logical multicast group to each logical port in the group.  For each member of the  group,  OVN  executes
       the logical pipeline as described, with the logical output port replaced by the group member.

       Pipeline Stages

       ovn-northd  is  responsible  for  populating  the Logical_Flow table, so the stages are an implementation
       detail and subject to change.  This section describes the current logical flow table.

       The ingress pipeline consists of the following stages:

              •      Port Security (Table 0): Validates the source address, drops packets with a VLAN tag,  and,
                     if configured, verifies that the logical port is allowed to send with the source address.

              •      L2  Destination  Lookup  (Table  1):  Forwards  known  unicast addresses to the appropriate
                     logical port.  Unicast packets to unknown hosts are forwarded to logical  ports  configured
                     with the special unknown mac address.  Broadcast, and multicast are flooded to all ports in
                     the logical switch.

       The egress pipeline consists of the following stages:

              •      ACL (Table 0): Applies any specified access control lists.

              •      Port  Security  (Table  1):  If  configured,  verifies  that the logical port is allowed to
                     receive packets with the destination address.

   Summary:
       logical_datapath              Datapath_Binding
       pipeline                      string, either ingress or egress
       table_id                      integer, in range 0 to 15
       priority                      integer, in range 0 to 65,535
       match                         string
       actions                       string
       external_ids : stage-name     optional string
       Common Columns:
         external_ids                map of string-string pairs

   Details:
       logical_datapath: Datapath_Binding
              The logical datapath to which the logical flow belongs.

       pipeline: string, either ingress or egress
              The primary flows used for deciding on a packet’s destination are the ingress flows.   The  egress
              flows implement ACLs.  See Logical Life Cycle of a Packet, above, for details.

       table_id: integer, in range 0 to 15
              The stage in the logical pipeline, analogous to an OpenFlow table number.

       priority: integer, in range 0 to 65,535
              The  flow’s  priority.   Flows  with  numerically  higher priority take precedence over those with
              lower.  If two logical datapath flows with the same priority both match,  then  the  one  actually
              applied to the packet is undefined.

       match: string
              A  matching expression.  OVN provides a superset of OpenFlow matching capabilities, using a syntax
              similar to Boolean expressions in a programming language.

              The most important components of match expression are comparisons between symbols  and  constants,
              e.g.  ip4.dst  ==  192.168.0.1,  ip.proto  ==  6, arp.op == 1, eth.type == 0x800.  The logical AND
              operator && and logical OR operator || can combine comparisons into a larger expression.

              Matching expressions also support parentheses for grouping, the logical NOT prefix operator !, and
              literals 0 and 1 to express ``false’’ or ``true,’’ respectively.  The latter is useful  by  itself
              as a catch-all expression that matches every packet.

              Symbols

              Type.  Symbols have integer or string type.  Integer symbols have a width in bits.

              Kinds.  There are three kinds of symbols:

              •      Fields.  A field symbol represents a packet header or metadata field.  For example, a field
                     named vlan.tci might represent the VLAN TCI field in a packet.

                     A  field  symbol can have integer or string type.  Integer fields can be nominal or ordinal
                     (see Level of Measurement, below).

              •      Subfields.  A subfield represents a subset of bits from a larger  field.   For  example,  a
                     field  vlan.vid  might  be defined as an alias for vlan.tci[0..11].  Subfields are provided
                     for syntactic convenience, because it is always possible to instead refer to  a  subset  of
                     bits from a field directly.

                     Only  ordinal  fields  (see Level of Measurement, below) may have subfields.  Subfields are
                     always ordinal.

              •      Predicates.  A predicate is shorthand for a Boolean expression.   Predicates  may  be  used
                     much  like  1-bit  fields.  For example, ip4 might expand to eth.type == 0x800.  Predicates
                     are provided for syntactic convenience, because it is always possible  to  instead  specify
                     the underlying expression directly.

                     A  predicate  whose  expansion  refers  to  any  nominal  field  or predicate (see Level of
                     Measurement, below) is nominal; other predicates have Boolean level of measurement.

              Level of Measurement.  See http://en.wikipedia.org/wiki/Level_of_measurement for  the  statistical
              concept on which this classification is based.  There are three levels:

              •      Ordinal.   In  statistics, ordinal values can be ordered on a scale.  OVN considers a field
                     (or subfield) to be ordinal if its bits can be examined individually.  This is true for the
                     OpenFlow fields that OpenFlow or Open vSwitch makes ``maskable.’’

                     Any use  of  a  ordinal  field  may  specify  a  single  bit  or  a  range  of  bits,  e.g.
                     vlan.tci[13..15] refers to the PCP field within the VLAN TCI, and eth.dst[40] refers to the
                     multicast bit in the Ethernet destination address.

                     OVN  supports  all  the  usual  arithmetic  relations (==, !=, <, <=, >, and >=) on ordinal
                     fields and their subfields, because OVN can implement these in OpenFlow and Open vSwitch as
                     collections of bitwise tests.

              •      Nominal.  In statistics, nominal values cannot be usefully compared  except  for  equality.
                     This  is  true of OpenFlow port numbers, Ethernet types, and IP protocols are examples: all
                     of these are just identifiers assigned arbitrarily with no deeper meaning.  In OpenFlow and
                     Open vSwitch, bits in these fields generally aren’t individually addressable.

                     OVN only supports arithmetic tests for equality on nominal  fields,  because  OpenFlow  and
                     Open  vSwitch provide no way for a flow to efficiently implement other comparisons on them.
                     (A test for inequality can be sort of built out of two flows with different priorities, but
                     OVN matching expressions always generate flows with a single priority.)

                     String fields are always nominal.

              •      Boolean.  A nominal field that has only two values, 0 and 1, is somewhat exceptional, since
                     it is easy to support both equality and inequality tests on such a field: either one can be
                     implemented as a test for 0 or 1.

                     Only predicates (see above) have a Boolean level of measurement.

                     This isn’t a standard level of measurement.

              Prerequisites.  Any symbol can have prerequisites, which are additional condition implied  by  the
              use  of  the  symbol.   For example, For example, icmp4.type symbol might have prerequisite icmp4,
              which would cause an expression icmp4.type == 0 to be interpreted as icmp4.type  ==  0  &&  icmp4,
              which  would  in  turn  expand to icmp4.type == 0 && eth.type == 0x800 && ip4.proto == 1 (assuming
              icmp4 is a predicate defined as suggested under Types above).

              Relational operators

              All of the standard relational operators ==, !=, <, <=, >, and >= are supported.   Nominal  fields
              support  only  ==  and  !=, and only in a positive sense when outer ! are taken into account, e.g.
              given string field inport, inport == "eth0" and !(inport != "eth0") are acceptable, but not inport
              != "eth0".

              The implementation of == (or != when it is negated), is more efficient  than  that  of  the  other
              relational operators.

              Constants

              Integer  constants may be expressed in decimal, hexadecimal prefixed by 0x, or as dotted-quad IPv4
              addresses, IPv6 addresses in their standard forms, or Ethernet addresses  as  colon-separated  hex
              digits.   A  constant  in any of these forms may be followed by a slash and a second constant (the
              mask) in the same form, to form a masked constant.  IPv4 and IPv6 masks may be given as  integers,
              to express CIDR prefixes.

              String constants have the same syntax as quoted strings in JSON (thus, they are Unicode strings).

              Some  operators  support  sets  of  constants written inside curly braces { ... }.  Commas between
              elements of a set, and after the last elements, are optional.  With ==, ``field  ==  {  constant1,
              constant2,  ...  }’’  is  syntactic  sugar  for ``field == constant1 || field == constant2 || ....
              Similarly, ``field != { constant1, constant2, ... }’’ is equivalent to  ``field  !=  constant1  &&
              field != constant2 && ...’’.

              Miscellaneous

              Comparisons  may  name  the symbol or the constant first, e.g. tcp.src == 80 and 80 == tcp.src are
              both acceptable.

              Tests for a range may be expressed using a  syntax  like  1024  <=  tcp.src  <=  49151,  which  is
              equivalent to 1024 <= tcp.src && tcp.src <= 49151.

              For  a  one-bit  field  or  predicate,  a  mention  of its name is equivalent to symobl == 1, e.g.
              vlan.present is equivalent to vlan.present == 1.  The same is true  for  one-bit  subfields,  e.g.
              vlan.tci[12].  There is no technical limitation to implementing the same for ordinal fields of all
              widths,  but  the  implementation  is  expensive enough that the syntax parser requires writing an
              explicit comparison against zero to make mistakes less likely, e.g. in tcp.src != 0 the comparison
              against 0 is required.

              Operator precedence is as shown below, from highest to lowest.  There  are  two  exceptions  where
              parentheses  are required even though the table would suggest that they are not: && and || require
              parentheses when  used  together,  and  !  requires  parentheses  when  applied  to  a  relational
              expression.   Thus,  in  (eth.type == 0x800 || eth.type == 0x86dd) && ip.proto == 6 or !(arp.op ==
              1), the parentheses are mandatory.

              •      ()==   !=   <   <=   >   >=!&&   ||

              Comments may be introduced by //, which extends to the next new-line.  Comments within a line  may
              be bracketed by /* and */.  Multiline comments are not supported.

              Symbols

              Most  of  the  symbols below have integer type.  Only inport and outport have string type.  inport
              names a logical port.  Thus, its value  is  a  logical_port  name  from  the  Port_Binding  table.
              outport  may  name  a  logical  port,  as  inport,  or  a  logical  multicast group defined in the
              Multicast_Group table.  For both symbols, only names within the flow’s  logical  datapath  may  be
              used.

              •      reg0...reg4inport outporteth.src eth.dst eth.typevlan.tci vlan.vid vlan.pcp vlan.presentip.proto ip.dscp ip.ecn ip.ttl ip.fragip4.src ip4.dstip6.src ip6.dst ip6.labelarp.op arp.spa arp.tpa arp.sha arp.thatcp.src tcp.dst tcp.flagsudp.src udp.dstsctp.src sctp.dsticmp4.type icmp4.codeicmp6.type icmp6.codend.target nd.sll nd.tllct_state, which has the following Boolean subfields:

                     •      ct.new: True for a new flow

                     •      ct.est: True for an established flow

                     •      ct.rel: True for a related flow

                     •      ct.rpl: True for a reply flow

                     •      ct.inv: True for a connection entry in a bad state

                     ct_state and its subfields are initialized by the ct_next action, described below.

              The following predicates are supported:

              •      eth.bcast expands to eth.dst == ff:ff:ff:ff:ff:ffeth.mcast expands to eth.dst[40]vlan.present expands to vlan.tci[12]ip4 expands to eth.type == 0x800ip4.mcast expands to ip4.dst[28..31] == 0xeip6 expands to eth.type == 0x86ddip expands to ip4 || ip6icmp4 expands to ip4 && ip.proto == 1icmp6 expands to ip6 && ip.proto == 58icmp expands to icmp4 || icmp6ip.is_frag expands to ip.frag[0]ip.later_frag expands to ip.frag[1]ip.first_frag expands to ip.is_frag && !ip.later_fragarp expands to eth.type == 0x806nd expands to icmp6.type == {135, 136} && icmp6.code == 0tcp expands to ip.proto == 6udp expands to ip.proto == 17sctp expands to ip.proto == 132

       actions: string
              Logical  datapath  actions,  to  be  executed when the logical flow represented by this row is the
              highest-priority match.

              Actions share lexical syntax with the match column.  An empty set of actions (or one that contains
              just white space or comments), or a set of actions that consists of just drop;, causes the matched
              packets to be dropped.   Otherwise,  the  column  should  contain  a  sequence  of  actions,  each
              terminated by a semicolon.

              The following actions are defined:

              output;
                     In  the  ingress  pipeline,  this  action executes the egress pipeline as a subroutine.  If
                     outport names a logical port, the egress pipeline executes  once;  if  it  is  a  multicast
                     group, the egress pipeline runs once for each logical port in the group.

                     In the egress pipeline, this action performs the actual output to the outport logical port.
                     (In the egress pipeline, outport never names a multicast group.)

                     Output  to the input port is implicitly dropped, that is, output becomes a no-op if outport
                     == inport.  Occasionally it may be useful to override this behavior, e.g. to  send  an  ARP
                     reply  to  an  ARP  request; to do so, use inport = ""; to set the logical input port to an
                     empty string (which should not be used as the name of any logical port).

              next;
              next(table);
                   Executes another logical datapath table as a subroutine.  By default,  the  table  after  the
                   current one is executed.  Specify table to jump to a specific table in the same pipeline.

              field = constant;
                   Sets  data  or metadata field field to constant value constant, e.g. outport = "vif0"; to set
                   the logical output port.  To set only a subset of bits in a field,  specify  a  subfield  for
                   field  or  a masked constant, e.g. one may use vlan.pcp[2] = 1; or vlan.pcp = 4/4; to set the
                   most sigificant bit of the VLAN PCP.

                   Assigning to a field with prerequisites implicitly adds those prerequisites to  match;  thus,
                   for  example,  a  flow that sets tcp.dst applies only to TCP flows, regardless of whether its
                   match mentions any TCP field.

                   Not all fields are modifiable (e.g.  eth.type  and  ip.proto  are  read-only),  and  not  all
                   modifiable  fields  may  be  partially  modified (e.g. ip.ttl must assigned as a whole).  The
                   outport field is modifiable in the ingress pipeline but not in the egress pipeline.

              field1 = field2;
                   Sets data or metadata field field1 to the value of data or metadata field field2, e.g. reg0 =
                   ip4.src; copies ip4.src into reg0.  To modify only a subset of  a  field’s  bits,  specify  a
                   subfield  for  field1  or  field2  or  both,  e.g.  vlan.pcp  = reg0[0..2]; copies the least-
                   significant bits of reg0 into the VLAN PCP.

                   field1 and field2 must be the same type, either both string or both integer fields.  If  they
                   are both integer fields, they must have the same width.

                   If field1 or field2 has prerequisites, they are added implicitly to match.  It is possible to
                   write  an assignment with contradictory prerequisites, such as ip4.src = ip6.src[0..31];, but
                   the contradiction means that a logical flow with such an assignment will never be matched.

              field1 <-> field2;
                   Similar to field1 = field2; except that the two values are exchanged instead of copied.  Both
                   field1 and field2 must modifiable.

              ip.ttl--;
                   Decrements the IPv4 or IPv6 TTL.   If  this  would  make  the  TTL  zero  or  negative,  then
                   processing  of  the packet halts; no further actions are processed.  (To properly handle such
                   cases, a higher-priority flow should match on ip.ttl == {0, 1};.)

                   Prerequisite: ip

              ct_next;
                   Apply connection tracking to the flow, initializing ct_state for matching  in  later  tables.
                   Automatically moves on to the next table, as if followed by next.

                   As  a  side effect, IP fragments will be reassembled for matching.  If a fragmented packet is
                   output, then it will be  sent  with  any  overlapping  fragments  squashed.   The  connection
                   tracking state is scoped by the logical port, so overlapping addresses may be used.  To allow
                   traffic related to the matched flow, execute ct_commit.

                   It  is  possible  to have actions follow ct_next, but they will not have access to any of its
                   side-effects and is not generally useful.

              ct_commit;
                   Commit the flow to the connection tracking entry associated with it by  a  previous  call  to
                   ct_next.

              The following actions will likely be useful later, but they have not been thought out carefully.

              arp { action; ... };
                     Temporarily  replaces  the  IPv4  packet being processed by an ARP packet and executes each
                     nested action on the ARP packet.  Actions following the arp action, if any,  apply  to  the
                     original, unmodified packet.

                     The  ARP  packet that this action operates on is initialized based on the IPv4 packet being
                     processed, as follows.  These are default values that the nested actions will probably want
                     to change:

                     •      eth.src unchanged

                     •      eth.dst unchanged

                     •      eth.type = 0x0806arp.op = 1 (ARP request)

                     •      arp.sha copied from eth.srcarp.spa copied from ip4.srcarp.tha = 00:00:00:00:00:00arp.tpa copied from ip4.dst

                     Prerequisite: ip4

              icmp4 { action; ... };
                     Temporarily replaces the IPv4 packet being processed by an ICMPv4 packet and executes  each
                     nested  action  on the ICMPv4 packet.  Actions following the icmp4 action, if any, apply to
                     the original, unmodified packet.

                     The ICMPv4 packet that this action operates on is initialized  based  on  the  IPv4  packet
                     being  processed,  as  follows.   These  are  default  values  that the nested actions will
                     probably want to change.  Ethernet and IPv4 fields not listed here are not changed:

                     •      ip.proto = 1 (ICMPv4)

                     •      ip.frag = 0 (not a fragment)

                     •      icmp4.type = 3 (destination unreachable)

                     •      icmp4.code = 1 (host unreachable)

                     Details TBD.

                     Prerequisite: ip4

              tcp_reset;
                     This action transforms the current TCP packet according to the following pseudocode:

                     if (tcp.ack) {
                             tcp.seq = tcp.ack;
                     } else {
                             tcp.ack = tcp.seq + length(tcp.payload);
                             tcp.seq = 0;
                     }
                     tcp.flags = RST;

                     Then, the action drops all TCP options and payload data, and updates the TCP checksum.

                     Details TBD.

                     Prerequisite: tcp

       external_ids : stage-name: optional string
              Human-readable name for this flow’s stage in the pipeline.

     Common Columns:
       The overall purpose of these columns is described under Common Columns at the beginning of this document.

       external_ids: map of string-string pairs

Multicast_Group TABLE

       The rows in this table define multicast groups of logical ports.  Multicast groups allow a single  packet
       transmitted  over a tunnel to a hypervisor to be delivered to multiple VMs on that hypervisor, which uses
       bandwidth more efficiently.

       Each row in this table defines a logical multicast  group  numbered  tunnel_key  within  datapath,  whose
       logical ports are listed in the ports column.

   Summary:
       datapath                      Datapath_Binding
       tunnel_key                    integer, in range 32,768 to 65,535
       name                          string
       ports                         set of 1 or more weak reference to Port_Bindings

   Details:
       datapath: Datapath_Binding
              The logical datapath in which the multicast group resides.

       tunnel_key: integer, in range 32,768 to 65,535
              The  value  used  to designate this logical egress port in tunnel encapsulations.  An index forces
              the key to be unique within the datapath.  The unusual range ensures that multicast group  IDs  do
              not overlap with logical port IDs.

       name: string
              The  logical  multicast  group’s name.  An index forces the name to be unique within the datapath.
              Logical flows in the ingress pipeline may output to the  group  just  as  for  individual  logical
              ports, by assigning the group’s name to outport and executing an output action.

              Multicast  group names and logical port names share a single namespace and thus should not overlap
              (but the database schema cannot enforce this).  To try to avoid conflicts, ovn-northd  uses  names
              that begin with _MC_.

       ports: set of 1 or more weak reference to Port_Bindings
              The  logical  ports  included  in the multicast group.  All of these ports must be in the datapath
              logical datapath (but the database schema cannot enforce this).

Datapath_Binding TABLE

       Each row in this table identifies physical bindings of a logical datapath.  A logical datapath implements
       a logical pipeline among the ports in the Port_Binding  table  associated  with  it.   In  practice,  the
       pipeline in a given logical datapath implements either a logical switch or a logical router.

   Summary:
       tunnel_key                    integer, in range 1 to 16,777,215 (must be unique within table)
       OVN_Northbound Relationship:
         external_ids : logical-switch
                                     optional string, containing an uuid
         external_ids : logical-router
                                     optional string, containing an uuid
       Common Columns:
         external_ids                map of string-string pairs

   Details:
       tunnel_key: integer, in range 1 to 16,777,215 (must be unique within table)
              The  tunnel key value to which the logical datapath is bound.  The Tunnel Encapsulation section in
              ovn-architecture(7) describes how tunnel keys are constructed for each supported encapsulation.

     OVN_Northbound Relationship:
       Each row in Datapath_Binding is associated with some logical datapath.  ovn-northd  uses  these  keys  to
       track the association of a logical datapath with concepts in the OVN_Northbound database.

       external_ids : logical-switch: optional string, containing an uuid
              For a logical datapath that represents a logical switch, ovn-northd stores in this key the UUID of
              the corresponding Logical_Switch row in the OVN_Northbound database.

       external_ids : logical-router: optional string, containing an uuid
              For a logical datapath that represents a logical router, ovn-northd stores in this key the UUID of
              the corresponding Logical_Router row in the OVN_Northbound database.

     Common Columns:
       The overall purpose of these columns is described under Common Columns at the beginning of this document.

       external_ids: map of string-string pairs

Port_Binding TABLE

       Most  rows  in  this table identify the physical location of a logical port.  (The exceptions are logical
       patch ports, which do not have any physical location.)

       For every Logical_Port record in OVN_Northbound database, ovn-northd creates  a  record  in  this  table.
       ovn-northd  populates  and maintains every column except the chassis column, which it leaves empty in new
       records.

       ovn-controller/ovn-controller-vtep populates the chassis column for the records that identify the logical
       ports that are located on its hypervisor/gateway, which ovn-controller/ovn-controller-vtep in turn  finds
       out  by  monitoring  the local hypervisor’s Open_vSwitch database, which identifies logical ports via the
       conventions described in IntegrationGuide.md.

       When a chassis shuts down gracefully, it should clean up  the  chassis  column  that  it  previously  had
       populated.   (This  is  not  critical  because  resources  hosted  on the chassis are equally unreachable
       regardless of whether their rows are present.)  To handle the case where a VM is shut  down  abruptly  on
       one  chassis, then brought up again on a different one, ovn-controller/ovn-controller-vtep must overwrite
       the chassis column with new information.

   Summary:
       Core Features:
         datapath                    Datapath_Binding
         logical_port                string (must be unique within table)
         chassis                     optional weak reference to Chassis
         tunnel_key                  integer, in range 1 to 32,767
         mac                         set of strings
         type                        string
       Patch Options:
         options : peer              optional string
       Localnet Options:
         options : network_name      optional string
         tag                         optional integer, in range 1 to 4,095
       VTEP Options:
         options : vtep-physical-switch
                                     optional string
         options : vtep-logical-switch
                                     optional string
       Nested Containers:
         parent_port                 optional string
         tag                         optional integer, in range 1 to 4,095

   Details:
     Core Features:

       datapath: Datapath_Binding
              The logical datapath to which the logical port belongs.

       logical_port: string (must be unique within table)
              A logical port, taken from name in the OVN_Northbound database’s Logical_Port table.  OVN does not
              prescribe a particular format for the logical port ID.

       chassis: optional weak reference to Chassis
              The physical location of the logical port.  To successfully identify a chassis, this  column  must
              be a Chassis record.  This is populated by ovn-controller/ovn-controller-vtep.

       tunnel_key: integer, in range 1 to 32,767
              A  number  that  represents the logical port in the key (e.g. STT key or Geneve TLV) field carried
              within tunnel protocol packets.

              The tunnel ID must be unique within the scope of a logical datapath.

       mac: set of strings
              The Ethernet address or addresses used as a source address on the logical port, each in  the  form
              xx:xx:xx:xx:xx:xx.   The  string  unknown is also allowed to indicate that the logical port has an
              unknown set of (additional) source addresses.

              A VM interface would ordinarily have a single Ethernet address.  A gateway  port  might  initially
              only have unknown, and then add MAC addresses to the set as it learns new source addresses.

       type: string
              A type for this logical port.  Logical ports can be used to model other types of connectivity into
              an OVN logical switch.  The following types are defined:

              (empty string)
                     VM (or VIF) interface.

              patch  One  of  a  pair  of  logical  ports that act as if connected by a patch cable.  Useful for
                     connecting two logical datapaths, e.g. to connect a logical router to a logical  switch  or
                     to another logical router.

              localnet
                     A  connection to a locally accessible network from each ovn-controller instance.  A logical
                     switch can only have a single localnet port attached and at most one regular logical  port.
                     This is used to model direct connectivity to an existing network.

              vtep   A  port to a logical switch on a VTEP gateway chassis.  In order to get this port correctly
                     recognized    by    the    OVN    controller,    the    options:vtep-physical-switch    and
                     options:vtep-logical-switch must also be defined.

     Patch Options:
       These options apply to logical ports with type of patch.

       options : peer: optional string
              The  logical_port  in  the  Port_Binding  record  for  the  other  side  of  the patch.  The named
              logical_port must specify this logical_port in its own  peer  option.   That  is,  the  two  patch
              logical ports must have reversed logical_port and peer values.

     Localnet Options:
       These options apply to logical ports with type of localnet.

       options : network_name: optional string
              Required.   ovn-controller  uses  the  configuration entry ovn-bridge-mappings to determine how to
              connect to this network.  ovn-bridge-mappings is a list of network names mapped  to  a  local  OVS
              bridge  that provides access to that network.  An example of configuring ovn-bridge-mappings would
              be:

              $ ovs-vsctl set open . external-ids:ovn-bridge-mappings=physnet1:br-eth0,physnet2:br-eth1

              When a logical switch has a localnet port attached, every  chassis  that  may  have  a  local  vif
              attached  to  that  logical  switch  must have a bridge mapping configured to reach that localnet.
              Traffic that arrives on a localnet port is never forwarded over a tunnel to another chassis.

       tag: optional integer, in range 1 to 4,095
              If set, indicates that the port represents a connection to a specific VLAN on a locally accessible
              network. The VLAN ID is used to match incoming traffic and is also added to outgoing traffic.

     VTEP Options:
       These options apply to logical ports with type of vtep.

       options : vtep-physical-switch: optional string
              Required. The name of the VTEP gateway.

       options : vtep-logical-switch: optional string
              Required.  A logical switch name connected by the VTEP gateway.  Must be set when type is vtep.

     Nested Containers:
       These columns support containers nested within a VM.  Specifically, they are used when type is empty  and
       logical_port  identifies the interface of a container spawned inside a VM.  They are empty for containers
       or VMs that run directly on a hypervisor.

       parent_port: optional string
              This is taken from parent_name in the OVN_Northbound database’s Logical_Port table.

       tag: optional integer, in range 1 to 4,095
              Identifies the VLAN tag in the network traffic associated with that container’s network interface.

              This column is used for a different purpose when type is localnet (see Localnet Options, above).

Open vSwitch 2.5.9                               DB Schema 1.0.0                                       ovn-sb(5)