Provided by: libfabric-dev_1.17.0-3_amd64 bug

NAME

       fabric - Fabric Interface Library

SYNOPSIS

              #include <rdma/fabric.h>

       Libfabric  is  a  high-performance fabric software library designed to provide low-latency
       interfaces to fabric hardware.

OVERVIEW

       Libfabric provides `process direct  I/O'  to  application  software  communicating  across
       fabric  software  and  hardware.   Process  direct  I/O, historically referred to as RDMA,
       allows an application to  directly  access  network  resources  without  operating  system
       interventions.  Data transfers can occur directly to and from application memory.

       There are two components to the libfabric software:

       Fabric Providers
              Conceptually,  a  fabric  provider  may  be  viewed as a local hardware NIC driver,
              though a provider is not limited  by  this  definition.   The  first  component  of
              libfabric  is  a  general  purpose  framework that is capable of handling different
              types of fabric hardware.  All fabric hardware devices and their  software  drivers
              are required to support this framework.  Devices and the drivers that plug into the
              libfabric framework are referred to  as  fabric  providers,  or  simply  providers.
              Provider details may be found in fi_provider(7).

       Fabric Interfaces
              The  second  component  is  a  set  of communication operations.  Libfabric defines
              several sets of communication functions that providers  can  support.   It  is  not
              required  that  providers  implement  all the interfaces that are defined; however,
              providers clearly indicate which interfaces they do support.

FABRIC INTERFACES

       The fabric interfaces are designed such that they are cohesive and not simply a  union  of
       disjoint  interfaces.   The  interfaces  are  logically  divided  into two groups: control
       interfaces and communication operations.  The control  interfaces  are  a  common  set  of
       operations  that  provide access to local communication resources, such as address vectors
       and event queues.  The communication operations expose particular models of  communication
       and  fabric  functionality,  such  as  message  queues,  remote  memory access, and atomic
       operations.  Communication operations are associated with fabric endpoints.

       Applications will typically use the control interfaces to discover local capabilities  and
       allocate  necessary  resources.   They  will  then  allocate and configure a communication
       endpoint to send and receive data, or perform other types of data transfers,  with  remote
       endpoints.

CONTROL INTERFACES

       The  control  interfaces  APIs  provide  applications  access  to network resources.  This
       involves  listing  all  the  interfaces  available,  obtaining  the  capabilities  of  the
       interfaces and opening a provider.

       fi_getinfo - Fabric Information
              The  fi_getinfo  call is the base call used to discover and request fabric services
              offered by the system.  Applications can use this call  to  indicate  the  type  of
              communication  that they desire.  The results from fi_getinfo, fi_info, are used to
              reserve and configure fabric resources.

       fi_getinfo returns a list of fi_info  structures.   Each  structure  references  a  single
       fabric  provider, indicating the interfaces that the provider supports, along with a named
       set of resources.  A fabric provider  may  include  multiple  fi_info  structures  in  the
       returned list.

       fi_fabric - Fabric Domain
              A  fabric  domain  represents  a collection of hardware and software resources that
              access a single physical or virtual network.  All network ports on  a  system  that
              can  communicate  with  each  other  through  the  fabric belong to the same fabric
              domain.  A fabric domain shares network addresses and can span multiple  providers.
              libfabric supports systems connected to multiple fabrics.

       fi_domain - Access Domains
              An  access domain represents a single logical connection into a fabric.  It may map
              to a single physical or virtual NIC or  a  port.   An  access  domain  defines  the
              boundary  across  which  fabric  resources  may  be associated.  Each access domain
              belongs to a single fabric domain.

       fi_endpoint - Fabric Endpoint
              A fabric endpoint is a communication portal.  An endpoint may be either  active  or
              passive.   Passive  endpoints  are  used to listen for connection requests.  Active
              endpoints can perform data  transfers.   Endpoints  are  configured  with  specific
              communication capabilities and data transfer interfaces.

       fi_eq - Event Queue
              Event  queues,  are  used  to  collect  and  report  the completion of asynchronous
              operations and events.  Event queues report events that are not directly associated
              with data transfer operations.

       fi_cq - Completion Queue
              Completion  queues  are high-performance event queues used to report the completion
              of data transfer operations.

       fi_cntr - Event Counters
              Event counters are used to report the number of completed asynchronous  operations.
              Event  counters are considered light-weight, in that a completion simply increments
              a counter, rather than placing an entry into an event queue.

       fi_mr - Memory Region
              Memory regions describe application local memory  buffers.   In  order  for  fabric
              resources to access application memory, the application must first grant permission
              to the fabric provider  by  constructing  a  memory  region.   Memory  regions  are
              required for specific types of data transfer operations, such as RMA transfers (see
              below).

       fi_av - Address Vector
              Address vectors are used to map higher level addresses, such as IP addresses, which
              may be more natural for an application to use, into fabric specific addresses.  The
              use of address vectors allows providers to reduce the amount of memory required  to
              maintain  large  address look-up tables, and eliminate expensive address resolution
              and look-up methods during data transfer operations.

DATA TRANSFER INTERFACES

       Fabric endpoints are associated with multiple data transfer  interfaces.   Each  interface
       set  is  designed  to support a specific style of communication, with an endpoint allowing
       the different  interfaces  to  be  used  in  conjunction.   The  following  data  transfer
       interfaces are defined by libfabric.

       fi_msg - Message Queue
              Message  queues  expose  a  simple,  message-based  FIFO  queue  interface  to  the
              application.  Message data transfers allow applications to send  and  receive  data
              with message boundaries being maintained.

       fi_tagged - Tagged Message Queues
              Tagged  message  lists  expose  send/receive  data transfer operations built on the
              concept of tagged messaging.  The tagged message queue is conceptually  similar  to
              standard  message  queues,  but  with the addition of 64-bit tags for each message.
              Sent messages are matched with receive buffers  that  are  tagged  with  a  similar
              value.

       fi_rma - Remote Memory Access
              RMA transfers are one-sided operations that read or write data directly to a remote
              memory region.  Other than defining the appropriate memory region,  RMA  operations
              do not require interaction at the target side for the data transfer to complete.

       fi_atomic - Atomic
              Atomic  operations can perform one of several operations on a remote memory region.
              Atomic operations include well-known functionality, such as atomic-add and compare-
              and-swap,  plus  several  other  pre-defined  calls.   Unlike  other  data transfer
              interfaces, atomic operations are aware of the data formatting at the target memory
              region.

LOGGING INTERFACE

       Logging   can  be  controlled  using  the  FI_LOG_LEVEL,  FI_LOG_PROV,  and  FI_LOG_SUBSYS
       environment variables.

       FI_LOG_LEVEL
              FI_LOG_LEVEL controls the amount of logging data that is output.  The following log
              levels are defined.

       - Warn Warn is the least verbose setting and is intended for reporting errors or warnings.

       - Trace
              Trace  is  more  verbose  and  is  meant  to include non-detailed output helpful to
              tracing program execution.

       - Info Info is high traffic and meant for detailed output.

       - Debug
              Debug is high traffic and is  likely  to  impact  application  performance.   Debug
              output is only available if the library has been compiled with debugging enabled.

       FI_LOG_PROV
              The  FI_LOG_PROV  environment  variable  enables  or disables logging from specific
              providers.  Providers can be enabled by listing them in a comma separated  fashion.
              If  the list begins with the `^' symbol, then the list will be negated.  By default
              all providers are enabled.

       Example: To enable logging from the psm and sockets provider: FI_LOG_PROV=“psm,sockets”

       Example: To enable logging from providers other than psm: FI_LOG_PROV=“^psm”

       FI_LOG_SUBSYS
              The FI_LOG_SUBSYS environment variable enables or disables logging at the subsystem
              level.  The syntax for enabling or disabling subsystems is similar to that used for
              FI_LOG_PROV.  The following subsystems are defined.

       - core Provides output related to the core framework and its management of providers.

       - fabric
              Provides output specific to interactions associated with the fabric object.

       - domain
              Provides output specific to interactions associated with the domain object.

       - ep_ctrl
              Provides output specific to endpoint  non-data  transfer  operations,  such  as  CM
              operations.

       - ep_data
              Provides output specific to endpoint data transfer operations.

       - av   Provides output specific to address vector operations.

       - cq   Provides output specific to completion queue operations.

       - eq   Provides output specific to event queue operations.

       - mr   Provides output specific to memory registration.

PROVIDER INSTALLATION AND SELECTION

       The  libfabric  build  scripts  will  install  all  providers  that  are  supported by the
       installation system.  Providers that are missing build  prerequisites  will  be  disabled.
       Installed   providers   will   dynamically   check   for  necessary  hardware  on  library
       initialization and respond appropriately to application queries.

       Users can enable or disable available providers through build configuration options.   See
       `configure  –help'  for  details.  In general, a specific provider can be controlled using
       the configure option `–enable-'.  For example, `–enable-udp' (or  `–enable-udp=yes')  will
       add the udp provider to the build.  To disable the provider, `–enable-udp=no' can be used.

       Providers  can  also  be  enable or disabled at run time using the FI_PROVIDER environment
       variable.  The FI_PROVIDER variable is set to a  comma  separated  list  of  providers  to
       include.  If the list begins with the `^' symbol, then the list will be negated.

       Example: To enable the udp and tcp providers only, set: FI_PROVIDER=“udp,tcp”

       The  fi_info  utility,  which is included as part of the libfabric package, can be used to
       retrieve information about which providers are available in the system.  Additionally,  it
       can  retrieve  a list of all environment variables that may be used to configure libfabric
       and each provider.  See fi_info(1) for more details.

ENVIRONMENT VARIABLE CONTROLS

       Core features of libfabric and its providers may be configured by an administrator through
       the  use  of  environment  variables.   Man  pages will usually describe the most commonly
       accessed variables, such as those mentioned above.  However, libfabric defines  interfaces
       for publishing and obtaining environment variables.  These are targeted for providers, but
       allow applications and users to obtain the full list of variables that may be  set,  along
       with a brief description of their use.

       A  full  list  of  variables available may be obtained by running the fi_info application,
       with the -e or –env command line option.

NOTES

   System Calls
       Because libfabric is designed to provide applications direct access  to  fabric  hardware,
       there  are limits on how libfabric resources may be used in conjunction with system calls.
       These limitations are notable for developers  who  may  be  familiar  programming  to  the
       sockets  interface.   Although  limits  are  provider specific, the following restrictions
       apply to many providers and should be adhered  to  by  applications  desiring  portability
       across providers.

       fork   Fabric  resources  are  not  guaranteed  to  be available by child processes.  This
              includes objects, such as endpoints and completion queues, as well  as  application
              controlled data buffers which have been assigned to the network.  For example, data
              buffers that have been registered with a fabric domain may not be  available  in  a
              child process because of copy on write restrictions.

   CUDA deadlock
       In  some  cases,  calls  to  cudaMemcpy  within  libfabric may result in a deadlock.  This
       typically occurs when a CUDA kernel blocks until a cudaMemcpy on the host  completes.   To
       avoid this deadlock, cudaMemcpy may be disabled by setting FI_HMEM_CUDA_ENABLE_XFER=0.  If
       this environment variable is set and there is a  call  to  cudaMemcpy  with  libfabric,  a
       warning  will be emitted and no copy will occur.  Note that not all providers support this
       option.

       Another mechanism which can be used to avoid deadlock is Nvidia’s gdrcopy.  Using  gdrcopy
       requires     an     external     library     and     kernel     module     available    at
       https://github.com/NVIDIA/gdrcopy.  Libfabric must  be  configured  with  gdrcopy  support
       using  the --with-gdrcopy option, and be run with FI_HMEM_CUDA_USE_GDRCOPY=1.  This may be
       used in conjunction with the above option to provide a method  for  copying  to/from  CUDA
       device  memory  when  cudaMemcpy  cannot be used.  Again, this may not be supported by all
       providers.

ABI CHANGES

       libfabric  releases  maintain  compatibility  with  older  releases,  so   that   compiled
       applications  can continue to work as-is, and previously written applications will compile
       against newer versions of the library without needing source code  changes.   The  changes
       below  describe  ABI updates that have occurred and which libfabric release corresponds to
       the changes.

       Note that because most functions  called  by  applications  actually  call  static  inline
       functions,  which  in  turn  reference  function  pointers  in order to call directly into
       providers, libfabric only exports a  handful  of  functions  directly.   ABI  changes  are
       limited  to  those  functions, most notably the fi_getinfo call and its returned attribute
       structures.

       The ABI version is independent from the libfabric release version.

   ABI 1.0
       The initial libfabric release (1.0.0) also corresponds to ABI version 1.0.   The  1.0  ABI
       was unchanged for libfabric major.minor versions 1.0, 1.1, 1.2, 1.3, and 1.4.

   ABI 1.1
       A  number  of  external data structures were appended starting with libfabric version 1.5.
       These changes included adding the fields to the following data structures.   The  1.1  ABI
       was exported by libfabric versions 1.5 and 1.6.

       fi_fabric_attr
              Added api_version

       fi_domain_attr
              Added  cntr_cnt,  mr_iov_limit,  caps, mode, auth_key, auth_key_size, max_err_data,
              and mr_cnt fields.  The mr_mode field was also changed from an enum to  an  integer
              flag field.

       fi_ep_attr
              Added auth_key_size and auth_key fields.

   ABI 1.2
       The  1.2  ABI  version  was  exported  by libfabric versions 1.7 and 1.8, and expanded the
       following structure.

       fi_info
              The fi_info structure was expanded to reference a new fabric object, fid_nic.  When
              available,  the  fid_nic  references  a  new  set  of attributes related to network
              hardware details.

   ABI 1.3
       The 1.3 ABI version was exported by libfabric versions 1.9, 1.10,  and  1.11.   Added  new
       fields to the following attributes:

       fi_domain_attr
              Added tclass

       fi_tx_attr
              Added tclass

   ABI 1.4
       The  1.4 ABI version was exported by libfabric 1.12.  Added fi_tostr_r, a thread-safe (re-
       entrant) version of fi_tostr.

   ABI 1.5
       ABI version starting with libfabric 1.13.  Added new fi_open API call.

   ABI 1.6
       ABI version starting with libfabric 1.14.  Added fi_log_ready for providers.

SEE ALSO

       fi_info(1),  fi_provider(7),  fi_getinfo(3),   fi_endpoint(3),   fi_domain(3),   fi_av(3),
       fi_eq(3), fi_cq(3), fi_cntr(3), fi_mr(3)

AUTHORS

       OpenFabrics.