Provided by: manpages_5.13-1_all bug

NAME

       numa - overview of Non-Uniform Memory Architecture

DESCRIPTION

       Non-Uniform  Memory Access (NUMA) refers to multiprocessor systems whose memory is divided
       into multiple memory nodes.  The access time of a memory  node  depends  on  the  relative
       locations  of  the  accessing CPU and the accessed node.  (This contrasts with a symmetric
       multiprocessor system, where the access time for all of the memory is  the  same  for  all
       CPUs.)   Normally, each CPU on a NUMA system has a local memory node whose contents can be
       accessed faster than the memory in the node local to another CPU or the memory  on  a  bus
       shared by all CPUs.

   NUMA system calls
       The  Linux  kernel  implements  the following NUMA-related system calls: get_mempolicy(2),
       mbind(2), migrate_pages(2), move_pages(2), and  set_mempolicy(2).   However,  applications
       should normally use the interface provided by libnuma; see "Library Support" below.

   /proc/[number]/numa_maps (since Linux 2.6.14)
       This file displays information about a process's NUMA memory policy and allocation.

       Each  line contains information about a memory range used by the process, displaying—among
       other information—the effective memory policy for that memory range and on which nodes the
       pages have been allocated.

       numa_maps  is  a read-only file.  When /proc/<pid>/numa_maps is read, the kernel will scan
       the virtual address space of the process and report how  memory  is  used.   One  line  is
       displayed for each unique memory range of the process.

       The  first  field of each line shows the starting address of the memory range.  This field
       allows a correlation with the contents of the /proc/<pid>/maps file,  which  contains  the
       end  address  of  the  range  and  other  information,  such as the access permissions and
       sharing.

       The second field shows the memory policy currently in effect for the memory  range.   Note
       that  the effective policy is not necessarily the policy installed by the process for that
       memory range.  Specifically, if the process installed a "default" policy for  that  range,
       the  effective  policy  for that range will be the process policy, which may or may not be
       "default".

       The rest of the line contains information about the pages allocated in the  memory  range,
       as follows:

       N<node>=<nr_pages>
              The  number of pages allocated on <node>.  <nr_pages> includes only pages currently
              mapped by the process.  Page migration and  memory  reclaim  may  have  temporarily
              unmapped  pages  associated  with this memory range.  These pages may show up again
              only after the process has attempted  to  reference  them.   If  the  memory  range
              represents a shared memory area or file mapping, other processes may currently have
              additional pages mapped in a corresponding memory range.

       file=<filename>
              The file backing the memory range.   If  the  file  is  mapped  as  private,  write
              accesses  may have generated COW (Copy-On-Write) pages in this memory range.  These
              pages are displayed as anonymous pages.

       heap   Memory range is used for the heap.

       stack  Memory range is used for the stack.

       huge   Huge memory range.  The page counts shown are huge  pages  and  not  regular  sized
              pages.

       anon=<pages>
              The number of anonymous page in the range.

       dirty=<pages>
              Number of dirty pages.

       mapped=<pages>
              Total number of mapped pages, if different from dirty and anon pages.

       mapmax=<count>
              Maximum mapcount (number of processes mapping a single page) encountered during the
              scan.  This may be used as an indicator of the degree of  sharing  occurring  in  a
              given memory range.

       swapcache=<count>
              Number of pages that have an associated entry on a swap device.

       active=<pages>
              The number of pages on the active list.  This field is shown only if different from
              the number of pages in this range.  This means that some inactive  pages  exist  in
              the memory range that may be removed from memory by the swapper soon.

       writeback=<pages>
              Number of pages that are currently being written out to disk.

CONFORMING TO

       No standards govern NUMA interfaces.

NOTES

       The  Linux  NUMA  system  calls  and  /proc interface are available only if the kernel was
       configured and built with the CONFIG_NUMA option.

   Library support
       Link with -lnuma to get the system call definitions.  libnuma and the required  <numaif.h>
       header are available in the numactl package.

       However,  applications  should  not  use these system calls directly.  Instead, the higher
       level interface provided by the numa(3) functions in the numactl package  is  recommended.
       The  numactl  package  is available at ⟨ftp://oss.sgi.com/www/projects/libnuma/download/⟩.
       The package is also included in some Linux distributions.  Some distributions include  the
       development library and header in the separate numactl-devel package.

SEE ALSO

       get_mempolicy(2),   mbind(2),   move_pages(2),   set_mempolicy(2),   numa(3),   cpuset(7),
       numactl(8)

COLOPHON

       This page is part of release 5.13 of the Linux man-pages project.  A  description  of  the
       project,  information  about  reporting  bugs, and the latest version of this page, can be
       found at https://www.kernel.org/doc/man-pages/.