cgroups(7) - phpMan

Command: man perldoc info search(apropos)  


CGROUPS(7)                          Linux Programmer's Manual                          CGROUPS(7)

NAME
       cgroups - Linux control groups

DESCRIPTION
       Control  groups,  usually  referred  to as cgroups, are a Linux kernel feature which allow
       processes to be organized into hierarchical groups whose usage of  various  types  of  re-
       sources  can  then  be  limited  and monitored.  The kernel's cgroup interface is provided
       through a pseudo-filesystem called cgroupfs.  Grouping is implemented in the  core  cgroup
       kernel  code, while resource tracking and limits are implemented in a set of per-resource-
       type subsystems (memory, CPU, and so on).

   Terminology
       A cgroup is a collection of processes that are bound to a set of limits or parameters  de-
       fined via the cgroup filesystem.

       A subsystem is a kernel component that modifies the behavior of the processes in a cgroup.
       Various subsystems have been implemented, making it possible to do things such as limiting
       the  amount of CPU time and memory available to a cgroup, accounting for the CPU time used
       by a cgroup, and freezing and resuming execution of the processes in a cgroup.  Subsystems
       are sometimes also known as resource controllers (or simply, controllers).

       The  cgroups  for  a controller are arranged in a hierarchy.  This hierarchy is defined by
       creating, removing, and renaming subdirectories within the  cgroup  filesystem.   At  each
       level  of  the  hierarchy, attributes (e.g., limits) can be defined.  The limits, control,
       and accounting provided by cgroups generally have effect throughout the  subhierarchy  un-
       derneath  the  cgroup  where  the  attributes  are defined.  Thus, for example, the limits
       placed on a cgroup at a higher level in the hierarchy cannot  be  exceeded  by  descendant
       cgroups.

   Cgroups version 1 and version 2
       The initial release of the cgroups implementation was in Linux 2.6.24.  Over time, various
       cgroup controllers have been added to allow the management of various types of  resources.
       However,  the  development of these controllers was largely uncoordinated, with the result
       that many inconsistencies arose between controllers and management of the  cgroup  hierar-
       chies  became  rather complex.  A longer description of these problems can be found in the
       kernel source file Documentation/admin-guide/cgroup-v2.rst (or Documentation/cgroup-v2.txt
       in Linux 4.17 and earlier).

       Because  of  the  problems  with  the  initial cgroups implementation (cgroups version 1),
       starting in Linux 3.10, work began on a new, orthogonal  implementation  to  remedy  these
       problems.   Initially marked experimental, and hidden behind the -o __DEVEL__sane_behavior
       mount option, the new version (cgroups version 2) was eventually made  official  with  the
       release  of Linux 4.5.  Differences between the two versions are described in the text be-
       low.  The file cgroup.sane_behavior, present in cgroups v1, is a relic of this  mount  op-
       tion. The file always reports "0" and is only retained for backward compatibility.

       Although  cgroups v2 is intended as a replacement for cgroups v1, the older system contin-
       ues to exist (and for compatibility  reasons  is  unlikely  to  be  removed).   Currently,
       cgroups  v2  implements only a subset of the controllers available in cgroups v1.  The two
       systems are implemented so that both v1 controllers and v2 controllers can be  mounted  on
       the same system.  Thus, for example, it is possible to use those controllers that are sup-
       ported under version 2, while also using version 1 controllers where version  2  does  not
       yet  support  those  controllers.  The only restriction here is that a controller can't be
       simultaneously employed in both a cgroups v1 hierarchy and in the cgroups v2 hierarchy.

CGROUPS VERSION 1
       Under cgroups v1, each controller may be mounted against a separate cgroup filesystem that
       provides  its  own  hierarchical  organization of the processes on the system.  It is also
       possible to comount multiple (or even all) cgroups v1 controllers against the same  cgroup
       filesystem,  meaning that the comounted controllers manage the same hierarchical organiza-
       tion of processes.

       For each mounted hierarchy, the directory tree mirrors the control group hierarchy.   Each
       control group is represented by a directory, with each of its child control cgroups repre-
       sented as a child directory.  For instance, /user/joe/1.session represents  control  group
       1.session,  which  is a child of cgroup joe, which is a child of /user.  Under each cgroup
       directory is a set of files which can be read or written to,  reflecting  resource  limits
       and a few general cgroup properties.

   Tasks (threads) versus processes
       In  cgroups  v1,  a  distinction  is  drawn  between processes and tasks.  In this view, a
       process can consist of multiple tasks (more commonly called  threads,  from  a  user-space
       perspective,  and  called  such  in the remainder of this man page).  In cgroups v1, it is
       possible to independently manipulate the cgroup memberships of the threads in a process.

       The cgroups v1 ability to split threads across different cgroups caused problems  in  some
       cases.   For example, it made no sense for the memory controller, since all of the threads
       of a process share a single address space.  Because of these problems, the ability to  in-
       dependently  manipulate  the cgroup memberships of the threads in a process was removed in
       the initial cgroups v2 implementation, and subsequently restored in a  more  limited  form
       (see the discussion of "thread mode" below).

   Mounting v1 controllers
       The  use  of  cgroups requires a kernel built with the CONFIG_CGROUP option.  In addition,
       each of the v1 controllers has an associated configuration option that must be set in  or-
       der to employ that controller.

       In  order  to  use  a  v1 controller, it must be mounted against a cgroup filesystem.  The
       usual place for such mounts is under a  tmpfs(5)  filesystem  mounted  at  /sys/fs/cgroup.
       Thus, one might mount the cpu controller as follows:

           mount -t cgroup -o cpu none /sys/fs/cgroup/cpu

       It  is  possible to comount multiple controllers against the same hierarchy.  For example,
       here the cpu and cpuacct controllers are comounted against a single hierarchy:

           mount -t cgroup -o cpu,cpuacct none /sys/fs/cgroup/cpu,cpuacct

       Comounting controllers has the effect that a process is in the same cgroup for all of  the
       comounted  controllers.   Separately mounting controllers allows a process to be in cgroup
       /foo1 for one controller while being in /foo2/foo3 for another.

       It is possible to comount all v1 controllers against the same hierarchy:

           mount -t cgroup -o all cgroup /sys/fs/cgroup

       (One can achieve the same result by omitting -o all, since it is the default  if  no  con-
       trollers are explicitly specified.)

       It  is not possible to mount the same controller against multiple cgroup hierarchies.  For
       example, it is not possible to mount both the cpu and cpuacct controllers against one  hi-
       erarchy,  and to mount the cpu controller alone against another hierarchy.  It is possible
       to create multiple mount points with exactly the same set of comounted controllers.   How-
       ever,  in this case all that results is multiple mount points providing a view of the same
       hierarchy.

       Note  that  on  many  systems,  the  v1  controllers  are  automatically   mounted   under
       /sys/fs/cgroup; in particular, systemd(1) automatically creates such mount points.

   Unmounting v1 controllers
       A  mounted  cgroup filesystem can be unmounted using the umount(8) command, as in the fol-
       lowing example:

           umount /sys/fs/cgroup/pids

       But note well: a cgroup filesystem is unmounted only if it is not busy, that is, it has no
       child  cgroups.  If this is not the case, then the only effect of the umount(8) is to make
       the mount invisible.  Thus, to ensure that the mount point is  really  removed,  one  must
       first  remove all child cgroups, which in turn can be done only after all member processes
       have been moved from those cgroups to the root cgroup.

   Cgroups version 1 controllers
       Each of the cgroups version 1 controllers is governed by  a  kernel  configuration  option
       (listed  below).  Additionally, the availability of the cgroups feature is governed by the
       CONFIG_CGROUPS kernel configuration option.

       cpu (since Linux 2.6.24; CONFIG_CGROUP_SCHED)
              Cgroups can be guaranteed a minimum number of "CPU shares" when a system  is  busy.
              This does not limit a cgroup's CPU usage if the CPUs are not busy.  For further in-
              formation,   see   Documentation/scheduler/sched-design-CFS.rst   (or    Documenta-
              tion/scheduler/sched-design-CFS.txt in Linux 5.2 and earlier).

              In  Linux 3.2, this controller was extended to provide CPU "bandwidth" control.  If
              the kernel is configured with CONFIG_CFS_BANDWIDTH, then within each scheduling pe-
              riod  (defined via a file in the cgroup directory), it is possible to define an up-
              per limit on the CPU time allocated to the processes in a cgroup.  This upper limit
              applies even if there is no other competition for the CPU.  Further information can
              be found in the kernel source file Documentation/scheduler/sched-bwc.rst (or  Docu-
              mentation/scheduler/sched-bwc.txt in Linux 5.2 and earlier).

       cpuacct (since Linux 2.6.24; CONFIG_CGROUP_CPUACCT)
              This provides accounting for CPU usage by groups of processes.

              Further  information  can  be  found  in  the  kernel source file Documentation/ad-
              min-guide/cgroup-v1/cpuacct.rst (or  Documentation/cgroup-v1/cpuacct.txt  in  Linux
              5.2 and earlier).

       cpuset (since Linux 2.6.24; CONFIG_CPUSETS)
              This  cgroup  can  be  used to bind the processes in a cgroup to a specified set of
              CPUs and NUMA nodes.

              Further information can be  found  in  the  kernel  source  file  Documentation/ad-
              min-guide/cgroup-v1/cpusets.rst  (or  Documentation/cgroup-v1/cpusets.txt  in Linux
              5.2 and earlier).

       memory (since Linux 2.6.25; CONFIG_MEMCG)
              The memory controller supports reporting and limiting  of  process  memory,  kernel
              memory, and swap used by cgroups.

              Further  information  can  be  found  in  the  kernel source file Documentation/ad-
              min-guide/cgroup-v1/memory.rst (or Documentation/cgroup-v1/memory.txt in Linux  5.2
              and earlier).

       devices (since Linux 2.6.26; CONFIG_CGROUP_DEVICE)
              This  supports  controlling  which  processes may create (mknod) devices as well as
              open them for reading or writing.  The policies may be specified as allow-lists and
              deny-lists.   Hierarchy  is  enforced, so new rules must not violate existing rules
              for the target or ancestor cgroups.

              Further information can be  found  in  the  kernel  source  file  Documentation/ad-
              min-guide/cgroup-v1/devices.rst  (or  Documentation/cgroup-v1/devices.txt  in Linux
              5.2 and earlier).

       freezer (since Linux 2.6.28; CONFIG_CGROUP_FREEZER)
              The freezer cgroup can suspend and restore (resume)  all  processes  in  a  cgroup.
              Freezing  a  cgroup /A also causes its children, for example, processes in /A/B, to
              be frozen.

              Further information can be  found  in  the  kernel  source  file  Documentation/ad-
              min-guide/cgroup-v1/freezer-subsystem.rst  (or Documentation/cgroup-v1/freezer-sub-
              system.txt in Linux 5.2 and earlier).

       net_cls (since Linux 2.6.29; CONFIG_CGROUP_NET_CLASSID)
              This places a classid, specified for the cgroup, on network packets  created  by  a
              cgroup.   These  classids  can  then  be used in firewall rules, as well as used to
              shape traffic using tc(8).  This applies only to packets leaving the cgroup, not to
              traffic arriving at the cgroup.

              Further  information  can  be  found  in  the  kernel source file Documentation/ad-
              min-guide/cgroup-v1/net_cls.rst (or  Documentation/cgroup-v1/net_cls.txt  in  Linux
              5.2 and earlier).

       blkio (since Linux 2.6.33; CONFIG_BLK_CGROUP)
              The  blkio cgroup controls and limits access to specified block devices by applying
              IO control in the form of throttling and upper limits against leaf nodes and inter-
              mediate nodes in the storage hierarchy.

              Two policies are available.  The first is a proportional-weight time-based division
              of disk implemented with CFQ.  This is in effect for leaf  nodes  using  CFQ.   The
              second is a throttling policy which specifies upper I/O rate limits on a device.

              Further  information  can  be  found  in  the  kernel source file Documentation/ad-
              min-guide/cgroup-v1/blkio-controller.rst   (or   Documentation/cgroup-v1/blkio-con-
              troller.txt in Linux 5.2 and earlier).

       perf_event (since Linux 2.6.39; CONFIG_CGROUP_PERF)
              This controller allows perf monitoring of the set of processes grouped in a cgroup.

              Further information can be found in the kernel source files

       net_prio (since Linux 3.3; CONFIG_CGROUP_NET_PRIO)
              This allows priorities to be specified, per network interface, for cgroups.

              Further  information  can  be  found  in  the  kernel source file Documentation/ad-
              min-guide/cgroup-v1/net_prio.rst (or Documentation/cgroup-v1/net_prio.txt in  Linux
              5.2 and earlier).

       hugetlb (since Linux 3.5; CONFIG_CGROUP_HUGETLB)
              This supports limiting the use of huge pages by cgroups.

              Further  information  can  be  found  in  the  kernel source file Documentation/ad-
              min-guide/cgroup-v1/hugetlb.rst (or  Documentation/cgroup-v1/hugetlb.txt  in  Linux
              5.2 and earlier).

       pids (since Linux 4.3; CONFIG_CGROUP_PIDS)
              This  controller  permits  limiting  the number of process that may be created in a
              cgroup (and its descendants).

              Further information can be  found  in  the  kernel  source  file  Documentation/ad-
              min-guide/cgroup-v1/pids.rst  (or Documentation/cgroup-v1/pids.txt in Linux 5.2 and
              earlier).

       rdma (since Linux 4.11; CONFIG_CGROUP_RDMA)
              The RDMA controller permits limiting the  use  of  RDMA/IB-specific  resources  per
              cgroup.

              Further  information  can  be  found  in  the  kernel source file Documentation/ad-
              min-guide/cgroup-v1/rdma.rst (or Documentation/cgroup-v1/rdma.txt in Linux 5.2  and
              earlier).

   Creating cgroups and moving processes
       A  cgroup filesystem initially contains a single root cgroup, '/', which all processes be-
       long to.  A new cgroup is created by creating a directory in the cgroup filesystem:

           mkdir /sys/fs/cgroup/cpu/cg1

       This creates a new empty cgroup.

       A process may be moved to this cgroup by writing its PID into  the  cgroup's  cgroup.procs
       file:

           echo $$ > /sys/fs/cgroup/cpu/cg1/cgroup.procs

       Only one PID at a time should be written to this file.

       Writing  the  value 0 to a cgroup.procs file causes the writing process to be moved to the
       corresponding cgroup.

       When writing a PID into the cgroup.procs, all threads in the process are  moved  into  the
       new cgroup at once.

       Within  a hierarchy, a process can be a member of exactly one cgroup.  Writing a process's
       PID to a cgroup.procs file automatically removes it from the cgroup of which it was previ-
       ously a member.

       The  cgroup.procs file can be read to obtain a list of the processes that are members of a
       cgroup.  The returned list of PIDs is not guaranteed to be in order.  Nor is it guaranteed
       to  be  free  of  duplicates.   (For example, a PID may be recycled while reading from the
       list.)

       In cgroups v1, an individual thread can be moved to another cgroup by writing  its  thread
       ID  (i.e., the kernel thread ID returned by clone(2) and gettid(2)) to the tasks file in a
       cgroup directory.  This file can be read to discover the set of threads that  are  members
       of the cgroup.

   Removing cgroups
       To  remove  a  cgroup, it must first have no child cgroups and contain no (nonzombie) pro-
       cesses.  So long as that is the case, one can simply remove  the  corresponding  directory
       pathname.  Note that files in a cgroup directory cannot and need not be removed.

   Cgroups v1 release notification
       Two files can be used to determine whether the kernel provides notifications when a cgroup
       becomes empty.  A cgroup is considered to be empty when it contains no child  cgroups  and
       no member processes.

       A  special file in the root directory of each cgroup hierarchy, release_agent, can be used
       to register the pathname of a program that may be invoked when a cgroup in  the  hierarchy
       becomes  empty.   The  pathname  of  the  newly empty cgroup (relative to the cgroup mount
       point) is provided as the sole command-line argument when the release_agent program is in-
       voked.  The release_agent program might remove the cgroup directory, or perhaps repopulate
       it with a process.

       The default value of the release_agent file is empty, meaning that no release agent is in-
       voked.

       The  content  of  the release_agent file can also be specified via a mount option when the
       cgroup filesystem is mounted:

           mount -o release_agent=pathname ...

       Whether or not the release_agent program is invoked when a particular cgroup becomes empty
       is  determined  by the value in the notify_on_release file in the corresponding cgroup di-
       rectory.  If this file contains the value 0, then the release_agent  program  is  not  in-
       voked.   If  it  contains  the value 1, the release_agent program is invoked.  The default
       value for this file in the root cgroup is 0.  At the time when a new  cgroup  is  created,
       the value in this file is inherited from the corresponding file in the parent cgroup.

   Cgroup v1 named hierarchies
       In  cgroups  v1,  it  is  possible  to  mount a cgroup hierarchy that has no attached con-
       trollers:

           mount -t cgroup -o none,name=somename none /some/mount/point

       Multiple instances of such hierarchies can be mounted; each hierarchy must have  a  unique
       name.  The only purpose of such hierarchies is to track processes.  (See the discussion of
       release notification below.)  An example of this is the name=systemd cgroup hierarchy that
       is used by systemd(1) to track services and user sessions.

       Since Linux 5.0, the cgroup_no_v1 kernel boot option (described below) can be used to dis-
       able cgroup v1 named hierarchies, by specifying cgroup_no_v1=named.

CGROUPS VERSION 2
       In cgroups v2, all mounted controllers reside in a single unified hierarchy.  While  (dif-
       ferent)  controllers  may be simultaneously mounted under the v1 and v2 hierarchies, it is
       not possible to mount the same controller simultaneously under both the v1 and the v2  hi-
       erarchies.

       The  new  behaviors in cgroups v2 are summarized here, and in some cases elaborated in the
       following subsections.

       1. Cgroups v2 provides a unified hierarchy against which all controllers are mounted.

       2. "Internal" processes are not permitted.  With the exception of the  root  cgroup,  pro-
          cesses  may  reside  only  in  leaf nodes (cgroups that do not themselves contain child
          cgroups).  The details are somewhat more subtle than this, and are described below.

       3. Active cgroups must be specified  via  the  files  cgroup.controllers  and  cgroup.sub-
          tree_control.

       4. The  tasks  file has been removed.  In addition, the cgroup.clone_children file that is
          employed by the cpuset controller has been removed.

       5. An  improved  mechanism  for  notification  of  empty  cgroups  is  provided   by   the
          cgroup.events file.

       For  more  changes,  see  the  Documentation/admin-guide/cgroup-v2.rst  file in the kernel
       source (or Documentation/cgroup-v2.txt in Linux 4.17 and earlier).

       Some of the new behaviors listed above saw subsequent modification with  the  addition  in
       Linux 4.14 of "thread mode" (described below).

   Cgroups v2 unified hierarchy
       In  cgroups  v1,  the ability to mount different controllers against different hierarchies
       was intended to allow great flexibility for application design.  In practice, though,  the
       flexibility  turned  out to be less useful than expected, and in many cases added complex-
       ity.  Therefore, in cgroups v2, all available controllers are mounted against a single hi-
       erarchy.  The available controllers are automatically mounted, meaning that it is not nec-
       essary (or possible) to specify the controllers when mounting the cgroup v2 filesystem us-
       ing a command such as the following:

           mount -t cgroup2 none /mnt/cgroup2

       A cgroup v2 controller is available only if it is not currently in use via a mount against
       a cgroup v1 hierarchy.  Or, to put things another way, it is not possible  to  employ  the
       same controller against both a v1 hierarchy and the unified v2 hierarchy.  This means that
       it may be necessary first to unmount a v1 controller (as described above) before that con-
       troller  is  available  in v2.  Since systemd(1) makes heavy use of some v1 controllers by
       default, it can in some cases be simpler to boot the system with selected  v1  controllers
       disabled.   To  do  this,  specify the cgroup_no_v1=list option on the kernel boot command
       line; list is a comma-separated list of the names of the controllers to  disable,  or  the
       word  all  to  disable  all  v1 controllers.  (This situation is correctly handled by sys-
       temd(1), which falls back to operating without the specified controllers.)

       Note that on many modern systems, systemd(1) automatically mounts the  cgroup2  filesystem
       at /sys/fs/cgroup/unified during the boot process.

   Cgroups v2 mount options
       The following options (mount -o) can be specified when mounting the group v2 filesystem:

       nsdelegate (since Linux 4.15)
              Treat cgroup namespaces as delegation boundaries.  For details, see below.

       memory_localevents (since Linux 5.2)
              The  memory.events  should  show statistics only for the cgroup itself, and not for
              any descendant cgroups.  This was the behavior before Linux 5.2.  Starting in Linux
              5.2,  the  default behavior is to include statistics for descendant cgroups in mem-
              ory.events, and this mount option can be used to revert  to  the  legacy  behavior.
              This option is system wide and can be set on mount or modified through remount only
              from the initial mount namespace; it is silently ignored in noninitial namespaces.

   Cgroups v2 controllers
       The  following  controllers,  documented  in  the  kernel  source  file  Documentation/ad-
       min-guide/cgroup-v2.rst  (or  Documentation/cgroup-v2.txt  in Linux 4.17 and earlier), are
       supported in cgroups version 2:

       cpu (since Linux 4.15)
              This is the successor to the version 1 cpu and cpuacct controllers.

       cpuset (since Linux 5.0)
              This is the successor of the version 1 cpuset controller.

       freezer (since Linux 5.2)
              This is the successor of the version 1 freezer controller.

       hugetlb (since Linux 5.6)
              This is the successor of the version 1 hugetlb controller.

       io (since Linux 4.5)
              This is the successor of the version 1 blkio controller.

       memory (since Linux 4.5)
              This is the successor of the version 1 memory controller.

       perf_event (since Linux 4.11)
              This is the same as the version 1 perf_event controller.

       pids (since Linux 4.5)
              This is the same as the version 1 pids controller.

       rdma (since Linux 4.11)
              This is the same as the version 1 rdma controller.

       There is no direct equivalent of the net_cls and net_prio controllers from cgroups version
       1.   Instead,  support  has  been  added to iptables(8) to allow eBPF filters that hook on
       cgroup v2 pathnames to make decisions about network traffic on a per-cgroup basis.

       The v2 devices controller provides no interface files; instead, device control is gated by
       attaching an eBPF (BPF_CGROUP_DEVICE) program to a v2 cgroup.

   Cgroups v2 subtree control
       Each cgroup in the v2 hierarchy contains the following two files:

       cgroup.controllers
              This  read-only  file  exposes a list of the controllers that are available in this
              cgroup.  The contents of this file match the contents of the cgroup.subtree_control
              file in the parent cgroup.

       cgroup.subtree_control
              This  is a list of controllers that are active (enabled) in the cgroup.  The set of
              controllers in this file is a subset of the set in the cgroup.controllers  of  this
              cgroup.   The set of active controllers is modified by writing strings to this file
              containing space-delimited controller names, each preceded by '+' (to enable a con-
              troller) or '-' (to disable a controller), as in the following example:

                  echo '+pids -memory' > x/y/cgroup.subtree_control

              An  attempt  to enable a controller that is not present in cgroup.controllers leads
              to an ENOENT error when writing to the cgroup.subtree_control file.

       Because the list of controllers in cgroup.subtree_control is a subset of those cgroup.con-
       trollers,  a controller that has been disabled in one cgroup in the hierarchy can never be
       re-enabled in the subtree below that cgroup.

       A cgroup's cgroup.subtree_control file determines the set of controllers  that  are  exer-
       cised  in the child cgroups.  When a controller (e.g., pids) is present in the cgroup.sub-
       tree_control file of a parent cgroup, then the  corresponding  controller-interface  files
       (e.g.,  pids.max) are automatically created in the children of that cgroup and can be used
       to exert resource control in the child cgroups.

   Cgroups v2 "no internal processes" rule
       Cgroups v2 enforces a so-called "no internal processes" rule.  Roughly speaking, this rule
       means that, with the exception of the root cgroup, processes may reside only in leaf nodes
       (cgroups that do not themselves contain child cgroups).  This avoids the  need  to  decide
       how  to  partition resources between processes which are members of cgroup A and processes
       in child cgroups of A.

       For instance, if cgroup /cg1/cg2 exists, then a process may reside in /cg1/cg2, but not in
       /cg1.   This  is to avoid an ambiguity in cgroups v1 with respect to the delegation of re-
       sources between processes in /cg1 and its child  cgroups.   The  recommended  approach  in
       cgroups  v2  is  to  create a subdirectory called leaf for any nonleaf cgroup which should
       contain processes, but no child cgroups.  Thus, processes which previously would have gone
       into  /cg1 would now go into /cg1/leaf.  This has the advantage of making explicit the re-
       lationship between processes in /cg1/leaf and /cg1's other children.

       The "no internal processes" rule is in fact more subtle  than  stated  above.   More  pre-
       cisely,  the rule is that a (nonroot) cgroup can't both (1) have member processes, and (2)
       distribute resources into child cgroups--that is, have a  nonempty  cgroup.subtree_control
       file.   Thus, it is possible for a cgroup to have both member processes and child cgroups,
       but before controllers can be enabled for that cgroup, the member processes must be  moved
       out of the cgroup (e.g., perhaps into the child cgroups).

       With  the  Linux  4.14  addition of "thread mode" (described below), the "no internal pro-
       cesses" rule has been relaxed in some cases.

   Cgroups v2 cgroup.events file
       Each nonroot cgroup in the v2 hierarchy contains a read-only  file,  cgroup.events,  whose
       contents are key-value pairs (delimited by newline characters, with the key and value sep-
       arated by spaces) providing state information about the cgroup:

           $ cat mygrp/cgroup.events
           populated 1
           frozen 0

       The following keys may appear in this file:

       populated
              The value of this key is either 1, if this cgroup or any  of  its  descendants  has
              member processes, or otherwise 0.

       frozen (since Linux 5.2)
              The value of this key is 1 if this cgroup is currently frozen, or 0 if it is not.

       The  cgroup.events  file can be monitored, in order to receive notification when the value
       of one of its keys changes.  Such monitoring can be done using inotify(7), which  notifies
       changes  as  IN_MODIFY events, or poll(2), which notifies changes by returning the POLLPRI
       and POLLERR bits in the revents field.

   Cgroup v2 release notification
       Cgroups v2 provides a new mechanism for  obtaining  notification  when  a  cgroup  becomes
       empty.  The cgroups v1 release_agent and notify_on_release files are removed, and replaced
       by the populated key in the cgroup.events file.  This key either has the value 0,  meaning
       that the cgroup (and its descendants) contain no (nonzombie) member processes, or 1, mean-
       ing that the cgroup (or one of its descendants) contains member processes.

       The cgroups v2 release-notification mechanism offers the  following  advantages  over  the
       cgroups v1 release_agent mechanism:

       *  It  allows  for  cheaper  notification,  since  a  single  process can monitor multiple
          cgroup.events files (using the techniques described earlier).  By contrast, the cgroups
          v1 mechanism requires the expense of creating a process for each notification.

       *  Notification  for  different  cgroup  subhierarchies can be delegated to different pro-
          cesses.  By contrast, the cgroups v1 mechanism allows only one release agent for an en-
          tire hierarchy.

   Cgroups v2 cgroup.stat file
       Each cgroup in the v2 hierarchy contains a read-only cgroup.stat file (first introduced in
       Linux 4.14) that consists of lines containing key-value pairs.  The  following  keys  cur-
       rently appear in this file:

       nr_descendants
              This  is  the  total number of visible (i.e., living) descendant cgroups underneath
              this cgroup.

       nr_dying_descendants
              This is the total number of dying descendant cgroups  underneath  this  cgroup.   A
              cgroup enters the dying state after being deleted.  It remains in that state for an
              undefined period (which will depend on system load) while resources are  freed  be-
              fore  the cgroup is destroyed.  Note that the presence of some cgroups in the dying
              state is normal, and is not indicative of any problem.

              A process can't be made a member of a dying cgroup, and a  dying  cgroup  can't  be
              brought back to life.

   Limiting the number of descendant cgroups
       Each  cgroup  in  the v2 hierarchy contains the following files, which can be used to view
       and set limits on the number of descendant cgroups under that cgroup:

       cgroup.max.depth (since Linux 4.14)
              This file defines a limit on the depth of nesting of descendant cgroups.   A  value
              of  0  in this file means that no descendant cgroups can be created.  An attempt to
              create a descendant whose nesting level exceeds the  limit  fails  (mkdir(2)  fails
              with the error EAGAIN).

              Writing  the string "max" to this file means that no limit is imposed.  The default
              value in this file is "max".

       cgroup.max.descendants (since Linux 4.14)
              This file defines a limit on the number of live descendant cgroups that this cgroup
              may  have.   An  attempt to create more descendants than allowed by the limit fails
              (mkdir(2) fails with the error EAGAIN).

              Writing the string "max" to this file means that no limit is imposed.  The  default
              value in this file is "max".

CGROUPS DELEGATION: DELEGATING A HIERARCHY TO A LESS PRIVILEGED USER
       In  the  context  of  cgroups,  delegation means passing management of some subtree of the
       cgroup hierarchy to a nonprivileged user.  Cgroups  v1  provides  support  for  delegation
       based  on  file permissions in the cgroup hierarchy but with less strict containment rules
       than v2 (as noted below).  Cgroups v2 supports delegation with containment by explicit de-
       sign.   The  focus  of the discussion in this section is on delegation in cgroups v2, with
       some differences for cgroups v1 noted along the way.

       Some terminology is required in order to describe delegation.  A delegater is a privileged
       user  (i.e., root) who owns a parent cgroup.  A delegatee is a nonprivileged user who will
       be granted the permissions needed to manage some subhierarchy under  that  parent  cgroup,
       known as the delegated subtree.

       To  perform  delegation, the delegater makes certain directories and files writable by the
       delegatee, typically by changing the ownership of the objects to be the  user  ID  of  the
       delegatee.   Assuming that we want to delegate the hierarchy rooted at (say) /dlgt_grp and
       that there are not yet any child cgroups under that cgroup, the ownership of the following
       is changed to the user ID of the delegatee:

       /dlgt_grp
              Changing  the  ownership of the root of the subtree means that any new cgroups cre-
              ated under the subtree (and the files they contain) will also be owned by the dele-
              gatee.

       /dlgt_grp/cgroup.procs
              Changing  the  ownership  of  this file means that the delegatee can move processes
              into the root of the delegated subtree.

       /dlgt_grp/cgroup.subtree_control (cgroups v2 only)
              Changing the ownership of this file means that the delegatee can enable controllers
              (that are present in /dlgt_grp/cgroup.controllers) in order to further redistribute
              resources at lower levels in the subtree.  (As an alternative to changing the  own-
              ership  of  this file, the delegater might instead add selected controllers to this
              file.)

       /dlgt_grp/cgroup.threads (cgroups v2 only)
              Changing the ownership of this file is necessary if a  threaded  subtree  is  being
              delegated  (see the description of "thread mode", below).  This permits the delega-
              tee to write thread IDs to the file.  (The ownership  of  this  file  can  also  be
              changed  when  delegating  a  domain subtree, but currently this serves no purpose,
              since, as described below, it is not possible  to  move  a  thread  between  domain
              cgroups by writing its thread ID to the cgroup.threads file.)

              In cgroups v1, the corresponding file that should instead be delegated is the tasks
              file.

       The delegater should not change the ownership of any of the  controller  interfaces  files
       (e.g., pids.max, memory.high) in dlgt_grp.  Those files are used from the next level above
       the delegated subtree in order to distribute resources into the subtree, and the delegatee
       should not have permission to change the resources that are distributed into the delegated
       subtree.

       See also the discussion of the /sys/kernel/cgroup/delegate file in NOTES  for  information
       about further delegatable files in cgroups v2.

       After the aforementioned steps have been performed, the delegatee can create child cgroups
       within the delegated subtree (the cgroup subdirectories and the files they contain will be
       owned  by  the delegatee) and move processes between cgroups in the subtree.  If some con-
       trollers are present in dlgt_grp/cgroup.subtree_control, or the ownership of that file was
       passed  to the delegatee, the delegatee can also control the further redistribution of the
       corresponding resources into the delegated subtree.

   Cgroups v2 delegation: nsdelegate and cgroup namespaces
       Starting with Linux 4.13, there is a second  way  to  perform  cgroup  delegation  in  the
       cgroups  v2  hierarchy.   This  is done by mounting or remounting the cgroup v2 filesystem
       with the nsdelegate mount option.  For example, if the cgroup v2  filesystem  has  already
       been mounted, we can remount it with the nsdelegate option as follows:

           mount -t cgroup2 -o remount,nsdelegate \
                            none /sys/fs/cgroup/unified

       The effect of this mount option is to cause cgroup namespaces to automatically become del-
       egation boundaries.  More specifically, the following restrictions apply for processes in-
       side the cgroup namespace:

       *  Writes  to  controller interface files in the root directory of the namespace will fail
          with the error EPERM.  Processes inside the cgroup namespace can still write  to  dele-
          gatable  files  in  the root directory of the cgroup namespace such as cgroup.procs and
          cgroup.subtree_control, and can create subhierarchy underneath the root directory.

       *  Attempts to migrate processes across the namespace boundary are denied (with the  error
          ENOENT).   Processes  inside the cgroup namespace can still (subject to the containment
          rules described below) move processes between cgroups within the subhierarchy under the
          namespace root.

       The  ability  to define cgroup namespaces as delegation boundaries makes cgroup namespaces
       more useful.  To understand why, suppose that we already have one  cgroup  hierarchy  that
       has  been delegated to a nonprivileged user, cecilia, using the older delegation technique
       described above.  Suppose further that cecilia wanted to further delegate  a  subhierarchy
       under  the  existing  delegated hierarchy.  (For example, the delegated hierarchy might be
       associated with an unprivileged container run by cecilia.)  Even if a cgroup namespace was
       employed, because both hierarchies are owned by the unprivileged user cecilia, the follow-
       ing illegitimate actions could be performed:

       *  A process in the inferior hierarchy could change the resource  controller  settings  in
          the root directory of that hierarchy.  (These resource controller settings are intended
          to allow control to be exercised from the parent cgroup; a  process  inside  the  child
          cgroup should not be allowed to modify them.)

       *  A  process inside the inferior hierarchy could move processes into and out of the infe-
          rior hierarchy if the cgroups in the superior hierarchy were somehow visible.

       Employing the nsdelegate mount option prevents both of these possibilities.

       The nsdelegate mount option only has an effect when performed in the initial  mount  name-
       space; in other mount namespaces, the option is silently ignored.

       Note: On some systems, systemd(1) automatically mounts the cgroup v2 filesystem.  In order
       to experiment with the nsdelegate operation, it may be useful to boot the kernel with  the
       following command-line options:

           cgroup_no_v1=all systemd.legacy_systemd_cgroup_controller

       These  options  cause the kernel to boot with the cgroups v1 controllers disabled (meaning
       that the controllers are available in the v2 hierarchy), and tells systemd(1) not to mount
       and use the cgroup v2 hierarchy, so that the v2 hierarchy can be manually mounted with the
       desired options after boot-up.

   Cgroup delegation containment rules
       Some delegation containment rules ensure that the delegatee  can  move  processes  between
       cgroups  within the delegated subtree, but can't move processes from outside the delegated
       subtree into the subtree or vice versa.  A nonprivileged process (i.e., the delegatee) can
       write  the PID of a "target" process into a cgroup.procs file only if all of the following
       are true:

       *  The writer has write permission on the cgroup.procs file in the destination cgroup.

       *  The writer has write permission on the cgroup.procs file in the nearest common ancestor
          of the source and destination cgroups.  Note that in some cases, the nearest common an-
          cestor may be the source or destination cgroup itself.  This  requirement  is  not  en-
          forced  for cgroups v1 hierarchies, with the consequence that containment in v1 is less
          strict than in v2.  (For example, in cgroups v1 the user that owns two  distinct  dele-
          gated subhierarchies can move a process between the hierarchies.)

       *  If  the cgroup v2 filesystem was mounted with the nsdelegate option, the writer must be
          able to see the source and destination cgroups from its cgroup namespace.

       *  In cgroups v1: the effective UID of the writer (i.e., the delegatee) matches  the  real
          user  ID  or  the saved set-user-ID of the target process.  Before Linux 4.11, this re-
          quirement also applied in cgroups v2 (This was a historical requirement inherited  from
          cgroups  v1  that  was later deemed unnecessary, since the other rules suffice for con-
          tainment in cgroups v2.)

       Note: one consequence of these delegation containment rules is that the unprivileged dele-
       gatee  can't  place  the  first process into the delegated subtree; instead, the delegater
       must place the first process (a process owned by the delegatee) into  the  delegated  sub-
       tree.

CGROUPS VERSION 2 THREAD MODE
       Among  the  restrictions imposed by cgroups v2 that were not present in cgroups v1 are the
       following:

       *  No thread-granularity control: all of the threads of a process  must  be  in  the  same
          cgroup.

       *  No  internal  processes:  a  cgroup  can't both have member processes and exercise con-
          trollers on child cgroups.

       Both of these restrictions were added because the lack of these  restrictions  had  caused
       problems in cgroups v1.  In particular, the cgroups v1 ability to allow thread-level gran-
       ularity for cgroup membership made no sense for some controllers.  (A notable example  was
       the  memory  controller:  since  threads share an address space, it made no sense to split
       threads across different memory cgroups.)

       Notwithstanding the initial design decision in cgroups v2, there were use cases  for  cer-
       tain  controllers,  notably the cpu controller, for which thread-level granularity of con-
       trol was meaningful and useful.  To accommodate such use cases, Linux  4.14  added  thread
       mode for cgroups v2.

       Thread mode allows the following:

       *  The  creation  of  threaded  subtrees  in  which the threads of a process may be spread
          across cgroups inside the tree.  (A threaded subtree may contain multiple multithreaded
          processes.)

       *  The  concept of threaded controllers, which can distribute resources across the cgroups
          in a threaded subtree.

       *  A relaxation of the "no internal processes rule", so that, within a threaded subtree, a
          cgroup  can  both  contain  member  threads  and  exercise  resource control over child
          cgroups.

       With the  addition  of  thread  mode,  each  nonroot  cgroup  now  contains  a  new  file,
       cgroup.type,  that exposes, and in some circumstances can be used to change, the "type" of
       a cgroup.  This file contains one of the following type values:

       domain This is a normal v2 cgroup that provides process-granularity control.  If a process
              is  a member of this cgroup, then all threads of the process are (by definition) in
              the same cgroup.  This is the default cgroup type, and provides the  same  behavior
              that was provided for cgroups in the initial cgroups v2 implementation.

       threaded
              This  cgroup  is  a  member  of  a  threaded subtree.  Threads can be added to this
              cgroup, and controllers can be enabled for the cgroup.

       domain threaded
              This is a domain cgroup that serves as the root of a threaded subtree.  This cgroup
              type is also known as "threaded root".

       domain invalid
              This  is  a  cgroup  inside a threaded subtree that is in an "invalid" state.  Pro-
              cesses can't be added to the cgroup, and  controllers  can't  be  enabled  for  the
              cgroup.   The only thing that can be done with this cgroup (other than deleting it)
              is to convert it to a threaded cgroup by  writing  the  string  "threaded"  to  the
              cgroup.type file.

              The  rationale  for  the  existence of this "interim" type during the creation of a
              threaded subtree (rather than the kernel simply immediately converting all  cgroups
              under  the  threaded root to the type threaded) is to allow for possible future ex-
              tensions to the thread mode model

   Threaded versus domain controllers
       With the addition of threads mode, cgroups v2 now distinguishes two types of resource con-
       trollers:

       *  Threaded controllers: these controllers support thread-granularity for resource control
          and can be enabled inside threaded subtrees, with the  result  that  the  corresponding
          controller-interface  files  appear  inside the cgroups in the threaded subtree.  As at
          Linux 4.19, the following controllers are threaded: cpu, perf_event, and pids.

       *  Domain controllers: these controllers support only  process  granularity  for  resource
          control.  From the perspective of a domain controller, all threads of a process are al-
          ways in the same cgroup.  Domain controllers can't be enabled inside  a  threaded  sub-
          tree.

   Creating a threaded subtree
       There are two pathways that lead to the creation of a threaded subtree.  The first pathway
       proceeds as follows:

       1. We write the string "threaded" to the cgroup.type file of a cgroup y/z  that  currently
          has the type domain.  This has the following effects:

          *  The type of the cgroup y/z becomes threaded.

          *  The type of the parent cgroup, y, becomes domain threaded.  The parent cgroup is the
             root of a threaded subtree (also known as the "threaded root").

          *  All other cgroups under y that were not already of type threaded (because they  were
             inside already existing threaded subtrees under the new threaded root) are converted
             to type domain invalid.  Any subsequently created cgroups under y will also have the
             type domain invalid.

       2. We  write the string "threaded" to each of the domain invalid cgroups under y, in order
          to convert them to the type threaded.  As a consequence of this step, all threads under
          the  threaded root now have the type threaded and the threaded subtree is now fully us-
          able.  The requirement to write "threaded" to each of these cgroups is somewhat cumber-
          some, but allows for possible future extensions to the thread-mode model.

       The second way of creating a threaded subtree is as follows:

       1. In an existing cgroup, z, that currently has the type domain, we (1) enable one or more
          threaded controllers and (2) make a process a member of z.  (These  two  steps  can  be
          done in either order.)  This has the following consequences:

          *  The type of z becomes domain threaded.

          *  All  of  the descendant cgroups of x that were not already of type threaded are con-
             verted to type domain invalid.

       2. As before, we make the threaded subtree usable by writing the string "threaded" to each
          of the domain invalid cgroups under y, in order to convert them to the type threaded.

       One  of  the consequences of the above pathways to creating a threaded subtree is that the
       threaded root cgroup can be a parent only to threaded (and domain invalid)  cgroups.   The
       threaded  root  cgroup  can't be a parent of a domain cgroups, and a threaded cgroup can't
       have a sibling that is a domain cgroup.

   Using a threaded subtree
       Within a threaded subtree, threaded controllers can be enabled in each subgroup whose type
       has  been changed to threaded; upon doing so, the corresponding controller interface files
       appear in the children of that cgroup.

       A process can be moved into a threaded subtree by writing its PID to the cgroup.procs file
       in  one  of the cgroups inside the tree.  This has the effect of making all of the threads
       in the process members of the corresponding cgroup and makes the process a member  of  the
       threaded  subtree.  The threads of the process can then be spread across the threaded sub-
       tree by writing their thread IDs (see gettid(2)) to the cgroup.threads files in  different
       cgroups inside the subtree.  The threads of a process must all reside in the same threaded
       subtree.

       As with writing to  cgroup.procs,  some  containment  rules  apply  when  writing  to  the
       cgroup.threads file:

       *  The  writer  must  have  write permission on the cgroup.threads file in the destination
          cgroup.

       *  The writer must have write permission on the cgroup.procs file in the  common  ancestor
          of  the source and destination cgroups.  (In some cases, the common ancestor may be the
          source or destination cgroup itself.)

       *  The source and destination cgroups must be in the same threaded  subtree.   (Outside  a
          threaded  subtree,  an  attempt  to  move  a  thread  by  writing  its thread ID to the
          cgroup.threads file in a different domain cgroup fails with the error EOPNOTSUPP.)

       The cgroup.threads file is present in each cgroup (including domain cgroups)  and  can  be
       read  in  order  to discover the set of threads that is present in the cgroup.  The set of
       thread IDs obtained when reading this file is not guaranteed to be ordered or free of  du-
       plicates.

       The  cgroup.procs  file in the threaded root shows the PIDs of all processes that are mem-
       bers of the threaded subtree.  The cgroup.procs files in the other cgroups in the  subtree
       are not readable.

       Domain  controllers  can't be enabled in a threaded subtree; no controller-interface files
       appear inside the cgroups underneath the threaded root.  From the point of view of  a  do-
       main  controller,  threaded  subtrees  are  invisible:  a  multithreaded  process inside a
       threaded subtree appears to a domain controller as a process that resides in the  threaded
       root cgroup.

       Within  a  threaded subtree, the "no internal processes" rule does not apply: a cgroup can
       both contain member processes (or thread) and exercise controllers on child cgroups.

   Rules for writing to cgroup.type and creating threaded subtrees
       A number of rules apply when writing to the cgroup.type file:

       *  Only the string "threaded" may be written.  In other words, the only  explicit  transi-
          tion that is possible is to convert a domain cgroup to type threaded.

       *  The  effect  of writing "threaded" depends on the current value in cgroup.type, as fol-
          lows:

          o  domain or domain threaded: start the creation of a threaded subtree (whose  root  is
             the parent of this cgroup) via the first of the pathways described above;

          o  domain invalid: convert this cgroup (which is inside a threaded subtree) to a usable
             (i.e., threaded) state;

          o  threaded: no effect (a "no-op").

       *  We can't write to a cgroup.type file if the parent's type is domain invalid.  In  other
          words,  the  cgroups of a threaded subtree must be converted to the threaded state in a
          top-down manner.

       There are also some constraints that must be satisfied in order to create a threaded  sub-
       tree rooted at the cgroup x:

       *  There can be no member processes in the descendant cgroups of x.  (The cgroup x can it-
          self have member processes.)

       *  No domain controllers may be enabled in x's cgroup.subtree_control file.

       If any of the above constraints is violated, then an attempt  to  write  "threaded"  to  a
       cgroup.type file fails with the error ENOTSUP.

   The "domain threaded" cgroup type
       According  to  the  pathways  described  above,  the type of a cgroup can change to domain
       threaded in either of the following cases:

       *  The string "threaded" is written to a child cgroup.

       *  A threaded controller is enabled inside the cgroup and a process is made  a  member  of
          the cgroup.

       A  domain  threaded  cgroup,  x,  can revert to the type domain if the above conditions no
       longer hold true--that is, if all threaded child cgroups of x are removed and either x  no
       longer has threaded controllers enabled or no longer has member processes.

       When a domain threaded cgroup x reverts to the type domain:

       *  All  domain  invalid descendants of x that are not in lower-level threaded subtrees re-
          vert to the type domain.

       *  The root cgroups in any  lower-level  threaded  subtrees  revert  to  the  type  domain
          threaded.

   Exceptions for the root cgroup
       The root cgroup of the v2 hierarchy is treated exceptionally: it can be the parent of both
       domain and threaded cgroups.  If the string "threaded" is written to the cgroup.type  file
       of one of the children of the root cgroup, then

       *  The type of that cgroup becomes threaded.

       *  The  type  of  any descendants of that cgroup that are not part of lower-level threaded
          subtrees changes to domain invalid.

       Note that in this case, there is no cgroup whose type becomes domain  threaded.   (Notion-
       ally, the root cgroup can be considered as the threaded root for the cgroup whose type was
       changed to threaded.)

       The aim of this exceptional treatment for the root cgroup is to allow  a  threaded  cgroup
       that  employs  the cpu controller to be placed as high as possible in the hierarchy, so as
       to minimize the (small) cost of traversing the cgroup hierarchy.

   The cgroups v2 "cpu" controller and realtime threads
       As at Linux 4.19, the cgroups v2 cpu controller  does  not  support  control  of  realtime
       threads  (specifically  threads  scheduled under any of the policies SCHED_FIFO, SCHED_RR,
       described SCHED_DEADLINE; see sched(7)).  Therefore, the cpu controller can be enabled  in
       the  root cgroup only if all realtime threads are in the root cgroup.  (If there are real-
       time threads in nonroot cgroups, then a write(2) of the string "+cpu" to  the  cgroup.sub-
       tree_control file fails with the error EINVAL.)

       On  some  systems, systemd(1) places certain realtime threads in nonroot cgroups in the v2
       hierarchy.  On such systems, these threads must first be moved to the root  cgroup  before
       the cpu controller can be enabled.

ERRORS
       The following errors can occur for mount(2):

       EBUSY  An  attempt  to mount a cgroup version 1 filesystem specified neither the name= op-
              tion (to mount a named hierarchy) nor a controller name (or all).

NOTES
       A child process created via fork(2) inherits its parent's cgroup memberships.  A process's
       cgroup memberships are preserved across execve(2).

       The clone3(2) CLONE_INTO_CGROUP flag can be used to create a child process that begins its
       life in a different version 2 cgroup from the parent process.

   /proc files
       /proc/cgroups (since Linux 2.6.24)
              This file contains information about the controllers that  are  compiled  into  the
              kernel.   An  example of the contents of this file (reformatted for readability) is
              the following:

                  #subsys_name    hierarchy      num_cgroups    enabled
                  cpuset          4              1              1
                  cpu             8              1              1
                  cpuacct         8              1              1
                  blkio           6              1              1
                  memory          3              1              1
                  devices         10             84             1
                  freezer         7              1              1
                  net_cls         9              1              1
                  perf_event      5              1              1
                  net_prio        9              1              1
                  hugetlb         0              1              0
                  pids            2              1              1

              The fields in this file are, from left to right:

              1. The name of the controller.

              2. The unique ID of the cgroup hierarchy on which this controller is  mounted.   If
                 multiple  cgroups v1 controllers are bound to the same hierarchy, then each will
                 show the same hierarchy ID in this field.  The value in this field will be 0 if:

                   a) the controller is not mounted on a cgroups v1 hierarchy;

                   b) the controller is bound to the cgroups v2 single unified hierarchy; or

                   c) the controller is disabled (see below).

              3. The number of control groups in this hierarchy using this controller.

              4. This field contains the value 1 if this controller is enabled, or 0  if  it  has
                 been disabled (via the cgroup_disable kernel command-line boot parameter).

       /proc/[pid]/cgroup (since Linux 2.6.24)
              This  file describes control groups to which the process with the corresponding PID
              belongs.  The displayed information differs for cgroups version 1 and version 2 hi-
              erarchies.

              For each cgroup hierarchy of which the process is a member, there is one entry con-
              taining three colon-separated fields:

                  hierarchy-ID:controller-list:cgroup-path

              For example:

                  5:cpuacct,cpu,cpuset:/daemons

              The colon-separated fields are, from left to right:

              1. For cgroups version 1 hierarchies, this field contains  a  unique  hierarchy  ID
                 number  that can be matched to a hierarchy ID in /proc/cgroups.  For the cgroups
                 version 2 hierarchy, this field contains the value 0.

              2. For cgroups version 1 hierarchies, this field contains a comma-separated list of
                 the  controllers  bound  to the hierarchy.  For the cgroups version 2 hierarchy,
                 this field is empty.

              3. This field contains the pathname of the control group in the hierarchy to  which
                 the  process belongs.  This pathname is relative to the mount point of the hier-
                 archy.

   /sys/kernel/cgroup files
       /sys/kernel/cgroup/delegate (since Linux 4.15)
              This file exports a list of the cgroups v2 files (one per line) that  are  delegat-
              able (i.e., whose ownership should be changed to the user ID of the delegatee).  In
              the future, the set of delegatable files may change or grow, and this file provides
              a way for the kernel to inform user-space applications of which files must be dele-
              gated.  As at Linux 4.15, one sees the following when inspecting this file:

                  $ cat /sys/kernel/cgroup/delegate
                  cgroup.procs
                  cgroup.subtree_control
                  cgroup.threads

       /sys/kernel/cgroup/features (since Linux 4.15)
              Over time, the set of cgroups v2 features that  are  provided  by  the  kernel  may
              change or grow, or some features may not be enabled by default.  This file provides
              a way for user-space applications to discover what features the running kernel sup-
              ports and has enabled.  Features are listed one per line:

                  $ cat /sys/kernel/cgroup/features
                  nsdelegate
                  memory_localevents

              The entries that can appear in this file are:

              memory_localevents (since Linux 5.2)
                     The kernel supports the memory_localevents mount option.

              nsdelegate (since Linux 4.15)
                     The kernel supports the nsdelegate mount option.

SEE ALSO
       prlimit(1),   systemd(1),   systemd-cgls(1),  systemd-cgtop(1),  clone(2),  ioprio_set(2),
       perf_event_open(2),   setrlimit(2),   cgroup_namespaces(7),   cpuset(7),    namespaces(7),
       sched(7), user_namespaces(7)

       The kernel source file Documentation/admin-guide/cgroup-v2.rst.

COLOPHON
       This  page  is  part of release 5.10 of the Linux man-pages project.  A description of the
       project, information about reporting bugs, and the latest version of  this  page,  can  be
       found at https://www.kernel.org/doc/man-pages/.

Linux                                       2020-08-13                                 CGROUPS(7)

Generated by $Id: phpMan.php,v 4.55 2007/09/05 04:42:51 chedong Exp $ Author: Che Dong
On Apache
Under GNU General Public License
2025-02-21 10:18 @18.218.42.199 CrawledBy Mozilla/5.0 AppleWebKit/537.36 (KHTML, like Gecko; compatible; ClaudeBot/1.0; +claudebot@anthropic.com)
Valid XHTML 1.0!Valid CSS!