[RHEL7,COMMIT] mm/mem_cgroup_iter: Provide _iter_invalidate() the dying memcg as an argument

Submitted by Vasily Averin on March 3, 2021, 6:26 a.m.


Message ID 202103030626.1236QCv9009262@vz7build.vvs.sw.ru
State New
Series "mm/mem_cgroup_iter: Reduce the number of iterator restarts upon cgroup removals"
Headers show

Commit Message

Vasily Averin March 3, 2021, 6:26 a.m.
The commit is pushed to "branch-rh7-3.10.0-1160.15.2.vz7.173.x-ovz" and will appear at https://src.openvz.org/scm/ovz/vzkernel.git
after rh7-3.10.0-1160.15.2.vz7.173.1
commit 398ccdab44c6959a26b79132c2e69c93acef327a
Author: Konstantin Khorenko <khorenko@virtuozzo.com>
Date:   Wed Mar 3 09:26:12 2021 +0300

    mm/mem_cgroup_iter: Provide _iter_invalidate() the dying memcg as an argument
    Patch-set description:
    May thanks to Kirill Tkhai for his bright ideas and review!
    Problem description from the user point of view:
      * the Node is slow
      * the Node has a lot of free RAM
      * the Node has a lot of swapin/swapout
      * kswapd is always running
    Problem in a nutshell from technical point of view:
      * kswapd is looping in shrink_zone() inside the loop
          do {} while ((memcg = mem_cgroup_iter(root, memcg, &reclaim)));
        (and never goes trough the outer loop)
      * there are a quite a number of memory cgroups of the Node (~1000)
      * some cgroups are hard to reclaim (reclaim may take ~3 seconds),
        this is because of very busy disk due to permanent swapin/swapout
      * mem_cgroup_iter() does not have success scanning all cgroups
        in a row, it restarts from the root cgroup one time after
        another (after different number of cgroups scanned)
    Q: Why does mem_cgroup_iter() restart from the root memcg?
    A: Because it is invalidated once some memory cgroup is
       destroyed on the Node.
       Note: ANY memory cgroup destroy on the Node leads to iter
    The following patchset solves this problem in the following way:
    there is no need to restart the iter until we see the iter has
    the position which is exactly the memory cgroup being destroyed.
    The patchset ensures the iter->last_visited is NULL-ified on
    invalidation and thus restarts only in the unlikely case when
    the iter points to the memcg being destroyed.
    Testing: i've tested this patchset using modified kernel which breaks
    the memcg iterator in case of global reclaim with probability of 2%.
    3 kernels have been tested: "release", KASAN-only, "debug" kernels.
    Each worked for 12 hours, no issues, from 12000 to 26000 races were
    caught during this period (i.e. dying memcg was found in some iterator
    and wiped).
    The testing scenario is documented in the jira issue.
    +++ Current patch description:
    It will be used by next patches when we search for this pointer
    stored in iterators.
    Signed-off-by: Konstantin Khorenko <khorenko@virtuozzo.com>
    Reviewed-by: Kirill Tkhai <ktkhai@virtuozzo.com>
 mm/memcontrol.c | 7 ++++---
 1 file changed, 4 insertions(+), 3 deletions(-)

Patch hide | download patch | download mbox

diff --git a/mm/memcontrol.c b/mm/memcontrol.c
index 4d25cd3..393d927 100644
--- a/mm/memcontrol.c
+++ b/mm/memcontrol.c
@@ -1577,7 +1577,8 @@  skip_node:
 	return NULL;
-static void mem_cgroup_iter_invalidate(struct mem_cgroup *root)
+static void mem_cgroup_iter_invalidate(struct mem_cgroup *root,
+				       struct mem_cgroup *dead_memcg)
 	 * When a group in the hierarchy below root is destroyed, the
@@ -6874,14 +6875,14 @@  static void mem_cgroup_invalidate_reclaim_iterators(struct mem_cgroup *memcg)
 	struct mem_cgroup *parent = memcg;
 	while ((parent = parent_mem_cgroup(parent)))
-		mem_cgroup_iter_invalidate(parent);
+		mem_cgroup_iter_invalidate(parent, memcg);
 	 * if the root memcg is not hierarchical we have to check it
 	 * explicitely.
 	if (!root_mem_cgroup->use_hierarchy)
-		mem_cgroup_iter_invalidate(root_mem_cgroup);
+		mem_cgroup_iter_invalidate(root_mem_cgroup, memcg);
 static void mem_cgroup_free_all(struct mem_cgroup *memcg)