Google Groups no longer supports new Usenet posts or subscriptions. Historical content remains viewable.
Dismiss

[PATCH] mm/slub: fix accumulate per cpu partial cache objects

10 views
Skip to first unread message

Wanpeng Li

unread,
Dec 27, 2013, 4:50:02 AM12/27/13
to
SLUB per cpu partial cache is a list of slab caches to accelerate objects
allocation. However, current codes just accumulate the objects number of
the first slab cache of per cpu partial cache instead of traverse the whole
list.

Signed-off-by: Wanpeng Li <liw...@linux.vnet.ibm.com>
---
mm/slub.c | 32 +++++++++++++++++++++++---------
1 files changed, 23 insertions(+), 9 deletions(-)

diff --git a/mm/slub.c b/mm/slub.c
index 545a170..799bfdc 100644
--- a/mm/slub.c
+++ b/mm/slub.c
@@ -4280,7 +4280,7 @@ static ssize_t show_slab_objects(struct kmem_cache *s,
struct kmem_cache_cpu *c = per_cpu_ptr(s->cpu_slab,
cpu);
int node;
- struct page *page;
+ struct page *page, *p;

page = ACCESS_ONCE(c->page);
if (!page)
@@ -4298,8 +4298,9 @@ static ssize_t show_slab_objects(struct kmem_cache *s,
nodes[node] += x;

page = ACCESS_ONCE(c->partial);
- if (page) {
- x = page->pobjects;
+ while ((p = page)) {
+ page = p->next;
+ x = p->pobjects;
total += x;
nodes[node] += x;
}
@@ -4520,13 +4521,15 @@ static ssize_t slabs_cpu_partial_show(struct kmem_cache *s, char *buf)
int pages = 0;
int cpu;
int len;
+ struct page *p;

for_each_online_cpu(cpu) {
struct page *page = per_cpu_ptr(s->cpu_slab, cpu)->partial;

- if (page) {
- pages += page->pages;
- objects += page->pobjects;
+ while ((p = page)) {
+ page = p->next;
+ pages += p->pages;
+ objects += p->pobjects;
}
}

@@ -4535,10 +4538,21 @@ static ssize_t slabs_cpu_partial_show(struct kmem_cache *s, char *buf)
#ifdef CONFIG_SMP
for_each_online_cpu(cpu) {
struct page *page = per_cpu_ptr(s->cpu_slab, cpu) ->partial;
+ objects = 0;
+ pages = 0;
+
+ if (!page)
+ continue;
+
+ while ((p = page)) {
+ page = p->next;
+ pages += p->pages;
+ objects += p->pobjects;
+ }

- if (page && len < PAGE_SIZE - 20)
- len += sprintf(buf + len, " C%d=%d(%d)", cpu,
- page->pobjects, page->pages);
+ if (len < PAGE_SIZE - 20)
+ len += sprintf(buf + len, " C%d=%d(%d)", cpu,
+ objects, pages);
}
#endif
return len + sprintf(buf + len, "\n");
--
1.7.7.6

--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majo...@vger.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/

Li Zefan

unread,
Dec 27, 2013, 9:00:01 PM12/27/13
to
Can we apply this patch first? It was sent month ago, but Pekka was not responsive.

=============================

[PATCH] slub: Fix calculation of cpu slabs

/sys/kernel/slab/:t-0000048 # cat cpu_slabs
231 N0=16 N1=215
/sys/kernel/slab/:t-0000048 # cat slabs
145 N0=36 N1=109

See, the number of slabs is smaller than that of cpu slabs.

The bug was introduced by commit 49e2258586b423684f03c278149ab46d8f8b6700
("slub: per cpu cache for partial pages").

We should use page->pages instead of page->pobjects when calculating
the number of cpu partial slabs. This also fixes the mapping of slabs
and nodes.

As there's no variable storing the number of total/active objects in
cpu partial slabs, and we don't have user interfaces requiring those
statistics, I just add WARN_ON for those cases.

Cc: <sta...@vger.kernel.org> # 3.2+
Signed-off-by: Li Zefan <liz...@huawei.com>
Acked-by: Christoph Lameter <c...@linux.com>
Reviewed-by: Wanpeng Li <liw...@linux.vnet.ibm.com>
---
mm/slub.c | 8 +++++++-
1 file changed, 7 insertions(+), 1 deletion(-)

diff --git a/mm/slub.c b/mm/slub.c
index e3ba1f2..6ea461d 100644
--- a/mm/slub.c
+++ b/mm/slub.c
@@ -4300,7 +4300,13 @@ static ssize_t show_slab_objects(struct kmem_cache *s,

page = ACCESS_ONCE(c->partial);
if (page) {
- x = page->pobjects;
+ node = page_to_nid(page);
+ if (flags & SO_TOTAL)
+ WARN_ON_ONCE(1);
+ else if (flags & SO_OBJECTS)
+ WARN_ON_ONCE(1);
+ else
+ x = page->pages;
total += x;
nodes[node] += x;
}
-- 1.8.0.2

Pekka Enberg

unread,
Dec 29, 2013, 6:50:01 AM12/29/13
to
Applied. Wanpeng, care to resend your patch?

Pekka Enberg

unread,
Dec 30, 2013, 5:00:01 AM12/30/13
to
On 12/30/2013 03:08 AM, Wanpeng Li wrote:
> Zefan's patch is good enough, mine doesn't need any more.

OK, thanks guys!

Pekka
0 new messages