slub: Enable sysfs support for !CONFIG_SLUB_DEBUG
Currently disabling CONFIG_SLUB_DEBUG also disabled SYSFS support meaning that the slabs cannot be tuned without DEBUG. Make SYSFS support independent of CONFIG_SLUB_DEBUG Signed-off-by: Christoph Lameter <cl@linux.com> Signed-off-by: Pekka Enberg <penberg@kernel.org>
This commit is contained in:
Родитель
15b7c51420
Коммит
ab4d5ed5ee
|
@ -87,7 +87,7 @@ struct kmem_cache {
|
||||||
unsigned long min_partial;
|
unsigned long min_partial;
|
||||||
const char *name; /* Name (only for display!) */
|
const char *name; /* Name (only for display!) */
|
||||||
struct list_head list; /* List of slab caches */
|
struct list_head list; /* List of slab caches */
|
||||||
#ifdef CONFIG_SLUB_DEBUG
|
#ifdef CONFIG_SYSFS
|
||||||
struct kobject kobj; /* For sysfs */
|
struct kobject kobj; /* For sysfs */
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
|
|
|
@ -353,7 +353,7 @@ config SLUB_DEBUG_ON
|
||||||
config SLUB_STATS
|
config SLUB_STATS
|
||||||
default n
|
default n
|
||||||
bool "Enable SLUB performance statistics"
|
bool "Enable SLUB performance statistics"
|
||||||
depends on SLUB && SLUB_DEBUG && SYSFS
|
depends on SLUB && SYSFS
|
||||||
help
|
help
|
||||||
SLUB statistics are useful to debug SLUBs allocation behavior in
|
SLUB statistics are useful to debug SLUBs allocation behavior in
|
||||||
order find ways to optimize the allocator. This should never be
|
order find ways to optimize the allocator. This should never be
|
||||||
|
|
40
mm/slub.c
40
mm/slub.c
|
@ -198,7 +198,7 @@ struct track {
|
||||||
|
|
||||||
enum track_item { TRACK_ALLOC, TRACK_FREE };
|
enum track_item { TRACK_ALLOC, TRACK_FREE };
|
||||||
|
|
||||||
#ifdef CONFIG_SLUB_DEBUG
|
#ifdef CONFIG_SYSFS
|
||||||
static int sysfs_slab_add(struct kmem_cache *);
|
static int sysfs_slab_add(struct kmem_cache *);
|
||||||
static int sysfs_slab_alias(struct kmem_cache *, const char *);
|
static int sysfs_slab_alias(struct kmem_cache *, const char *);
|
||||||
static void sysfs_slab_remove(struct kmem_cache *);
|
static void sysfs_slab_remove(struct kmem_cache *);
|
||||||
|
@ -1102,7 +1102,7 @@ static inline void slab_free_hook(struct kmem_cache *s, void *x) {}
|
||||||
static inline void slab_free_hook_irq(struct kmem_cache *s,
|
static inline void slab_free_hook_irq(struct kmem_cache *s,
|
||||||
void *object) {}
|
void *object) {}
|
||||||
|
|
||||||
#endif
|
#endif /* CONFIG_SLUB_DEBUG */
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Slab allocation and freeing
|
* Slab allocation and freeing
|
||||||
|
@ -3373,7 +3373,7 @@ void *__kmalloc_node_track_caller(size_t size, gfp_t gfpflags,
|
||||||
}
|
}
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
#ifdef CONFIG_SLUB_DEBUG
|
#ifdef CONFIG_SYSFS
|
||||||
static int count_inuse(struct page *page)
|
static int count_inuse(struct page *page)
|
||||||
{
|
{
|
||||||
return page->inuse;
|
return page->inuse;
|
||||||
|
@ -3383,7 +3383,9 @@ static int count_total(struct page *page)
|
||||||
{
|
{
|
||||||
return page->objects;
|
return page->objects;
|
||||||
}
|
}
|
||||||
|
#endif
|
||||||
|
|
||||||
|
#ifdef CONFIG_SLUB_DEBUG
|
||||||
static int validate_slab(struct kmem_cache *s, struct page *page,
|
static int validate_slab(struct kmem_cache *s, struct page *page,
|
||||||
unsigned long *map)
|
unsigned long *map)
|
||||||
{
|
{
|
||||||
|
@ -3474,6 +3476,7 @@ static long validate_slab_cache(struct kmem_cache *s)
|
||||||
kfree(map);
|
kfree(map);
|
||||||
return count;
|
return count;
|
||||||
}
|
}
|
||||||
|
#endif
|
||||||
|
|
||||||
#ifdef SLUB_RESILIENCY_TEST
|
#ifdef SLUB_RESILIENCY_TEST
|
||||||
static void resiliency_test(void)
|
static void resiliency_test(void)
|
||||||
|
@ -3532,9 +3535,12 @@ static void resiliency_test(void)
|
||||||
validate_slab_cache(kmalloc_caches[9]);
|
validate_slab_cache(kmalloc_caches[9]);
|
||||||
}
|
}
|
||||||
#else
|
#else
|
||||||
|
#ifdef CONFIG_SYSFS
|
||||||
static void resiliency_test(void) {};
|
static void resiliency_test(void) {};
|
||||||
#endif
|
#endif
|
||||||
|
#endif
|
||||||
|
|
||||||
|
#ifdef CONFIG_DEBUG
|
||||||
/*
|
/*
|
||||||
* Generate lists of code addresses where slabcache objects are allocated
|
* Generate lists of code addresses where slabcache objects are allocated
|
||||||
* and freed.
|
* and freed.
|
||||||
|
@ -3763,7 +3769,9 @@ static int list_locations(struct kmem_cache *s, char *buf,
|
||||||
len += sprintf(buf, "No data\n");
|
len += sprintf(buf, "No data\n");
|
||||||
return len;
|
return len;
|
||||||
}
|
}
|
||||||
|
#endif
|
||||||
|
|
||||||
|
#ifdef CONFIG_SYSFS
|
||||||
enum slab_stat_type {
|
enum slab_stat_type {
|
||||||
SL_ALL, /* All slabs */
|
SL_ALL, /* All slabs */
|
||||||
SL_PARTIAL, /* Only partially allocated slabs */
|
SL_PARTIAL, /* Only partially allocated slabs */
|
||||||
|
@ -3816,6 +3824,8 @@ static ssize_t show_slab_objects(struct kmem_cache *s,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
down_read(&slub_lock);
|
||||||
|
#ifdef CONFIG_SLUB_DEBUG
|
||||||
if (flags & SO_ALL) {
|
if (flags & SO_ALL) {
|
||||||
for_each_node_state(node, N_NORMAL_MEMORY) {
|
for_each_node_state(node, N_NORMAL_MEMORY) {
|
||||||
struct kmem_cache_node *n = get_node(s, node);
|
struct kmem_cache_node *n = get_node(s, node);
|
||||||
|
@ -3832,7 +3842,9 @@ static ssize_t show_slab_objects(struct kmem_cache *s,
|
||||||
nodes[node] += x;
|
nodes[node] += x;
|
||||||
}
|
}
|
||||||
|
|
||||||
} else if (flags & SO_PARTIAL) {
|
} else
|
||||||
|
#endif
|
||||||
|
if (flags & SO_PARTIAL) {
|
||||||
for_each_node_state(node, N_NORMAL_MEMORY) {
|
for_each_node_state(node, N_NORMAL_MEMORY) {
|
||||||
struct kmem_cache_node *n = get_node(s, node);
|
struct kmem_cache_node *n = get_node(s, node);
|
||||||
|
|
||||||
|
@ -3857,6 +3869,7 @@ static ssize_t show_slab_objects(struct kmem_cache *s,
|
||||||
return x + sprintf(buf + x, "\n");
|
return x + sprintf(buf + x, "\n");
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#ifdef CONFIG_SLUB_DEBUG
|
||||||
static int any_slab_objects(struct kmem_cache *s)
|
static int any_slab_objects(struct kmem_cache *s)
|
||||||
{
|
{
|
||||||
int node;
|
int node;
|
||||||
|
@ -3872,6 +3885,7 @@ static int any_slab_objects(struct kmem_cache *s)
|
||||||
}
|
}
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
#endif
|
||||||
|
|
||||||
#define to_slab_attr(n) container_of(n, struct slab_attribute, attr)
|
#define to_slab_attr(n) container_of(n, struct slab_attribute, attr)
|
||||||
#define to_slab(n) container_of(n, struct kmem_cache, kobj);
|
#define to_slab(n) container_of(n, struct kmem_cache, kobj);
|
||||||
|
@ -3973,11 +3987,13 @@ static ssize_t aliases_show(struct kmem_cache *s, char *buf)
|
||||||
}
|
}
|
||||||
SLAB_ATTR_RO(aliases);
|
SLAB_ATTR_RO(aliases);
|
||||||
|
|
||||||
|
#ifdef CONFIG_SLUB_DEBUG
|
||||||
static ssize_t slabs_show(struct kmem_cache *s, char *buf)
|
static ssize_t slabs_show(struct kmem_cache *s, char *buf)
|
||||||
{
|
{
|
||||||
return show_slab_objects(s, buf, SO_ALL);
|
return show_slab_objects(s, buf, SO_ALL);
|
||||||
}
|
}
|
||||||
SLAB_ATTR_RO(slabs);
|
SLAB_ATTR_RO(slabs);
|
||||||
|
#endif
|
||||||
|
|
||||||
static ssize_t partial_show(struct kmem_cache *s, char *buf)
|
static ssize_t partial_show(struct kmem_cache *s, char *buf)
|
||||||
{
|
{
|
||||||
|
@ -4003,6 +4019,7 @@ static ssize_t objects_partial_show(struct kmem_cache *s, char *buf)
|
||||||
}
|
}
|
||||||
SLAB_ATTR_RO(objects_partial);
|
SLAB_ATTR_RO(objects_partial);
|
||||||
|
|
||||||
|
#ifdef CONFIG_SLUB_DEBUG
|
||||||
static ssize_t total_objects_show(struct kmem_cache *s, char *buf)
|
static ssize_t total_objects_show(struct kmem_cache *s, char *buf)
|
||||||
{
|
{
|
||||||
return show_slab_objects(s, buf, SO_ALL|SO_TOTAL);
|
return show_slab_objects(s, buf, SO_ALL|SO_TOTAL);
|
||||||
|
@ -4055,6 +4072,7 @@ static ssize_t failslab_store(struct kmem_cache *s, const char *buf,
|
||||||
}
|
}
|
||||||
SLAB_ATTR(failslab);
|
SLAB_ATTR(failslab);
|
||||||
#endif
|
#endif
|
||||||
|
#endif
|
||||||
|
|
||||||
static ssize_t reclaim_account_show(struct kmem_cache *s, char *buf)
|
static ssize_t reclaim_account_show(struct kmem_cache *s, char *buf)
|
||||||
{
|
{
|
||||||
|
@ -4091,6 +4109,7 @@ static ssize_t destroy_by_rcu_show(struct kmem_cache *s, char *buf)
|
||||||
}
|
}
|
||||||
SLAB_ATTR_RO(destroy_by_rcu);
|
SLAB_ATTR_RO(destroy_by_rcu);
|
||||||
|
|
||||||
|
#ifdef CONFIG_SLUB_DEBUG
|
||||||
static ssize_t red_zone_show(struct kmem_cache *s, char *buf)
|
static ssize_t red_zone_show(struct kmem_cache *s, char *buf)
|
||||||
{
|
{
|
||||||
return sprintf(buf, "%d\n", !!(s->flags & SLAB_RED_ZONE));
|
return sprintf(buf, "%d\n", !!(s->flags & SLAB_RED_ZONE));
|
||||||
|
@ -4166,6 +4185,7 @@ static ssize_t validate_store(struct kmem_cache *s,
|
||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
SLAB_ATTR(validate);
|
SLAB_ATTR(validate);
|
||||||
|
#endif
|
||||||
|
|
||||||
static ssize_t shrink_show(struct kmem_cache *s, char *buf)
|
static ssize_t shrink_show(struct kmem_cache *s, char *buf)
|
||||||
{
|
{
|
||||||
|
@ -4186,6 +4206,7 @@ static ssize_t shrink_store(struct kmem_cache *s,
|
||||||
}
|
}
|
||||||
SLAB_ATTR(shrink);
|
SLAB_ATTR(shrink);
|
||||||
|
|
||||||
|
#ifdef CONFIG_SLUB_DEBUG
|
||||||
static ssize_t alloc_calls_show(struct kmem_cache *s, char *buf)
|
static ssize_t alloc_calls_show(struct kmem_cache *s, char *buf)
|
||||||
{
|
{
|
||||||
if (!(s->flags & SLAB_STORE_USER))
|
if (!(s->flags & SLAB_STORE_USER))
|
||||||
|
@ -4201,6 +4222,7 @@ static ssize_t free_calls_show(struct kmem_cache *s, char *buf)
|
||||||
return list_locations(s, buf, TRACK_FREE);
|
return list_locations(s, buf, TRACK_FREE);
|
||||||
}
|
}
|
||||||
SLAB_ATTR_RO(free_calls);
|
SLAB_ATTR_RO(free_calls);
|
||||||
|
#endif
|
||||||
|
|
||||||
#ifdef CONFIG_NUMA
|
#ifdef CONFIG_NUMA
|
||||||
static ssize_t remote_node_defrag_ratio_show(struct kmem_cache *s, char *buf)
|
static ssize_t remote_node_defrag_ratio_show(struct kmem_cache *s, char *buf)
|
||||||
|
@ -4307,25 +4329,33 @@ static struct attribute *slab_attrs[] = {
|
||||||
&min_partial_attr.attr,
|
&min_partial_attr.attr,
|
||||||
&objects_attr.attr,
|
&objects_attr.attr,
|
||||||
&objects_partial_attr.attr,
|
&objects_partial_attr.attr,
|
||||||
|
#ifdef CONFIG_SLUB_DEBUG
|
||||||
&total_objects_attr.attr,
|
&total_objects_attr.attr,
|
||||||
&slabs_attr.attr,
|
&slabs_attr.attr,
|
||||||
|
#endif
|
||||||
&partial_attr.attr,
|
&partial_attr.attr,
|
||||||
&cpu_slabs_attr.attr,
|
&cpu_slabs_attr.attr,
|
||||||
&ctor_attr.attr,
|
&ctor_attr.attr,
|
||||||
&aliases_attr.attr,
|
&aliases_attr.attr,
|
||||||
&align_attr.attr,
|
&align_attr.attr,
|
||||||
|
#ifdef CONFIG_SLUB_DEBUG
|
||||||
&sanity_checks_attr.attr,
|
&sanity_checks_attr.attr,
|
||||||
&trace_attr.attr,
|
&trace_attr.attr,
|
||||||
|
#endif
|
||||||
&hwcache_align_attr.attr,
|
&hwcache_align_attr.attr,
|
||||||
&reclaim_account_attr.attr,
|
&reclaim_account_attr.attr,
|
||||||
&destroy_by_rcu_attr.attr,
|
&destroy_by_rcu_attr.attr,
|
||||||
|
#ifdef CONFIG_SLUB_DEBUG
|
||||||
&red_zone_attr.attr,
|
&red_zone_attr.attr,
|
||||||
&poison_attr.attr,
|
&poison_attr.attr,
|
||||||
&store_user_attr.attr,
|
&store_user_attr.attr,
|
||||||
&validate_attr.attr,
|
&validate_attr.attr,
|
||||||
|
#endif
|
||||||
&shrink_attr.attr,
|
&shrink_attr.attr,
|
||||||
|
#ifdef CONFIG_SLUB_DEBUG
|
||||||
&alloc_calls_attr.attr,
|
&alloc_calls_attr.attr,
|
||||||
&free_calls_attr.attr,
|
&free_calls_attr.attr,
|
||||||
|
#endif
|
||||||
#ifdef CONFIG_ZONE_DMA
|
#ifdef CONFIG_ZONE_DMA
|
||||||
&cache_dma_attr.attr,
|
&cache_dma_attr.attr,
|
||||||
#endif
|
#endif
|
||||||
|
@ -4608,7 +4638,7 @@ static int __init slab_sysfs_init(void)
|
||||||
}
|
}
|
||||||
|
|
||||||
__initcall(slab_sysfs_init);
|
__initcall(slab_sysfs_init);
|
||||||
#endif
|
#endif /* CONFIG_SYSFS */
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* The /proc/slabinfo ABI
|
* The /proc/slabinfo ABI
|
||||||
|
|
Загрузка…
Ссылка в новой задаче