License cleanup: add SPDX GPL-2.0 license identifier to files with no license
Many source files in the tree are missing licensing information, which
makes it harder for compliance tools to determine the correct license.
By default all files without license information are under the default
license of the kernel, which is GPL version 2.
Update the files which contain no license information with the 'GPL-2.0'
SPDX license identifier. The SPDX identifier is a legally binding
shorthand, which can be used instead of the full boiler plate text.
This patch is based on work done by Thomas Gleixner and Kate Stewart and
Philippe Ombredanne.
How this work was done:
Patches were generated and checked against linux-4.14-rc6 for a subset of
the use cases:
- file had no licensing information it it.
- file was a */uapi/* one with no licensing information in it,
- file was a */uapi/* one with existing licensing information,
Further patches will be generated in subsequent months to fix up cases
where non-standard license headers were used, and references to license
had to be inferred by heuristics based on keywords.
The analysis to determine which SPDX License Identifier to be applied to
a file was done in a spreadsheet of side by side results from of the
output of two independent scanners (ScanCode & Windriver) producing SPDX
tag:value files created by Philippe Ombredanne. Philippe prepared the
base worksheet, and did an initial spot review of a few 1000 files.
The 4.13 kernel was the starting point of the analysis with 60,537 files
assessed. Kate Stewart did a file by file comparison of the scanner
results in the spreadsheet to determine which SPDX license identifier(s)
to be applied to the file. She confirmed any determination that was not
immediately clear with lawyers working with the Linux Foundation.
Criteria used to select files for SPDX license identifier tagging was:
- Files considered eligible had to be source code files.
- Make and config files were included as candidates if they contained >5
lines of source
- File already had some variant of a license header in it (even if <5
lines).
All documentation files were explicitly excluded.
The following heuristics were used to determine which SPDX license
identifiers to apply.
- when both scanners couldn't find any license traces, file was
considered to have no license information in it, and the top level
COPYING file license applied.
For non */uapi/* files that summary was:
SPDX license identifier # files
---------------------------------------------------|-------
GPL-2.0 11139
and resulted in the first patch in this series.
If that file was a */uapi/* path one, it was "GPL-2.0 WITH
Linux-syscall-note" otherwise it was "GPL-2.0". Results of that was:
SPDX license identifier # files
---------------------------------------------------|-------
GPL-2.0 WITH Linux-syscall-note 930
and resulted in the second patch in this series.
- if a file had some form of licensing information in it, and was one
of the */uapi/* ones, it was denoted with the Linux-syscall-note if
any GPL family license was found in the file or had no licensing in
it (per prior point). Results summary:
SPDX license identifier # files
---------------------------------------------------|------
GPL-2.0 WITH Linux-syscall-note 270
GPL-2.0+ WITH Linux-syscall-note 169
((GPL-2.0 WITH Linux-syscall-note) OR BSD-2-Clause) 21
((GPL-2.0 WITH Linux-syscall-note) OR BSD-3-Clause) 17
LGPL-2.1+ WITH Linux-syscall-note 15
GPL-1.0+ WITH Linux-syscall-note 14
((GPL-2.0+ WITH Linux-syscall-note) OR BSD-3-Clause) 5
LGPL-2.0+ WITH Linux-syscall-note 4
LGPL-2.1 WITH Linux-syscall-note 3
((GPL-2.0 WITH Linux-syscall-note) OR MIT) 3
((GPL-2.0 WITH Linux-syscall-note) AND MIT) 1
and that resulted in the third patch in this series.
- when the two scanners agreed on the detected license(s), that became
the concluded license(s).
- when there was disagreement between the two scanners (one detected a
license but the other didn't, or they both detected different
licenses) a manual inspection of the file occurred.
- In most cases a manual inspection of the information in the file
resulted in a clear resolution of the license that should apply (and
which scanner probably needed to revisit its heuristics).
- When it was not immediately clear, the license identifier was
confirmed with lawyers working with the Linux Foundation.
- If there was any question as to the appropriate license identifier,
the file was flagged for further research and to be revisited later
in time.
In total, over 70 hours of logged manual review was done on the
spreadsheet to determine the SPDX license identifiers to apply to the
source files by Kate, Philippe, Thomas and, in some cases, confirmation
by lawyers working with the Linux Foundation.
Kate also obtained a third independent scan of the 4.13 code base from
FOSSology, and compared selected files where the other two scanners
disagreed against that SPDX file, to see if there was new insights. The
Windriver scanner is based on an older version of FOSSology in part, so
they are related.
Thomas did random spot checks in about 500 files from the spreadsheets
for the uapi headers and agreed with SPDX license identifier in the
files he inspected. For the non-uapi files Thomas did random spot checks
in about 15000 files.
In initial set of patches against 4.14-rc6, 3 files were found to have
copy/paste license identifier errors, and have been fixed to reflect the
correct identifier.
Additionally Philippe spent 10 hours this week doing a detailed manual
inspection and review of the 12,461 patched files from the initial patch
version early this week with:
- a full scancode scan run, collecting the matched texts, detected
license ids and scores
- reviewing anything where there was a license detected (about 500+
files) to ensure that the applied SPDX license was correct
- reviewing anything where there was no detection but the patch license
was not GPL-2.0 WITH Linux-syscall-note to ensure that the applied
SPDX license was correct
This produced a worksheet with 20 files needing minor correction. This
worksheet was then exported into 3 different .csv files for the
different types of files to be modified.
These .csv files were then reviewed by Greg. Thomas wrote a script to
parse the csv files and add the proper SPDX tag to the file, in the
format that the file expected. This script was further refined by Greg
based on the output to detect more types of files automatically and to
distinguish between header and source .c files (which need different
comment types.) Finally Greg ran the script using the .csv files to
generate the patches.
Reviewed-by: Kate Stewart <kstewart@linuxfoundation.org>
Reviewed-by: Philippe Ombredanne <pombredanne@nexb.com>
Reviewed-by: Thomas Gleixner <tglx@linutronix.de>
Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
2017-11-01 17:07:57 +03:00
|
|
|
/* SPDX-License-Identifier: GPL-2.0 */
|
2007-05-07 01:49:36 +04:00
|
|
|
#ifndef _LINUX_SLUB_DEF_H
|
|
|
|
#define _LINUX_SLUB_DEF_H
|
|
|
|
|
|
|
|
/*
|
|
|
|
* SLUB : A Slab allocator without object queues.
|
|
|
|
*
|
2008-07-04 20:59:22 +04:00
|
|
|
* (C) 2007 SGI, Christoph Lameter
|
2007-05-07 01:49:36 +04:00
|
|
|
*/
|
|
|
|
#include <linux/kobject.h>
|
|
|
|
|
2008-02-08 04:47:41 +03:00
|
|
|
enum stat_item {
|
|
|
|
ALLOC_FASTPATH, /* Allocation from cpu slab */
|
|
|
|
ALLOC_SLOWPATH, /* Allocation by getting a new cpu slab */
|
2013-11-08 16:47:36 +04:00
|
|
|
FREE_FASTPATH, /* Free to cpu slab */
|
2008-02-08 04:47:41 +03:00
|
|
|
FREE_SLOWPATH, /* Freeing not to cpu slab */
|
|
|
|
FREE_FROZEN, /* Freeing to frozen slab */
|
|
|
|
FREE_ADD_PARTIAL, /* Freeing moves slab to partial list */
|
|
|
|
FREE_REMOVE_PARTIAL, /* Freeing removes last object */
|
2012-02-03 19:34:56 +04:00
|
|
|
ALLOC_FROM_PARTIAL, /* Cpu slab acquired from node partial list */
|
2008-02-08 04:47:41 +03:00
|
|
|
ALLOC_SLAB, /* Cpu slab acquired from page allocator */
|
|
|
|
ALLOC_REFILL, /* Refill cpu slab from slab freelist */
|
2011-06-01 21:25:57 +04:00
|
|
|
ALLOC_NODE_MISMATCH, /* Switching cpu slab */
|
2008-02-08 04:47:41 +03:00
|
|
|
FREE_SLAB, /* Slab freed to the page allocator */
|
|
|
|
CPUSLAB_FLUSH, /* Abandoning of the cpu slab */
|
|
|
|
DEACTIVATE_FULL, /* Cpu slab was full when deactivated */
|
|
|
|
DEACTIVATE_EMPTY, /* Cpu slab was empty when deactivated */
|
|
|
|
DEACTIVATE_TO_HEAD, /* Cpu slab was moved to the head of partials */
|
|
|
|
DEACTIVATE_TO_TAIL, /* Cpu slab was moved to the tail of partials */
|
|
|
|
DEACTIVATE_REMOTE_FREES,/* Slab contained remotely freed objects */
|
2011-06-01 21:25:58 +04:00
|
|
|
DEACTIVATE_BYPASS, /* Implicit deactivation */
|
2008-04-14 20:11:40 +04:00
|
|
|
ORDER_FALLBACK, /* Number of times fallback was necessary */
|
2011-03-22 21:35:00 +03:00
|
|
|
CMPXCHG_DOUBLE_CPU_FAIL,/* Failure of this_cpu_cmpxchg_double */
|
2011-06-01 21:25:49 +04:00
|
|
|
CMPXCHG_DOUBLE_FAIL, /* Number of times that cmpxchg double did not match */
|
2011-08-10 01:12:27 +04:00
|
|
|
CPU_PARTIAL_ALLOC, /* Used cpu partial on alloc */
|
2012-02-03 19:34:56 +04:00
|
|
|
CPU_PARTIAL_FREE, /* Refill cpu partial on free */
|
|
|
|
CPU_PARTIAL_NODE, /* Refill cpu partial from node partial */
|
|
|
|
CPU_PARTIAL_DRAIN, /* Drain cpu partial to node partial */
|
2008-02-08 04:47:41 +03:00
|
|
|
NR_SLUB_STAT_ITEMS };
|
|
|
|
|
2007-10-16 12:26:05 +04:00
|
|
|
struct kmem_cache_cpu {
|
2011-02-25 20:38:54 +03:00
|
|
|
void **freelist; /* Pointer to next available object */
|
|
|
|
unsigned long tid; /* Globally unique transaction id */
|
2008-01-08 10:20:31 +03:00
|
|
|
struct page *page; /* The slab from which we are allocating */
|
2017-07-07 01:36:31 +03:00
|
|
|
#ifdef CONFIG_SLUB_CPU_PARTIAL
|
2011-08-10 01:12:27 +04:00
|
|
|
struct page *partial; /* Partially allocated frozen slabs */
|
2017-07-07 01:36:31 +03:00
|
|
|
#endif
|
2008-02-08 04:47:41 +03:00
|
|
|
#ifdef CONFIG_SLUB_STATS
|
|
|
|
unsigned stat[NR_SLUB_STAT_ITEMS];
|
|
|
|
#endif
|
2007-10-16 12:26:08 +04:00
|
|
|
};
|
2007-10-16 12:26:05 +04:00
|
|
|
|
2017-07-07 01:36:31 +03:00
|
|
|
#ifdef CONFIG_SLUB_CPU_PARTIAL
|
|
|
|
#define slub_percpu_partial(c) ((c)->partial)
|
|
|
|
|
|
|
|
#define slub_set_percpu_partial(c, p) \
|
|
|
|
({ \
|
|
|
|
slub_percpu_partial(c) = (p)->next; \
|
|
|
|
})
|
|
|
|
|
|
|
|
#define slub_percpu_partial_read_once(c) READ_ONCE(slub_percpu_partial(c))
|
|
|
|
#else
|
|
|
|
#define slub_percpu_partial(c) NULL
|
|
|
|
|
|
|
|
#define slub_set_percpu_partial(c, p)
|
|
|
|
|
|
|
|
#define slub_percpu_partial_read_once(c) NULL
|
|
|
|
#endif // CONFIG_SLUB_CPU_PARTIAL
|
|
|
|
|
2008-04-14 20:11:31 +04:00
|
|
|
/*
|
|
|
|
* Word size structure that can be atomically updated or read and that
|
|
|
|
* contains both the order and the number of objects that a slab of the
|
|
|
|
* given order would contain.
|
|
|
|
*/
|
|
|
|
struct kmem_cache_order_objects {
|
2018-04-06 02:21:39 +03:00
|
|
|
unsigned int x;
|
2008-04-14 20:11:31 +04:00
|
|
|
};
|
|
|
|
|
2007-05-07 01:49:36 +04:00
|
|
|
/*
|
|
|
|
* Slab cache management.
|
|
|
|
*/
|
|
|
|
struct kmem_cache {
|
2010-08-07 16:29:22 +04:00
|
|
|
struct kmem_cache_cpu __percpu *cpu_slab;
|
2019-03-06 02:42:07 +03:00
|
|
|
/* Used for retrieving partial slabs, etc. */
|
2017-11-16 04:32:18 +03:00
|
|
|
slab_flags_t flags;
|
2011-02-25 20:38:51 +03:00
|
|
|
unsigned long min_partial;
|
2019-03-06 02:42:07 +03:00
|
|
|
unsigned int size; /* The size of an object including metadata */
|
|
|
|
unsigned int object_size;/* The size of an object without metadata */
|
|
|
|
unsigned int offset; /* Free pointer offset */
|
2017-07-07 01:36:34 +03:00
|
|
|
#ifdef CONFIG_SLUB_CPU_PARTIAL
|
2018-04-06 02:21:10 +03:00
|
|
|
/* Number of per cpu partial objects to keep around */
|
|
|
|
unsigned int cpu_partial;
|
2017-07-07 01:36:34 +03:00
|
|
|
#endif
|
2008-04-14 20:11:31 +04:00
|
|
|
struct kmem_cache_order_objects oo;
|
2007-05-07 01:49:36 +04:00
|
|
|
|
|
|
|
/* Allocation and freeing of slabs */
|
2008-04-14 20:11:40 +04:00
|
|
|
struct kmem_cache_order_objects max;
|
2008-04-14 20:11:40 +04:00
|
|
|
struct kmem_cache_order_objects min;
|
2008-02-15 01:21:32 +03:00
|
|
|
gfp_t allocflags; /* gfp flags to use on each alloc */
|
2007-05-07 01:49:36 +04:00
|
|
|
int refcount; /* Refcount for slab cache destroy */
|
2008-07-26 06:45:34 +04:00
|
|
|
void (*ctor)(void *);
|
2018-04-06 02:21:06 +03:00
|
|
|
unsigned int inuse; /* Offset to metadata */
|
2018-04-06 02:21:02 +03:00
|
|
|
unsigned int align; /* Alignment */
|
2018-04-06 02:20:55 +03:00
|
|
|
unsigned int red_left_pad; /* Left redzone padding size */
|
2007-05-07 01:49:36 +04:00
|
|
|
const char *name; /* Name (only for display!) */
|
|
|
|
struct list_head list; /* List of slab caches */
|
2010-10-05 22:57:26 +04:00
|
|
|
#ifdef CONFIG_SYSFS
|
2007-05-07 01:49:36 +04:00
|
|
|
struct kobject kobj; /* For sysfs */
|
slub: make sysfs file removal asynchronous
Commit bf5eb3de3847 ("slub: separate out sysfs_slab_release() from
sysfs_slab_remove()") made slub sysfs file removals synchronous to
kmem_cache shutdown.
Unfortunately, this created a possible ABBA deadlock between slab_mutex
and sysfs draining mechanism triggering the following lockdep warning.
======================================================
[ INFO: possible circular locking dependency detected ]
4.10.0-test+ #48 Not tainted
-------------------------------------------------------
rmmod/1211 is trying to acquire lock:
(s_active#120){++++.+}, at: [<ffffffff81308073>] kernfs_remove+0x23/0x40
but task is already holding lock:
(slab_mutex){+.+.+.}, at: [<ffffffff8120f691>] kmem_cache_destroy+0x41/0x2d0
which lock already depends on the new lock.
the existing dependency chain (in reverse order) is:
-> #1 (slab_mutex){+.+.+.}:
lock_acquire+0xf6/0x1f0
__mutex_lock+0x75/0x950
mutex_lock_nested+0x1b/0x20
slab_attr_store+0x75/0xd0
sysfs_kf_write+0x45/0x60
kernfs_fop_write+0x13c/0x1c0
__vfs_write+0x28/0x120
vfs_write+0xc8/0x1e0
SyS_write+0x49/0xa0
entry_SYSCALL_64_fastpath+0x1f/0xc2
-> #0 (s_active#120){++++.+}:
__lock_acquire+0x10ed/0x1260
lock_acquire+0xf6/0x1f0
__kernfs_remove+0x254/0x320
kernfs_remove+0x23/0x40
sysfs_remove_dir+0x51/0x80
kobject_del+0x18/0x50
__kmem_cache_shutdown+0x3e6/0x460
kmem_cache_destroy+0x1fb/0x2d0
kvm_exit+0x2d/0x80 [kvm]
vmx_exit+0x19/0xa1b [kvm_intel]
SyS_delete_module+0x198/0x1f0
entry_SYSCALL_64_fastpath+0x1f/0xc2
other info that might help us debug this:
Possible unsafe locking scenario:
CPU0 CPU1
---- ----
lock(slab_mutex);
lock(s_active#120);
lock(slab_mutex);
lock(s_active#120);
*** DEADLOCK ***
2 locks held by rmmod/1211:
#0: (cpu_hotplug.dep_map){++++++}, at: [<ffffffff810a7877>] get_online_cpus+0x37/0x80
#1: (slab_mutex){+.+.+.}, at: [<ffffffff8120f691>] kmem_cache_destroy+0x41/0x2d0
stack backtrace:
CPU: 3 PID: 1211 Comm: rmmod Not tainted 4.10.0-test+ #48
Hardware name: Hewlett-Packard HP Compaq Pro 6300 SFF/339A, BIOS K01 v02.05 05/07/2012
Call Trace:
print_circular_bug+0x1be/0x210
__lock_acquire+0x10ed/0x1260
lock_acquire+0xf6/0x1f0
__kernfs_remove+0x254/0x320
kernfs_remove+0x23/0x40
sysfs_remove_dir+0x51/0x80
kobject_del+0x18/0x50
__kmem_cache_shutdown+0x3e6/0x460
kmem_cache_destroy+0x1fb/0x2d0
kvm_exit+0x2d/0x80 [kvm]
vmx_exit+0x19/0xa1b [kvm_intel]
SyS_delete_module+0x198/0x1f0
? SyS_delete_module+0x5/0x1f0
entry_SYSCALL_64_fastpath+0x1f/0xc2
It'd be the cleanest to deal with the issue by removing sysfs files
without holding slab_mutex before the rest of shutdown; however, given
the current code structure, it is pretty difficult to do so.
This patch punts sysfs file removal to a work item. Before commit
bf5eb3de3847, the removal was punted to a RCU delayed work item which is
executed after release. Now, we're punting to a different work item on
shutdown which still maintains the goal removing the sysfs files earlier
when destroying kmem_caches.
Link: http://lkml.kernel.org/r/20170620204512.GI21326@htj.duckdns.org
Fixes: bf5eb3de3847 ("slub: separate out sysfs_slab_release() from sysfs_slab_remove()")
Signed-off-by: Tejun Heo <tj@kernel.org>
Reported-by: Steven Rostedt (VMware) <rostedt@goodmis.org>
Tested-by: Steven Rostedt (VMware) <rostedt@goodmis.org>
Cc: Christoph Lameter <cl@linux.com>
Cc: Pekka Enberg <penberg@kernel.org>
Cc: David Rientjes <rientjes@google.com>
Cc: Joonsoo Kim <iamjoonsoo.kim@lge.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2017-06-24 01:08:52 +03:00
|
|
|
struct work_struct kobj_remove_work;
|
2007-07-17 15:03:24 +04:00
|
|
|
#endif
|
2016-01-21 02:02:32 +03:00
|
|
|
#ifdef CONFIG_MEMCG
|
2015-02-13 01:59:20 +03:00
|
|
|
struct memcg_cache_params memcg_params;
|
2019-03-06 02:42:07 +03:00
|
|
|
/* For propagation, maximum size of a stored attr */
|
2018-04-06 02:20:51 +03:00
|
|
|
unsigned int max_attr_size;
|
2014-04-08 02:39:31 +04:00
|
|
|
#ifdef CONFIG_SYSFS
|
|
|
|
struct kset *memcg_kset;
|
|
|
|
#endif
|
2012-12-19 02:22:27 +04:00
|
|
|
#endif
|
2007-05-07 01:49:36 +04:00
|
|
|
|
2017-09-07 02:19:18 +03:00
|
|
|
#ifdef CONFIG_SLAB_FREELIST_HARDENED
|
|
|
|
unsigned long random;
|
|
|
|
#endif
|
|
|
|
|
2007-05-07 01:49:36 +04:00
|
|
|
#ifdef CONFIG_NUMA
|
2008-01-08 10:20:26 +03:00
|
|
|
/*
|
|
|
|
* Defragmentation by allocating from a remote node.
|
|
|
|
*/
|
2018-04-06 02:20:48 +03:00
|
|
|
unsigned int remote_node_defrag_ratio;
|
2007-05-07 01:49:36 +04:00
|
|
|
#endif
|
2016-07-27 01:21:59 +03:00
|
|
|
|
|
|
|
#ifdef CONFIG_SLAB_FREELIST_RANDOM
|
|
|
|
unsigned int *random_seq;
|
|
|
|
#endif
|
|
|
|
|
2016-07-29 01:49:07 +03:00
|
|
|
#ifdef CONFIG_KASAN
|
|
|
|
struct kasan_cache kasan_info;
|
|
|
|
#endif
|
|
|
|
|
2018-04-06 02:21:31 +03:00
|
|
|
unsigned int useroffset; /* Usercopy region offset */
|
|
|
|
unsigned int usersize; /* Usercopy region size */
|
usercopy: Prepare for usercopy whitelisting
This patch prepares the slab allocator to handle caches having annotations
(useroffset and usersize) defining usercopy regions.
This patch is modified from Brad Spengler/PaX Team's PAX_USERCOPY
whitelisting code in the last public patch of grsecurity/PaX based on
my understanding of the code. Changes or omissions from the original
code are mine and don't reflect the original grsecurity/PaX code.
Currently, hardened usercopy performs dynamic bounds checking on slab
cache objects. This is good, but still leaves a lot of kernel memory
available to be copied to/from userspace in the face of bugs. To further
restrict what memory is available for copying, this creates a way to
whitelist specific areas of a given slab cache object for copying to/from
userspace, allowing much finer granularity of access control. Slab caches
that are never exposed to userspace can declare no whitelist for their
objects, thereby keeping them unavailable to userspace via dynamic copy
operations. (Note, an implicit form of whitelisting is the use of constant
sizes in usercopy operations and get_user()/put_user(); these bypass
hardened usercopy checks since these sizes cannot change at runtime.)
To support this whitelist annotation, usercopy region offset and size
members are added to struct kmem_cache. The slab allocator receives a
new function, kmem_cache_create_usercopy(), that creates a new cache
with a usercopy region defined, suitable for declaring spans of fields
within the objects that get copied to/from userspace.
In this patch, the default kmem_cache_create() marks the entire allocation
as whitelisted, leaving it semantically unchanged. Once all fine-grained
whitelists have been added (in subsequent patches), this will be changed
to a usersize of 0, making caches created with kmem_cache_create() not
copyable to/from userspace.
After the entire usercopy whitelist series is applied, less than 15%
of the slab cache memory remains exposed to potential usercopy bugs
after a fresh boot:
Total Slab Memory: 48074720
Usercopyable Memory: 6367532 13.2%
task_struct 0.2% 4480/1630720
RAW 0.3% 300/96000
RAWv6 2.1% 1408/64768
ext4_inode_cache 3.0% 269760/8740224
dentry 11.1% 585984/5273856
mm_struct 29.1% 54912/188448
kmalloc-8 100.0% 24576/24576
kmalloc-16 100.0% 28672/28672
kmalloc-32 100.0% 81920/81920
kmalloc-192 100.0% 96768/96768
kmalloc-128 100.0% 143360/143360
names_cache 100.0% 163840/163840
kmalloc-64 100.0% 167936/167936
kmalloc-256 100.0% 339968/339968
kmalloc-512 100.0% 350720/350720
kmalloc-96 100.0% 455616/455616
kmalloc-8192 100.0% 655360/655360
kmalloc-1024 100.0% 812032/812032
kmalloc-4096 100.0% 819200/819200
kmalloc-2048 100.0% 1310720/1310720
After some kernel build workloads, the percentage (mainly driven by
dentry and inode caches expanding) drops under 10%:
Total Slab Memory: 95516184
Usercopyable Memory: 8497452 8.8%
task_struct 0.2% 4000/1456000
RAW 0.3% 300/96000
RAWv6 2.1% 1408/64768
ext4_inode_cache 3.0% 1217280/39439872
dentry 11.1% 1623200/14608800
mm_struct 29.1% 73216/251264
kmalloc-8 100.0% 24576/24576
kmalloc-16 100.0% 28672/28672
kmalloc-32 100.0% 94208/94208
kmalloc-192 100.0% 96768/96768
kmalloc-128 100.0% 143360/143360
names_cache 100.0% 163840/163840
kmalloc-64 100.0% 245760/245760
kmalloc-256 100.0% 339968/339968
kmalloc-512 100.0% 350720/350720
kmalloc-96 100.0% 563520/563520
kmalloc-8192 100.0% 655360/655360
kmalloc-1024 100.0% 794624/794624
kmalloc-4096 100.0% 819200/819200
kmalloc-2048 100.0% 1257472/1257472
Signed-off-by: David Windsor <dave@nullcore.net>
[kees: adjust commit log, split out a few extra kmalloc hunks]
[kees: add field names to function declarations]
[kees: convert BUGs to WARNs and fail closed]
[kees: add attack surface reduction analysis to commit log]
Cc: Pekka Enberg <penberg@kernel.org>
Cc: David Rientjes <rientjes@google.com>
Cc: Joonsoo Kim <iamjoonsoo.kim@lge.com>
Cc: Andrew Morton <akpm@linux-foundation.org>
Cc: linux-mm@kvack.org
Cc: linux-xfs@vger.kernel.org
Signed-off-by: Kees Cook <keescook@chromium.org>
Acked-by: Christoph Lameter <cl@linux.com>
2017-06-11 05:50:28 +03:00
|
|
|
|
2010-09-28 17:10:26 +04:00
|
|
|
struct kmem_cache_node *node[MAX_NUMNODES];
|
2007-05-07 01:49:36 +04:00
|
|
|
};
|
|
|
|
|
2017-07-07 01:36:34 +03:00
|
|
|
#ifdef CONFIG_SLUB_CPU_PARTIAL
|
|
|
|
#define slub_cpu_partial(s) ((s)->cpu_partial)
|
|
|
|
#define slub_set_cpu_partial(s, n) \
|
|
|
|
({ \
|
|
|
|
slub_cpu_partial(s) = (n); \
|
|
|
|
})
|
|
|
|
#else
|
|
|
|
#define slub_cpu_partial(s) (0)
|
|
|
|
#define slub_set_cpu_partial(s, n)
|
2019-03-06 02:42:07 +03:00
|
|
|
#endif /* CONFIG_SLUB_CPU_PARTIAL */
|
2017-07-07 01:36:34 +03:00
|
|
|
|
2014-05-06 23:50:08 +04:00
|
|
|
#ifdef CONFIG_SYSFS
|
|
|
|
#define SLAB_SUPPORTS_SYSFS
|
2018-06-28 09:26:09 +03:00
|
|
|
void sysfs_slab_unlink(struct kmem_cache *);
|
2017-02-23 02:41:11 +03:00
|
|
|
void sysfs_slab_release(struct kmem_cache *);
|
2014-05-06 23:50:08 +04:00
|
|
|
#else
|
2018-06-28 09:26:09 +03:00
|
|
|
static inline void sysfs_slab_unlink(struct kmem_cache *s)
|
|
|
|
{
|
|
|
|
}
|
2017-02-23 02:41:11 +03:00
|
|
|
static inline void sysfs_slab_release(struct kmem_cache *s)
|
2014-05-06 23:50:08 +04:00
|
|
|
{
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
2015-02-14 01:39:35 +03:00
|
|
|
void object_err(struct kmem_cache *s, struct page *page,
|
|
|
|
u8 *object, char *reason);
|
|
|
|
|
2016-07-29 01:49:04 +03:00
|
|
|
void *fixup_red_left(struct kmem_cache *s, void *p);
|
|
|
|
|
2016-03-26 00:21:59 +03:00
|
|
|
static inline void *nearest_obj(struct kmem_cache *cache, struct page *page,
|
|
|
|
void *x) {
|
|
|
|
void *object = x - (x - page_address(page)) % cache->size;
|
|
|
|
void *last_object = page_address(page) +
|
|
|
|
(page->objects - 1) * cache->size;
|
2016-07-29 01:49:04 +03:00
|
|
|
void *result = (unlikely(object > last_object)) ? last_object : object;
|
|
|
|
|
|
|
|
result = fixup_red_left(cache, result);
|
|
|
|
return result;
|
2016-03-26 00:21:59 +03:00
|
|
|
}
|
|
|
|
|
2007-05-07 01:49:36 +04:00
|
|
|
#endif /* _LINUX_SLUB_DEF_H */
|