Merge git://git.kernel.org/pub/scm/linux/kernel/git/dhowells/linux-2.6-nommu
* git://git.kernel.org/pub/scm/linux/kernel/git/dhowells/linux-2.6-nommu: NOMMU: Support XIP on initramfs NOMMU: Teach kobjsize() about VMA regions. FLAT: Don't attempt to expand the userspace stack to fill the space allocated FDPIC: Don't attempt to expand the userspace stack to fill the space allocated NOMMU: Improve procfs output using per-MM VMAs NOMMU: Make mmap allocation page trimming behaviour configurable. NOMMU: Make VMAs per MM as for MMU-mode linux NOMMU: Delete askedalloc and realalloc variables NOMMU: Rename ARM's struct vm_region NOMMU: Fix cleanup handling in ramfs_nommu_get_umapped_area()
This commit is contained in:
Коммит
c40f6f8bbc
|
@ -109,12 +109,18 @@ and it's also much more restricted in the latter case:
|
|||
FURTHER NOTES ON NO-MMU MMAP
|
||||
============================
|
||||
|
||||
(*) A request for a private mapping of less than a page in size may not return
|
||||
a page-aligned buffer. This is because the kernel calls kmalloc() to
|
||||
allocate the buffer, not get_free_page().
|
||||
(*) A request for a private mapping of a file may return a buffer that is not
|
||||
page-aligned. This is because XIP may take place, and the data may not be
|
||||
paged aligned in the backing store.
|
||||
|
||||
(*) A list of all the mappings on the system is visible through /proc/maps in
|
||||
no-MMU mode.
|
||||
(*) A request for an anonymous mapping will always be page aligned. If
|
||||
possible the size of the request should be a power of two otherwise some
|
||||
of the space may be wasted as the kernel must allocate a power-of-2
|
||||
granule but will only discard the excess if appropriately configured as
|
||||
this has an effect on fragmentation.
|
||||
|
||||
(*) A list of all the private copy and anonymous mappings on the system is
|
||||
visible through /proc/maps in no-MMU mode.
|
||||
|
||||
(*) A list of all the mappings in use by a process is visible through
|
||||
/proc/<pid>/maps in no-MMU mode.
|
||||
|
@ -242,3 +248,18 @@ PROVIDING SHAREABLE BLOCK DEVICE SUPPORT
|
|||
Provision of shared mappings on block device files is exactly the same as for
|
||||
character devices. If there isn't a real device underneath, then the driver
|
||||
should allocate sufficient contiguous memory to honour any supported mapping.
|
||||
|
||||
|
||||
=================================
|
||||
ADJUSTING PAGE TRIMMING BEHAVIOUR
|
||||
=================================
|
||||
|
||||
NOMMU mmap automatically rounds up to the nearest power-of-2 number of pages
|
||||
when performing an allocation. This can have adverse effects on memory
|
||||
fragmentation, and as such, is left configurable. The default behaviour is to
|
||||
aggressively trim allocations and discard any excess pages back in to the page
|
||||
allocator. In order to retain finer-grained control over fragmentation, this
|
||||
behaviour can either be disabled completely, or bumped up to a higher page
|
||||
watermark where trimming begins.
|
||||
|
||||
Page trimming behaviour is configurable via the sysctl `vm.nr_trim_pages'.
|
||||
|
|
|
@ -38,6 +38,7 @@ Currently, these files are in /proc/sys/vm:
|
|||
- numa_zonelist_order
|
||||
- nr_hugepages
|
||||
- nr_overcommit_hugepages
|
||||
- nr_trim_pages (only if CONFIG_MMU=n)
|
||||
|
||||
==============================================================
|
||||
|
||||
|
@ -348,3 +349,20 @@ Change the maximum size of the hugepage pool. The maximum is
|
|||
nr_hugepages + nr_overcommit_hugepages.
|
||||
|
||||
See Documentation/vm/hugetlbpage.txt
|
||||
|
||||
==============================================================
|
||||
|
||||
nr_trim_pages
|
||||
|
||||
This is available only on NOMMU kernels.
|
||||
|
||||
This value adjusts the excess page trimming behaviour of power-of-2 aligned
|
||||
NOMMU mmap allocations.
|
||||
|
||||
A value of 0 disables trimming of allocations entirely, while a value of 1
|
||||
trims excess pages aggressively. Any value >= 1 acts as the watermark where
|
||||
trimming of allocations is initiated.
|
||||
|
||||
The default value is 1.
|
||||
|
||||
See Documentation/nommu-mmap.txt for more information.
|
||||
|
|
|
@ -24,7 +24,6 @@ typedef struct {
|
|||
* modified for 2.6 by Hyok S. Choi <hyok.choi@samsung.com>
|
||||
*/
|
||||
typedef struct {
|
||||
struct vm_list_struct *vmlist;
|
||||
unsigned long end_brk;
|
||||
} mm_context_t;
|
||||
|
||||
|
|
|
@ -71,7 +71,7 @@ static DEFINE_SPINLOCK(consistent_lock);
|
|||
* the amount of RAM found at boot time.) I would imagine that get_vm_area()
|
||||
* would have to initialise this each time prior to calling vm_region_alloc().
|
||||
*/
|
||||
struct vm_region {
|
||||
struct arm_vm_region {
|
||||
struct list_head vm_list;
|
||||
unsigned long vm_start;
|
||||
unsigned long vm_end;
|
||||
|
@ -79,20 +79,20 @@ struct vm_region {
|
|||
int vm_active;
|
||||
};
|
||||
|
||||
static struct vm_region consistent_head = {
|
||||
static struct arm_vm_region consistent_head = {
|
||||
.vm_list = LIST_HEAD_INIT(consistent_head.vm_list),
|
||||
.vm_start = CONSISTENT_BASE,
|
||||
.vm_end = CONSISTENT_END,
|
||||
};
|
||||
|
||||
static struct vm_region *
|
||||
vm_region_alloc(struct vm_region *head, size_t size, gfp_t gfp)
|
||||
static struct arm_vm_region *
|
||||
arm_vm_region_alloc(struct arm_vm_region *head, size_t size, gfp_t gfp)
|
||||
{
|
||||
unsigned long addr = head->vm_start, end = head->vm_end - size;
|
||||
unsigned long flags;
|
||||
struct vm_region *c, *new;
|
||||
struct arm_vm_region *c, *new;
|
||||
|
||||
new = kmalloc(sizeof(struct vm_region), gfp);
|
||||
new = kmalloc(sizeof(struct arm_vm_region), gfp);
|
||||
if (!new)
|
||||
goto out;
|
||||
|
||||
|
@ -127,9 +127,9 @@ vm_region_alloc(struct vm_region *head, size_t size, gfp_t gfp)
|
|||
return NULL;
|
||||
}
|
||||
|
||||
static struct vm_region *vm_region_find(struct vm_region *head, unsigned long addr)
|
||||
static struct arm_vm_region *arm_vm_region_find(struct arm_vm_region *head, unsigned long addr)
|
||||
{
|
||||
struct vm_region *c;
|
||||
struct arm_vm_region *c;
|
||||
|
||||
list_for_each_entry(c, &head->vm_list, vm_list) {
|
||||
if (c->vm_active && c->vm_start == addr)
|
||||
|
@ -149,7 +149,7 @@ __dma_alloc(struct device *dev, size_t size, dma_addr_t *handle, gfp_t gfp,
|
|||
pgprot_t prot)
|
||||
{
|
||||
struct page *page;
|
||||
struct vm_region *c;
|
||||
struct arm_vm_region *c;
|
||||
unsigned long order;
|
||||
u64 mask = ISA_DMA_THRESHOLD, limit;
|
||||
|
||||
|
@ -214,7 +214,7 @@ __dma_alloc(struct device *dev, size_t size, dma_addr_t *handle, gfp_t gfp,
|
|||
/*
|
||||
* Allocate a virtual address in the consistent mapping region.
|
||||
*/
|
||||
c = vm_region_alloc(&consistent_head, size,
|
||||
c = arm_vm_region_alloc(&consistent_head, size,
|
||||
gfp & ~(__GFP_DMA | __GFP_HIGHMEM));
|
||||
if (c) {
|
||||
pte_t *pte;
|
||||
|
@ -311,13 +311,13 @@ static int dma_mmap(struct device *dev, struct vm_area_struct *vma,
|
|||
void *cpu_addr, dma_addr_t dma_addr, size_t size)
|
||||
{
|
||||
unsigned long flags, user_size, kern_size;
|
||||
struct vm_region *c;
|
||||
struct arm_vm_region *c;
|
||||
int ret = -ENXIO;
|
||||
|
||||
user_size = (vma->vm_end - vma->vm_start) >> PAGE_SHIFT;
|
||||
|
||||
spin_lock_irqsave(&consistent_lock, flags);
|
||||
c = vm_region_find(&consistent_head, (unsigned long)cpu_addr);
|
||||
c = arm_vm_region_find(&consistent_head, (unsigned long)cpu_addr);
|
||||
spin_unlock_irqrestore(&consistent_lock, flags);
|
||||
|
||||
if (c) {
|
||||
|
@ -359,7 +359,7 @@ EXPORT_SYMBOL(dma_mmap_writecombine);
|
|||
*/
|
||||
void dma_free_coherent(struct device *dev, size_t size, void *cpu_addr, dma_addr_t handle)
|
||||
{
|
||||
struct vm_region *c;
|
||||
struct arm_vm_region *c;
|
||||
unsigned long flags, addr;
|
||||
pte_t *ptep;
|
||||
int idx;
|
||||
|
@ -378,7 +378,7 @@ void dma_free_coherent(struct device *dev, size_t size, void *cpu_addr, dma_addr
|
|||
size = PAGE_ALIGN(size);
|
||||
|
||||
spin_lock_irqsave(&consistent_lock, flags);
|
||||
c = vm_region_find(&consistent_head, (unsigned long)cpu_addr);
|
||||
c = arm_vm_region_find(&consistent_head, (unsigned long)cpu_addr);
|
||||
if (!c)
|
||||
goto no_area;
|
||||
|
||||
|
|
|
@ -10,7 +10,6 @@ struct sram_list_struct {
|
|||
};
|
||||
|
||||
typedef struct {
|
||||
struct vm_list_struct *vmlist;
|
||||
unsigned long end_brk;
|
||||
unsigned long stack_start;
|
||||
|
||||
|
|
|
@ -160,15 +160,15 @@ put_reg(struct task_struct *task, int regno, unsigned long data)
|
|||
static inline int is_user_addr_valid(struct task_struct *child,
|
||||
unsigned long start, unsigned long len)
|
||||
{
|
||||
struct vm_list_struct *vml;
|
||||
struct vm_area_struct *vma;
|
||||
struct sram_list_struct *sraml;
|
||||
|
||||
/* overflow */
|
||||
if (start + len < start)
|
||||
return -EIO;
|
||||
|
||||
for (vml = child->mm->context.vmlist; vml; vml = vml->next)
|
||||
if (start >= vml->vma->vm_start && start + len < vml->vma->vm_end)
|
||||
vma = find_vma(child->mm, start);
|
||||
if (vma && start >= vma->vm_start && start + len <= vma->vm_end)
|
||||
return 0;
|
||||
|
||||
for (sraml = child->mm->context.sram_list; sraml; sraml = sraml->next)
|
||||
|
|
|
@ -32,6 +32,7 @@
|
|||
#include <linux/module.h>
|
||||
#include <linux/kallsyms.h>
|
||||
#include <linux/fs.h>
|
||||
#include <linux/rbtree.h>
|
||||
#include <asm/traps.h>
|
||||
#include <asm/cacheflush.h>
|
||||
#include <asm/cplb.h>
|
||||
|
@ -83,6 +84,7 @@ static void decode_address(char *buf, unsigned long address)
|
|||
struct mm_struct *mm;
|
||||
unsigned long flags, offset;
|
||||
unsigned char in_atomic = (bfin_read_IPEND() & 0x10) || in_atomic();
|
||||
struct rb_node *n;
|
||||
|
||||
#ifdef CONFIG_KALLSYMS
|
||||
unsigned long symsize;
|
||||
|
@ -128,9 +130,10 @@ static void decode_address(char *buf, unsigned long address)
|
|||
if (!mm)
|
||||
continue;
|
||||
|
||||
vml = mm->context.vmlist;
|
||||
while (vml) {
|
||||
struct vm_area_struct *vma = vml->vma;
|
||||
for (n = rb_first(&mm->mm_rb); n; n = rb_next(n)) {
|
||||
struct vm_area_struct *vma;
|
||||
|
||||
vma = rb_entry(n, struct vm_area_struct, vm_rb);
|
||||
|
||||
if (address >= vma->vm_start && address < vma->vm_end) {
|
||||
char _tmpbuf[256];
|
||||
|
@ -176,8 +179,6 @@ static void decode_address(char *buf, unsigned long address)
|
|||
|
||||
goto done;
|
||||
}
|
||||
|
||||
vml = vml->next;
|
||||
}
|
||||
if (!in_atomic)
|
||||
mmput(mm);
|
||||
|
|
|
@ -69,7 +69,8 @@ static inline int put_reg(struct task_struct *task, int regno,
|
|||
}
|
||||
|
||||
/*
|
||||
* check that an address falls within the bounds of the target process's memory mappings
|
||||
* check that an address falls within the bounds of the target process's memory
|
||||
* mappings
|
||||
*/
|
||||
static inline int is_user_addr_valid(struct task_struct *child,
|
||||
unsigned long start, unsigned long len)
|
||||
|
@ -79,11 +80,11 @@ static inline int is_user_addr_valid(struct task_struct *child,
|
|||
return -EIO;
|
||||
return 0;
|
||||
#else
|
||||
struct vm_list_struct *vml;
|
||||
struct vm_area_struct *vma;
|
||||
|
||||
for (vml = child->mm->context.vmlist; vml; vml = vml->next)
|
||||
if (start >= vml->vma->vm_start && start + len <= vml->vma->vm_end)
|
||||
return 0;
|
||||
vma = find_vma(child->mm, start);
|
||||
if (vma && start >= vma->vm_start && start + len <= vma->vm_end)
|
||||
return 0;
|
||||
|
||||
return -EIO;
|
||||
#endif
|
||||
|
|
|
@ -4,7 +4,6 @@
|
|||
/* Copyright (C) 2002, David McCullough <davidm@snapgear.com> */
|
||||
|
||||
typedef struct {
|
||||
struct vm_list_struct *vmlist;
|
||||
unsigned long end_brk;
|
||||
} mm_context_t;
|
||||
|
||||
|
|
|
@ -4,7 +4,6 @@
|
|||
/* Copyright (C) 2002, David McCullough <davidm@snapgear.com> */
|
||||
|
||||
typedef struct {
|
||||
struct vm_list_struct *vmlist;
|
||||
unsigned long end_brk;
|
||||
} mm_context_t;
|
||||
|
||||
|
|
|
@ -9,7 +9,6 @@ typedef struct {
|
|||
mm_context_id_t id;
|
||||
void *vdso;
|
||||
#else
|
||||
struct vm_list_struct *vmlist;
|
||||
unsigned long end_brk;
|
||||
#endif
|
||||
#ifdef CONFIG_BINFMT_ELF_FDPIC
|
||||
|
|
|
@ -168,9 +168,6 @@ static int load_elf_fdpic_binary(struct linux_binprm *bprm,
|
|||
struct elf_fdpic_params exec_params, interp_params;
|
||||
struct elf_phdr *phdr;
|
||||
unsigned long stack_size, entryaddr;
|
||||
#ifndef CONFIG_MMU
|
||||
unsigned long fullsize;
|
||||
#endif
|
||||
#ifdef ELF_FDPIC_PLAT_INIT
|
||||
unsigned long dynaddr;
|
||||
#endif
|
||||
|
@ -390,11 +387,6 @@ static int load_elf_fdpic_binary(struct linux_binprm *bprm,
|
|||
goto error_kill;
|
||||
}
|
||||
|
||||
/* expand the stack mapping to use up the entire allocation granule */
|
||||
fullsize = kobjsize((char *) current->mm->start_brk);
|
||||
if (!IS_ERR_VALUE(do_mremap(current->mm->start_brk, stack_size,
|
||||
fullsize, 0, 0)))
|
||||
stack_size = fullsize;
|
||||
up_write(¤t->mm->mmap_sem);
|
||||
|
||||
current->mm->brk = current->mm->start_brk;
|
||||
|
@ -1567,11 +1559,9 @@ end_coredump:
|
|||
static int elf_fdpic_dump_segments(struct file *file, size_t *size,
|
||||
unsigned long *limit, unsigned long mm_flags)
|
||||
{
|
||||
struct vm_list_struct *vml;
|
||||
|
||||
for (vml = current->mm->context.vmlist; vml; vml = vml->next) {
|
||||
struct vm_area_struct *vma = vml->vma;
|
||||
struct vm_area_struct *vma;
|
||||
|
||||
for (vma = current->mm->mmap; vma; vma = vma->vm_next) {
|
||||
if (!maydump(vma, mm_flags))
|
||||
continue;
|
||||
|
||||
|
@ -1617,9 +1607,6 @@ static int elf_fdpic_core_dump(long signr, struct pt_regs *regs,
|
|||
elf_fpxregset_t *xfpu = NULL;
|
||||
#endif
|
||||
int thread_status_size = 0;
|
||||
#ifndef CONFIG_MMU
|
||||
struct vm_list_struct *vml;
|
||||
#endif
|
||||
elf_addr_t *auxv;
|
||||
unsigned long mm_flags;
|
||||
|
||||
|
@ -1685,13 +1672,7 @@ static int elf_fdpic_core_dump(long signr, struct pt_regs *regs,
|
|||
fill_prstatus(prstatus, current, signr);
|
||||
elf_core_copy_regs(&prstatus->pr_reg, regs);
|
||||
|
||||
#ifdef CONFIG_MMU
|
||||
segs = current->mm->map_count;
|
||||
#else
|
||||
segs = 0;
|
||||
for (vml = current->mm->context.vmlist; vml; vml = vml->next)
|
||||
segs++;
|
||||
#endif
|
||||
#ifdef ELF_CORE_EXTRA_PHDRS
|
||||
segs += ELF_CORE_EXTRA_PHDRS;
|
||||
#endif
|
||||
|
@ -1766,20 +1747,10 @@ static int elf_fdpic_core_dump(long signr, struct pt_regs *regs,
|
|||
mm_flags = current->mm->flags;
|
||||
|
||||
/* write program headers for segments dump */
|
||||
for (
|
||||
#ifdef CONFIG_MMU
|
||||
vma = current->mm->mmap; vma; vma = vma->vm_next
|
||||
#else
|
||||
vml = current->mm->context.vmlist; vml; vml = vml->next
|
||||
#endif
|
||||
) {
|
||||
for (vma = current->mm->mmap; vma; vma = vma->vm_next) {
|
||||
struct elf_phdr phdr;
|
||||
size_t sz;
|
||||
|
||||
#ifndef CONFIG_MMU
|
||||
vma = vml->vma;
|
||||
#endif
|
||||
|
||||
sz = vma->vm_end - vma->vm_start;
|
||||
|
||||
phdr.p_type = PT_LOAD;
|
||||
|
|
|
@ -417,8 +417,8 @@ static int load_flat_file(struct linux_binprm * bprm,
|
|||
unsigned long textpos = 0, datapos = 0, result;
|
||||
unsigned long realdatastart = 0;
|
||||
unsigned long text_len, data_len, bss_len, stack_len, flags;
|
||||
unsigned long len, reallen, memp = 0;
|
||||
unsigned long extra, rlim;
|
||||
unsigned long len, memp = 0;
|
||||
unsigned long memp_size, extra, rlim;
|
||||
unsigned long *reloc = 0, *rp;
|
||||
struct inode *inode;
|
||||
int i, rev, relocs = 0;
|
||||
|
@ -543,17 +543,10 @@ static int load_flat_file(struct linux_binprm * bprm,
|
|||
}
|
||||
|
||||
len = data_len + extra + MAX_SHARED_LIBS * sizeof(unsigned long);
|
||||
len = PAGE_ALIGN(len);
|
||||
down_write(¤t->mm->mmap_sem);
|
||||
realdatastart = do_mmap(0, 0, len,
|
||||
PROT_READ|PROT_WRITE|PROT_EXEC, MAP_PRIVATE, 0);
|
||||
/* Remap to use all availabe slack region space */
|
||||
if (realdatastart && (realdatastart < (unsigned long)-4096)) {
|
||||
reallen = kobjsize((void *)realdatastart);
|
||||
if (reallen > len) {
|
||||
realdatastart = do_mremap(realdatastart, len,
|
||||
reallen, MREMAP_FIXED, realdatastart);
|
||||
}
|
||||
}
|
||||
up_write(¤t->mm->mmap_sem);
|
||||
|
||||
if (realdatastart == 0 || realdatastart >= (unsigned long)-4096) {
|
||||
|
@ -591,21 +584,14 @@ static int load_flat_file(struct linux_binprm * bprm,
|
|||
|
||||
reloc = (unsigned long *) (datapos+(ntohl(hdr->reloc_start)-text_len));
|
||||
memp = realdatastart;
|
||||
|
||||
memp_size = len;
|
||||
} else {
|
||||
|
||||
len = text_len + data_len + extra + MAX_SHARED_LIBS * sizeof(unsigned long);
|
||||
len = PAGE_ALIGN(len);
|
||||
down_write(¤t->mm->mmap_sem);
|
||||
textpos = do_mmap(0, 0, len,
|
||||
PROT_READ | PROT_EXEC | PROT_WRITE, MAP_PRIVATE, 0);
|
||||
/* Remap to use all availabe slack region space */
|
||||
if (textpos && (textpos < (unsigned long) -4096)) {
|
||||
reallen = kobjsize((void *)textpos);
|
||||
if (reallen > len) {
|
||||
textpos = do_mremap(textpos, len, reallen,
|
||||
MREMAP_FIXED, textpos);
|
||||
}
|
||||
}
|
||||
up_write(¤t->mm->mmap_sem);
|
||||
|
||||
if (!textpos || textpos >= (unsigned long) -4096) {
|
||||
|
@ -622,7 +608,7 @@ static int load_flat_file(struct linux_binprm * bprm,
|
|||
reloc = (unsigned long *) (textpos + ntohl(hdr->reloc_start) +
|
||||
MAX_SHARED_LIBS * sizeof(unsigned long));
|
||||
memp = textpos;
|
||||
|
||||
memp_size = len;
|
||||
#ifdef CONFIG_BINFMT_ZFLAT
|
||||
/*
|
||||
* load it all in and treat it like a RAM load from now on
|
||||
|
@ -680,10 +666,12 @@ static int load_flat_file(struct linux_binprm * bprm,
|
|||
* set up the brk stuff, uses any slack left in data/bss/stack
|
||||
* allocation. We put the brk after the bss (between the bss
|
||||
* and stack) like other platforms.
|
||||
* Userspace code relies on the stack pointer starting out at
|
||||
* an address right at the end of a page.
|
||||
*/
|
||||
current->mm->start_brk = datapos + data_len + bss_len;
|
||||
current->mm->brk = (current->mm->start_brk + 3) & ~3;
|
||||
current->mm->context.end_brk = memp + kobjsize((void *) memp) - stack_len;
|
||||
current->mm->context.end_brk = memp + memp_size - stack_len;
|
||||
}
|
||||
|
||||
if (flags & FLAT_FLAG_KTRACE)
|
||||
|
@ -790,8 +778,8 @@ static int load_flat_file(struct linux_binprm * bprm,
|
|||
|
||||
/* zero the BSS, BRK and stack areas */
|
||||
memset((void*)(datapos + data_len), 0, bss_len +
|
||||
(memp + kobjsize((void *) memp) - stack_len - /* end brk */
|
||||
libinfo->lib_list[id].start_brk) + /* start brk */
|
||||
(memp + memp_size - stack_len - /* end brk */
|
||||
libinfo->lib_list[id].start_brk) + /* start brk */
|
||||
stack_len);
|
||||
|
||||
return 0;
|
||||
|
|
|
@ -41,8 +41,6 @@ do { \
|
|||
(vmi)->used = 0; \
|
||||
(vmi)->largest_chunk = 0; \
|
||||
} while(0)
|
||||
|
||||
extern int nommu_vma_show(struct seq_file *, struct vm_area_struct *);
|
||||
#endif
|
||||
|
||||
extern int proc_tid_stat(struct seq_file *m, struct pid_namespace *ns,
|
||||
|
|
|
@ -73,6 +73,9 @@ static int meminfo_proc_show(struct seq_file *m, void *v)
|
|||
"HighFree: %8lu kB\n"
|
||||
"LowTotal: %8lu kB\n"
|
||||
"LowFree: %8lu kB\n"
|
||||
#endif
|
||||
#ifndef CONFIG_MMU
|
||||
"MmapCopy: %8lu kB\n"
|
||||
#endif
|
||||
"SwapTotal: %8lu kB\n"
|
||||
"SwapFree: %8lu kB\n"
|
||||
|
@ -115,6 +118,9 @@ static int meminfo_proc_show(struct seq_file *m, void *v)
|
|||
K(i.freehigh),
|
||||
K(i.totalram-i.totalhigh),
|
||||
K(i.freeram-i.freehigh),
|
||||
#endif
|
||||
#ifndef CONFIG_MMU
|
||||
K((unsigned long) atomic_read(&mmap_pages_allocated)),
|
||||
#endif
|
||||
K(i.totalswap),
|
||||
K(i.freeswap),
|
||||
|
|
|
@ -33,33 +33,33 @@
|
|||
#include "internal.h"
|
||||
|
||||
/*
|
||||
* display a single VMA to a sequenced file
|
||||
* display a single region to a sequenced file
|
||||
*/
|
||||
int nommu_vma_show(struct seq_file *m, struct vm_area_struct *vma)
|
||||
static int nommu_region_show(struct seq_file *m, struct vm_region *region)
|
||||
{
|
||||
unsigned long ino = 0;
|
||||
struct file *file;
|
||||
dev_t dev = 0;
|
||||
int flags, len;
|
||||
|
||||
flags = vma->vm_flags;
|
||||
file = vma->vm_file;
|
||||
flags = region->vm_flags;
|
||||
file = region->vm_file;
|
||||
|
||||
if (file) {
|
||||
struct inode *inode = vma->vm_file->f_path.dentry->d_inode;
|
||||
struct inode *inode = region->vm_file->f_path.dentry->d_inode;
|
||||
dev = inode->i_sb->s_dev;
|
||||
ino = inode->i_ino;
|
||||
}
|
||||
|
||||
seq_printf(m,
|
||||
"%08lx-%08lx %c%c%c%c %08llx %02x:%02x %lu %n",
|
||||
vma->vm_start,
|
||||
vma->vm_end,
|
||||
region->vm_start,
|
||||
region->vm_end,
|
||||
flags & VM_READ ? 'r' : '-',
|
||||
flags & VM_WRITE ? 'w' : '-',
|
||||
flags & VM_EXEC ? 'x' : '-',
|
||||
flags & VM_MAYSHARE ? flags & VM_SHARED ? 'S' : 's' : 'p',
|
||||
((loff_t)vma->vm_pgoff) << PAGE_SHIFT,
|
||||
((loff_t)region->vm_pgoff) << PAGE_SHIFT,
|
||||
MAJOR(dev), MINOR(dev), ino, &len);
|
||||
|
||||
if (file) {
|
||||
|
@ -75,61 +75,54 @@ int nommu_vma_show(struct seq_file *m, struct vm_area_struct *vma)
|
|||
}
|
||||
|
||||
/*
|
||||
* display a list of all the VMAs the kernel knows about
|
||||
* display a list of all the REGIONs the kernel knows about
|
||||
* - nommu kernals have a single flat list
|
||||
*/
|
||||
static int nommu_vma_list_show(struct seq_file *m, void *v)
|
||||
static int nommu_region_list_show(struct seq_file *m, void *_p)
|
||||
{
|
||||
struct vm_area_struct *vma;
|
||||
struct rb_node *p = _p;
|
||||
|
||||
vma = rb_entry((struct rb_node *) v, struct vm_area_struct, vm_rb);
|
||||
return nommu_vma_show(m, vma);
|
||||
return nommu_region_show(m, rb_entry(p, struct vm_region, vm_rb));
|
||||
}
|
||||
|
||||
static void *nommu_vma_list_start(struct seq_file *m, loff_t *_pos)
|
||||
static void *nommu_region_list_start(struct seq_file *m, loff_t *_pos)
|
||||
{
|
||||
struct rb_node *_rb;
|
||||
struct rb_node *p;
|
||||
loff_t pos = *_pos;
|
||||
void *next = NULL;
|
||||
|
||||
down_read(&nommu_vma_sem);
|
||||
down_read(&nommu_region_sem);
|
||||
|
||||
for (_rb = rb_first(&nommu_vma_tree); _rb; _rb = rb_next(_rb)) {
|
||||
if (pos == 0) {
|
||||
next = _rb;
|
||||
break;
|
||||
}
|
||||
pos--;
|
||||
}
|
||||
|
||||
return next;
|
||||
for (p = rb_first(&nommu_region_tree); p; p = rb_next(p))
|
||||
if (pos-- == 0)
|
||||
return p;
|
||||
return NULL;
|
||||
}
|
||||
|
||||
static void nommu_vma_list_stop(struct seq_file *m, void *v)
|
||||
static void nommu_region_list_stop(struct seq_file *m, void *v)
|
||||
{
|
||||
up_read(&nommu_vma_sem);
|
||||
up_read(&nommu_region_sem);
|
||||
}
|
||||
|
||||
static void *nommu_vma_list_next(struct seq_file *m, void *v, loff_t *pos)
|
||||
static void *nommu_region_list_next(struct seq_file *m, void *v, loff_t *pos)
|
||||
{
|
||||
(*pos)++;
|
||||
return rb_next((struct rb_node *) v);
|
||||
}
|
||||
|
||||
static const struct seq_operations proc_nommu_vma_list_seqop = {
|
||||
.start = nommu_vma_list_start,
|
||||
.next = nommu_vma_list_next,
|
||||
.stop = nommu_vma_list_stop,
|
||||
.show = nommu_vma_list_show
|
||||
static struct seq_operations proc_nommu_region_list_seqop = {
|
||||
.start = nommu_region_list_start,
|
||||
.next = nommu_region_list_next,
|
||||
.stop = nommu_region_list_stop,
|
||||
.show = nommu_region_list_show
|
||||
};
|
||||
|
||||
static int proc_nommu_vma_list_open(struct inode *inode, struct file *file)
|
||||
static int proc_nommu_region_list_open(struct inode *inode, struct file *file)
|
||||
{
|
||||
return seq_open(file, &proc_nommu_vma_list_seqop);
|
||||
return seq_open(file, &proc_nommu_region_list_seqop);
|
||||
}
|
||||
|
||||
static const struct file_operations proc_nommu_vma_list_operations = {
|
||||
.open = proc_nommu_vma_list_open,
|
||||
static const struct file_operations proc_nommu_region_list_operations = {
|
||||
.open = proc_nommu_region_list_open,
|
||||
.read = seq_read,
|
||||
.llseek = seq_lseek,
|
||||
.release = seq_release,
|
||||
|
@ -137,7 +130,7 @@ static const struct file_operations proc_nommu_vma_list_operations = {
|
|||
|
||||
static int __init proc_nommu_init(void)
|
||||
{
|
||||
proc_create("maps", S_IRUGO, NULL, &proc_nommu_vma_list_operations);
|
||||
proc_create("maps", S_IRUGO, NULL, &proc_nommu_region_list_operations);
|
||||
return 0;
|
||||
}
|
||||
|
||||
|
|
|
@ -15,25 +15,32 @@
|
|||
*/
|
||||
void task_mem(struct seq_file *m, struct mm_struct *mm)
|
||||
{
|
||||
struct vm_list_struct *vml;
|
||||
unsigned long bytes = 0, sbytes = 0, slack = 0;
|
||||
struct vm_area_struct *vma;
|
||||
struct vm_region *region;
|
||||
struct rb_node *p;
|
||||
unsigned long bytes = 0, sbytes = 0, slack = 0, size;
|
||||
|
||||
down_read(&mm->mmap_sem);
|
||||
for (vml = mm->context.vmlist; vml; vml = vml->next) {
|
||||
if (!vml->vma)
|
||||
continue;
|
||||
for (p = rb_first(&mm->mm_rb); p; p = rb_next(p)) {
|
||||
vma = rb_entry(p, struct vm_area_struct, vm_rb);
|
||||
|
||||
bytes += kobjsize(vml);
|
||||
if (atomic_read(&mm->mm_count) > 1 ||
|
||||
atomic_read(&vml->vma->vm_usage) > 1
|
||||
) {
|
||||
sbytes += kobjsize((void *) vml->vma->vm_start);
|
||||
sbytes += kobjsize(vml->vma);
|
||||
bytes += kobjsize(vma);
|
||||
|
||||
region = vma->vm_region;
|
||||
if (region) {
|
||||
size = kobjsize(region);
|
||||
size += region->vm_end - region->vm_start;
|
||||
} else {
|
||||
bytes += kobjsize((void *) vml->vma->vm_start);
|
||||
bytes += kobjsize(vml->vma);
|
||||
slack += kobjsize((void *) vml->vma->vm_start) -
|
||||
(vml->vma->vm_end - vml->vma->vm_start);
|
||||
size = vma->vm_end - vma->vm_start;
|
||||
}
|
||||
|
||||
if (atomic_read(&mm->mm_count) > 1 ||
|
||||
vma->vm_flags & VM_MAYSHARE) {
|
||||
sbytes += size;
|
||||
} else {
|
||||
bytes += size;
|
||||
if (region)
|
||||
slack = region->vm_end - vma->vm_end;
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -70,13 +77,14 @@ void task_mem(struct seq_file *m, struct mm_struct *mm)
|
|||
|
||||
unsigned long task_vsize(struct mm_struct *mm)
|
||||
{
|
||||
struct vm_list_struct *tbp;
|
||||
struct vm_area_struct *vma;
|
||||
struct rb_node *p;
|
||||
unsigned long vsize = 0;
|
||||
|
||||
down_read(&mm->mmap_sem);
|
||||
for (tbp = mm->context.vmlist; tbp; tbp = tbp->next) {
|
||||
if (tbp->vma)
|
||||
vsize += kobjsize((void *) tbp->vma->vm_start);
|
||||
for (p = rb_first(&mm->mm_rb); p; p = rb_next(p)) {
|
||||
vma = rb_entry(p, struct vm_area_struct, vm_rb);
|
||||
vsize += vma->vm_end - vma->vm_start;
|
||||
}
|
||||
up_read(&mm->mmap_sem);
|
||||
return vsize;
|
||||
|
@ -85,15 +93,19 @@ unsigned long task_vsize(struct mm_struct *mm)
|
|||
int task_statm(struct mm_struct *mm, int *shared, int *text,
|
||||
int *data, int *resident)
|
||||
{
|
||||
struct vm_list_struct *tbp;
|
||||
struct vm_area_struct *vma;
|
||||
struct vm_region *region;
|
||||
struct rb_node *p;
|
||||
int size = kobjsize(mm);
|
||||
|
||||
down_read(&mm->mmap_sem);
|
||||
for (tbp = mm->context.vmlist; tbp; tbp = tbp->next) {
|
||||
size += kobjsize(tbp);
|
||||
if (tbp->vma) {
|
||||
size += kobjsize(tbp->vma);
|
||||
size += kobjsize((void *) tbp->vma->vm_start);
|
||||
for (p = rb_first(&mm->mm_rb); p; p = rb_next(p)) {
|
||||
vma = rb_entry(p, struct vm_area_struct, vm_rb);
|
||||
size += kobjsize(vma);
|
||||
region = vma->vm_region;
|
||||
if (region) {
|
||||
size += kobjsize(region);
|
||||
size += region->vm_end - region->vm_start;
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -104,21 +116,63 @@ int task_statm(struct mm_struct *mm, int *shared, int *text,
|
|||
return size;
|
||||
}
|
||||
|
||||
/*
|
||||
* display a single VMA to a sequenced file
|
||||
*/
|
||||
static int nommu_vma_show(struct seq_file *m, struct vm_area_struct *vma)
|
||||
{
|
||||
unsigned long ino = 0;
|
||||
struct file *file;
|
||||
dev_t dev = 0;
|
||||
int flags, len;
|
||||
|
||||
flags = vma->vm_flags;
|
||||
file = vma->vm_file;
|
||||
|
||||
if (file) {
|
||||
struct inode *inode = vma->vm_file->f_path.dentry->d_inode;
|
||||
dev = inode->i_sb->s_dev;
|
||||
ino = inode->i_ino;
|
||||
}
|
||||
|
||||
seq_printf(m,
|
||||
"%08lx-%08lx %c%c%c%c %08lx %02x:%02x %lu %n",
|
||||
vma->vm_start,
|
||||
vma->vm_end,
|
||||
flags & VM_READ ? 'r' : '-',
|
||||
flags & VM_WRITE ? 'w' : '-',
|
||||
flags & VM_EXEC ? 'x' : '-',
|
||||
flags & VM_MAYSHARE ? flags & VM_SHARED ? 'S' : 's' : 'p',
|
||||
vma->vm_pgoff << PAGE_SHIFT,
|
||||
MAJOR(dev), MINOR(dev), ino, &len);
|
||||
|
||||
if (file) {
|
||||
len = 25 + sizeof(void *) * 6 - len;
|
||||
if (len < 1)
|
||||
len = 1;
|
||||
seq_printf(m, "%*c", len, ' ');
|
||||
seq_path(m, &file->f_path, "");
|
||||
}
|
||||
|
||||
seq_putc(m, '\n');
|
||||
return 0;
|
||||
}
|
||||
|
||||
/*
|
||||
* display mapping lines for a particular process's /proc/pid/maps
|
||||
*/
|
||||
static int show_map(struct seq_file *m, void *_vml)
|
||||
static int show_map(struct seq_file *m, void *_p)
|
||||
{
|
||||
struct vm_list_struct *vml = _vml;
|
||||
struct rb_node *p = _p;
|
||||
|
||||
return nommu_vma_show(m, vml->vma);
|
||||
return nommu_vma_show(m, rb_entry(p, struct vm_area_struct, vm_rb));
|
||||
}
|
||||
|
||||
static void *m_start(struct seq_file *m, loff_t *pos)
|
||||
{
|
||||
struct proc_maps_private *priv = m->private;
|
||||
struct vm_list_struct *vml;
|
||||
struct mm_struct *mm;
|
||||
struct rb_node *p;
|
||||
loff_t n = *pos;
|
||||
|
||||
/* pin the task and mm whilst we play with them */
|
||||
|
@ -134,9 +188,9 @@ static void *m_start(struct seq_file *m, loff_t *pos)
|
|||
}
|
||||
|
||||
/* start from the Nth VMA */
|
||||
for (vml = mm->context.vmlist; vml; vml = vml->next)
|
||||
for (p = rb_first(&mm->mm_rb); p; p = rb_next(p))
|
||||
if (n-- == 0)
|
||||
return vml;
|
||||
return p;
|
||||
return NULL;
|
||||
}
|
||||
|
||||
|
@ -152,12 +206,12 @@ static void m_stop(struct seq_file *m, void *_vml)
|
|||
}
|
||||
}
|
||||
|
||||
static void *m_next(struct seq_file *m, void *_vml, loff_t *pos)
|
||||
static void *m_next(struct seq_file *m, void *_p, loff_t *pos)
|
||||
{
|
||||
struct vm_list_struct *vml = _vml;
|
||||
struct rb_node *p = _p;
|
||||
|
||||
(*pos)++;
|
||||
return vml ? vml->next : NULL;
|
||||
return p ? rb_next(p) : NULL;
|
||||
}
|
||||
|
||||
static const struct seq_operations proc_pid_maps_ops = {
|
||||
|
|
|
@ -262,11 +262,11 @@ unsigned long ramfs_nommu_get_unmapped_area(struct file *file,
|
|||
ret = -ENOMEM;
|
||||
pages = kzalloc(lpages * sizeof(struct page *), GFP_KERNEL);
|
||||
if (!pages)
|
||||
goto out;
|
||||
goto out_free;
|
||||
|
||||
nr = find_get_pages(inode->i_mapping, pgoff, lpages, pages);
|
||||
if (nr != lpages)
|
||||
goto out; /* leave if some pages were missing */
|
||||
goto out_free_pages; /* leave if some pages were missing */
|
||||
|
||||
/* check the pages for physical adjacency */
|
||||
ptr = pages;
|
||||
|
@ -274,19 +274,18 @@ unsigned long ramfs_nommu_get_unmapped_area(struct file *file,
|
|||
page++;
|
||||
for (loop = lpages; loop > 1; loop--)
|
||||
if (*ptr++ != page++)
|
||||
goto out;
|
||||
goto out_free_pages;
|
||||
|
||||
/* okay - all conditions fulfilled */
|
||||
ret = (unsigned long) page_address(pages[0]);
|
||||
|
||||
out:
|
||||
if (pages) {
|
||||
ptr = pages;
|
||||
for (loop = lpages; loop > 0; loop--)
|
||||
put_page(*ptr++);
|
||||
kfree(pages);
|
||||
}
|
||||
|
||||
out_free_pages:
|
||||
ptr = pages;
|
||||
for (loop = nr; loop > 0; loop--)
|
||||
put_page(*ptr++);
|
||||
out_free:
|
||||
kfree(pages);
|
||||
out:
|
||||
return ret;
|
||||
}
|
||||
|
||||
|
|
|
@ -22,7 +22,6 @@ typedef struct {
|
|||
unsigned long dtlb_ptd_mapping; /* [DAMR5] PTD mapping for dtlb cached PGE */
|
||||
|
||||
#else
|
||||
struct vm_list_struct *vmlist;
|
||||
unsigned long end_brk;
|
||||
|
||||
#endif
|
||||
|
|
|
@ -4,7 +4,6 @@
|
|||
#if !defined(CONFIG_MMU)
|
||||
|
||||
typedef struct {
|
||||
struct vm_list_struct *vmlist;
|
||||
unsigned long end_brk;
|
||||
} mm_context_t;
|
||||
|
||||
|
|
|
@ -56,19 +56,9 @@ extern unsigned long mmap_min_addr;
|
|||
|
||||
extern struct kmem_cache *vm_area_cachep;
|
||||
|
||||
/*
|
||||
* This struct defines the per-mm list of VMAs for uClinux. If CONFIG_MMU is
|
||||
* disabled, then there's a single shared list of VMAs maintained by the
|
||||
* system, and mm's subscribe to these individually
|
||||
*/
|
||||
struct vm_list_struct {
|
||||
struct vm_list_struct *next;
|
||||
struct vm_area_struct *vma;
|
||||
};
|
||||
|
||||
#ifndef CONFIG_MMU
|
||||
extern struct rb_root nommu_vma_tree;
|
||||
extern struct rw_semaphore nommu_vma_sem;
|
||||
extern struct rb_root nommu_region_tree;
|
||||
extern struct rw_semaphore nommu_region_sem;
|
||||
|
||||
extern unsigned int kobjsize(const void *objp);
|
||||
#endif
|
||||
|
@ -1061,6 +1051,7 @@ extern void memmap_init_zone(unsigned long, int, unsigned long,
|
|||
unsigned long, enum memmap_context);
|
||||
extern void setup_per_zone_pages_min(void);
|
||||
extern void mem_init(void);
|
||||
extern void __init mmap_init(void);
|
||||
extern void show_mem(void);
|
||||
extern void si_meminfo(struct sysinfo * val);
|
||||
extern void si_meminfo_node(struct sysinfo *val, int nid);
|
||||
|
@ -1072,6 +1063,9 @@ extern void setup_per_cpu_pageset(void);
|
|||
static inline void setup_per_cpu_pageset(void) {}
|
||||
#endif
|
||||
|
||||
/* nommu.c */
|
||||
extern atomic_t mmap_pages_allocated;
|
||||
|
||||
/* prio_tree.c */
|
||||
void vma_prio_tree_add(struct vm_area_struct *, struct vm_area_struct *old);
|
||||
void vma_prio_tree_insert(struct vm_area_struct *, struct prio_tree_root *);
|
||||
|
|
|
@ -96,6 +96,23 @@ struct page {
|
|||
#endif /* WANT_PAGE_VIRTUAL */
|
||||
};
|
||||
|
||||
/*
|
||||
* A region containing a mapping of a non-memory backed file under NOMMU
|
||||
* conditions. These are held in a global tree and are pinned by the VMAs that
|
||||
* map parts of them.
|
||||
*/
|
||||
struct vm_region {
|
||||
struct rb_node vm_rb; /* link in global region tree */
|
||||
unsigned long vm_flags; /* VMA vm_flags */
|
||||
unsigned long vm_start; /* start address of region */
|
||||
unsigned long vm_end; /* region initialised to here */
|
||||
unsigned long vm_top; /* region allocated to here */
|
||||
unsigned long vm_pgoff; /* the offset in vm_file corresponding to vm_start */
|
||||
struct file *vm_file; /* the backing file or NULL */
|
||||
|
||||
atomic_t vm_usage; /* region usage count */
|
||||
};
|
||||
|
||||
/*
|
||||
* This struct defines a memory VMM memory area. There is one of these
|
||||
* per VM-area/task. A VM area is any part of the process virtual memory
|
||||
|
@ -152,7 +169,7 @@ struct vm_area_struct {
|
|||
unsigned long vm_truncate_count;/* truncate_count or restart_addr */
|
||||
|
||||
#ifndef CONFIG_MMU
|
||||
atomic_t vm_usage; /* refcount (VMAs shared if !MMU) */
|
||||
struct vm_region *vm_region; /* NOMMU mapping region */
|
||||
#endif
|
||||
#ifdef CONFIG_NUMA
|
||||
struct mempolicy *vm_policy; /* NUMA policy for the VMA */
|
||||
|
|
|
@ -317,6 +317,7 @@ static int __init do_name(void)
|
|||
if (wfd >= 0) {
|
||||
sys_fchown(wfd, uid, gid);
|
||||
sys_fchmod(wfd, mode);
|
||||
sys_ftruncate(wfd, body_len);
|
||||
vcollected = kstrdup(collected, GFP_KERNEL);
|
||||
state = CopyFile;
|
||||
}
|
||||
|
|
12
ipc/shm.c
12
ipc/shm.c
|
@ -990,6 +990,7 @@ asmlinkage long sys_shmdt(char __user *shmaddr)
|
|||
*/
|
||||
vma = find_vma(mm, addr);
|
||||
|
||||
#ifdef CONFIG_MMU
|
||||
while (vma) {
|
||||
next = vma->vm_next;
|
||||
|
||||
|
@ -1034,6 +1035,17 @@ asmlinkage long sys_shmdt(char __user *shmaddr)
|
|||
vma = next;
|
||||
}
|
||||
|
||||
#else /* CONFIG_MMU */
|
||||
/* under NOMMU conditions, the exact address to be destroyed must be
|
||||
* given */
|
||||
retval = -EINVAL;
|
||||
if (vma->vm_start == addr && vma->vm_ops == &shm_vm_ops) {
|
||||
do_munmap(mm, vma->vm_start, vma->vm_end - vma->vm_start);
|
||||
retval = 0;
|
||||
}
|
||||
|
||||
#endif
|
||||
|
||||
up_write(&mm->mmap_sem);
|
||||
return retval;
|
||||
}
|
||||
|
|
|
@ -1481,12 +1481,10 @@ void __init proc_caches_init(void)
|
|||
fs_cachep = kmem_cache_create("fs_cache",
|
||||
sizeof(struct fs_struct), 0,
|
||||
SLAB_HWCACHE_ALIGN|SLAB_PANIC, NULL);
|
||||
vm_area_cachep = kmem_cache_create("vm_area_struct",
|
||||
sizeof(struct vm_area_struct), 0,
|
||||
SLAB_PANIC, NULL);
|
||||
mm_cachep = kmem_cache_create("mm_struct",
|
||||
sizeof(struct mm_struct), ARCH_MIN_MMSTRUCT_ALIGN,
|
||||
SLAB_HWCACHE_ALIGN|SLAB_PANIC, NULL);
|
||||
mmap_init();
|
||||
}
|
||||
|
||||
/*
|
||||
|
|
|
@ -82,6 +82,9 @@ extern int percpu_pagelist_fraction;
|
|||
extern int compat_log;
|
||||
extern int latencytop_enabled;
|
||||
extern int sysctl_nr_open_min, sysctl_nr_open_max;
|
||||
#ifndef CONFIG_MMU
|
||||
extern int sysctl_nr_trim_pages;
|
||||
#endif
|
||||
#ifdef CONFIG_RCU_TORTURE_TEST
|
||||
extern int rcutorture_runnable;
|
||||
#endif /* #ifdef CONFIG_RCU_TORTURE_TEST */
|
||||
|
@ -1102,6 +1105,17 @@ static struct ctl_table vm_table[] = {
|
|||
.mode = 0644,
|
||||
.proc_handler = &proc_dointvec
|
||||
},
|
||||
#else
|
||||
{
|
||||
.ctl_name = CTL_UNNUMBERED,
|
||||
.procname = "nr_trim_pages",
|
||||
.data = &sysctl_nr_trim_pages,
|
||||
.maxlen = sizeof(sysctl_nr_trim_pages),
|
||||
.mode = 0644,
|
||||
.proc_handler = &proc_dointvec_minmax,
|
||||
.strategy = &sysctl_intvec,
|
||||
.extra1 = &zero,
|
||||
},
|
||||
#endif
|
||||
{
|
||||
.ctl_name = VM_LAPTOP_MODE,
|
||||
|
|
|
@ -512,6 +512,13 @@ config DEBUG_VIRTUAL
|
|||
|
||||
If unsure, say N.
|
||||
|
||||
config DEBUG_NOMMU_REGIONS
|
||||
bool "Debug the global anon/private NOMMU mapping region tree"
|
||||
depends on DEBUG_KERNEL && !MMU
|
||||
help
|
||||
This option causes the global tree of anonymous and private mapping
|
||||
regions to be regularly checked for invalid topology.
|
||||
|
||||
config DEBUG_WRITECOUNT
|
||||
bool "Debug filesystem writers count"
|
||||
depends on DEBUG_KERNEL
|
||||
|
|
10
mm/mmap.c
10
mm/mmap.c
|
@ -2472,3 +2472,13 @@ void mm_drop_all_locks(struct mm_struct *mm)
|
|||
|
||||
mutex_unlock(&mm_all_locks_mutex);
|
||||
}
|
||||
|
||||
/*
|
||||
* initialise the VMA slab
|
||||
*/
|
||||
void __init mmap_init(void)
|
||||
{
|
||||
vm_area_cachep = kmem_cache_create("vm_area_struct",
|
||||
sizeof(struct vm_area_struct), 0,
|
||||
SLAB_PANIC, NULL);
|
||||
}
|
||||
|
|
1079
mm/nommu.c
1079
mm/nommu.c
Разница между файлами не показана из-за своего большого размера
Загрузить разницу
Загрузка…
Ссылка в новой задаче