242 строки
6.5 KiB
C
242 строки
6.5 KiB
C
/* init.c: memory initialisation for FRV
|
|
*
|
|
* Copyright (C) 2004 Red Hat, Inc. All Rights Reserved.
|
|
* Written by David Howells (dhowells@redhat.com)
|
|
*
|
|
* This program is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU General Public License
|
|
* as published by the Free Software Foundation; either version
|
|
* 2 of the License, or (at your option) any later version.
|
|
*
|
|
* Derived from:
|
|
* - linux/arch/m68knommu/mm/init.c
|
|
* - Copyright (C) 1998 D. Jeff Dionne <jeff@lineo.ca>, Kenneth Albanowski <kjahds@kjahds.com>,
|
|
* - Copyright (C) 2000 Lineo, Inc. (www.lineo.com)
|
|
* - linux/arch/m68k/mm/init.c
|
|
* - Copyright (C) 1995 Hamish Macdonald
|
|
*/
|
|
|
|
#include <linux/signal.h>
|
|
#include <linux/sched.h>
|
|
#include <linux/pagemap.h>
|
|
#include <linux/swap.h>
|
|
#include <linux/mm.h>
|
|
#include <linux/kernel.h>
|
|
#include <linux/string.h>
|
|
#include <linux/types.h>
|
|
#include <linux/bootmem.h>
|
|
#include <linux/highmem.h>
|
|
#include <linux/module.h>
|
|
|
|
#include <asm/setup.h>
|
|
#include <asm/segment.h>
|
|
#include <asm/page.h>
|
|
#include <asm/pgtable.h>
|
|
#include <asm/system.h>
|
|
#include <asm/mmu_context.h>
|
|
#include <asm/virtconvert.h>
|
|
#include <asm/sections.h>
|
|
#include <asm/tlb.h>
|
|
|
|
#undef DEBUG
|
|
|
|
DEFINE_PER_CPU(struct mmu_gather, mmu_gathers);
|
|
|
|
/*
|
|
* BAD_PAGE is the page that is used for page faults when linux
|
|
* is out-of-memory. Older versions of linux just did a
|
|
* do_exit(), but using this instead means there is less risk
|
|
* for a process dying in kernel mode, possibly leaving a inode
|
|
* unused etc..
|
|
*
|
|
* BAD_PAGETABLE is the accompanying page-table: it is initialized
|
|
* to point to BAD_PAGE entries.
|
|
*
|
|
* ZERO_PAGE is a special page that is used for zero-initialized
|
|
* data and COW.
|
|
*/
|
|
static unsigned long empty_bad_page_table;
|
|
static unsigned long empty_bad_page;
|
|
|
|
unsigned long empty_zero_page;
|
|
EXPORT_SYMBOL(empty_zero_page);
|
|
|
|
/*****************************************************************************/
|
|
/*
|
|
*
|
|
*/
|
|
void show_mem(void)
|
|
{
|
|
unsigned long i;
|
|
int free = 0, total = 0, reserved = 0, shared = 0;
|
|
|
|
printk("\nMem-info:\n");
|
|
show_free_areas();
|
|
i = max_mapnr;
|
|
while (i-- > 0) {
|
|
struct page *page = &mem_map[i];
|
|
|
|
total++;
|
|
if (PageReserved(page))
|
|
reserved++;
|
|
else if (!page_count(page))
|
|
free++;
|
|
else
|
|
shared += page_count(page) - 1;
|
|
}
|
|
|
|
printk("%d pages of RAM\n",total);
|
|
printk("%d free pages\n",free);
|
|
printk("%d reserved pages\n",reserved);
|
|
printk("%d pages shared\n",shared);
|
|
|
|
} /* end show_mem() */
|
|
|
|
/*****************************************************************************/
|
|
/*
|
|
* paging_init() continues the virtual memory environment setup which
|
|
* was begun by the code in arch/head.S.
|
|
* The parameters are pointers to where to stick the starting and ending
|
|
* addresses of available kernel virtual memory.
|
|
*/
|
|
void __init paging_init(void)
|
|
{
|
|
unsigned long zones_size[MAX_NR_ZONES] = {0, };
|
|
|
|
/* allocate some pages for kernel housekeeping tasks */
|
|
empty_bad_page_table = (unsigned long) alloc_bootmem_pages(PAGE_SIZE);
|
|
empty_bad_page = (unsigned long) alloc_bootmem_pages(PAGE_SIZE);
|
|
empty_zero_page = (unsigned long) alloc_bootmem_pages(PAGE_SIZE);
|
|
|
|
memset((void *) empty_zero_page, 0, PAGE_SIZE);
|
|
|
|
#ifdef CONFIG_HIGHMEM
|
|
if (num_physpages - num_mappedpages) {
|
|
pgd_t *pge;
|
|
pud_t *pue;
|
|
pmd_t *pme;
|
|
|
|
pkmap_page_table = alloc_bootmem_pages(PAGE_SIZE);
|
|
|
|
memset(pkmap_page_table, 0, PAGE_SIZE);
|
|
|
|
pge = swapper_pg_dir + pgd_index_k(PKMAP_BASE);
|
|
pue = pud_offset(pge, PKMAP_BASE);
|
|
pme = pmd_offset(pue, PKMAP_BASE);
|
|
__set_pmd(pme, virt_to_phys(pkmap_page_table) | _PAGE_TABLE);
|
|
}
|
|
#endif
|
|
|
|
/* distribute the allocatable pages across the various zones and pass them to the allocator
|
|
*/
|
|
zones_size[ZONE_NORMAL] = max_low_pfn - min_low_pfn;
|
|
#ifdef CONFIG_HIGHMEM
|
|
zones_size[ZONE_HIGHMEM] = num_physpages - num_mappedpages;
|
|
#endif
|
|
|
|
free_area_init(zones_size);
|
|
|
|
#ifdef CONFIG_MMU
|
|
/* initialise init's MMU context */
|
|
init_new_context(&init_task, &init_mm);
|
|
#endif
|
|
|
|
} /* end paging_init() */
|
|
|
|
/*****************************************************************************/
|
|
/*
|
|
*
|
|
*/
|
|
void __init mem_init(void)
|
|
{
|
|
unsigned long npages = (memory_end - memory_start) >> PAGE_SHIFT;
|
|
unsigned long tmp;
|
|
#ifdef CONFIG_MMU
|
|
unsigned long loop, pfn;
|
|
int datapages = 0;
|
|
#endif
|
|
int codek = 0, datak = 0;
|
|
|
|
/* this will put all memory onto the freelists */
|
|
totalram_pages = free_all_bootmem();
|
|
|
|
#ifdef CONFIG_MMU
|
|
for (loop = 0 ; loop < npages ; loop++)
|
|
if (PageReserved(&mem_map[loop]))
|
|
datapages++;
|
|
|
|
#ifdef CONFIG_HIGHMEM
|
|
for (pfn = num_physpages - 1; pfn >= num_mappedpages; pfn--) {
|
|
struct page *page = &mem_map[pfn];
|
|
|
|
ClearPageReserved(page);
|
|
init_page_count(page);
|
|
__free_page(page);
|
|
totalram_pages++;
|
|
}
|
|
#endif
|
|
|
|
codek = ((unsigned long) &_etext - (unsigned long) &_stext) >> 10;
|
|
datak = datapages << (PAGE_SHIFT - 10);
|
|
|
|
#else
|
|
codek = (_etext - _stext) >> 10;
|
|
datak = 0; //(_ebss - _sdata) >> 10;
|
|
#endif
|
|
|
|
tmp = nr_free_pages() << PAGE_SHIFT;
|
|
printk("Memory available: %luKiB/%luKiB RAM, %luKiB/%luKiB ROM (%dKiB kernel code, %dKiB data)\n",
|
|
tmp >> 10,
|
|
npages << (PAGE_SHIFT - 10),
|
|
(rom_length > 0) ? ((rom_length >> 10) - codek) : 0,
|
|
rom_length >> 10,
|
|
codek,
|
|
datak
|
|
);
|
|
|
|
} /* end mem_init() */
|
|
|
|
/*****************************************************************************/
|
|
/*
|
|
* free the memory that was only required for initialisation
|
|
*/
|
|
void free_initmem(void)
|
|
{
|
|
#if defined(CONFIG_RAMKERNEL) && !defined(CONFIG_PROTECT_KERNEL)
|
|
unsigned long start, end, addr;
|
|
|
|
start = PAGE_ALIGN((unsigned long) &__init_begin); /* round up */
|
|
end = ((unsigned long) &__init_end) & PAGE_MASK; /* round down */
|
|
|
|
/* next to check that the page we free is not a partial page */
|
|
for (addr = start; addr < end; addr += PAGE_SIZE) {
|
|
ClearPageReserved(virt_to_page(addr));
|
|
init_page_count(virt_to_page(addr));
|
|
free_page(addr);
|
|
totalram_pages++;
|
|
}
|
|
|
|
printk("Freeing unused kernel memory: %ldKiB freed (0x%lx - 0x%lx)\n",
|
|
(end - start) >> 10, start, end);
|
|
#endif
|
|
} /* end free_initmem() */
|
|
|
|
/*****************************************************************************/
|
|
/*
|
|
* free the initial ramdisk memory
|
|
*/
|
|
#ifdef CONFIG_BLK_DEV_INITRD
|
|
void __init free_initrd_mem(unsigned long start, unsigned long end)
|
|
{
|
|
int pages = 0;
|
|
for (; start < end; start += PAGE_SIZE) {
|
|
ClearPageReserved(virt_to_page(start));
|
|
init_page_count(virt_to_page(start));
|
|
free_page(start);
|
|
totalram_pages++;
|
|
pages++;
|
|
}
|
|
printk("Freeing initrd memory: %dKiB freed\n", (pages * PAGE_SIZE) >> 10);
|
|
} /* end free_initrd_mem() */
|
|
#endif
|