305 строки
7.9 KiB
C
305 строки
7.9 KiB
C
/*
|
|
* Copyright 2010 Tilera Corporation. All Rights Reserved.
|
|
*
|
|
* This program is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU General Public License
|
|
* as published by the Free Software Foundation, version 2.
|
|
*
|
|
* This program is distributed in the hope that it will be useful, but
|
|
* WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE, GOOD TITLE or
|
|
* NON INFRINGEMENT. See the GNU General Public License for
|
|
* more details.
|
|
*
|
|
* based on machine_kexec.c from other architectures in linux-2.6.18
|
|
*/
|
|
|
|
#include <linux/mm.h>
|
|
#include <linux/kexec.h>
|
|
#include <linux/delay.h>
|
|
#include <linux/reboot.h>
|
|
#include <linux/errno.h>
|
|
#include <linux/vmalloc.h>
|
|
#include <linux/cpumask.h>
|
|
#include <linux/kernel.h>
|
|
#include <linux/elf.h>
|
|
#include <linux/highmem.h>
|
|
#include <linux/mmu_context.h>
|
|
#include <linux/io.h>
|
|
#include <linux/timex.h>
|
|
#include <asm/pgtable.h>
|
|
#include <asm/pgalloc.h>
|
|
#include <asm/cacheflush.h>
|
|
#include <asm/checksum.h>
|
|
#include <asm/tlbflush.h>
|
|
#include <asm/homecache.h>
|
|
#include <hv/hypervisor.h>
|
|
|
|
|
|
/*
|
|
* This stuff is not in elf.h and is not in any other kernel include.
|
|
* This stuff is needed below in the little boot notes parser to
|
|
* extract the command line so we can pass it to the hypervisor.
|
|
*/
|
|
struct Elf32_Bhdr {
|
|
Elf32_Word b_signature;
|
|
Elf32_Word b_size;
|
|
Elf32_Half b_checksum;
|
|
Elf32_Half b_records;
|
|
};
|
|
#define ELF_BOOT_MAGIC 0x0E1FB007
|
|
#define EBN_COMMAND_LINE 0x00000004
|
|
#define roundupsz(X) (((X) + 3) & ~3)
|
|
|
|
/* - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - */
|
|
|
|
|
|
void machine_shutdown(void)
|
|
{
|
|
/*
|
|
* Normally we would stop all the other processors here, but
|
|
* the check in machine_kexec_prepare below ensures we'll only
|
|
* get this far if we've been booted with "nosmp" on the
|
|
* command line or without CONFIG_SMP so there's nothing to do
|
|
* here (for now).
|
|
*/
|
|
}
|
|
|
|
void machine_crash_shutdown(struct pt_regs *regs)
|
|
{
|
|
/*
|
|
* Cannot happen. This type of kexec is disabled on this
|
|
* architecture (and enforced in machine_kexec_prepare below).
|
|
*/
|
|
}
|
|
|
|
|
|
int machine_kexec_prepare(struct kimage *image)
|
|
{
|
|
if (num_online_cpus() > 1) {
|
|
pr_warning("%s: detected attempt to kexec "
|
|
"with num_online_cpus() > 1\n",
|
|
__func__);
|
|
return -ENOSYS;
|
|
}
|
|
if (image->type != KEXEC_TYPE_DEFAULT) {
|
|
pr_warning("%s: detected attempt to kexec "
|
|
"with unsupported type: %d\n",
|
|
__func__,
|
|
image->type);
|
|
return -ENOSYS;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
void machine_kexec_cleanup(struct kimage *image)
|
|
{
|
|
/*
|
|
* We did nothing in machine_kexec_prepare,
|
|
* so we have nothing to do here.
|
|
*/
|
|
}
|
|
|
|
/*
|
|
* If we can find elf boot notes on this page, return the command
|
|
* line. Otherwise, silently return null. Somewhat kludgy, but no
|
|
* good way to do this without significantly rearchitecting the
|
|
* architecture-independent kexec code.
|
|
*/
|
|
|
|
static unsigned char *kexec_bn2cl(void *pg)
|
|
{
|
|
struct Elf32_Bhdr *bhdrp;
|
|
Elf32_Nhdr *nhdrp;
|
|
unsigned char *desc;
|
|
unsigned char *command_line;
|
|
__sum16 csum;
|
|
|
|
bhdrp = (struct Elf32_Bhdr *) pg;
|
|
|
|
/*
|
|
* This routine is invoked for every source page, so make
|
|
* sure to quietly ignore every impossible page.
|
|
*/
|
|
if (bhdrp->b_signature != ELF_BOOT_MAGIC ||
|
|
bhdrp->b_size > PAGE_SIZE)
|
|
return 0;
|
|
|
|
/*
|
|
* If we get a checksum mismatch, warn with the checksum
|
|
* so we can diagnose better.
|
|
*/
|
|
csum = ip_compute_csum(pg, bhdrp->b_size);
|
|
if (csum != 0) {
|
|
pr_warning("%s: bad checksum %#x (size %d)\n",
|
|
__func__, csum, bhdrp->b_size);
|
|
return 0;
|
|
}
|
|
|
|
nhdrp = (Elf32_Nhdr *) (bhdrp + 1);
|
|
|
|
while (nhdrp->n_type != EBN_COMMAND_LINE) {
|
|
|
|
desc = (unsigned char *) (nhdrp + 1);
|
|
desc += roundupsz(nhdrp->n_descsz);
|
|
|
|
nhdrp = (Elf32_Nhdr *) desc;
|
|
|
|
/* still in bounds? */
|
|
if ((unsigned char *) (nhdrp + 1) >
|
|
((unsigned char *) pg) + bhdrp->b_size) {
|
|
|
|
pr_info("%s: out of bounds\n", __func__);
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
command_line = (unsigned char *) (nhdrp + 1);
|
|
desc = command_line;
|
|
|
|
while (*desc != '\0') {
|
|
desc++;
|
|
if (((unsigned long)desc & PAGE_MASK) != (unsigned long)pg) {
|
|
pr_info("%s: ran off end of page\n",
|
|
__func__);
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
return command_line;
|
|
}
|
|
|
|
static void kexec_find_and_set_command_line(struct kimage *image)
|
|
{
|
|
kimage_entry_t *ptr, entry;
|
|
|
|
unsigned char *command_line = 0;
|
|
unsigned char *r;
|
|
HV_Errno hverr;
|
|
|
|
for (ptr = &image->head;
|
|
(entry = *ptr) && !(entry & IND_DONE);
|
|
ptr = (entry & IND_INDIRECTION) ?
|
|
phys_to_virt((entry & PAGE_MASK)) : ptr + 1) {
|
|
|
|
if ((entry & IND_SOURCE)) {
|
|
void *va =
|
|
kmap_atomic_pfn(entry >> PAGE_SHIFT);
|
|
r = kexec_bn2cl(va);
|
|
if (r) {
|
|
command_line = r;
|
|
break;
|
|
}
|
|
kunmap_atomic(va);
|
|
}
|
|
}
|
|
|
|
if (command_line != 0) {
|
|
pr_info("setting new command line to \"%s\"\n",
|
|
command_line);
|
|
|
|
hverr = hv_set_command_line(
|
|
(HV_VirtAddr) command_line, strlen(command_line));
|
|
kunmap_atomic(command_line);
|
|
} else {
|
|
pr_info("%s: no command line found; making empty\n",
|
|
__func__);
|
|
hverr = hv_set_command_line((HV_VirtAddr) command_line, 0);
|
|
}
|
|
if (hverr)
|
|
pr_warning("%s: hv_set_command_line returned error: %d\n",
|
|
__func__, hverr);
|
|
}
|
|
|
|
/*
|
|
* The kexec code range-checks all its PAs, so to avoid having it run
|
|
* amok and allocate memory and then sequester it from every other
|
|
* controller, we force it to come from controller zero. We also
|
|
* disable the oom-killer since if we do end up running out of memory,
|
|
* that almost certainly won't help.
|
|
*/
|
|
struct page *kimage_alloc_pages_arch(gfp_t gfp_mask, unsigned int order)
|
|
{
|
|
gfp_mask |= __GFP_THISNODE | __GFP_NORETRY;
|
|
return alloc_pages_node(0, gfp_mask, order);
|
|
}
|
|
|
|
/*
|
|
* Address range in which pa=va mapping is set in setup_quasi_va_is_pa().
|
|
* For tilepro, PAGE_OFFSET is used since this is the largest possbile value
|
|
* for tilepro, while for tilegx, we limit it to entire middle level page
|
|
* table which we assume has been allocated and is undoubtedly large enough.
|
|
*/
|
|
#ifndef __tilegx__
|
|
#define QUASI_VA_IS_PA_ADDR_RANGE PAGE_OFFSET
|
|
#else
|
|
#define QUASI_VA_IS_PA_ADDR_RANGE PGDIR_SIZE
|
|
#endif
|
|
|
|
static void setup_quasi_va_is_pa(void)
|
|
{
|
|
HV_PTE pte;
|
|
unsigned long i;
|
|
|
|
/*
|
|
* Flush our TLB to prevent conflicts between the previous contents
|
|
* and the new stuff we're about to add.
|
|
*/
|
|
local_flush_tlb_all();
|
|
|
|
/*
|
|
* setup VA is PA, at least up to QUASI_VA_IS_PA_ADDR_RANGE.
|
|
* Note here we assume that level-1 page table is defined by
|
|
* HPAGE_SIZE.
|
|
*/
|
|
pte = hv_pte(_PAGE_KERNEL | _PAGE_HUGE_PAGE);
|
|
pte = hv_pte_set_mode(pte, HV_PTE_MODE_CACHE_NO_L3);
|
|
for (i = 0; i < (QUASI_VA_IS_PA_ADDR_RANGE >> HPAGE_SHIFT); i++) {
|
|
unsigned long vaddr = i << HPAGE_SHIFT;
|
|
pgd_t *pgd = pgd_offset(current->mm, vaddr);
|
|
pud_t *pud = pud_offset(pgd, vaddr);
|
|
pte_t *ptep = (pte_t *) pmd_offset(pud, vaddr);
|
|
unsigned long pfn = i << (HPAGE_SHIFT - PAGE_SHIFT);
|
|
|
|
if (pfn_valid(pfn))
|
|
__set_pte(ptep, pfn_pte(pfn, pte));
|
|
}
|
|
}
|
|
|
|
|
|
void machine_kexec(struct kimage *image)
|
|
{
|
|
void *reboot_code_buffer;
|
|
pte_t *ptep;
|
|
void (*rnk)(unsigned long, void *, unsigned long)
|
|
__noreturn;
|
|
|
|
/* Mask all interrupts before starting to reboot. */
|
|
interrupt_mask_set_mask(~0ULL);
|
|
|
|
kexec_find_and_set_command_line(image);
|
|
|
|
/*
|
|
* Adjust the home caching of the control page to be cached on
|
|
* this cpu, and copy the assembly helper into the control
|
|
* code page, which we map in the vmalloc area.
|
|
*/
|
|
homecache_change_page_home(image->control_code_page, 0,
|
|
smp_processor_id());
|
|
reboot_code_buffer = page_address(image->control_code_page);
|
|
BUG_ON(reboot_code_buffer == NULL);
|
|
ptep = virt_to_pte(NULL, (unsigned long)reboot_code_buffer);
|
|
__set_pte(ptep, pte_mkexec(*ptep));
|
|
memcpy(reboot_code_buffer, relocate_new_kernel,
|
|
relocate_new_kernel_size);
|
|
__flush_icache_range(
|
|
(unsigned long) reboot_code_buffer,
|
|
(unsigned long) reboot_code_buffer + relocate_new_kernel_size);
|
|
|
|
setup_quasi_va_is_pa();
|
|
|
|
/* now call it */
|
|
rnk = reboot_code_buffer;
|
|
(*rnk)(image->head, reboot_code_buffer, image->start);
|
|
}
|