mirror of https://github.com/crash-utility/crash
5444 lines
157 KiB
C
5444 lines
157 KiB
C
/*
|
|
* arm64.c - core analysis suite
|
|
*
|
|
* Copyright (C) 2012-2020 David Anderson
|
|
* Copyright (C) 2012-2020 Red Hat, Inc. All rights reserved.
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify
|
|
* it under the terms of the GNU General Public License as published by
|
|
* the Free Software Foundation; either version 2 of the License, or
|
|
* (at your option) any later version.
|
|
*
|
|
* This program is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
* GNU General Public License for more details.
|
|
*/
|
|
|
|
#ifdef ARM64
|
|
|
|
#include "defs.h"
|
|
#include <elf.h>
|
|
#include <endian.h>
|
|
#include <math.h>
|
|
#include <sys/ioctl.h>
|
|
|
|
#define NOT_IMPLEMENTED(X) error((X), "%s: function not implemented\n", __func__)
|
|
|
|
static struct machine_specific arm64_machine_specific = { 0 };
|
|
static int arm64_verify_symbol(const char *, ulong, char);
|
|
static void arm64_parse_cmdline_args(void);
|
|
static int arm64_search_for_kimage_voffset(ulong);
|
|
static int verify_kimage_voffset(void);
|
|
static void arm64_calc_kimage_voffset(void);
|
|
static void arm64_calc_phys_offset(void);
|
|
static void arm64_calc_physvirt_offset(void);
|
|
static void arm64_calc_virtual_memory_ranges(void);
|
|
static void arm64_get_section_size_bits(void);
|
|
static int arm64_kdump_phys_base(ulong *);
|
|
static ulong arm64_processor_speed(void);
|
|
static void arm64_init_kernel_pgd(void);
|
|
static int arm64_kvtop(struct task_context *, ulong, physaddr_t *, int);
|
|
static int arm64_uvtop(struct task_context *, ulong, physaddr_t *, int);
|
|
static int arm64_vtop_2level_64k(ulong, ulong, physaddr_t *, int);
|
|
static int arm64_vtop_3level_64k(ulong, ulong, physaddr_t *, int);
|
|
static int arm64_vtop_2level_16k(ulong, ulong, physaddr_t *, int);
|
|
static int arm64_vtop_3level_16k(ulong, ulong, physaddr_t *, int);
|
|
static int arm64_vtop_4level_16k(ulong, ulong, physaddr_t *, int);
|
|
static int arm64_vtop_3level_4k(ulong, ulong, physaddr_t *, int);
|
|
static int arm64_vtop_4level_4k(ulong, ulong, physaddr_t *, int);
|
|
static ulong arm64_get_task_pgd(ulong);
|
|
static void arm64_irq_stack_init(void);
|
|
static void arm64_overflow_stack_init(void);
|
|
static void arm64_stackframe_init(void);
|
|
static int arm64_eframe_search(struct bt_info *);
|
|
static int arm64_is_kernel_exception_frame(struct bt_info *, ulong);
|
|
static int arm64_in_exception_text(ulong);
|
|
static int arm64_in_exp_entry(ulong);
|
|
static void arm64_back_trace_cmd(struct bt_info *);
|
|
static void arm64_back_trace_cmd_v2(struct bt_info *);
|
|
static void arm64_print_text_symbols(struct bt_info *, struct arm64_stackframe *, FILE *);
|
|
static int arm64_print_stackframe_entry(struct bt_info *, int, struct arm64_stackframe *, FILE *);
|
|
static int arm64_print_stackframe_entry_v2(struct bt_info *, int, struct arm64_stackframe *, FILE *);
|
|
static void arm64_display_full_frame(struct bt_info *, ulong);
|
|
static void arm64_display_full_frame_v2(struct bt_info *, struct arm64_stackframe *, struct arm64_stackframe *);
|
|
static int arm64_unwind_frame(struct bt_info *, struct arm64_stackframe *);
|
|
static int arm64_unwind_frame_v2(struct bt_info *, struct arm64_stackframe *, FILE *);
|
|
static int arm64_get_dumpfile_stackframe(struct bt_info *, struct arm64_stackframe *);
|
|
static int arm64_in_kdump_text(struct bt_info *, struct arm64_stackframe *);
|
|
static int arm64_in_kdump_text_on_irq_stack(struct bt_info *);
|
|
static int arm64_switch_stack(struct bt_info *, struct arm64_stackframe *, FILE *);
|
|
static int arm64_switch_stack_from_overflow(struct bt_info *, struct arm64_stackframe *, FILE *);
|
|
static int arm64_get_stackframe(struct bt_info *, struct arm64_stackframe *);
|
|
static void arm64_get_stack_frame(struct bt_info *, ulong *, ulong *);
|
|
static void arm64_gen_hidden_frame(struct bt_info *bt, ulong, struct arm64_stackframe *);
|
|
static void arm64_print_exception_frame(struct bt_info *, ulong, int, FILE *);
|
|
static void arm64_do_bt_reference_check(struct bt_info *, ulong, char *);
|
|
static int arm64_translate_pte(ulong, void *, ulonglong);
|
|
static ulong arm64_vmalloc_start(void);
|
|
static int arm64_is_task_addr(ulong);
|
|
static int arm64_dis_filter(ulong, char *, unsigned int);
|
|
static void arm64_cmd_mach(void);
|
|
static void arm64_display_machine_stats(void);
|
|
static int arm64_get_smp_cpus(void);
|
|
static void arm64_clear_machdep_cache(void);
|
|
static int arm64_on_process_stack(struct bt_info *, ulong);
|
|
static int arm64_in_alternate_stack(int, ulong);
|
|
static int arm64_in_alternate_stackv(int cpu, ulong stkptr, ulong *stacks, ulong stack_size);
|
|
static int arm64_on_irq_stack(int, ulong);
|
|
static int arm64_on_overflow_stack(int, ulong);
|
|
static void arm64_set_irq_stack(struct bt_info *);
|
|
static void arm64_set_overflow_stack(struct bt_info *);
|
|
static void arm64_set_process_stack(struct bt_info *);
|
|
static int arm64_get_kvaddr_ranges(struct vaddr_range *);
|
|
static void arm64_get_crash_notes(void);
|
|
static void arm64_calc_VA_BITS(void);
|
|
static int arm64_is_uvaddr(ulong, struct task_context *);
|
|
static void arm64_calc_KERNELPACMASK(void);
|
|
static void arm64_recalc_KERNELPACMASK(void);
|
|
static int arm64_get_vmcoreinfo(unsigned long *vaddr, const char *label, int base);
|
|
static ulong arm64_set_irq_stack_size(void);
|
|
|
|
struct kernel_range {
|
|
unsigned long modules_vaddr, modules_end;
|
|
unsigned long vmalloc_start_addr, vmalloc_end;
|
|
unsigned long vmemmap_vaddr, vmemmap_end;
|
|
};
|
|
static struct kernel_range *arm64_get_va_range(struct machine_specific *ms);
|
|
static void arm64_get_struct_page_size(struct machine_specific *ms);
|
|
|
|
/* mte tag shift bit */
|
|
#define MTE_TAG_SHIFT 56
|
|
/* native kernel pointers tag */
|
|
#define KASAN_TAG_KERNEL 0xFF
|
|
/* minimum value for random tags */
|
|
#define KASAN_TAG_MIN 0xF0
|
|
/* right shift the tag to MTE_TAG_SHIFT bit */
|
|
#define mte_tag_shifted(tag) ((ulong)(tag) << MTE_TAG_SHIFT)
|
|
/* get the top byte value of the original kvaddr */
|
|
#define mte_tag_get(addr) (unsigned char)((ulong)(addr) >> MTE_TAG_SHIFT)
|
|
/* reset the top byte to get an untaggged kvaddr */
|
|
#define mte_tag_reset(addr) (((ulong)addr & ~mte_tag_shifted(KASAN_TAG_KERNEL)) | \
|
|
mte_tag_shifted(KASAN_TAG_KERNEL))
|
|
|
|
struct user_regs_bitmap_struct {
|
|
struct arm64_pt_regs ur;
|
|
ulong bitmap[32];
|
|
};
|
|
|
|
static inline bool is_mte_kvaddr(ulong addr)
|
|
{
|
|
/* check for ARM64_MTE enabled */
|
|
if (!(machdep->flags & ARM64_MTE))
|
|
return false;
|
|
|
|
/* check the validity of HW Tag-Based kvaddr */
|
|
if (mte_tag_get(addr) >= KASAN_TAG_MIN && mte_tag_get(addr) < KASAN_TAG_KERNEL)
|
|
return true;
|
|
|
|
return false;
|
|
}
|
|
|
|
static int arm64_is_kvaddr(ulong addr)
|
|
{
|
|
if (is_mte_kvaddr(addr))
|
|
return (mte_tag_reset(addr) >= (ulong)(machdep->kvbase));
|
|
|
|
return (addr >= (ulong)(machdep->kvbase));
|
|
}
|
|
|
|
static void arm64_calc_kernel_start(void)
|
|
{
|
|
struct machine_specific *ms = machdep->machspec;
|
|
struct syment *sp;
|
|
|
|
if (THIS_KERNEL_VERSION >= LINUX(5,11,0))
|
|
sp = kernel_symbol_search("_stext");
|
|
else
|
|
sp = kernel_symbol_search("_text");
|
|
|
|
ms->kimage_text = (sp ? sp->value : 0);
|
|
sp = kernel_symbol_search("_end");
|
|
ms->kimage_end = (sp ? sp->value : 0);
|
|
}
|
|
|
|
static int
|
|
arm64_vmemmap_is_page_ptr(ulong addr, physaddr_t *phys)
|
|
{
|
|
ulong size = SIZE(page);
|
|
ulong pfn, nr;
|
|
|
|
if (IS_SPARSEMEM() && (machdep->flags & VMEMMAP) &&
|
|
(addr >= VMEMMAP_VADDR && addr <= VMEMMAP_END) &&
|
|
!((addr - VMEMMAP_VADDR) % size)) {
|
|
|
|
pfn = (addr - machdep->machspec->vmemmap) / size;
|
|
nr = pfn_to_section_nr(pfn);
|
|
if (valid_section_nr(nr)) {
|
|
if (phys)
|
|
*phys = PTOB(pfn);
|
|
return TRUE;
|
|
}
|
|
}
|
|
return FALSE;
|
|
}
|
|
|
|
static void arm64_get_vmemmap_page_ptr(void)
|
|
{
|
|
struct machine_specific *ms = machdep->machspec;
|
|
|
|
/* If vmemmap exists, it means kernel enabled CONFIG_SPARSEMEM_VMEMMAP */
|
|
if (arm64_get_vmcoreinfo(&ms->vmemmap, "SYMBOL(vmemmap)", NUM_HEX))
|
|
goto out;
|
|
|
|
/* The global symbol of vmemmap is removed since kernel commit 7bc1a0f9e1765 */
|
|
if (kernel_symbol_exists("vmemmap"))
|
|
ms->vmemmap = symbol_value("vmemmap");
|
|
else
|
|
ms->vmemmap = ms->vmemmap_vaddr - ((ms->phys_offset >> machdep->pageshift) * ms->struct_page_size);
|
|
|
|
out:
|
|
if (ms->vmemmap)
|
|
machdep->is_page_ptr = arm64_vmemmap_is_page_ptr;
|
|
}
|
|
|
|
static int
|
|
arm64_get_current_task_reg(int regno, const char *name,
|
|
int size, void *value)
|
|
{
|
|
struct bt_info bt_info, bt_setup;
|
|
struct task_context *tc;
|
|
struct user_regs_bitmap_struct *ur_bitmap;
|
|
ulong ip, sp;
|
|
bool ret = FALSE;
|
|
|
|
switch (regno) {
|
|
case X0_REGNUM ... PC_REGNUM:
|
|
break;
|
|
default:
|
|
return FALSE;
|
|
}
|
|
|
|
tc = CURRENT_CONTEXT();
|
|
if (!tc)
|
|
return FALSE;
|
|
BZERO(&bt_setup, sizeof(struct bt_info));
|
|
clone_bt_info(&bt_setup, &bt_info, tc);
|
|
fill_stackbuf(&bt_info);
|
|
|
|
get_dumpfile_regs(&bt_info, &sp, &ip);
|
|
if (bt_info.stackbuf)
|
|
FREEBUF(bt_info.stackbuf);
|
|
ur_bitmap = (struct user_regs_bitmap_struct *)bt_info.machdep;
|
|
if (!ur_bitmap)
|
|
return FALSE;
|
|
|
|
if (!bt_info.need_free) {
|
|
goto get_all;
|
|
}
|
|
|
|
switch (regno) {
|
|
case X0_REGNUM ... X30_REGNUM:
|
|
if (!NUM_IN_BITMAP(ur_bitmap->bitmap,
|
|
REG_SEQ(arm64_pt_regs, regs[0]) + regno - X0_REGNUM)) {
|
|
FREEBUF(ur_bitmap);
|
|
return FALSE;
|
|
}
|
|
break;
|
|
case SP_REGNUM:
|
|
if (!NUM_IN_BITMAP(ur_bitmap->bitmap,
|
|
REG_SEQ(arm64_pt_regs, sp))) {
|
|
FREEBUF(ur_bitmap);
|
|
return FALSE;
|
|
}
|
|
break;
|
|
case PC_REGNUM:
|
|
if (!NUM_IN_BITMAP(ur_bitmap->bitmap,
|
|
REG_SEQ(arm64_pt_regs, pc))) {
|
|
FREEBUF(ur_bitmap);
|
|
return FALSE;
|
|
}
|
|
break;
|
|
}
|
|
|
|
get_all:
|
|
switch (regno) {
|
|
case X0_REGNUM ... X30_REGNUM:
|
|
if (size != sizeof(ur_bitmap->ur.regs[regno]))
|
|
break;
|
|
memcpy(value, &ur_bitmap->ur.regs[regno], size);
|
|
ret = TRUE;
|
|
break;
|
|
case SP_REGNUM:
|
|
if (size != sizeof(ur_bitmap->ur.sp))
|
|
break;
|
|
memcpy(value, &ur_bitmap->ur.sp, size);
|
|
ret = TRUE;
|
|
break;
|
|
case PC_REGNUM:
|
|
if (size != sizeof(ur_bitmap->ur.pc))
|
|
break;
|
|
memcpy(value, &ur_bitmap->ur.pc, size);
|
|
ret = TRUE;
|
|
break;
|
|
}
|
|
|
|
if (bt_info.need_free) {
|
|
FREEBUF(ur_bitmap);
|
|
bt_info.need_free = FALSE;
|
|
}
|
|
return ret;
|
|
}
|
|
|
|
/*
|
|
* Do all necessary machine-specific setup here. This is called several times
|
|
* during initialization.
|
|
*/
|
|
void
|
|
arm64_init(int when)
|
|
{
|
|
ulong value;
|
|
struct machine_specific *ms;
|
|
|
|
#if defined(__x86_64__)
|
|
if (ACTIVE())
|
|
error(FATAL, "compiled for the ARM64 architecture\n");
|
|
#endif
|
|
|
|
switch (when) {
|
|
case SETUP_ENV:
|
|
machdep->process_elf_notes = process_elf64_notes;
|
|
break;
|
|
|
|
case PRE_SYMTAB:
|
|
machdep->machspec = &arm64_machine_specific;
|
|
machdep->verify_symbol = arm64_verify_symbol;
|
|
if (pc->flags & KERNEL_DEBUG_QUERY)
|
|
return;
|
|
machdep->verify_paddr = generic_verify_paddr;
|
|
if (machdep->cmdline_args[0])
|
|
arm64_parse_cmdline_args();
|
|
machdep->flags |= MACHDEP_BT_TEXT;
|
|
|
|
ms = machdep->machspec;
|
|
|
|
/*
|
|
* The st->_stext_vmlinux is needed in arm64_init(PRE_GDB) when a
|
|
* dumpfile does not have vmcoreinfo and we use -m vabits_actual
|
|
* option, e.g. a raw RAM dumpfile.
|
|
*/
|
|
if (ms->VA_BITS_ACTUAL)
|
|
st->_stext_vmlinux = UNINITIALIZED;
|
|
|
|
if (!ms->kimage_voffset && STREQ(pc->live_memsrc, "/dev/crash"))
|
|
ioctl(pc->mfd, DEV_CRASH_ARCH_DATA, &ms->kimage_voffset);
|
|
|
|
if (!ms->kimage_voffset)
|
|
arm64_get_vmcoreinfo(&ms->kimage_voffset, "NUMBER(kimage_voffset)", NUM_HEX);
|
|
|
|
if (ms->kimage_voffset ||
|
|
(ACTIVE() && (symbol_value_from_proc_kallsyms("kimage_voffset") != BADVAL))) {
|
|
machdep->flags |= NEW_VMEMMAP;
|
|
|
|
/*
|
|
* Even if CONFIG_RANDOMIZE_BASE is not configured,
|
|
* derive_kaslr_offset() should work and set
|
|
* kt->relocate to 0
|
|
*/
|
|
if (!kt->relocate && !(kt->flags2 & (RELOC_AUTO|KASLR)))
|
|
kt->flags2 |= (RELOC_AUTO|KASLR);
|
|
}
|
|
|
|
break;
|
|
|
|
case PRE_GDB:
|
|
if (kernel_symbol_exists("kimage_voffset"))
|
|
machdep->flags |= NEW_VMEMMAP;
|
|
|
|
if (kernel_symbol_exists("cpu_enable_mte"))
|
|
machdep->flags |= ARM64_MTE;
|
|
|
|
if (!machdep->pagesize && arm64_get_vmcoreinfo(&value, "PAGESIZE", NUM_DEC))
|
|
machdep->pagesize = (unsigned int)value;
|
|
|
|
if (!machdep->pagesize) {
|
|
/*
|
|
* Kerneldoc Documentation/arm64/booting.txt describes
|
|
* the kernel image header flags field.
|
|
*/
|
|
value = machdep->machspec->kernel_flags;
|
|
value = (value >> 1) & 3;
|
|
|
|
switch(value)
|
|
{
|
|
case 0:
|
|
break;
|
|
case 1:
|
|
machdep->pagesize = 4096;
|
|
break;
|
|
case 2:
|
|
machdep->pagesize = 16384;
|
|
break;
|
|
case 3:
|
|
machdep->pagesize = 65536;
|
|
break;
|
|
}
|
|
|
|
}
|
|
|
|
/*
|
|
* This code section will only be executed if the kernel is
|
|
* earlier than Linux 4.4 (if there is no vmcoreinfo)
|
|
*/
|
|
if (!machdep->pagesize &&
|
|
kernel_symbol_exists("swapper_pg_dir") &&
|
|
kernel_symbol_exists("idmap_pg_dir")) {
|
|
value = symbol_value("swapper_pg_dir") -
|
|
symbol_value("idmap_pg_dir");
|
|
|
|
/*
|
|
* idmap_pg_dir is 2 pages prior to 4.1,
|
|
* and 3 pages thereafter. Only 4K and 64K
|
|
* page sizes are supported.
|
|
*/
|
|
switch (value)
|
|
{
|
|
case (4096 * 2):
|
|
case (4096 * 3):
|
|
machdep->pagesize = 4096;
|
|
break;
|
|
case (65536 * 2):
|
|
case (65536 * 3):
|
|
machdep->pagesize = 65536;
|
|
break;
|
|
}
|
|
} else if (ACTIVE())
|
|
machdep->pagesize = memory_page_size(); /* host */
|
|
|
|
machdep->pageshift = ffs(machdep->pagesize) - 1;
|
|
machdep->pageoffset = machdep->pagesize - 1;
|
|
machdep->pagemask = ~((ulonglong)machdep->pageoffset);
|
|
|
|
ms = machdep->machspec;
|
|
arm64_get_struct_page_size(ms);
|
|
arm64_calc_VA_BITS();
|
|
arm64_calc_KERNELPACMASK();
|
|
|
|
/* vabits_actual introduced after mm flip, so it should be flipped layout */
|
|
if (ms->VA_BITS_ACTUAL) {
|
|
ms->page_offset = ARM64_FLIP_PAGE_OFFSET;
|
|
/* useless on arm64 */
|
|
machdep->identity_map_base = ARM64_FLIP_PAGE_OFFSET;
|
|
machdep->kvbase = ARM64_FLIP_PAGE_OFFSET;
|
|
ms->userspace_top = ARM64_USERSPACE_TOP_ACTUAL;
|
|
} else {
|
|
ms->page_offset = ARM64_PAGE_OFFSET;
|
|
machdep->identity_map_base = ARM64_PAGE_OFFSET;
|
|
machdep->kvbase = ARM64_VA_START;
|
|
ms->userspace_top = ARM64_USERSPACE_TOP;
|
|
}
|
|
machdep->is_kvaddr = arm64_is_kvaddr;
|
|
machdep->kvtop = arm64_kvtop;
|
|
|
|
/* The defaults */
|
|
ms->vmalloc_end = ARM64_VMALLOC_END;
|
|
ms->vmemmap_vaddr = ARM64_VMEMMAP_VADDR;
|
|
ms->vmemmap_end = ARM64_VMEMMAP_END;
|
|
|
|
if (machdep->flags & NEW_VMEMMAP) {
|
|
struct syment *sp;
|
|
struct kernel_range *r;
|
|
|
|
/* It is finally decided in arm64_calc_kernel_start() */
|
|
sp = kernel_symbol_search("_text");
|
|
ms->kimage_text = (sp ? sp->value : 0);
|
|
sp = kernel_symbol_search("_end");
|
|
ms->kimage_end = (sp ? sp->value : 0);
|
|
|
|
if (ms->struct_page_size && (r = arm64_get_va_range(ms))) {
|
|
/* We can get all the MODULES/VMALLOC/VMEMMAP ranges now.*/
|
|
ms->modules_vaddr = r->modules_vaddr;
|
|
ms->modules_end = r->modules_end - 1;
|
|
ms->vmalloc_start_addr = r->vmalloc_start_addr;
|
|
ms->vmalloc_end = r->vmalloc_end - 1;
|
|
ms->vmemmap_vaddr = r->vmemmap_vaddr;
|
|
ms->vmemmap_end = r->vmemmap_end - 1;
|
|
} else if (ms->VA_BITS_ACTUAL) {
|
|
ms->modules_vaddr = (st->_stext_vmlinux & TEXT_OFFSET_MASK) - ARM64_MODULES_VSIZE;
|
|
ms->modules_end = ms->modules_vaddr + ARM64_MODULES_VSIZE -1;
|
|
ms->vmalloc_start_addr = ms->modules_end + 1;
|
|
} else {
|
|
ms->modules_vaddr = ARM64_VA_START;
|
|
if (kernel_symbol_exists("kasan_init"))
|
|
ms->modules_vaddr += ARM64_KASAN_SHADOW_SIZE;
|
|
ms->modules_end = ms->modules_vaddr + ARM64_MODULES_VSIZE -1;
|
|
ms->vmalloc_start_addr = ms->modules_end + 1;
|
|
}
|
|
|
|
arm64_calc_kimage_voffset();
|
|
} else {
|
|
ms->modules_vaddr = ARM64_PAGE_OFFSET - MEGABYTES(64);
|
|
ms->modules_end = ARM64_PAGE_OFFSET - 1;
|
|
ms->vmalloc_start_addr = ARM64_VA_START;
|
|
}
|
|
|
|
switch (machdep->pagesize)
|
|
{
|
|
case 4096:
|
|
machdep->ptrs_per_pgd = PTRS_PER_PGD_L3_4K;
|
|
if ((machdep->pgd =
|
|
(char *)malloc(PTRS_PER_PGD_L3_4K * 8)) == NULL)
|
|
error(FATAL, "cannot malloc pgd space.");
|
|
if (machdep->machspec->VA_BITS > PGDIR_SHIFT_L4_4K) {
|
|
machdep->flags |= VM_L4_4K;
|
|
if ((machdep->pud =
|
|
(char *)malloc(PTRS_PER_PUD_L4_4K * 8))
|
|
== NULL)
|
|
error(FATAL, "cannot malloc pud space.");
|
|
} else {
|
|
machdep->flags |= VM_L3_4K;
|
|
machdep->pud = NULL; /* not used */
|
|
}
|
|
if ((machdep->pmd =
|
|
(char *)malloc(PTRS_PER_PMD_L3_4K * 8)) == NULL)
|
|
error(FATAL, "cannot malloc pmd space.");
|
|
if ((machdep->ptbl =
|
|
(char *)malloc(PTRS_PER_PTE_L3_4K * 8)) == NULL)
|
|
error(FATAL, "cannot malloc ptbl space.");
|
|
break;
|
|
|
|
case 16384:
|
|
if (machdep->machspec->VA_BITS == 48) {
|
|
machdep->flags |= VM_L4_16K;
|
|
if (!machdep->ptrs_per_pgd)
|
|
machdep->ptrs_per_pgd = PTRS_PER_PGD_L4_16K;
|
|
if ((machdep->pgd =
|
|
(char *)malloc(machdep->ptrs_per_pgd * 8)) == NULL)
|
|
error(FATAL, "cannot malloc pgd space.");
|
|
if ((machdep->pud =
|
|
(char *)malloc(PTRS_PER_PUD_L4_16K * 8)) == NULL)
|
|
error(FATAL, "cannot malloc pud space.");
|
|
if ((machdep->pmd =
|
|
(char *)malloc(PTRS_PER_PMD_L4_16K * 8)) == NULL)
|
|
error(FATAL, "cannot malloc pmd space.");
|
|
if ((machdep->ptbl =
|
|
(char *)malloc(PTRS_PER_PTE_L4_16K * 8)) == NULL)
|
|
error(FATAL, "cannot malloc ptbl space.");
|
|
}
|
|
else if (machdep->machspec->VA_BITS == 47) {
|
|
machdep->flags |= VM_L3_16K;
|
|
if (!machdep->ptrs_per_pgd)
|
|
machdep->ptrs_per_pgd = PTRS_PER_PGD_L3_16K;
|
|
if ((machdep->pgd =
|
|
(char *)malloc(machdep->ptrs_per_pgd * 8)) == NULL)
|
|
error(FATAL, "cannot malloc pgd space.");
|
|
if ((machdep->pmd =
|
|
(char *)malloc(PTRS_PER_PMD_L3_16K * 8)) == NULL)
|
|
error(FATAL, "cannot malloc pmd space.");
|
|
if ((machdep->ptbl =
|
|
(char *)malloc(PTRS_PER_PTE_L3_16K * 8)) == NULL)
|
|
error(FATAL, "cannot malloc ptbl space.");
|
|
machdep->pud = NULL; /* not used */
|
|
} else if (machdep->machspec->VA_BITS == 36) {
|
|
machdep->flags |= VM_L2_16K;
|
|
if (!machdep->ptrs_per_pgd)
|
|
machdep->ptrs_per_pgd = PTRS_PER_PGD_L2_16K;
|
|
if ((machdep->pgd =
|
|
(char *)malloc(machdep->ptrs_per_pgd * 8)) == NULL)
|
|
error(FATAL, "cannot malloc pgd space.");
|
|
if ((machdep->ptbl =
|
|
(char *)malloc(PTRS_PER_PTE_L2_16K * 8)) == NULL)
|
|
error(FATAL, "cannot malloc ptbl space.");
|
|
machdep->pmd = NULL; /* not used */
|
|
machdep->pud = NULL; /* not used */
|
|
} else {
|
|
error(FATAL, "Do not support 52 bits, 4-level for 16K page now.");
|
|
}
|
|
break;
|
|
|
|
case 65536:
|
|
/*
|
|
* idmap_ptrs_per_pgd has been removed since Linux-v6.0-rc1, see:
|
|
* commit ebd9aea1f27e ("arm64: head: drop idmap_ptrs_per_pgd")
|
|
*/
|
|
if (kernel_symbol_exists("idmap_ptrs_per_pgd") &&
|
|
readmem(symbol_value("idmap_ptrs_per_pgd"), KVADDR,
|
|
&value, sizeof(ulong), "idmap_ptrs_per_pgd", QUIET|RETURN_ON_ERROR))
|
|
machdep->ptrs_per_pgd = value;
|
|
|
|
if (machdep->machspec->VA_BITS > PGDIR_SHIFT_L3_64K) {
|
|
machdep->flags |= VM_L3_64K;
|
|
if (!machdep->ptrs_per_pgd) {
|
|
if (machdep->machspec->VA_BITS == 52)
|
|
machdep->ptrs_per_pgd = PTRS_PER_PGD_L3_64K_52;
|
|
else if (machdep->machspec->VA_BITS == 48)
|
|
machdep->ptrs_per_pgd = PTRS_PER_PGD_L3_64K_48;
|
|
else
|
|
error(FATAL, "wrong VA_BITS for 64K page.");
|
|
}
|
|
if ((machdep->pgd =
|
|
(char *)malloc(machdep->ptrs_per_pgd * 8)) == NULL)
|
|
error(FATAL, "cannot malloc pgd space.");
|
|
if ((machdep->pmd =
|
|
(char *)malloc(PTRS_PER_PMD_L3_64K * 8)) == NULL)
|
|
error(FATAL, "cannot malloc pmd space.");
|
|
if ((machdep->ptbl =
|
|
(char *)malloc(PTRS_PER_PTE_L3_64K * 8)) == NULL)
|
|
error(FATAL, "cannot malloc ptbl space.");
|
|
} else {
|
|
machdep->flags |= VM_L2_64K;
|
|
if (!machdep->ptrs_per_pgd)
|
|
machdep->ptrs_per_pgd = PTRS_PER_PGD_L2_64K;
|
|
if ((machdep->pgd =
|
|
(char *)malloc(machdep->ptrs_per_pgd * 8)) == NULL)
|
|
error(FATAL, "cannot malloc pgd space.");
|
|
if ((machdep->ptbl =
|
|
(char *)malloc(PTRS_PER_PTE_L2_64K * 8)) == NULL)
|
|
error(FATAL, "cannot malloc ptbl space.");
|
|
machdep->pmd = NULL; /* not used */
|
|
}
|
|
machdep->pud = NULL; /* not used */
|
|
break;
|
|
|
|
default:
|
|
if (machdep->pagesize)
|
|
error(FATAL, "invalid/unsupported page size: %d\n",
|
|
machdep->pagesize);
|
|
else
|
|
error(FATAL, "cannot determine page size\n");
|
|
}
|
|
|
|
machdep->last_pgd_read = 0;
|
|
machdep->last_pud_read = 0;
|
|
machdep->last_pmd_read = 0;
|
|
machdep->last_ptbl_read = 0;
|
|
machdep->clear_machdep_cache = arm64_clear_machdep_cache;
|
|
|
|
machdep->stacksize = ARM64_STACK_SIZE;
|
|
machdep->flags |= VMEMMAP;
|
|
machdep->uvtop = arm64_uvtop;
|
|
machdep->is_uvaddr = arm64_is_uvaddr;
|
|
machdep->eframe_search = arm64_eframe_search;
|
|
machdep->back_trace = arm64_back_trace_cmd;
|
|
machdep->in_alternate_stack = arm64_in_alternate_stack;
|
|
machdep->processor_speed = arm64_processor_speed;
|
|
machdep->get_task_pgd = arm64_get_task_pgd;
|
|
machdep->get_stack_frame = arm64_get_stack_frame;
|
|
machdep->get_stackbase = generic_get_stackbase;
|
|
machdep->get_stacktop = generic_get_stacktop;
|
|
machdep->translate_pte = arm64_translate_pte;
|
|
machdep->memory_size = generic_memory_size;
|
|
machdep->vmalloc_start = arm64_vmalloc_start;
|
|
machdep->get_kvaddr_ranges = arm64_get_kvaddr_ranges;
|
|
machdep->is_task_addr = arm64_is_task_addr;
|
|
machdep->dis_filter = arm64_dis_filter;
|
|
machdep->cmd_mach = arm64_cmd_mach;
|
|
machdep->get_smp_cpus = arm64_get_smp_cpus;
|
|
machdep->line_number_hooks = NULL;
|
|
machdep->value_to_symbol = generic_machdep_value_to_symbol;
|
|
machdep->dump_irq = generic_dump_irq;
|
|
machdep->show_interrupts = generic_show_interrupts;
|
|
machdep->get_irq_affinity = generic_get_irq_affinity;
|
|
machdep->dumpfile_init = NULL;
|
|
machdep->verify_line_number = NULL;
|
|
machdep->init_kernel_pgd = arm64_init_kernel_pgd;
|
|
machdep->get_current_task_reg = arm64_get_current_task_reg;
|
|
|
|
/* use machdep parameters */
|
|
arm64_calc_phys_offset();
|
|
arm64_calc_physvirt_offset();
|
|
|
|
if (CRASHDEBUG(1)) {
|
|
if (machdep->flags & NEW_VMEMMAP)
|
|
fprintf(fp, "kimage_voffset: %lx\n",
|
|
machdep->machspec->kimage_voffset);
|
|
fprintf(fp, "phys_offset: %lx\n",
|
|
machdep->machspec->phys_offset);
|
|
fprintf(fp, "physvirt_offset: %lx\n", machdep->machspec->physvirt_offset);
|
|
}
|
|
|
|
break;
|
|
|
|
case POST_GDB:
|
|
/* Rely on kernel version to decide the kernel start address */
|
|
arm64_calc_kernel_start();
|
|
|
|
/* Can we get the size of struct page before POST_GDB */
|
|
ms = machdep->machspec;
|
|
if (!ms->struct_page_size)
|
|
arm64_calc_virtual_memory_ranges();
|
|
|
|
arm64_get_vmemmap_page_ptr();
|
|
arm64_get_section_size_bits();
|
|
|
|
if (!machdep->max_physmem_bits) {
|
|
if (arm64_get_vmcoreinfo(&machdep->max_physmem_bits, "NUMBER(MAX_PHYSMEM_BITS)", NUM_DEC)) {
|
|
/* nothing */
|
|
} else if (machdep->machspec->VA_BITS == 52) /* guess */
|
|
machdep->max_physmem_bits = _MAX_PHYSMEM_BITS_52;
|
|
else if (THIS_KERNEL_VERSION >= LINUX(3,17,0))
|
|
machdep->max_physmem_bits = _MAX_PHYSMEM_BITS_3_17;
|
|
else
|
|
machdep->max_physmem_bits = _MAX_PHYSMEM_BITS;
|
|
}
|
|
|
|
if (CRASHDEBUG(1)) {
|
|
if (ms->VA_BITS_ACTUAL) {
|
|
fprintf(fp, "CONFIG_ARM64_VA_BITS: %ld\n", ms->CONFIG_ARM64_VA_BITS);
|
|
fprintf(fp, " VA_BITS_ACTUAL: %ld\n", ms->VA_BITS_ACTUAL);
|
|
fprintf(fp, "(calculated) VA_BITS: %ld\n", ms->VA_BITS);
|
|
fprintf(fp, " PAGE_OFFSET: %lx\n", ARM64_FLIP_PAGE_OFFSET);
|
|
fprintf(fp, " VA_START: %lx\n", ms->VA_START);
|
|
fprintf(fp, " modules: %lx - %lx\n", ms->modules_vaddr, ms->modules_end);
|
|
fprintf(fp, " vmalloc: %lx - %lx\n", ms->vmalloc_start_addr, ms->vmalloc_end);
|
|
fprintf(fp, "kernel image: %lx - %lx\n", ms->kimage_text, ms->kimage_end);
|
|
fprintf(fp, " vmemmap: %lx - %lx\n\n", ms->vmemmap_vaddr, ms->vmemmap_end);
|
|
}
|
|
}
|
|
|
|
if (THIS_KERNEL_VERSION >= LINUX(5,19,0)) {
|
|
ms->__SWP_TYPE_BITS = 5;
|
|
ms->__SWP_TYPE_SHIFT = 3;
|
|
ms->__SWP_TYPE_MASK = ((1UL << ms->__SWP_TYPE_BITS) - 1);
|
|
ms->__SWP_OFFSET_SHIFT = (ms->__SWP_TYPE_BITS + ms->__SWP_TYPE_SHIFT);
|
|
ms->__SWP_OFFSET_BITS = 50;
|
|
ms->__SWP_OFFSET_MASK = ((1UL << ms->__SWP_OFFSET_BITS) - 1);
|
|
ms->PTE_PROT_NONE = (1UL << 58);
|
|
ms->PTE_FILE = 0; /* unused */
|
|
} else if (THIS_KERNEL_VERSION >= LINUX(4,0,0)) {
|
|
ms->__SWP_TYPE_BITS = 6;
|
|
ms->__SWP_TYPE_SHIFT = 2;
|
|
ms->__SWP_TYPE_MASK = ((1UL << ms->__SWP_TYPE_BITS) - 1);
|
|
ms->__SWP_OFFSET_SHIFT = (ms->__SWP_TYPE_BITS + ms->__SWP_TYPE_SHIFT);
|
|
ms->__SWP_OFFSET_BITS = 50;
|
|
ms->__SWP_OFFSET_MASK = ((1UL << ms->__SWP_OFFSET_BITS) - 1);
|
|
ms->PTE_PROT_NONE = (1UL << 58);
|
|
ms->PTE_FILE = 0; /* unused */
|
|
} else if (THIS_KERNEL_VERSION >= LINUX(3,13,0)) {
|
|
ms->__SWP_TYPE_BITS = 6;
|
|
ms->__SWP_TYPE_SHIFT = 3;
|
|
ms->__SWP_TYPE_MASK = ((1UL << ms->__SWP_TYPE_BITS) - 1);
|
|
ms->__SWP_OFFSET_SHIFT = (ms->__SWP_TYPE_BITS + ms->__SWP_TYPE_SHIFT);
|
|
ms->__SWP_OFFSET_BITS = 49;
|
|
ms->__SWP_OFFSET_MASK = ((1UL << ms->__SWP_OFFSET_BITS) - 1);
|
|
ms->PTE_PROT_NONE = (1UL << 58);
|
|
ms->PTE_FILE = (1UL << 2);
|
|
} else if (THIS_KERNEL_VERSION >= LINUX(3,11,0)) {
|
|
ms->__SWP_TYPE_BITS = 6;
|
|
ms->__SWP_TYPE_SHIFT = 4;
|
|
ms->__SWP_TYPE_MASK = ((1UL << ms->__SWP_TYPE_BITS) - 1);
|
|
ms->__SWP_OFFSET_SHIFT = (ms->__SWP_TYPE_BITS + ms->__SWP_TYPE_SHIFT);
|
|
ms->__SWP_OFFSET_BITS = 0; /* unused */
|
|
ms->__SWP_OFFSET_MASK = 0; /* unused */
|
|
ms->PTE_PROT_NONE = (1UL << 2);
|
|
ms->PTE_FILE = (1UL << 3);
|
|
} else {
|
|
ms->__SWP_TYPE_BITS = 6;
|
|
ms->__SWP_TYPE_SHIFT = 3;
|
|
ms->__SWP_TYPE_MASK = ((1UL << ms->__SWP_TYPE_BITS) - 1);
|
|
ms->__SWP_OFFSET_SHIFT = (ms->__SWP_TYPE_BITS + ms->__SWP_TYPE_SHIFT);
|
|
ms->__SWP_OFFSET_BITS = 0; /* unused */
|
|
ms->__SWP_OFFSET_MASK = 0; /* unused */
|
|
ms->PTE_PROT_NONE = (1UL << 1);
|
|
ms->PTE_FILE = (1UL << 2);
|
|
}
|
|
|
|
if (symbol_exists("irq_desc"))
|
|
ARRAY_LENGTH_INIT(machdep->nr_irqs, irq_desc,
|
|
"irq_desc", NULL, 0);
|
|
else if (kernel_symbol_exists("nr_irqs"))
|
|
get_symbol_data("nr_irqs", sizeof(unsigned int),
|
|
&machdep->nr_irqs);
|
|
|
|
if (!machdep->hz)
|
|
machdep->hz = 100;
|
|
|
|
|
|
/*
|
|
* Let's calculate the KERNELPACMASK value based on the
|
|
* vabits, see:
|
|
* arch/arm64/kernel/vmcore_info.c
|
|
* arch/arm64/include/asm/pointer_auth.h
|
|
*/
|
|
if(!machdep->machspec->CONFIG_ARM64_KERNELPACMASK)
|
|
arm64_recalc_KERNELPACMASK();
|
|
|
|
arm64_irq_stack_init();
|
|
arm64_overflow_stack_init();
|
|
arm64_stackframe_init();
|
|
break;
|
|
|
|
case POST_INIT:
|
|
/*
|
|
* crash_notes contains machine specific information about the
|
|
* crash. In particular, it contains CPU registers at the time
|
|
* of the crash. We need this information to extract correct
|
|
* backtraces from the panic task.
|
|
*/
|
|
if (!LIVE())
|
|
arm64_get_crash_notes();
|
|
gdb_change_thread_context();
|
|
break;
|
|
|
|
case LOG_ONLY:
|
|
machdep->machspec = &arm64_machine_specific;
|
|
arm64_calc_VA_BITS();
|
|
arm64_calc_KERNELPACMASK();
|
|
arm64_calc_phys_offset();
|
|
machdep->machspec->page_offset = ARM64_PAGE_OFFSET;
|
|
arm64_calc_physvirt_offset();
|
|
break;
|
|
}
|
|
}
|
|
|
|
struct kernel_va_range_handler {
|
|
unsigned long kernel_versions_start; /* include */
|
|
unsigned long kernel_versions_end; /* exclude */
|
|
struct kernel_range *(*get_range)(struct machine_specific *);
|
|
};
|
|
|
|
static struct kernel_range tmp_range;
|
|
#define _PAGE_END(va) (-(1UL << ((va) - 1)))
|
|
#define SZ_64K 0x00010000
|
|
#define SZ_2M 0x00200000
|
|
|
|
/*
|
|
* Get the max shift of the size of struct page.
|
|
* Most of the time, it is 64 bytes, but not sure.
|
|
*/
|
|
static int arm64_get_struct_page_max_shift(struct machine_specific *ms)
|
|
{
|
|
return (int)ceil(log2(ms->struct_page_size));
|
|
}
|
|
|
|
/* Return TRUE if we succeed, return FALSE on failure. */
|
|
static int
|
|
arm64_get_vmcoreinfo(unsigned long *vaddr, const char *label, int base)
|
|
{
|
|
int err = 0;
|
|
char *string = pc->read_vmcoreinfo(label);
|
|
|
|
if (!string)
|
|
return FALSE;
|
|
|
|
switch (base) {
|
|
case NUM_HEX:
|
|
*vaddr = strtoul(string, NULL, 16);
|
|
break;
|
|
case NUM_DEC:
|
|
*vaddr = strtoul(string, NULL, 10);
|
|
break;
|
|
default:
|
|
err++;
|
|
error(INFO, "Unknown type:%#x, (NUM_HEX|NUM_DEC)\n", base);
|
|
}
|
|
free(string);
|
|
return err ? FALSE: TRUE;
|
|
}
|
|
|
|
/*
|
|
* The change is caused by the kernel patch since v5.18-rc1:
|
|
* "arm64: crash_core: Export MODULES, VMALLOC, and VMEMMAP ranges"
|
|
*/
|
|
static struct kernel_range *arm64_get_range_v5_18(struct machine_specific *ms)
|
|
{
|
|
struct kernel_range *r = &tmp_range;
|
|
|
|
/* Get the MODULES_VADDR ~ MODULES_END */
|
|
if (!arm64_get_vmcoreinfo(&r->modules_vaddr, "NUMBER(MODULES_VADDR)", NUM_HEX))
|
|
return NULL;
|
|
if (!arm64_get_vmcoreinfo(&r->modules_end, "NUMBER(MODULES_END)", NUM_HEX))
|
|
return NULL;
|
|
|
|
/* Get the VMEMMAP_START ~ VMEMMAP_END */
|
|
if (!arm64_get_vmcoreinfo(&r->vmemmap_vaddr, "NUMBER(VMEMMAP_START)", NUM_HEX))
|
|
return NULL;
|
|
if (!arm64_get_vmcoreinfo(&r->vmemmap_end, "NUMBER(VMEMMAP_END)", NUM_HEX))
|
|
return NULL;
|
|
|
|
/* Get the VMALLOC_START ~ VMALLOC_END */
|
|
if (!arm64_get_vmcoreinfo(&r->vmalloc_start_addr, "NUMBER(VMALLOC_START)", NUM_HEX))
|
|
return NULL;
|
|
if (!arm64_get_vmcoreinfo(&r->vmalloc_end, "NUMBER(VMALLOC_END)", NUM_HEX))
|
|
return NULL;
|
|
|
|
return r;
|
|
}
|
|
|
|
/*
|
|
* The change is caused by the kernel patch since v5.17-rc1:
|
|
* "b89ddf4cca43 arm64/bpf: Remove 128MB limit for BPF JIT programs"
|
|
*/
|
|
static struct kernel_range *arm64_get_range_v5_17(struct machine_specific *ms)
|
|
{
|
|
struct kernel_range *r = &tmp_range;
|
|
unsigned long v = ms->CONFIG_ARM64_VA_BITS;
|
|
unsigned long vmem_shift, vmemmap_size;
|
|
|
|
/* Not initialized yet */
|
|
if (v == 0)
|
|
return NULL;
|
|
|
|
if (v > 48)
|
|
v = 48;
|
|
|
|
/* Get the MODULES_VADDR ~ MODULES_END */
|
|
r->modules_vaddr = _PAGE_END(v);
|
|
r->modules_end = r->modules_vaddr + MEGABYTES(128);
|
|
|
|
/* Get the VMEMMAP_START ~ VMEMMAP_END */
|
|
vmem_shift = machdep->pageshift - arm64_get_struct_page_max_shift(ms);
|
|
vmemmap_size = (_PAGE_END(v) - PAGE_OFFSET) >> vmem_shift;
|
|
|
|
r->vmemmap_vaddr = (-(1UL << (ms->CONFIG_ARM64_VA_BITS - vmem_shift)));
|
|
r->vmemmap_end = r->vmemmap_vaddr + vmemmap_size;
|
|
|
|
/* Get the VMALLOC_START ~ VMALLOC_END */
|
|
r->vmalloc_start_addr = r->modules_end;
|
|
r->vmalloc_end = r->vmemmap_vaddr - MEGABYTES(256);
|
|
return r;
|
|
}
|
|
|
|
/*
|
|
* The change is caused by the kernel patch since v5.11:
|
|
* "9ad7c6d5e75b arm64: mm: tidy up top of kernel VA space"
|
|
*/
|
|
static struct kernel_range *arm64_get_range_v5_11(struct machine_specific *ms)
|
|
{
|
|
struct kernel_range *r = &tmp_range;
|
|
unsigned long v = ms->CONFIG_ARM64_VA_BITS;
|
|
unsigned long vmem_shift, vmemmap_size, bpf_jit_size = MEGABYTES(128);
|
|
|
|
/* Not initialized yet */
|
|
if (v == 0)
|
|
return NULL;
|
|
|
|
if (v > 48)
|
|
v = 48;
|
|
|
|
/* Get the MODULES_VADDR ~ MODULES_END */
|
|
r->modules_vaddr = _PAGE_END(v) + bpf_jit_size;
|
|
r->modules_end = r->modules_vaddr + MEGABYTES(128);
|
|
|
|
/* Get the VMEMMAP_START ~ VMEMMAP_END */
|
|
vmem_shift = machdep->pageshift - arm64_get_struct_page_max_shift(ms);
|
|
vmemmap_size = (_PAGE_END(v) - PAGE_OFFSET) >> vmem_shift;
|
|
|
|
r->vmemmap_vaddr = (-(1UL << (ms->CONFIG_ARM64_VA_BITS - vmem_shift)));
|
|
r->vmemmap_end = r->vmemmap_vaddr + vmemmap_size;
|
|
|
|
/* Get the VMALLOC_START ~ VMALLOC_END */
|
|
r->vmalloc_start_addr = r->modules_end;
|
|
r->vmalloc_end = r->vmemmap_vaddr - MEGABYTES(256);
|
|
return r;
|
|
}
|
|
|
|
static unsigned long arm64_get_pud_size(void)
|
|
{
|
|
unsigned long PUD_SIZE = 0;
|
|
|
|
switch (machdep->pagesize) {
|
|
case 4096:
|
|
if (machdep->machspec->VA_BITS > PGDIR_SHIFT_L4_4K) {
|
|
PUD_SIZE = PUD_SIZE_L4_4K;
|
|
} else {
|
|
PUD_SIZE = PGDIR_SIZE_L3_4K;
|
|
}
|
|
break;
|
|
|
|
case 65536:
|
|
PUD_SIZE = PGDIR_SIZE_L2_64K;
|
|
default:
|
|
break;
|
|
}
|
|
return PUD_SIZE;
|
|
}
|
|
|
|
/*
|
|
* The change is caused by the kernel patches since v5.4, such as:
|
|
* "ce3aaed87344 arm64: mm: Modify calculation of VMEMMAP_SIZE"
|
|
* "14c127c957c1 arm64: mm: Flip kernel VA space"
|
|
*/
|
|
static struct kernel_range *arm64_get_range_v5_4(struct machine_specific *ms)
|
|
{
|
|
struct kernel_range *r = &tmp_range;
|
|
unsigned long v = ms->CONFIG_ARM64_VA_BITS;
|
|
unsigned long kasan_shadow_shift, kasan_shadow_offset, PUD_SIZE;
|
|
unsigned long vmem_shift, vmemmap_size, bpf_jit_size = MEGABYTES(128);
|
|
char *string;
|
|
int ret;
|
|
|
|
/* Not initialized yet */
|
|
if (v == 0)
|
|
return NULL;
|
|
|
|
if (v > 48)
|
|
v = 48;
|
|
|
|
/* Get the MODULES_VADDR ~ MODULES_END */
|
|
if (kernel_symbol_exists("kasan_init")) {
|
|
/* See the arch/arm64/Makefile */
|
|
ret = get_kernel_config("CONFIG_KASAN_SW_TAGS", NULL);
|
|
if (ret == IKCONFIG_N)
|
|
return NULL;
|
|
kasan_shadow_shift = (ret == IKCONFIG_Y) ? 4: 3;
|
|
|
|
/* See the arch/arm64/Kconfig*/
|
|
ret = get_kernel_config("CONFIG_KASAN_SHADOW_OFFSET", &string);
|
|
if (ret != IKCONFIG_STR)
|
|
return NULL;
|
|
kasan_shadow_offset = atol(string);
|
|
|
|
r->modules_vaddr = (1UL << (64 - kasan_shadow_shift)) + kasan_shadow_offset
|
|
+ bpf_jit_size;
|
|
} else {
|
|
r->modules_vaddr = _PAGE_END(v) + bpf_jit_size;
|
|
}
|
|
|
|
r->modules_end = r->modules_vaddr + MEGABYTES(128);
|
|
|
|
/* Get the VMEMMAP_START ~ VMEMMAP_END */
|
|
vmem_shift = machdep->pageshift - arm64_get_struct_page_max_shift(ms);
|
|
vmemmap_size = (_PAGE_END(v) - PAGE_OFFSET) >> vmem_shift;
|
|
|
|
r->vmemmap_vaddr = (-vmemmap_size - SZ_2M);
|
|
/*
|
|
* In the v5.7, the patch: "bbd6ec605c arm64/mm: Enable memory hot remove"
|
|
* adds the VMEMMAP_END.
|
|
*
|
|
* But before the VMEMMAP_END was added to kernel, we can also see
|
|
* the following in arch/arm64/mm/dump.c:
|
|
* { VMEMMAP_START + VMEMMAP_SIZE, "vmemmap end" },
|
|
*/
|
|
r->vmemmap_end = r->vmemmap_vaddr + vmemmap_size;
|
|
|
|
/* Get the VMALLOC_START ~ VMALLOC_END */
|
|
PUD_SIZE = arm64_get_pud_size();
|
|
r->vmalloc_start_addr = r->modules_end;
|
|
r->vmalloc_end = (-PUD_SIZE - vmemmap_size - SZ_64K);
|
|
return r;
|
|
}
|
|
|
|
/*
|
|
* The change is caused by the kernel patches since v5.0, such as:
|
|
* "91fc957c9b1d arm64/bpf: don't allocate BPF JIT programs in module memory"
|
|
*/
|
|
static struct kernel_range *arm64_get_range_v5_0(struct machine_specific *ms)
|
|
{
|
|
struct kernel_range *r = &tmp_range;
|
|
unsigned long v = ms->CONFIG_ARM64_VA_BITS;
|
|
unsigned long kasan_shadow_shift, PUD_SIZE;
|
|
unsigned long vmemmap_size, bpf_jit_size = MEGABYTES(128);
|
|
unsigned long va_start, page_offset;
|
|
int ret;
|
|
|
|
/* Not initialized yet */
|
|
if (v == 0)
|
|
return NULL;
|
|
|
|
va_start = (0xffffffffffffffffUL - (1UL << v) + 1);
|
|
page_offset = (0xffffffffffffffffUL - (1UL << (v - 1)) + 1);
|
|
|
|
/* Get the MODULES_VADDR ~ MODULES_END */
|
|
if (kernel_symbol_exists("kasan_init")) {
|
|
/* See the arch/arm64/Makefile */
|
|
ret = get_kernel_config("CONFIG_KASAN_SW_TAGS", NULL);
|
|
if (ret == IKCONFIG_N)
|
|
return NULL;
|
|
kasan_shadow_shift = (ret == IKCONFIG_Y) ? 4: 3;
|
|
|
|
r->modules_vaddr = va_start + (1UL << (v - kasan_shadow_shift)) + bpf_jit_size;
|
|
} else {
|
|
r->modules_vaddr = va_start + bpf_jit_size;
|
|
}
|
|
|
|
r->modules_end = r->modules_vaddr + MEGABYTES(128);
|
|
|
|
/* Get the VMEMMAP_START ~ VMEMMAP_END */
|
|
vmemmap_size = (1UL << (v - machdep->pageshift - 1 + arm64_get_struct_page_max_shift(ms)));
|
|
|
|
r->vmemmap_vaddr = page_offset - vmemmap_size;
|
|
r->vmemmap_end = r->vmemmap_vaddr + vmemmap_size; /* See the arch/arm64/mm/dump.c */
|
|
|
|
/* Get the VMALLOC_START ~ VMALLOC_END */
|
|
PUD_SIZE = arm64_get_pud_size();
|
|
|
|
r->vmalloc_start_addr = r->modules_end;
|
|
r->vmalloc_end = page_offset - PUD_SIZE - vmemmap_size - SZ_64K;
|
|
return r;
|
|
}
|
|
|
|
static struct kernel_va_range_handler kernel_va_range_handlers[] = {
|
|
{
|
|
LINUX(5,18,0),
|
|
LINUX(999,0,0), /* Just a boundary */
|
|
get_range: arm64_get_range_v5_18,
|
|
}, {
|
|
LINUX(5,17,0), LINUX(5,18,0),
|
|
get_range: arm64_get_range_v5_17,
|
|
}, {
|
|
LINUX(5,11,0), LINUX(5,17,0),
|
|
get_range: arm64_get_range_v5_11,
|
|
}, {
|
|
LINUX(5,4,0), LINUX(5,11,0),
|
|
get_range: arm64_get_range_v5_4,
|
|
}, {
|
|
LINUX(5,0,0), LINUX(5,4,0),
|
|
get_range: arm64_get_range_v5_0,
|
|
},
|
|
};
|
|
|
|
#define ARRAY_SIZE(a) (sizeof (a) / sizeof ((a)[0]))
|
|
|
|
static unsigned long arm64_get_kernel_version(void)
|
|
{
|
|
char *string;
|
|
|
|
if (THIS_KERNEL_VERSION)
|
|
return THIS_KERNEL_VERSION;
|
|
|
|
if ((string = pc->read_vmcoreinfo("OSRELEASE"))) {
|
|
parse_kernel_version(string);
|
|
free(string);
|
|
}
|
|
return THIS_KERNEL_VERSION;
|
|
}
|
|
|
|
/* Return NULL if we fail. */
|
|
static struct kernel_range *arm64_get_va_range(struct machine_specific *ms)
|
|
{
|
|
struct kernel_va_range_handler *h;
|
|
unsigned long kernel_version = arm64_get_kernel_version();
|
|
struct kernel_range *r = NULL;
|
|
int i;
|
|
|
|
if (!kernel_version)
|
|
goto range_failed;
|
|
|
|
for (i = 0; i < ARRAY_SIZE(kernel_va_range_handlers); i++) {
|
|
h = kernel_va_range_handlers + i;
|
|
|
|
/* Get the right hook for this kernel version */
|
|
if (h->kernel_versions_start <= kernel_version &&
|
|
kernel_version < h->kernel_versions_end) {
|
|
|
|
/* Get the correct virtual address ranges */
|
|
r = h->get_range(ms);
|
|
if (!r)
|
|
goto range_failed;
|
|
return r;
|
|
}
|
|
}
|
|
|
|
range_failed:
|
|
/* Reset ms->struct_page_size to 0 for arm64_calc_virtual_memory_ranges() */
|
|
ms->struct_page_size = 0;
|
|
return NULL;
|
|
}
|
|
|
|
/* Get the size of struct page {} */
|
|
static void arm64_get_struct_page_size(struct machine_specific *ms)
|
|
{
|
|
arm64_get_vmcoreinfo(&ms->struct_page_size, "SIZE(page)", NUM_DEC);
|
|
}
|
|
|
|
/*
|
|
* Accept or reject a symbol from the kernel namelist.
|
|
*/
|
|
static int
|
|
arm64_verify_symbol(const char *name, ulong value, char type)
|
|
{
|
|
if (!name || !strlen(name))
|
|
return FALSE;
|
|
|
|
if ((type == 'A') && STREQ(name, "_kernel_flags_le"))
|
|
machdep->machspec->kernel_flags = le64toh(value);
|
|
|
|
if ((type == 'A') && STREQ(name, "_kernel_flags_le_hi32"))
|
|
machdep->machspec->kernel_flags |= ((ulong)le32toh(value) << 32);
|
|
|
|
if ((type == 'A') && STREQ(name, "_kernel_flags_le_lo32"))
|
|
machdep->machspec->kernel_flags |= le32toh(value);
|
|
|
|
if (((type == 'A') || (type == 'a')) && (highest_bit_long(value) != 63))
|
|
return FALSE;
|
|
|
|
if ((value == 0) &&
|
|
((type == 'a') || (type == 'n') || (type == 'N') || (type == 'U')))
|
|
return FALSE;
|
|
|
|
if (STREQ(name, "$d") || STRNEQ(name, "$d.") ||
|
|
STREQ(name, "$x") || STRNEQ(name, "$x.") ||
|
|
STREQ(name, "$c") || STRNEQ(name, "$c."))
|
|
return FALSE;
|
|
|
|
if ((type == 'A') && STRNEQ(name, "__crc_"))
|
|
return FALSE;
|
|
|
|
if ((type == 'N') && strstr(name, "$d"))
|
|
return FALSE;
|
|
|
|
if (!(machdep->flags & KSYMS_START) && STREQ(name, "idmap_pg_dir"))
|
|
machdep->flags |= KSYMS_START;
|
|
|
|
return TRUE;
|
|
}
|
|
|
|
|
|
void
|
|
arm64_dump_machdep_table(ulong arg)
|
|
{
|
|
const struct machine_specific *ms = machdep->machspec;
|
|
int others, i;
|
|
|
|
others = 0;
|
|
fprintf(fp, " flags: %lx (", machdep->flags);
|
|
if (machdep->flags & KSYMS_START)
|
|
fprintf(fp, "%sKSYMS_START", others++ ? "|" : "");
|
|
if (machdep->flags & PHYS_OFFSET)
|
|
fprintf(fp, "%sPHYS_OFFSET", others++ ? "|" : "");
|
|
if (machdep->flags & VM_L2_64K)
|
|
fprintf(fp, "%sVM_L2_64K", others++ ? "|" : "");
|
|
if (machdep->flags & VM_L3_64K)
|
|
fprintf(fp, "%sVM_L3_64K", others++ ? "|" : "");
|
|
if (machdep->flags & VM_L2_16K)
|
|
fprintf(fp, "%sVM_L2_16K", others++ ? "|" : "");
|
|
if (machdep->flags & VM_L3_16K)
|
|
fprintf(fp, "%sVM_L3_16K", others++ ? "|" : "");
|
|
if (machdep->flags & VM_L4_16K)
|
|
fprintf(fp, "%sVM_L4_16K", others++ ? "|" : "");
|
|
if (machdep->flags & VM_L3_4K)
|
|
fprintf(fp, "%sVM_L3_4K", others++ ? "|" : "");
|
|
if (machdep->flags & VM_L4_4K)
|
|
fprintf(fp, "%sVM_L4_4K", others++ ? "|" : "");
|
|
if (machdep->flags & VMEMMAP)
|
|
fprintf(fp, "%sVMEMMAP", others++ ? "|" : "");
|
|
if (machdep->flags & KDUMP_ENABLED)
|
|
fprintf(fp, "%sKDUMP_ENABLED", others++ ? "|" : "");
|
|
if (machdep->flags & IRQ_STACKS)
|
|
fprintf(fp, "%sIRQ_STACKS", others++ ? "|" : "");
|
|
if (machdep->flags & UNW_4_14)
|
|
fprintf(fp, "%sUNW_4_14", others++ ? "|" : "");
|
|
if (machdep->flags & MACHDEP_BT_TEXT)
|
|
fprintf(fp, "%sMACHDEP_BT_TEXT", others++ ? "|" : "");
|
|
if (machdep->flags & NEW_VMEMMAP)
|
|
fprintf(fp, "%sNEW_VMEMMAP", others++ ? "|" : "");
|
|
if (machdep->flags & FLIPPED_VM)
|
|
fprintf(fp, "%sFLIPPED_VM", others++ ? "|" : "");
|
|
if (machdep->flags & HAS_PHYSVIRT_OFFSET)
|
|
fprintf(fp, "%sHAS_PHYSVIRT_OFFSET", others++ ? "|" : "");
|
|
if (machdep->flags & ARM64_MTE)
|
|
fprintf(fp, "%sARM64_MTE", others++ ? "|" : "");
|
|
fprintf(fp, ")\n");
|
|
|
|
fprintf(fp, " kvbase: %lx\n", machdep->kvbase);
|
|
fprintf(fp, " identity_map_base: %lx\n", machdep->identity_map_base);
|
|
fprintf(fp, " pagesize: %d\n", machdep->pagesize);
|
|
fprintf(fp, " pageshift: %d\n", machdep->pageshift);
|
|
fprintf(fp, " pagemask: %lx\n", (ulong)machdep->pagemask);
|
|
fprintf(fp, " pageoffset: %lx\n", machdep->pageoffset);
|
|
fprintf(fp, " stacksize: %ld\n", machdep->stacksize);
|
|
fprintf(fp, " hz: %d\n", machdep->hz);
|
|
fprintf(fp, " mhz: %ld\n", machdep->mhz);
|
|
fprintf(fp, " memsize: %lld (0x%llx)\n",
|
|
(ulonglong)machdep->memsize, (ulonglong)machdep->memsize);
|
|
fprintf(fp, " bits: %d\n", machdep->bits);
|
|
fprintf(fp, " nr_irqs: %d\n", machdep->nr_irqs);
|
|
fprintf(fp, " eframe_search: arm64_eframe_search()\n");
|
|
fprintf(fp, " back_trace: arm64_back_trace_cmd() (default: %s method)\n",
|
|
kt->flags & USE_OPT_BT ? "optional" : "original");
|
|
fprintf(fp, " in_alternate_stack: arm64_in_alternate_stack()\n");
|
|
fprintf(fp, " processor_speed: arm64_processor_speed()\n");
|
|
fprintf(fp, " uvtop: arm64_uvtop()->%s()\n",
|
|
machdep->flags & VM_L3_4K ?
|
|
"arm64_vtop_3level_4k" :
|
|
machdep->flags & VM_L4_4K ?
|
|
"arm64_vtop_4level_4k" :
|
|
machdep->flags & VM_L2_16K ?
|
|
"arm64_vtop_2level_16k" :
|
|
machdep->flags & VM_L3_16K ?
|
|
"arm64_vtop_3level_16k" :
|
|
machdep->flags & VM_L4_16K ?
|
|
"arm64_vtop_4level_16k" :
|
|
machdep->flags & VM_L3_64K ?
|
|
"arm64_vtop_3level_64k" : "arm64_vtop_2level_64k");
|
|
fprintf(fp, " kvtop: arm64_kvtop()->%s()\n",
|
|
machdep->flags & VM_L3_4K ?
|
|
"arm64_vtop_3level_4k" :
|
|
machdep->flags & VM_L4_4K ?
|
|
"arm64_vtop_4level_4k" :
|
|
machdep->flags & VM_L2_16K ?
|
|
"arm64_vtop_2level_16k" :
|
|
machdep->flags & VM_L3_16K ?
|
|
"arm64_vtop_3level_16k" :
|
|
machdep->flags & VM_L4_16K ?
|
|
"arm64_vtop_4level_16k" :
|
|
machdep->flags & VM_L3_64K ?
|
|
"arm64_vtop_3level_64k" : "arm64_vtop_2level_64k");
|
|
fprintf(fp, " get_task_pgd: arm64_get_task_pgd()\n");
|
|
fprintf(fp, " dump_irq: generic_dump_irq()\n");
|
|
fprintf(fp, " get_stack_frame: arm64_get_stack_frame()\n");
|
|
fprintf(fp, " get_stackbase: generic_get_stackbase()\n");
|
|
fprintf(fp, " get_stacktop: generic_get_stacktop()\n");
|
|
fprintf(fp, " translate_pte: arm64_translate_pte()\n");
|
|
fprintf(fp, " memory_size: generic_memory_size()\n");
|
|
fprintf(fp, " vmalloc_start: arm64_vmalloc_start()\n");
|
|
fprintf(fp, " get_kvaddr_ranges: arm64_get_kvaddr_ranges()\n");
|
|
fprintf(fp, " is_task_addr: arm64_is_task_addr()\n");
|
|
fprintf(fp, " verify_symbol: arm64_verify_symbol()\n");
|
|
fprintf(fp, " dis_filter: arm64_dis_filter()\n");
|
|
fprintf(fp, " cmd_mach: arm64_cmd_mach()\n");
|
|
fprintf(fp, " get_smp_cpus: arm64_get_smp_cpus()\n");
|
|
fprintf(fp, " is_kvaddr: arm64_is_kvaddr()\n");
|
|
fprintf(fp, " is_uvaddr: arm64_is_uvaddr()\n");
|
|
fprintf(fp, " value_to_symbol: generic_machdep_value_to_symbol()\n");
|
|
fprintf(fp, " init_kernel_pgd: arm64_init_kernel_pgd\n");
|
|
fprintf(fp, " verify_paddr: generic_verify_paddr()\n");
|
|
fprintf(fp, " show_interrupts: generic_show_interrupts()\n");
|
|
fprintf(fp, " get_irq_affinity: generic_get_irq_affinity()\n");
|
|
fprintf(fp, " dumpfile_init: (not used)\n");
|
|
fprintf(fp, " process_elf_notes: process_elf64_notes()\n");
|
|
fprintf(fp, " verify_line_number: (not used)\n");
|
|
|
|
fprintf(fp, " xendump_p2m_create: (n/a)\n");
|
|
fprintf(fp, "xen_kdump_p2m_create: (n/a)\n");
|
|
fprintf(fp, " xendump_panic_task: (n/a)\n");
|
|
fprintf(fp, " get_xendump_regs: (n/a)\n");
|
|
fprintf(fp, " line_number_hooks: (not used)\n");
|
|
fprintf(fp, " last_pgd_read: %lx\n", machdep->last_pgd_read);
|
|
fprintf(fp, " last_pud_read: ");
|
|
if ((!(machdep->flags & VM_L4_4K)) && (!(machdep->flags & VM_L4_16K)))
|
|
fprintf(fp, "(not used)\n");
|
|
else
|
|
fprintf(fp, "%lx\n", machdep->last_pud_read);
|
|
fprintf(fp, " last_pmd_read: ");
|
|
if ((machdep->flags & VM_L2_64K) || (machdep->flags & VM_L2_16K))
|
|
fprintf(fp, "(not used)\n");
|
|
else
|
|
fprintf(fp, "%lx\n", machdep->last_pmd_read);
|
|
fprintf(fp, " last_ptbl_read: %lx\n", machdep->last_ptbl_read);
|
|
fprintf(fp, " clear_machdep_cache: arm64_clear_machdep_cache()\n");
|
|
fprintf(fp, " pgd: %lx\n", (ulong)machdep->pgd);
|
|
fprintf(fp, " pud: %lx\n", (ulong)machdep->pud);
|
|
fprintf(fp, " pmd: %lx\n", (ulong)machdep->pmd);
|
|
fprintf(fp, " ptbl: %lx\n", (ulong)machdep->ptbl);
|
|
fprintf(fp, " ptrs_per_pgd: %d\n", machdep->ptrs_per_pgd);
|
|
fprintf(fp, " section_size_bits: %ld\n", machdep->section_size_bits);
|
|
fprintf(fp, " max_physmem_bits: %ld\n", machdep->max_physmem_bits);
|
|
fprintf(fp, " sections_per_root: %ld\n", machdep->sections_per_root);
|
|
|
|
for (i = 0; i < MAX_MACHDEP_ARGS; i++) {
|
|
fprintf(fp, " cmdline_args[%d]: %s\n",
|
|
i, machdep->cmdline_args[i] ?
|
|
machdep->cmdline_args[i] : "(unused)");
|
|
}
|
|
|
|
fprintf(fp, " machspec: %lx\n", (ulong)ms);
|
|
fprintf(fp, " struct_page_size: %ld\n", ms->struct_page_size);
|
|
fprintf(fp, " VA_BITS: %ld\n", ms->VA_BITS);
|
|
fprintf(fp, " CONFIG_ARM64_VA_BITS: %ld\n", ms->CONFIG_ARM64_VA_BITS);
|
|
fprintf(fp, " VA_START: ");
|
|
if (ms->VA_START)
|
|
fprintf(fp, "%lx\n", ms->VA_START);
|
|
else
|
|
fprintf(fp, "(unused)\n");
|
|
fprintf(fp, " VA_BITS_ACTUAL: ");
|
|
if (ms->VA_BITS_ACTUAL)
|
|
fprintf(fp, "%ld\n", ms->VA_BITS_ACTUAL);
|
|
else
|
|
fprintf(fp, "(unused)\n");
|
|
fprintf(fp, "CONFIG_ARM64_KERNELPACMASK: ");
|
|
if (ms->CONFIG_ARM64_KERNELPACMASK)
|
|
fprintf(fp, "%lx\n", ms->CONFIG_ARM64_KERNELPACMASK);
|
|
else
|
|
fprintf(fp, "(unused)\n");
|
|
fprintf(fp, " userspace_top: %016lx\n", ms->userspace_top);
|
|
fprintf(fp, " page_offset: %016lx\n", ms->page_offset);
|
|
fprintf(fp, " vmalloc_start_addr: %016lx\n", ms->vmalloc_start_addr);
|
|
fprintf(fp, " vmalloc_end: %016lx\n", ms->vmalloc_end);
|
|
fprintf(fp, " modules_vaddr: %016lx\n", ms->modules_vaddr);
|
|
fprintf(fp, " modules_end: %016lx\n", ms->modules_end);
|
|
fprintf(fp, " vmemmap_vaddr: %016lx\n", ms->vmemmap_vaddr);
|
|
fprintf(fp, " vmemmap_end: %016lx\n", ms->vmemmap_end);
|
|
if (machdep->flags & NEW_VMEMMAP) {
|
|
fprintf(fp, " vmemmap: %016lx\n", ms->vmemmap);
|
|
fprintf(fp, " kimage_text: %016lx\n", ms->kimage_text);
|
|
fprintf(fp, " kimage_end: %016lx\n", ms->kimage_end);
|
|
fprintf(fp, " kimage_voffset: %016lx\n", ms->kimage_voffset);
|
|
}
|
|
fprintf(fp, " phys_offset: %lx\n", ms->phys_offset);
|
|
fprintf(fp, " physvirt_offset: %lx\n", ms->physvirt_offset);
|
|
fprintf(fp, "__exception_text_start: %lx\n", ms->__exception_text_start);
|
|
fprintf(fp, " __exception_text_end: %lx\n", ms->__exception_text_end);
|
|
fprintf(fp, " __irqentry_text_start: %lx\n", ms->__irqentry_text_start);
|
|
fprintf(fp, " __irqentry_text_end: %lx\n", ms->__irqentry_text_end);
|
|
fprintf(fp, " exp_entry1_start: %lx\n", ms->exp_entry1_start);
|
|
fprintf(fp, " exp_entry1_end: %lx\n", ms->exp_entry1_end);
|
|
fprintf(fp, " exp_entry2_start: %lx\n", ms->exp_entry2_start);
|
|
fprintf(fp, " exp_entry2_end: %lx\n", ms->exp_entry2_end);
|
|
fprintf(fp, " panic_task_regs: %lx\n", (ulong)ms->panic_task_regs);
|
|
fprintf(fp, " user_eframe_offset: %ld\n", ms->user_eframe_offset);
|
|
fprintf(fp, " kern_eframe_offset: %ld\n", ms->kern_eframe_offset);
|
|
fprintf(fp, " PTE_PROT_NONE: %lx\n", ms->PTE_PROT_NONE);
|
|
fprintf(fp, " PTE_FILE: ");
|
|
if (ms->PTE_FILE)
|
|
fprintf(fp, "%lx\n", ms->PTE_FILE);
|
|
else
|
|
fprintf(fp, "(unused)\n");
|
|
fprintf(fp, " __SWP_TYPE_BITS: %ld\n", ms->__SWP_TYPE_BITS);
|
|
fprintf(fp, " __SWP_TYPE_SHIFT: %ld\n", ms->__SWP_TYPE_SHIFT);
|
|
fprintf(fp, " __SWP_TYPE_MASK: %lx\n", ms->__SWP_TYPE_MASK);
|
|
fprintf(fp, " __SWP_OFFSET_BITS: ");
|
|
if (ms->__SWP_OFFSET_BITS)
|
|
fprintf(fp, "%ld\n", ms->__SWP_OFFSET_BITS);
|
|
else
|
|
fprintf(fp, "(unused)\n");
|
|
fprintf(fp, " __SWP_OFFSET_SHIFT: %ld\n", ms->__SWP_OFFSET_SHIFT);
|
|
fprintf(fp, " __SWP_OFFSET_MASK: ");
|
|
if (ms->__SWP_OFFSET_MASK)
|
|
fprintf(fp, "%lx\n", ms->__SWP_OFFSET_MASK);
|
|
else
|
|
fprintf(fp, "(unused)\n");
|
|
fprintf(fp, " machine_kexec_start: %lx\n", ms->machine_kexec_start);
|
|
fprintf(fp, " machine_kexec_end: %lx\n", ms->machine_kexec_end);
|
|
fprintf(fp, " crash_kexec_start: %lx\n", ms->crash_kexec_start);
|
|
fprintf(fp, " crash_kexec_end: %lx\n", ms->crash_kexec_end);
|
|
fprintf(fp, " crash_save_cpu_start: %lx\n", ms->crash_save_cpu_start);
|
|
fprintf(fp, " crash_save_cpu_end: %lx\n", ms->crash_save_cpu_end);
|
|
fprintf(fp, " kernel_flags: %lx\n", ms->kernel_flags);
|
|
fprintf(fp, " irq_stackbuf: %lx\n", (ulong)ms->irq_stackbuf);
|
|
if (machdep->flags & IRQ_STACKS) {
|
|
fprintf(fp, " irq_stack_size: %ld\n", ms->irq_stack_size);
|
|
for (i = 0; i < kt->cpus; i++)
|
|
fprintf(fp, " irq_stacks[%d]: %lx\n",
|
|
i, ms->irq_stacks[i]);
|
|
} else {
|
|
fprintf(fp, " irq_stack_size: (unused)\n");
|
|
fprintf(fp, " irq_stacks: (unused)\n");
|
|
}
|
|
}
|
|
|
|
static int
|
|
arm64_parse_machdep_arg_l(char *argstring, char *param, ulong *value)
|
|
{
|
|
int len;
|
|
int megabytes = FALSE;
|
|
char *p;
|
|
|
|
len = strlen(param);
|
|
if (!STRNEQ(argstring, param) || (argstring[len] != '='))
|
|
return FALSE;
|
|
|
|
if ((LASTCHAR(argstring) == 'm') ||
|
|
(LASTCHAR(argstring) == 'M')) {
|
|
LASTCHAR(argstring) = NULLCHAR;
|
|
megabytes = TRUE;
|
|
}
|
|
|
|
p = argstring + len + 1;
|
|
if (strlen(p)) {
|
|
int flags = RETURN_ON_ERROR | QUIET;
|
|
int err = 0;
|
|
|
|
if (STRNEQ(argstring, "max_physmem_bits")) {
|
|
*value = dtol(p, flags, &err);
|
|
} else if (STRNEQ(argstring, "vabits_actual")) {
|
|
*value = dtol(p, flags, &err);
|
|
} else if (megabytes) {
|
|
*value = dtol(p, flags, &err);
|
|
if (!err)
|
|
*value = MEGABYTES(*value);
|
|
} else {
|
|
*value = htol(p, flags, &err);
|
|
}
|
|
|
|
if (!err)
|
|
return TRUE;
|
|
}
|
|
|
|
return FALSE;
|
|
}
|
|
|
|
/*
|
|
* Parse machine dependent command line arguments.
|
|
*
|
|
* Force the phys_offset address via:
|
|
*
|
|
* --machdep phys_offset=<address>
|
|
*/
|
|
static void
|
|
arm64_parse_cmdline_args(void)
|
|
{
|
|
int index, i, c;
|
|
char *arglist[MAXARGS];
|
|
char buf[BUFSIZE];
|
|
char *p;
|
|
|
|
for (index = 0; index < MAX_MACHDEP_ARGS; index++) {
|
|
if (!machdep->cmdline_args[index])
|
|
break;
|
|
|
|
if (!strstr(machdep->cmdline_args[index], "=")) {
|
|
error(WARNING, "ignoring --machdep option: %x\n",
|
|
machdep->cmdline_args[index]);
|
|
continue;
|
|
}
|
|
|
|
strcpy(buf, machdep->cmdline_args[index]);
|
|
|
|
for (p = buf; *p; p++) {
|
|
if (*p == ',')
|
|
*p = ' ';
|
|
}
|
|
|
|
c = parse_line(buf, arglist);
|
|
|
|
for (i = 0; i < c; i++) {
|
|
if (arm64_parse_machdep_arg_l(arglist[i], "phys_offset",
|
|
&machdep->machspec->phys_offset)) {
|
|
error(NOTE,
|
|
"setting phys_offset to: 0x%lx\n\n",
|
|
machdep->machspec->phys_offset);
|
|
machdep->flags |= PHYS_OFFSET;
|
|
continue;
|
|
} else if (arm64_parse_machdep_arg_l(arglist[i], "kimage_voffset",
|
|
&machdep->machspec->kimage_voffset)) {
|
|
error(NOTE,
|
|
"setting kimage_voffset to: 0x%lx\n\n",
|
|
machdep->machspec->kimage_voffset);
|
|
continue;
|
|
} else if (arm64_parse_machdep_arg_l(arglist[i], "max_physmem_bits",
|
|
&machdep->max_physmem_bits)) {
|
|
error(NOTE,
|
|
"setting max_physmem_bits to: %ld\n\n",
|
|
machdep->max_physmem_bits);
|
|
continue;
|
|
} else if (arm64_parse_machdep_arg_l(arglist[i], "vabits_actual",
|
|
&machdep->machspec->VA_BITS_ACTUAL)) {
|
|
error(NOTE,
|
|
"setting vabits_actual to: %ld\n\n",
|
|
machdep->machspec->VA_BITS_ACTUAL);
|
|
continue;
|
|
}
|
|
|
|
error(WARNING, "ignoring --machdep option: %s\n",
|
|
arglist[i]);
|
|
}
|
|
}
|
|
}
|
|
|
|
#define MIN_KIMG_ALIGN (0x00200000) /* kimage load address must be aligned 2M */
|
|
/*
|
|
* Traverse the entire dumpfile to find/verify kimage_voffset.
|
|
*/
|
|
static int
|
|
arm64_search_for_kimage_voffset(ulong phys_base)
|
|
{
|
|
ulong kimage_load_addr;
|
|
ulong phys_end;
|
|
struct machine_specific *ms = machdep->machspec;
|
|
|
|
if (!arm_kdump_phys_end(&phys_end))
|
|
return FALSE;
|
|
|
|
for (kimage_load_addr = phys_base;
|
|
kimage_load_addr <= phys_end; kimage_load_addr += MIN_KIMG_ALIGN) {
|
|
ms->kimage_voffset = ms->vmalloc_start_addr - kimage_load_addr;
|
|
|
|
if ((kt->flags2 & KASLR) && (kt->flags & RELOC_SET))
|
|
ms->kimage_voffset += (kt->relocate * - 1);
|
|
|
|
if (verify_kimage_voffset()) {
|
|
if (CRASHDEBUG(1))
|
|
error(INFO,
|
|
"dumpfile searched for kimage_voffset: %lx\n\n",
|
|
ms->kimage_voffset);
|
|
break;
|
|
}
|
|
}
|
|
|
|
if (kimage_load_addr > phys_end)
|
|
return FALSE;
|
|
|
|
return TRUE;
|
|
}
|
|
|
|
static int
|
|
verify_kimage_voffset(void)
|
|
{
|
|
ulong kimage_voffset;
|
|
|
|
if (!readmem(symbol_value("kimage_voffset"), KVADDR, &kimage_voffset,
|
|
sizeof(kimage_voffset), "verify kimage_voffset", QUIET|RETURN_ON_ERROR))
|
|
return FALSE;
|
|
|
|
return (machdep->machspec->kimage_voffset == kimage_voffset);
|
|
}
|
|
|
|
static void
|
|
arm64_calc_kimage_voffset(void)
|
|
{
|
|
struct machine_specific *ms = machdep->machspec;
|
|
ulong phys_addr = 0;
|
|
int errflag;
|
|
|
|
if (ms->kimage_voffset) /* vmcoreinfo, ioctl, or --machdep override */
|
|
return;
|
|
|
|
if (ACTIVE()) {
|
|
char buf[BUFSIZE];
|
|
char *p1;
|
|
FILE *iomem;
|
|
ulong kimage_voffset, vaddr;
|
|
|
|
if (pc->flags & PROC_KCORE) {
|
|
kimage_voffset = symbol_value_from_proc_kallsyms("kimage_voffset");
|
|
if ((kimage_voffset != BADVAL) &&
|
|
(READMEM(pc->mfd, &vaddr, sizeof(ulong),
|
|
kimage_voffset, KCORE_USE_VADDR) > 0)) {
|
|
ms->kimage_voffset = vaddr;
|
|
return;
|
|
}
|
|
}
|
|
|
|
if ((iomem = fopen("/proc/iomem", "r")) == NULL)
|
|
return;
|
|
|
|
errflag = 1;
|
|
while (fgets(buf, BUFSIZE, iomem)) {
|
|
if(strstr(buf, ": Kernel code")) {
|
|
errflag = 0;
|
|
break;
|
|
}
|
|
if (strstr(buf, ": System RAM")) {
|
|
clean_line(buf);
|
|
|
|
if (!(p1 = strstr(buf, "-")))
|
|
continue;
|
|
|
|
*p1 = NULLCHAR;
|
|
|
|
phys_addr = htol(buf, RETURN_ON_ERROR | QUIET, NULL);
|
|
if (phys_addr == BADADDR)
|
|
continue;
|
|
}
|
|
}
|
|
fclose(iomem);
|
|
|
|
if (errflag)
|
|
return;
|
|
|
|
} else if (KDUMP_DUMPFILE()) {
|
|
errflag = 1;
|
|
if (arm_kdump_phys_base(&phys_addr)) { /* Get start address of first memory block */
|
|
ms->kimage_voffset = ms->vmalloc_start_addr - phys_addr;
|
|
if ((kt->flags2 & KASLR) && (kt->flags & RELOC_SET))
|
|
ms->kimage_voffset += (kt->relocate * -1);
|
|
if (verify_kimage_voffset() || arm64_search_for_kimage_voffset(phys_addr))
|
|
errflag = 0;
|
|
}
|
|
|
|
if (errflag) {
|
|
error(WARNING,
|
|
"kimage_voffset cannot be determined from the dumpfile.\n");
|
|
error(CONT,
|
|
"Try using the command line option: --machdep kimage_voffset=<addr>\n");
|
|
}
|
|
return;
|
|
} else {
|
|
error(WARNING,
|
|
"kimage_voffset cannot be determined from the dumpfile.\n");
|
|
error(CONT,
|
|
"Using default value of 0. If this is not correct, then try\n");
|
|
error(CONT,
|
|
"using the command line option: --machdep kimage_voffset=<addr>\n");
|
|
return;
|
|
}
|
|
|
|
ms->kimage_voffset = ms->vmalloc_start_addr - phys_addr;
|
|
|
|
if ((kt->flags2 & KASLR) && (kt->flags & RELOC_SET))
|
|
ms->kimage_voffset += (kt->relocate * -1);
|
|
}
|
|
|
|
/*
|
|
* The physvirt_offset only exits in kernel [5.4, 5.10)
|
|
*
|
|
* 1) In kernel v5.4, the patch:
|
|
* "5383cc6efed137 arm64: mm: Introduce vabits_actual"
|
|
*
|
|
* introduced the physvirt_offset.
|
|
*
|
|
* 2) In kernel v5.10, the patch:
|
|
* "7bc1a0f9e17658 arm64: mm: use single quantity
|
|
* to represent the PA to VA translation"
|
|
* removed the physvirt_offset.
|
|
*/
|
|
static void
|
|
arm64_calc_physvirt_offset(void)
|
|
{
|
|
struct machine_specific *ms = machdep->machspec;
|
|
ulong physvirt_offset;
|
|
struct syment *sp;
|
|
ulong value;
|
|
|
|
if ((sp = kernel_symbol_search("physvirt_offset")) &&
|
|
machdep->machspec->kimage_voffset) {
|
|
if (pc->flags & PROC_KCORE) {
|
|
value = symbol_value_from_proc_kallsyms("physvirt_offset");
|
|
if ((value != BADVAL) &&
|
|
(READMEM(pc->mfd, &physvirt_offset, sizeof(ulong),
|
|
value, KCORE_USE_VADDR) > 0)) {
|
|
machdep->flags |= HAS_PHYSVIRT_OFFSET;
|
|
ms->physvirt_offset = physvirt_offset;
|
|
|
|
/* Update the ms->phys_offset which is wrong */
|
|
ms->phys_offset = ms->physvirt_offset + ms->page_offset;
|
|
return;
|
|
}
|
|
}
|
|
|
|
if (READMEM(pc->mfd, &physvirt_offset, sizeof(physvirt_offset),
|
|
sp->value, sp->value -
|
|
machdep->machspec->kimage_voffset) > 0) {
|
|
machdep->flags |= HAS_PHYSVIRT_OFFSET;
|
|
ms->physvirt_offset = physvirt_offset;
|
|
return;
|
|
}
|
|
}
|
|
|
|
/* Useless if no symbol 'physvirt_offset', just keep semantics */
|
|
ms->physvirt_offset = ms->phys_offset - ms->page_offset;
|
|
|
|
}
|
|
|
|
static void
|
|
arm64_calc_phys_offset(void)
|
|
{
|
|
struct machine_specific *ms = machdep->machspec;
|
|
ulong phys_offset;
|
|
|
|
if (machdep->flags & PHYS_OFFSET) /* --machdep override */
|
|
return;
|
|
|
|
/*
|
|
* Next determine suitable value for phys_offset. User can override this
|
|
* by passing valid '--machdep phys_offset=<addr>' option.
|
|
*/
|
|
ms->phys_offset = 0;
|
|
|
|
if (ACTIVE()) {
|
|
char buf[BUFSIZE];
|
|
char *p1;
|
|
int errflag;
|
|
FILE *iomem;
|
|
physaddr_t paddr;
|
|
ulong vaddr;
|
|
struct syment *sp;
|
|
|
|
if ((machdep->flags & NEW_VMEMMAP) &&
|
|
ms->kimage_voffset && (sp = kernel_symbol_search("memstart_addr"))) {
|
|
if (pc->flags & PROC_KCORE) {
|
|
if (arm64_get_vmcoreinfo(&ms->phys_offset, "NUMBER(PHYS_OFFSET)", NUM_HEX))
|
|
return;
|
|
vaddr = symbol_value_from_proc_kallsyms("memstart_addr");
|
|
if (vaddr == BADVAL)
|
|
vaddr = sp->value;
|
|
paddr = KCORE_USE_VADDR;
|
|
} else {
|
|
vaddr = sp->value;
|
|
paddr = sp->value - machdep->machspec->kimage_voffset;
|
|
}
|
|
if (READMEM(pc->mfd, &phys_offset, sizeof(phys_offset),
|
|
vaddr, paddr) > 0) {
|
|
ms->phys_offset = phys_offset;
|
|
|
|
return;
|
|
}
|
|
}
|
|
|
|
if ((iomem = fopen("/proc/iomem", "r")) == NULL)
|
|
return;
|
|
|
|
/*
|
|
* Memory regions are sorted in ascending order. We take the
|
|
* first region which should be correct for most uses.
|
|
*/
|
|
errflag = 1;
|
|
while (fgets(buf, BUFSIZE, iomem)) {
|
|
if (strstr(buf, ": System RAM")) {
|
|
clean_line(buf);
|
|
errflag = 0;
|
|
break;
|
|
}
|
|
}
|
|
fclose(iomem);
|
|
|
|
if (errflag)
|
|
return;
|
|
|
|
if (!(p1 = strstr(buf, "-")))
|
|
return;
|
|
|
|
*p1 = NULLCHAR;
|
|
|
|
phys_offset = htol(buf, RETURN_ON_ERROR | QUIET, &errflag);
|
|
if (errflag)
|
|
return;
|
|
|
|
ms->phys_offset = phys_offset;
|
|
} else if (DISKDUMP_DUMPFILE() && diskdump_phys_base(&phys_offset)) {
|
|
ms->phys_offset = phys_offset;
|
|
} else if (KDUMP_DUMPFILE() && arm64_kdump_phys_base(&phys_offset)) {
|
|
ms->phys_offset = phys_offset;
|
|
} else {
|
|
error(WARNING,
|
|
"phys_offset cannot be determined from the dumpfile.\n");
|
|
error(CONT,
|
|
"Using default value of 0. If this is not correct, then try\n");
|
|
error(CONT,
|
|
"using the command line option: --machdep phys_offset=<addr>\n");
|
|
}
|
|
|
|
if (CRASHDEBUG(1))
|
|
fprintf(fp, "using %lx as phys_offset\n", ms->phys_offset);
|
|
}
|
|
|
|
/*
|
|
* Determine SECTION_SIZE_BITS either by reading VMCOREINFO or the kernel
|
|
* config, otherwise use the 64-bit ARM default definiton.
|
|
*/
|
|
static void
|
|
arm64_get_section_size_bits(void)
|
|
{
|
|
int ret;
|
|
char *string;
|
|
|
|
if (THIS_KERNEL_VERSION >= LINUX(5,12,0)) {
|
|
if (machdep->pagesize == 65536)
|
|
machdep->section_size_bits = _SECTION_SIZE_BITS_5_12_64K;
|
|
else
|
|
machdep->section_size_bits = _SECTION_SIZE_BITS_5_12;
|
|
} else
|
|
machdep->section_size_bits = _SECTION_SIZE_BITS;
|
|
|
|
if (arm64_get_vmcoreinfo(&machdep->section_size_bits, "NUMBER(SECTION_SIZE_BITS)", NUM_DEC)) {
|
|
/* nothing */
|
|
} else if (kt->ikconfig_flags & IKCONFIG_AVAIL) {
|
|
if ((ret = get_kernel_config("CONFIG_MEMORY_HOTPLUG", NULL)) == IKCONFIG_Y) {
|
|
if ((ret = get_kernel_config("CONFIG_HOTPLUG_SIZE_BITS", &string)) == IKCONFIG_STR)
|
|
machdep->section_size_bits = atol(string);
|
|
}
|
|
}
|
|
|
|
if (CRASHDEBUG(1))
|
|
fprintf(fp, "SECTION_SIZE_BITS: %ld\n", machdep->section_size_bits);
|
|
}
|
|
|
|
/*
|
|
* Determine PHYS_OFFSET either by reading VMCOREINFO or the kernel
|
|
* symbol, otherwise borrow the 32-bit ARM functionality.
|
|
*/
|
|
static int
|
|
arm64_kdump_phys_base(ulong *phys_offset)
|
|
{
|
|
struct syment *sp;
|
|
physaddr_t paddr;
|
|
|
|
if (arm64_get_vmcoreinfo(phys_offset, "NUMBER(PHYS_OFFSET)", NUM_HEX))
|
|
return TRUE;
|
|
|
|
if ((machdep->flags & NEW_VMEMMAP) &&
|
|
machdep->machspec->kimage_voffset &&
|
|
(sp = kernel_symbol_search("memstart_addr"))) {
|
|
paddr = sp->value - machdep->machspec->kimage_voffset;
|
|
if (READMEM(-1, phys_offset, sizeof(*phys_offset),
|
|
sp->value, paddr) > 0)
|
|
return TRUE;
|
|
}
|
|
|
|
return arm_kdump_phys_base(phys_offset);
|
|
}
|
|
|
|
static void
|
|
arm64_init_kernel_pgd(void)
|
|
{
|
|
int i;
|
|
ulong value;
|
|
|
|
if (!kernel_symbol_exists("init_mm") ||
|
|
!readmem(symbol_value("init_mm") + OFFSET(mm_struct_pgd), KVADDR,
|
|
&value, sizeof(void *), "init_mm.pgd", RETURN_ON_ERROR)) {
|
|
if (kernel_symbol_exists("swapper_pg_dir"))
|
|
value = symbol_value("swapper_pg_dir");
|
|
else {
|
|
error(WARNING, "cannot determine kernel pgd location\n");
|
|
return;
|
|
}
|
|
}
|
|
|
|
for (i = 0; i < NR_CPUS; i++)
|
|
vt->kernel_pgd[i] = value;
|
|
}
|
|
|
|
ulong arm64_PTOV(ulong paddr)
|
|
{
|
|
struct machine_specific *ms = machdep->machspec;
|
|
|
|
/*
|
|
* Either older kernel before kernel has 'physvirt_offset' or newer
|
|
* kernel which removes 'physvirt_offset' has the same formula:
|
|
* #define __phys_to_virt(x) ((unsigned long)((x) - PHYS_OFFSET) | PAGE_OFFSET)
|
|
*/
|
|
if (!(machdep->flags & HAS_PHYSVIRT_OFFSET))
|
|
return (paddr - ms->phys_offset) | PAGE_OFFSET;
|
|
else
|
|
return paddr - ms->physvirt_offset;
|
|
}
|
|
|
|
ulong
|
|
arm64_VTOP(ulong addr)
|
|
{
|
|
if (is_mte_kvaddr(addr))
|
|
addr = mte_tag_reset(addr);
|
|
|
|
if (machdep->flags & NEW_VMEMMAP) {
|
|
if (machdep->machspec->VA_START &&
|
|
(addr >= machdep->machspec->kimage_text) &&
|
|
(addr <= machdep->machspec->kimage_end)) {
|
|
return addr - machdep->machspec->kimage_voffset;
|
|
}
|
|
|
|
if (addr >= machdep->machspec->page_offset) {
|
|
if (machdep->flags & HAS_PHYSVIRT_OFFSET) {
|
|
return addr + machdep->machspec->physvirt_offset;
|
|
} else {
|
|
/*
|
|
* Either older kernel before kernel has 'physvirt_offset' or newer
|
|
* kernel which removes 'physvirt_offset' has the same formula:
|
|
* #define __lm_to_phys(addr) (((addr) & ~PAGE_OFFSET) + PHYS_OFFSET)
|
|
*/
|
|
return (addr & ~PAGE_OFFSET) + machdep->machspec->phys_offset;
|
|
}
|
|
}
|
|
else if (machdep->machspec->kimage_voffset)
|
|
return addr - machdep->machspec->kimage_voffset;
|
|
else /* no randomness */
|
|
return machdep->machspec->phys_offset
|
|
+ (addr - machdep->machspec->vmalloc_start_addr);
|
|
} else {
|
|
return machdep->machspec->phys_offset
|
|
+ (addr - machdep->machspec->page_offset);
|
|
}
|
|
}
|
|
|
|
static int
|
|
arm64_kvtop(struct task_context *tc, ulong kvaddr, physaddr_t *paddr, int verbose)
|
|
{
|
|
ulong kernel_pgd;
|
|
|
|
if (!IS_KVADDR(kvaddr))
|
|
return FALSE;
|
|
|
|
if (!vt->vmalloc_start) {
|
|
*paddr = VTOP(kvaddr);
|
|
return TRUE;
|
|
}
|
|
|
|
if (!IS_VMALLOC_ADDR(kvaddr)) {
|
|
*paddr = VTOP(kvaddr);
|
|
if (!verbose)
|
|
return TRUE;
|
|
}
|
|
|
|
kernel_pgd = vt->kernel_pgd[0];
|
|
*paddr = 0;
|
|
|
|
switch (machdep->flags & (VM_L2_64K|VM_L3_64K|VM_L3_4K|VM_L4_4K|VM_L2_16K|VM_L3_16K|VM_L4_16K))
|
|
{
|
|
case VM_L2_64K:
|
|
return arm64_vtop_2level_64k(kernel_pgd, kvaddr, paddr, verbose);
|
|
case VM_L3_64K:
|
|
return arm64_vtop_3level_64k(kernel_pgd, kvaddr, paddr, verbose);
|
|
case VM_L3_4K:
|
|
return arm64_vtop_3level_4k(kernel_pgd, kvaddr, paddr, verbose);
|
|
case VM_L4_4K:
|
|
return arm64_vtop_4level_4k(kernel_pgd, kvaddr, paddr, verbose);
|
|
case VM_L2_16K:
|
|
return arm64_vtop_2level_16k(kernel_pgd, kvaddr, paddr, verbose);
|
|
case VM_L3_16K:
|
|
return arm64_vtop_3level_16k(kernel_pgd, kvaddr, paddr, verbose);
|
|
case VM_L4_16K:
|
|
return arm64_vtop_4level_16k(kernel_pgd, kvaddr, paddr, verbose);
|
|
default:
|
|
return FALSE;
|
|
}
|
|
}
|
|
|
|
static int
|
|
arm64_uvtop(struct task_context *tc, ulong uvaddr, physaddr_t *paddr, int verbose)
|
|
{
|
|
ulong user_pgd;
|
|
|
|
readmem(tc->mm_struct + OFFSET(mm_struct_pgd), KVADDR,
|
|
&user_pgd, sizeof(long), "user pgd", FAULT_ON_ERROR);
|
|
|
|
*paddr = 0;
|
|
|
|
switch (machdep->flags & (VM_L2_64K|VM_L3_64K|VM_L3_4K|VM_L4_4K|VM_L2_16K|VM_L3_16K|VM_L4_16K))
|
|
{
|
|
case VM_L2_64K:
|
|
return arm64_vtop_2level_64k(user_pgd, uvaddr, paddr, verbose);
|
|
case VM_L3_64K:
|
|
return arm64_vtop_3level_64k(user_pgd, uvaddr, paddr, verbose);
|
|
case VM_L3_4K:
|
|
return arm64_vtop_3level_4k(user_pgd, uvaddr, paddr, verbose);
|
|
case VM_L4_4K:
|
|
return arm64_vtop_4level_4k(user_pgd, uvaddr, paddr, verbose);
|
|
case VM_L2_16K:
|
|
return arm64_vtop_2level_16k(user_pgd, uvaddr, paddr, verbose);
|
|
case VM_L3_16K:
|
|
return arm64_vtop_3level_16k(user_pgd, uvaddr, paddr, verbose);
|
|
case VM_L4_16K:
|
|
return arm64_vtop_4level_16k(user_pgd, uvaddr, paddr, verbose);
|
|
default:
|
|
return FALSE;
|
|
}
|
|
}
|
|
|
|
#define PTE_ADDR_LOW ((((1UL) << (48 - machdep->pageshift)) - 1) << machdep->pageshift)
|
|
#define PTE_ADDR_HIGH ((0xfUL) << 12)
|
|
#define PTE_ADDR_HIGH_SHIFT 36
|
|
#define PTE_TO_PHYS(pteval) (machdep->max_physmem_bits == 52 ? \
|
|
(((pteval & PTE_ADDR_LOW) | ((pteval & PTE_ADDR_HIGH) << PTE_ADDR_HIGH_SHIFT))) : (pteval & PTE_ADDR_LOW))
|
|
|
|
#define PUD_TYPE_MASK 3
|
|
#define PUD_TYPE_SECT 1
|
|
#define PMD_TYPE_MASK 3
|
|
#define PMD_TYPE_SECT 1
|
|
#define PMD_TYPE_TABLE 2
|
|
#define SECTION_PAGE_MASK_2MB ((long)(~((MEGABYTES(2))-1)))
|
|
#define SECTION_PAGE_MASK_32MB ((long)(~((MEGABYTES(32))-1)))
|
|
#define SECTION_PAGE_MASK_512MB ((long)(~((MEGABYTES(512))-1)))
|
|
#define SECTION_PAGE_MASK_1GB ((long)(~((GIGABYTES(1))-1)))
|
|
#define SECTION_PAGE_MASK_64GB ((long)(~((GIGABYTES(64))-1)))
|
|
|
|
static int
|
|
arm64_vtop_2level_64k(ulong pgd, ulong vaddr, physaddr_t *paddr, int verbose)
|
|
{
|
|
ulong *pgd_base, *pgd_ptr, pgd_val;
|
|
ulong *pte_base, *pte_ptr, pte_val;
|
|
|
|
if (verbose)
|
|
fprintf(fp, "PAGE DIRECTORY: %lx\n", pgd);
|
|
|
|
pgd_base = (ulong *)pgd;
|
|
FILL_PGD(pgd_base, KVADDR, machdep->ptrs_per_pgd * sizeof(ulong));
|
|
pgd_ptr = pgd_base + (((vaddr) >> PGDIR_SHIFT_L2_64K) & (machdep->ptrs_per_pgd - 1));
|
|
pgd_val = ULONG(machdep->pgd + PAGEOFFSET(pgd_ptr));
|
|
if (verbose)
|
|
fprintf(fp, " PGD: %lx => %lx\n", (ulong)pgd_ptr, pgd_val);
|
|
if (!pgd_val)
|
|
goto no_page;
|
|
|
|
/*
|
|
* #define __PAGETABLE_PUD_FOLDED
|
|
* #define __PAGETABLE_PMD_FOLDED
|
|
*/
|
|
|
|
if ((pgd_val & PMD_TYPE_MASK) == PMD_TYPE_SECT) {
|
|
ulong sectionbase = PTE_TO_PHYS(pgd_val & SECTION_PAGE_MASK_512MB);
|
|
if (verbose) {
|
|
fprintf(fp, " PAGE: %lx (512MB%s)\n\n", sectionbase,
|
|
IS_ZEROPAGE(sectionbase) ? ", ZERO PAGE" : "");
|
|
arm64_translate_pte(pgd_val, 0, 0);
|
|
}
|
|
*paddr = sectionbase + (vaddr & ~SECTION_PAGE_MASK_512MB);
|
|
return TRUE;
|
|
}
|
|
|
|
pte_base = (ulong *)PTOV(PTE_TO_PHYS(pgd_val));
|
|
FILL_PTBL(pte_base, KVADDR, PTRS_PER_PTE_L2_64K * sizeof(ulong));
|
|
pte_ptr = pte_base + (((vaddr) >> machdep->pageshift) & (PTRS_PER_PTE_L2_64K - 1));
|
|
pte_val = ULONG(machdep->ptbl + PAGEOFFSET(pte_ptr));
|
|
if (verbose)
|
|
fprintf(fp, " PTE: %lx => %lx\n", (ulong)pte_ptr, pte_val);
|
|
if (!pte_val)
|
|
goto no_page;
|
|
|
|
if (pte_val & PTE_VALID) {
|
|
*paddr = PTE_TO_PHYS(pte_val) + PAGEOFFSET(vaddr);
|
|
if (verbose) {
|
|
fprintf(fp, " PAGE: %lx %s\n\n", PAGEBASE(*paddr),
|
|
IS_ZEROPAGE(PAGEBASE(*paddr)) ? "(ZERO PAGE)" : "");
|
|
arm64_translate_pte(pte_val, 0, 0);
|
|
}
|
|
} else {
|
|
if (IS_UVADDR(vaddr, NULL))
|
|
*paddr = pte_val;
|
|
if (verbose) {
|
|
fprintf(fp, "\n");
|
|
arm64_translate_pte(pte_val, 0, 0);
|
|
}
|
|
goto no_page;
|
|
}
|
|
|
|
return TRUE;
|
|
no_page:
|
|
return FALSE;
|
|
}
|
|
|
|
static int
|
|
arm64_vtop_3level_64k(ulong pgd, ulong vaddr, physaddr_t *paddr, int verbose)
|
|
{
|
|
ulong *pgd_base, *pgd_ptr, pgd_val;
|
|
ulong *pmd_base, *pmd_ptr, pmd_val;
|
|
ulong *pte_base, *pte_ptr, pte_val;
|
|
|
|
if (verbose)
|
|
fprintf(fp, "PAGE DIRECTORY: %lx\n", pgd);
|
|
|
|
pgd_base = (ulong *)pgd;
|
|
FILL_PGD(pgd_base, KVADDR, machdep->ptrs_per_pgd * sizeof(ulong));
|
|
/*
|
|
* Use machdep->ptrs_per_pgd to mask vaddr instead of using macro, because
|
|
* 48-bits and 52-bits have different size of ptrs_per_pgd.
|
|
*/
|
|
pgd_ptr = pgd_base + (((vaddr) >> PGDIR_SHIFT_L3_64K) & (machdep->ptrs_per_pgd - 1));
|
|
pgd_val = ULONG(machdep->pgd + PGDIR_OFFSET_L3_64K(pgd_ptr));
|
|
if (verbose)
|
|
fprintf(fp, " PGD: %lx => %lx\n", (ulong)pgd_ptr, pgd_val);
|
|
if (!pgd_val)
|
|
goto no_page;
|
|
|
|
/*
|
|
* #define __PAGETABLE_PUD_FOLDED
|
|
*/
|
|
|
|
pmd_base = (ulong *)PTOV(PTE_TO_PHYS(pgd_val));
|
|
FILL_PMD(pmd_base, KVADDR, PTRS_PER_PMD_L3_64K * sizeof(ulong));
|
|
pmd_ptr = pmd_base + (((vaddr) >> PMD_SHIFT_L3_64K) & (PTRS_PER_PMD_L3_64K - 1));
|
|
pmd_val = ULONG(machdep->pmd + PAGEOFFSET(pmd_ptr));
|
|
if (verbose)
|
|
fprintf(fp, " PMD: %lx => %lx\n", (ulong)pmd_ptr, pmd_val);
|
|
if (!pmd_val)
|
|
goto no_page;
|
|
|
|
if ((pmd_val & PMD_TYPE_MASK) == PMD_TYPE_SECT) {
|
|
ulong sectionbase = PTE_TO_PHYS(pmd_val & SECTION_PAGE_MASK_512MB);
|
|
if (verbose) {
|
|
fprintf(fp, " PAGE: %lx (512MB%s)\n\n", sectionbase,
|
|
IS_ZEROPAGE(sectionbase) ? ", ZERO PAGE" : "");
|
|
arm64_translate_pte(pmd_val, 0, 0);
|
|
}
|
|
*paddr = sectionbase + (vaddr & ~SECTION_PAGE_MASK_512MB);
|
|
return TRUE;
|
|
}
|
|
|
|
pte_base = (ulong *)PTOV(PTE_TO_PHYS(pmd_val));
|
|
FILL_PTBL(pte_base, KVADDR, PTRS_PER_PTE_L3_64K * sizeof(ulong));
|
|
pte_ptr = pte_base + (((vaddr) >> machdep->pageshift) & (PTRS_PER_PTE_L3_64K - 1));
|
|
pte_val = ULONG(machdep->ptbl + PAGEOFFSET(pte_ptr));
|
|
if (verbose)
|
|
fprintf(fp, " PTE: %lx => %lx\n", (ulong)pte_ptr, pte_val);
|
|
if (!pte_val)
|
|
goto no_page;
|
|
|
|
if (pte_val & PTE_VALID) {
|
|
*paddr = PTE_TO_PHYS(pte_val) + PAGEOFFSET(vaddr);
|
|
if (verbose) {
|
|
fprintf(fp, " PAGE: %lx %s\n\n", PAGEBASE(*paddr),
|
|
IS_ZEROPAGE(PAGEBASE(*paddr)) ? "(ZERO PAGE)" : "");
|
|
arm64_translate_pte(pte_val, 0, 0);
|
|
}
|
|
} else {
|
|
if (IS_UVADDR(vaddr, NULL))
|
|
*paddr = pte_val;
|
|
if (verbose) {
|
|
fprintf(fp, "\n");
|
|
arm64_translate_pte(pte_val, 0, 0);
|
|
}
|
|
goto no_page;
|
|
}
|
|
|
|
return TRUE;
|
|
no_page:
|
|
return FALSE;
|
|
}
|
|
|
|
static int
|
|
arm64_vtop_2level_16k(ulong pgd, ulong vaddr, physaddr_t *paddr, int verbose)
|
|
{
|
|
ulong *pgd_base, *pgd_ptr, pgd_val;
|
|
ulong *pte_base, *pte_ptr, pte_val;
|
|
|
|
if (verbose)
|
|
fprintf(fp, "PAGE DIRECTORY: %lx\n", pgd);
|
|
|
|
pgd_base = (ulong *)pgd;
|
|
FILL_PGD(pgd_base, KVADDR, machdep->ptrs_per_pgd * sizeof(ulong));
|
|
pgd_ptr = pgd_base + (((vaddr) >> PGDIR_SHIFT_L2_16K) & (machdep->ptrs_per_pgd - 1));
|
|
pgd_val = ULONG(machdep->pgd + PGDIR_OFFSET_L2_16K(pgd_ptr));
|
|
if (verbose)
|
|
fprintf(fp, " PGD: %lx => %lx\n", (ulong)pgd_ptr, pgd_val);
|
|
if (!pgd_val)
|
|
goto no_page;
|
|
|
|
/*
|
|
* #define __PAGETABLE_PUD_FOLDED
|
|
* #define __PAGETABLE_PMD_FOLDED
|
|
*/
|
|
|
|
if ((pgd_val & PMD_TYPE_MASK) == PMD_TYPE_SECT) {
|
|
ulong sectionbase = PTE_TO_PHYS(pgd_val & SECTION_PAGE_MASK_32MB);
|
|
if (verbose) {
|
|
fprintf(fp, " PAGE: %lx (32MB%s)\n\n", sectionbase,
|
|
IS_ZEROPAGE(sectionbase) ? ", ZERO PAGE" : "");
|
|
arm64_translate_pte(pgd_val, 0, 0);
|
|
}
|
|
*paddr = sectionbase + (vaddr & ~SECTION_PAGE_MASK_32MB);
|
|
return TRUE;
|
|
}
|
|
|
|
pte_base = (ulong *)PTOV(PTE_TO_PHYS(pgd_val));
|
|
FILL_PTBL(pte_base, KVADDR, PTRS_PER_PTE_L2_16K * sizeof(ulong));
|
|
pte_ptr = pte_base + (((vaddr) >> machdep->pageshift) & (PTRS_PER_PTE_L2_16K - 1));
|
|
pte_val = ULONG(machdep->ptbl + PAGEOFFSET(pte_ptr));
|
|
if (verbose)
|
|
fprintf(fp, " PTE: %lx => %lx\n", (ulong)pte_ptr, pte_val);
|
|
if (!pte_val)
|
|
goto no_page;
|
|
|
|
if (pte_val & PTE_VALID) {
|
|
*paddr = PTE_TO_PHYS(pte_val) + PAGEOFFSET(vaddr);
|
|
if (verbose) {
|
|
fprintf(fp, " PAGE: %lx %s\n\n", PAGEBASE(*paddr),
|
|
IS_ZEROPAGE(PAGEBASE(*paddr)) ? "(ZERO PAGE)" : "");
|
|
arm64_translate_pte(pte_val, 0, 0);
|
|
}
|
|
} else {
|
|
if (IS_UVADDR(vaddr, NULL))
|
|
*paddr = pte_val;
|
|
if (verbose) {
|
|
fprintf(fp, "\n");
|
|
arm64_translate_pte(pte_val, 0, 0);
|
|
}
|
|
goto no_page;
|
|
}
|
|
|
|
return TRUE;
|
|
no_page:
|
|
return FALSE;
|
|
}
|
|
static int
|
|
arm64_vtop_3level_16k(ulong pgd, ulong vaddr, physaddr_t *paddr, int verbose)
|
|
{
|
|
ulong *pgd_base, *pgd_ptr, pgd_val;
|
|
ulong *pmd_base, *pmd_ptr, pmd_val;
|
|
ulong *pte_base, *pte_ptr, pte_val;
|
|
|
|
if (verbose)
|
|
fprintf(fp, "PAGE DIRECTORY: %lx\n", pgd);
|
|
|
|
pgd_base = (ulong *)pgd;
|
|
FILL_PGD(pgd_base, KVADDR, machdep->ptrs_per_pgd * sizeof(ulong));
|
|
pgd_ptr = pgd_base + (((vaddr) >> PGDIR_SHIFT_L3_16K) & (machdep->ptrs_per_pgd - 1));
|
|
pgd_val = ULONG(machdep->pgd + PGDIR_OFFSET_L3_16K(pgd_ptr));
|
|
if (verbose)
|
|
fprintf(fp, " PGD: %lx => %lx\n", (ulong)pgd_ptr, pgd_val);
|
|
if (!pgd_val)
|
|
goto no_page;
|
|
|
|
/*
|
|
* #define __PAGETABLE_PUD_FOLDED
|
|
*/
|
|
|
|
pmd_base = (ulong *)PTOV(PTE_TO_PHYS(pgd_val));
|
|
FILL_PMD(pmd_base, KVADDR, PTRS_PER_PMD_L3_16K * sizeof(ulong));
|
|
pmd_ptr = pmd_base + (((vaddr) >> PMD_SHIFT_L3_16K) & (PTRS_PER_PMD_L3_16K - 1));
|
|
pmd_val = ULONG(machdep->pmd + PAGEOFFSET(pmd_ptr));
|
|
if (verbose)
|
|
fprintf(fp, " PMD: %lx => %lx\n", (ulong)pmd_ptr, pmd_val);
|
|
if (!pmd_val)
|
|
goto no_page;
|
|
|
|
if ((pmd_val & PMD_TYPE_MASK) == PMD_TYPE_SECT) {
|
|
ulong sectionbase = PTE_TO_PHYS(pmd_val & SECTION_PAGE_MASK_32MB);
|
|
if (verbose) {
|
|
fprintf(fp, " PAGE: %lx (32MB%s)\n\n", sectionbase,
|
|
IS_ZEROPAGE(sectionbase) ? ", ZERO PAGE" : "");
|
|
arm64_translate_pte(pmd_val, 0, 0);
|
|
}
|
|
*paddr = sectionbase + (vaddr & ~SECTION_PAGE_MASK_32MB);
|
|
return TRUE;
|
|
}
|
|
|
|
pte_base = (ulong *)PTOV(PTE_TO_PHYS(pmd_val));
|
|
FILL_PTBL(pte_base, KVADDR, PTRS_PER_PTE_L3_16K * sizeof(ulong));
|
|
pte_ptr = pte_base + (((vaddr) >> machdep->pageshift) & (PTRS_PER_PTE_L3_16K - 1));
|
|
pte_val = ULONG(machdep->ptbl + PAGEOFFSET(pte_ptr));
|
|
if (verbose)
|
|
fprintf(fp, " PTE: %lx => %lx\n", (ulong)pte_ptr, pte_val);
|
|
if (!pte_val)
|
|
goto no_page;
|
|
|
|
if (pte_val & PTE_VALID) {
|
|
*paddr = PTE_TO_PHYS(pte_val) + PAGEOFFSET(vaddr);
|
|
if (verbose) {
|
|
fprintf(fp, " PAGE: %lx %s\n\n", PAGEBASE(*paddr),
|
|
IS_ZEROPAGE(PAGEBASE(*paddr)) ? "(ZERO PAGE)" : "");
|
|
arm64_translate_pte(pte_val, 0, 0);
|
|
}
|
|
} else {
|
|
if (IS_UVADDR(vaddr, NULL))
|
|
*paddr = pte_val;
|
|
if (verbose) {
|
|
fprintf(fp, "\n");
|
|
arm64_translate_pte(pte_val, 0, 0);
|
|
}
|
|
goto no_page;
|
|
}
|
|
|
|
return TRUE;
|
|
no_page:
|
|
return FALSE;
|
|
}
|
|
|
|
static int
|
|
arm64_vtop_4level_16k(ulong pgd, ulong vaddr, physaddr_t *paddr, int verbose)
|
|
{
|
|
ulong *pgd_base, *pgd_ptr, pgd_val;
|
|
ulong *pud_base, *pud_ptr, pud_val;
|
|
ulong *pmd_base, *pmd_ptr, pmd_val;
|
|
ulong *pte_base, *pte_ptr, pte_val;
|
|
|
|
if (verbose)
|
|
fprintf(fp, "PAGE DIRECTORY: %lx\n", pgd);
|
|
|
|
pgd_base = (ulong *)pgd;
|
|
FILL_PGD(pgd_base, KVADDR, PTRS_PER_PGD_L4_16K * sizeof(ulong));
|
|
pgd_ptr = pgd_base + (((vaddr) >> PGDIR_SHIFT_L4_16K) & (PTRS_PER_PGD_L4_16K - 1));
|
|
pgd_val = ULONG(machdep->pgd + PGDIR_OFFSET_L4_16K(pgd_ptr));
|
|
if (verbose)
|
|
fprintf(fp, " PGD: %lx => %lx\n", (ulong)pgd_ptr, pgd_val);
|
|
if (!pgd_val)
|
|
goto no_page;
|
|
|
|
pud_base = (ulong *)PTOV(PTE_TO_PHYS(pgd_val));
|
|
FILL_PUD(pud_base, KVADDR, PTRS_PER_PUD_L4_16K * sizeof(ulong));
|
|
pud_ptr = pud_base + (((vaddr) >> PUD_SHIFT_L4_16K) & (PTRS_PER_PUD_L4_16K - 1));
|
|
pud_val = ULONG(machdep->pud + PAGEOFFSET(pud_ptr));
|
|
if (verbose)
|
|
fprintf(fp, " PUD: %lx => %lx\n", (ulong)pud_ptr, pud_val);
|
|
if (!pud_val)
|
|
goto no_page;
|
|
|
|
/*
|
|
* TODO:
|
|
* PUD Section mapping is only supported for LPA2 is enabled.
|
|
* LPA2 depends on CONFIG_ARM64_16K_PAGES and CONFIG_ARM64_PA_BITS_52.
|
|
*/
|
|
|
|
pmd_base = (ulong *)PTOV(PTE_TO_PHYS(pud_val));
|
|
FILL_PMD(pmd_base, KVADDR, PTRS_PER_PMD_L4_16K * sizeof(ulong));
|
|
pmd_ptr = pmd_base + (((vaddr) >> PMD_SHIFT_L4_16K) & (PTRS_PER_PMD_L4_16K - 1));
|
|
pmd_val = ULONG(machdep->pmd + PAGEOFFSET(pmd_ptr));
|
|
if (verbose)
|
|
fprintf(fp, " PMD: %lx => %lx\n", (ulong)pmd_ptr, pmd_val);
|
|
if (!pmd_val)
|
|
goto no_page;
|
|
|
|
if ((pmd_val & PMD_TYPE_MASK) == PMD_TYPE_SECT) {
|
|
ulong sectionbase = PTE_TO_PHYS(pmd_val) & SECTION_PAGE_MASK_32MB;
|
|
if (verbose) {
|
|
fprintf(fp, " PAGE: %lx (32MB%s)\n\n", sectionbase,
|
|
IS_ZEROPAGE(sectionbase) ? ", ZERO PAGE" : "");
|
|
arm64_translate_pte(pmd_val, 0, 0);
|
|
}
|
|
*paddr = sectionbase + (vaddr & ~SECTION_PAGE_MASK_32MB);
|
|
return TRUE;
|
|
}
|
|
|
|
pte_base = (ulong *)PTOV(PTE_TO_PHYS(pmd_val));
|
|
FILL_PTBL(pte_base, KVADDR, PTRS_PER_PTE_L4_16K * sizeof(ulong));
|
|
pte_ptr = pte_base + (((vaddr) >> machdep->pageshift) & (PTRS_PER_PTE_L4_16K - 1));
|
|
pte_val = ULONG(machdep->ptbl + PAGEOFFSET(pte_ptr));
|
|
if (verbose)
|
|
fprintf(fp, " PTE: %lx => %lx\n", (ulong)pte_ptr, pte_val);
|
|
if (!pte_val)
|
|
goto no_page;
|
|
|
|
if (pte_val & PTE_VALID) {
|
|
*paddr = PTE_TO_PHYS(pte_val) + PAGEOFFSET(vaddr);
|
|
if (verbose) {
|
|
fprintf(fp, " PAGE: %lx %s\n\n", PAGEBASE(*paddr),
|
|
IS_ZEROPAGE(PAGEBASE(*paddr)) ? "(ZERO PAGE)" : "");
|
|
arm64_translate_pte(pte_val, 0, 0);
|
|
}
|
|
} else {
|
|
if (IS_UVADDR(vaddr, NULL))
|
|
*paddr = pte_val;
|
|
if (verbose) {
|
|
fprintf(fp, "\n");
|
|
arm64_translate_pte(pte_val, 0, 0);
|
|
}
|
|
goto no_page;
|
|
}
|
|
|
|
return TRUE;
|
|
no_page:
|
|
return FALSE;
|
|
}
|
|
|
|
static int
|
|
arm64_vtop_3level_4k(ulong pgd, ulong vaddr, physaddr_t *paddr, int verbose)
|
|
{
|
|
ulong *pgd_base, *pgd_ptr, pgd_val;
|
|
ulong *pmd_base, *pmd_ptr, pmd_val;
|
|
ulong *pte_base, *pte_ptr, pte_val;
|
|
|
|
if (verbose)
|
|
fprintf(fp, "PAGE DIRECTORY: %lx\n", pgd);
|
|
|
|
pgd_base = (ulong *)pgd;
|
|
FILL_PGD(pgd_base, KVADDR, PTRS_PER_PGD_L3_4K * sizeof(ulong));
|
|
pgd_ptr = pgd_base + (((vaddr) >> PGDIR_SHIFT_L3_4K) & (PTRS_PER_PGD_L3_4K - 1));
|
|
pgd_val = ULONG(machdep->pgd + PAGEOFFSET(pgd_ptr));
|
|
if (verbose)
|
|
fprintf(fp, " PGD: %lx => %lx\n", (ulong)pgd_ptr, pgd_val);
|
|
if (!pgd_val)
|
|
goto no_page;
|
|
|
|
if ((pgd_val & PUD_TYPE_MASK) == PUD_TYPE_SECT) {
|
|
ulong sectionbase = PTE_TO_PHYS(pgd_val & SECTION_PAGE_MASK_1GB);
|
|
if (verbose) {
|
|
fprintf(fp, " PAGE: %lx (1GB)\n\n", sectionbase);
|
|
arm64_translate_pte(pgd_val, 0, 0);
|
|
}
|
|
*paddr = sectionbase + (vaddr & ~SECTION_PAGE_MASK_1GB);
|
|
return TRUE;
|
|
}
|
|
/*
|
|
* #define __PAGETABLE_PUD_FOLDED
|
|
*/
|
|
|
|
pmd_base = (ulong *)PTOV(PTE_TO_PHYS(pgd_val));
|
|
FILL_PMD(pmd_base, KVADDR, PTRS_PER_PMD_L3_4K * sizeof(ulong));
|
|
pmd_ptr = pmd_base + (((vaddr) >> PMD_SHIFT_L3_4K) & (PTRS_PER_PMD_L3_4K - 1));
|
|
pmd_val = ULONG(machdep->pmd + PAGEOFFSET(pmd_ptr));
|
|
if (verbose)
|
|
fprintf(fp, " PMD: %lx => %lx\n", (ulong)pmd_ptr, pmd_val);
|
|
if (!pmd_val)
|
|
goto no_page;
|
|
|
|
if ((pmd_val & PMD_TYPE_MASK) == PMD_TYPE_SECT) {
|
|
ulong sectionbase = PTE_TO_PHYS(pmd_val & SECTION_PAGE_MASK_2MB);
|
|
if (verbose) {
|
|
fprintf(fp, " PAGE: %lx (2MB%s)\n\n", sectionbase,
|
|
IS_ZEROPAGE(sectionbase) ? ", ZERO PAGE" : "");
|
|
arm64_translate_pte(pmd_val, 0, 0);
|
|
}
|
|
*paddr = sectionbase + (vaddr & ~SECTION_PAGE_MASK_2MB);
|
|
return TRUE;
|
|
}
|
|
|
|
pte_base = (ulong *)PTOV(PTE_TO_PHYS(pmd_val));
|
|
FILL_PTBL(pte_base, KVADDR, PTRS_PER_PTE_L3_4K * sizeof(ulong));
|
|
pte_ptr = pte_base + (((vaddr) >> machdep->pageshift) & (PTRS_PER_PTE_L3_4K - 1));
|
|
pte_val = ULONG(machdep->ptbl + PAGEOFFSET(pte_ptr));
|
|
if (verbose)
|
|
fprintf(fp, " PTE: %lx => %lx\n", (ulong)pte_ptr, pte_val);
|
|
if (!pte_val)
|
|
goto no_page;
|
|
|
|
if (pte_val & PTE_VALID) {
|
|
*paddr = PTE_TO_PHYS(pte_val) + PAGEOFFSET(vaddr);
|
|
if (verbose) {
|
|
fprintf(fp, " PAGE: %lx %s\n\n", PAGEBASE(*paddr),
|
|
IS_ZEROPAGE(PAGEBASE(*paddr)) ? "(ZERO PAGE)" : "");
|
|
arm64_translate_pte(pte_val, 0, 0);
|
|
}
|
|
} else {
|
|
if (IS_UVADDR(vaddr, NULL))
|
|
*paddr = pte_val;
|
|
if (verbose) {
|
|
fprintf(fp, "\n");
|
|
arm64_translate_pte(pte_val, 0, 0);
|
|
}
|
|
goto no_page;
|
|
}
|
|
|
|
return TRUE;
|
|
no_page:
|
|
return FALSE;
|
|
}
|
|
|
|
static int
|
|
arm64_vtop_4level_4k(ulong pgd, ulong vaddr, physaddr_t *paddr, int verbose)
|
|
{
|
|
ulong *pgd_base, *pgd_ptr, pgd_val;
|
|
ulong *pud_base, *pud_ptr, pud_val;
|
|
ulong *pmd_base, *pmd_ptr, pmd_val;
|
|
ulong *pte_base, *pte_ptr, pte_val;
|
|
|
|
if (verbose)
|
|
fprintf(fp, "PAGE DIRECTORY: %lx\n", pgd);
|
|
|
|
pgd_base = (ulong *)pgd;
|
|
FILL_PGD(pgd_base, KVADDR, PTRS_PER_PGD_L4_4K * sizeof(ulong));
|
|
pgd_ptr = pgd_base + (((vaddr) >> PGDIR_SHIFT_L4_4K) & (PTRS_PER_PGD_L4_4K - 1));
|
|
pgd_val = ULONG(machdep->pgd + PGDIR_OFFSET_48VA(pgd_ptr));
|
|
if (verbose)
|
|
fprintf(fp, " PGD: %lx => %lx\n", (ulong)pgd_ptr, pgd_val);
|
|
if (!pgd_val)
|
|
goto no_page;
|
|
|
|
pud_base = (ulong *)PTOV(PTE_TO_PHYS(pgd_val));
|
|
FILL_PUD(pud_base, KVADDR, PTRS_PER_PUD_L4_4K * sizeof(ulong));
|
|
pud_ptr = pud_base + (((vaddr) >> PUD_SHIFT_L4_4K) & (PTRS_PER_PUD_L4_4K - 1));
|
|
pud_val = ULONG(machdep->pud + PAGEOFFSET(pud_ptr));
|
|
if (verbose)
|
|
fprintf(fp, " PUD: %lx => %lx\n", (ulong)pud_ptr, pud_val);
|
|
if (!pud_val)
|
|
goto no_page;
|
|
|
|
if ((pud_val & PUD_TYPE_MASK) == PUD_TYPE_SECT) {
|
|
ulong sectionbase = PTE_TO_PHYS(pud_val & SECTION_PAGE_MASK_1GB);
|
|
if (verbose) {
|
|
fprintf(fp, " PAGE: %lx (1GB)\n\n", sectionbase);
|
|
arm64_translate_pte(pud_val, 0, 0);
|
|
}
|
|
*paddr = sectionbase + (vaddr & ~SECTION_PAGE_MASK_1GB);
|
|
return TRUE;
|
|
}
|
|
|
|
pmd_base = (ulong *)PTOV(PTE_TO_PHYS(pud_val));
|
|
FILL_PMD(pmd_base, KVADDR, PTRS_PER_PMD_L4_4K * sizeof(ulong));
|
|
pmd_ptr = pmd_base + (((vaddr) >> PMD_SHIFT_L4_4K) & (PTRS_PER_PMD_L4_4K - 1));
|
|
pmd_val = ULONG(machdep->pmd + PAGEOFFSET(pmd_ptr));
|
|
if (verbose)
|
|
fprintf(fp, " PMD: %lx => %lx\n", (ulong)pmd_ptr, pmd_val);
|
|
if (!pmd_val)
|
|
goto no_page;
|
|
|
|
if ((pmd_val & PMD_TYPE_MASK) == PMD_TYPE_SECT) {
|
|
ulong sectionbase = PTE_TO_PHYS(pmd_val & SECTION_PAGE_MASK_2MB);
|
|
if (verbose) {
|
|
fprintf(fp, " PAGE: %lx (2MB%s)\n\n", sectionbase,
|
|
IS_ZEROPAGE(sectionbase) ? ", ZERO PAGE" : "");
|
|
arm64_translate_pte(pmd_val, 0, 0);
|
|
}
|
|
*paddr = sectionbase + (vaddr & ~SECTION_PAGE_MASK_2MB);
|
|
return TRUE;
|
|
}
|
|
|
|
pte_base = (ulong *)PTOV(PTE_TO_PHYS(pmd_val));
|
|
FILL_PTBL(pte_base, KVADDR, PTRS_PER_PTE_L4_4K * sizeof(ulong));
|
|
pte_ptr = pte_base + (((vaddr) >> machdep->pageshift) & (PTRS_PER_PTE_L4_4K - 1));
|
|
pte_val = ULONG(machdep->ptbl + PAGEOFFSET(pte_ptr));
|
|
if (verbose)
|
|
fprintf(fp, " PTE: %lx => %lx\n", (ulong)pte_ptr, pte_val);
|
|
if (!pte_val)
|
|
goto no_page;
|
|
|
|
if (pte_val & PTE_VALID) {
|
|
*paddr = PTE_TO_PHYS(pte_val) + PAGEOFFSET(vaddr);
|
|
if (verbose) {
|
|
fprintf(fp, " PAGE: %lx %s\n\n", PAGEBASE(*paddr),
|
|
IS_ZEROPAGE(PAGEBASE(*paddr)) ? "(ZERO PAGE)" : "");
|
|
arm64_translate_pte(pte_val, 0, 0);
|
|
}
|
|
} else {
|
|
if (IS_UVADDR(vaddr, NULL))
|
|
*paddr = pte_val;
|
|
if (verbose) {
|
|
fprintf(fp, "\n");
|
|
arm64_translate_pte(pte_val, 0, 0);
|
|
}
|
|
goto no_page;
|
|
}
|
|
|
|
return TRUE;
|
|
no_page:
|
|
return FALSE;
|
|
}
|
|
|
|
static ulong
|
|
arm64_get_task_pgd(ulong task)
|
|
{
|
|
struct task_context *tc;
|
|
ulong pgd;
|
|
|
|
if ((tc = task_to_context(task)) &&
|
|
readmem(tc->mm_struct + OFFSET(mm_struct_pgd), KVADDR,
|
|
&pgd, sizeof(long), "user pgd", RETURN_ON_ERROR))
|
|
return pgd;
|
|
else
|
|
return NO_TASK;
|
|
}
|
|
|
|
static ulong
|
|
arm64_processor_speed(void)
|
|
{
|
|
return 0;
|
|
};
|
|
|
|
/*
|
|
* Gather IRQ stack values.
|
|
*/
|
|
static void
|
|
arm64_irq_stack_init(void)
|
|
{
|
|
int i;
|
|
struct syment *sp;
|
|
struct gnu_request request, *req;
|
|
struct machine_specific *ms = machdep->machspec;
|
|
ulong p, sz;
|
|
req = &request;
|
|
|
|
if (symbol_exists("irq_stack") &&
|
|
(sp = per_cpu_symbol_search("irq_stack")) &&
|
|
get_symbol_type("irq_stack", NULL, req)) {
|
|
/* before v4.14 or CONFIG_VMAP_STACK disabled */
|
|
if (CRASHDEBUG(1)) {
|
|
fprintf(fp, "irq_stack: \n");
|
|
fprintf(fp, " type: %s\n",
|
|
(req->typecode == TYPE_CODE_ARRAY) ?
|
|
"TYPE_CODE_ARRAY" : "other");
|
|
fprintf(fp, " target_typecode: %s\n",
|
|
req->target_typecode == TYPE_CODE_INT ?
|
|
"TYPE_CODE_INT" : "other");
|
|
fprintf(fp, " target_length: %ld\n",
|
|
req->target_length);
|
|
fprintf(fp, " length: %ld\n", req->length);
|
|
}
|
|
|
|
if (!(ms->irq_stacks = (ulong *)malloc((size_t)(kt->cpus * sizeof(ulong)))))
|
|
error(FATAL, "cannot malloc irq_stack addresses\n");
|
|
ms->irq_stack_size = req->length;
|
|
machdep->flags |= IRQ_STACKS;
|
|
|
|
for (i = 0; i < kt->cpus; i++)
|
|
ms->irq_stacks[i] = kt->__per_cpu_offset[i] + sp->value;
|
|
} else if (symbol_exists("irq_stack_ptr") &&
|
|
(sp = per_cpu_symbol_search("irq_stack_ptr")) &&
|
|
get_symbol_type("irq_stack_ptr", NULL, req)) {
|
|
/* v4.14 and later with CONFIG_VMAP_STACK enabled */
|
|
if (CRASHDEBUG(1)) {
|
|
fprintf(fp, "irq_stack_ptr: \n");
|
|
fprintf(fp, " type: %x, %s\n",
|
|
(int)req->typecode,
|
|
(req->typecode == TYPE_CODE_PTR) ?
|
|
"TYPE_CODE_PTR" : "other");
|
|
fprintf(fp, " target_typecode: %x, %s\n",
|
|
(int)req->target_typecode,
|
|
req->target_typecode == TYPE_CODE_INT ?
|
|
"TYPE_CODE_INT" : "other");
|
|
fprintf(fp, " target_length: %ld\n",
|
|
req->target_length);
|
|
fprintf(fp, " length: %ld\n", req->length);
|
|
}
|
|
|
|
if (!(ms->irq_stacks = (ulong *)malloc((size_t)(kt->cpus * sizeof(ulong)))))
|
|
error(FATAL, "cannot malloc irq_stack addresses\n");
|
|
|
|
/*
|
|
* Determining the IRQ_STACK_SIZE is tricky, but for now
|
|
* 4.14 kernel has:
|
|
*
|
|
* #define IRQ_STACK_SIZE THREAD_SIZE
|
|
*
|
|
* and finding a solid usage of THREAD_SIZE is hard, but:
|
|
*
|
|
* union thread_union {
|
|
* ...
|
|
* unsigned long stack[THREAD_SIZE/sizeof(long)];
|
|
* };
|
|
*/
|
|
if (MEMBER_EXISTS("thread_union", "stack")) {
|
|
if ((sz = MEMBER_SIZE("thread_union", "stack")) > 0)
|
|
ms->irq_stack_size = sz;
|
|
} else {
|
|
ulong res = arm64_set_irq_stack_size();
|
|
ms->irq_stack_size = (res > 0) ? res : ARM64_IRQ_STACK_SIZE;
|
|
}
|
|
|
|
machdep->flags |= IRQ_STACKS;
|
|
|
|
for (i = 0; i < kt->cpus; i++) {
|
|
p = kt->__per_cpu_offset[i] + sp->value;
|
|
readmem(p, KVADDR, &(ms->irq_stacks[i]), sizeof(ulong),
|
|
"IRQ stack pointer", RETURN_ON_ERROR);
|
|
}
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Gather Overflow stack values.
|
|
*
|
|
* Overflow stack supported since 4.14, in commit 872d8327c
|
|
*/
|
|
static void
|
|
arm64_overflow_stack_init(void)
|
|
{
|
|
int i;
|
|
struct syment *sp;
|
|
struct gnu_request request, *req;
|
|
struct machine_specific *ms = machdep->machspec;
|
|
req = &request;
|
|
|
|
if (symbol_exists("overflow_stack") &&
|
|
(sp = per_cpu_symbol_search("overflow_stack")) &&
|
|
get_symbol_type("overflow_stack", NULL, req)) {
|
|
if (CRASHDEBUG(1)) {
|
|
fprintf(fp, "overflow_stack: \n");
|
|
fprintf(fp, " type: %x, %s\n",
|
|
(int)req->typecode,
|
|
(req->typecode == TYPE_CODE_ARRAY) ?
|
|
"TYPE_CODE_ARRAY" : "other");
|
|
fprintf(fp, " target_typecode: %x, %s\n",
|
|
(int)req->target_typecode,
|
|
req->target_typecode == TYPE_CODE_INT ?
|
|
"TYPE_CODE_INT" : "other");
|
|
fprintf(fp, " target_length: %ld\n",
|
|
req->target_length);
|
|
fprintf(fp, " length: %ld\n", req->length);
|
|
}
|
|
|
|
if (!(ms->overflow_stacks = (ulong *)malloc((size_t)(kt->cpus * sizeof(ulong)))))
|
|
error(FATAL, "cannot malloc overflow_stack addresses\n");
|
|
|
|
ms->overflow_stack_size = ARM64_OVERFLOW_STACK_SIZE;
|
|
machdep->flags |= OVERFLOW_STACKS;
|
|
|
|
for (i = 0; i < kt->cpus; i++)
|
|
ms->overflow_stacks[i] = kt->__per_cpu_offset[i] + sp->value;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Gather and verify all of the backtrace requirements.
|
|
*/
|
|
static void
|
|
arm64_stackframe_init(void)
|
|
{
|
|
long task_struct_thread;
|
|
long thread_struct_cpu_context;
|
|
long context_sp, context_pc, context_fp;
|
|
struct syment *sp1, *sp1n, *sp2, *sp2n, *sp3, *sp3n;
|
|
|
|
STRUCT_SIZE_INIT(note_buf, "note_buf_t");
|
|
STRUCT_SIZE_INIT(elf_prstatus, "elf_prstatus");
|
|
MEMBER_OFFSET_INIT(elf_prstatus_pr_pid, "elf_prstatus", "pr_pid");
|
|
MEMBER_OFFSET_INIT(elf_prstatus_pr_reg, "elf_prstatus", "pr_reg");
|
|
|
|
if (MEMBER_EXISTS("pt_regs", "stackframe")) {
|
|
machdep->machspec->user_eframe_offset = SIZE(pt_regs);
|
|
machdep->machspec->kern_eframe_offset = SIZE(pt_regs) - 16;
|
|
} else {
|
|
machdep->machspec->user_eframe_offset = SIZE(pt_regs) + 16;
|
|
machdep->machspec->kern_eframe_offset = SIZE(pt_regs);
|
|
}
|
|
|
|
if ((sp1 = kernel_symbol_search("__exception_text_start")) &&
|
|
(sp2 = kernel_symbol_search("__exception_text_end"))) {
|
|
machdep->machspec->__exception_text_start = sp1->value;
|
|
machdep->machspec->__exception_text_end = sp2->value;
|
|
}
|
|
if ((sp1 = kernel_symbol_search("__irqentry_text_start")) &&
|
|
(sp2 = kernel_symbol_search("__irqentry_text_end"))) {
|
|
machdep->machspec->__irqentry_text_start = sp1->value;
|
|
machdep->machspec->__irqentry_text_end = sp2->value;
|
|
}
|
|
if ((sp1 = kernel_symbol_search("vectors")) &&
|
|
(sp1n = kernel_symbol_search("cpu_switch_to")) &&
|
|
(sp2 = kernel_symbol_search("ret_fast_syscall")) &&
|
|
(sp2n = kernel_symbol_search("sys_rt_sigreturn_wrapper"))) {
|
|
machdep->machspec->exp_entry1_start = sp1->value;
|
|
machdep->machspec->exp_entry1_end = sp1n->value;
|
|
machdep->machspec->exp_entry2_start = sp2->value;
|
|
machdep->machspec->exp_entry2_end = sp2n->value;
|
|
}
|
|
|
|
if ((sp1 = kernel_symbol_search("crash_kexec")) &&
|
|
(sp1n = next_symbol(NULL, sp1)) &&
|
|
(sp2 = kernel_symbol_search("crash_save_cpu")) &&
|
|
(sp2n = next_symbol(NULL, sp2)) &&
|
|
(sp3 = kernel_symbol_search("machine_kexec")) &&
|
|
(sp3n = next_symbol(NULL, sp3))) {
|
|
machdep->machspec->crash_kexec_start = sp1->value;
|
|
machdep->machspec->crash_kexec_end = sp1n->value;
|
|
machdep->machspec->crash_save_cpu_start = sp2->value;
|
|
machdep->machspec->crash_save_cpu_end = sp2n->value;
|
|
machdep->machspec->machine_kexec_start = sp3->value;
|
|
machdep->machspec->machine_kexec_end = sp3n->value;
|
|
machdep->flags |= KDUMP_ENABLED;
|
|
}
|
|
|
|
task_struct_thread = MEMBER_OFFSET("task_struct", "thread");
|
|
thread_struct_cpu_context = MEMBER_OFFSET("thread_struct", "cpu_context");
|
|
|
|
if ((task_struct_thread == INVALID_OFFSET) ||
|
|
(thread_struct_cpu_context == INVALID_OFFSET)) {
|
|
error(INFO,
|
|
"cannot determine task_struct.thread.context offset\n");
|
|
return;
|
|
}
|
|
|
|
/*
|
|
* Pay for the convenience of using a hardcopy of a kernel structure.
|
|
*/
|
|
if (offsetof(struct arm64_stackframe, sp) !=
|
|
MEMBER_OFFSET("stackframe", "sp")) {
|
|
if (CRASHDEBUG(1))
|
|
error(INFO, "builtin stackframe.sp offset differs from kernel version\n");
|
|
}
|
|
if (offsetof(struct arm64_stackframe, fp) !=
|
|
MEMBER_OFFSET("stackframe", "fp")) {
|
|
if (CRASHDEBUG(1))
|
|
error(INFO, "builtin stackframe.fp offset differs from kernel version\n");
|
|
}
|
|
if (offsetof(struct arm64_stackframe, pc) !=
|
|
MEMBER_OFFSET("stackframe", "pc")) {
|
|
if (CRASHDEBUG(1))
|
|
error(INFO, "builtin stackframe.pc offset differs from kernel version\n");
|
|
}
|
|
if (!MEMBER_EXISTS("stackframe", "sp"))
|
|
machdep->flags |= UNW_4_14;
|
|
|
|
context_sp = MEMBER_OFFSET("cpu_context", "sp");
|
|
context_fp = MEMBER_OFFSET("cpu_context", "fp");
|
|
context_pc = MEMBER_OFFSET("cpu_context", "pc");
|
|
if (context_sp == INVALID_OFFSET) {
|
|
error(INFO, "cannot determine cpu_context.sp offset\n");
|
|
return;
|
|
}
|
|
if (context_fp == INVALID_OFFSET) {
|
|
error(INFO, "cannot determine cpu_context.fp offset\n");
|
|
return;
|
|
}
|
|
if (context_pc == INVALID_OFFSET) {
|
|
error(INFO, "cannot determine cpu_context.pc offset\n");
|
|
return;
|
|
}
|
|
ASSIGN_OFFSET(task_struct_thread_context_sp) =
|
|
task_struct_thread + thread_struct_cpu_context + context_sp;
|
|
ASSIGN_OFFSET(task_struct_thread_context_fp) =
|
|
task_struct_thread + thread_struct_cpu_context + context_fp;
|
|
ASSIGN_OFFSET(task_struct_thread_context_pc) =
|
|
task_struct_thread + thread_struct_cpu_context + context_pc;
|
|
}
|
|
|
|
#define KERNEL_MODE (1)
|
|
#define USER_MODE (2)
|
|
|
|
#define USER_EFRAME_OFFSET (machdep->machspec->user_eframe_offset)
|
|
#define KERN_EFRAME_OFFSET (machdep->machspec->kern_eframe_offset)
|
|
|
|
/*
|
|
* PSR bits
|
|
*/
|
|
#define PSR_MODE_EL0t 0x00000000
|
|
#define PSR_MODE_EL1t 0x00000004
|
|
#define PSR_MODE_EL1h 0x00000005
|
|
#define PSR_MODE_EL2t 0x00000008
|
|
#define PSR_MODE_EL2h 0x00000009
|
|
#define PSR_MODE_EL3t 0x0000000c
|
|
#define PSR_MODE_EL3h 0x0000000d
|
|
#define PSR_MODE_MASK 0x0000000f
|
|
|
|
/* Architecturally defined mapping between AArch32 and AArch64 registers */
|
|
#define compat_usr(x) regs[(x)]
|
|
#define compat_fp regs[11]
|
|
#define compat_sp regs[13]
|
|
#define compat_lr regs[14]
|
|
|
|
#define user_mode(ptregs) \
|
|
(((ptregs)->pstate & PSR_MODE_MASK) == PSR_MODE_EL0t)
|
|
|
|
#define compat_user_mode(ptregs) \
|
|
(((ptregs)->pstate & (PSR_MODE32_BIT | PSR_MODE_MASK)) == \
|
|
(PSR_MODE32_BIT | PSR_MODE_EL0t))
|
|
|
|
#define user_stack_pointer(ptregs) \
|
|
(!compat_user_mode(ptregs) ? (ptregs)->sp : (ptregs)->compat_sp)
|
|
|
|
#define user_frame_pointer(ptregs) \
|
|
(!compat_user_mode(ptregs) ? (ptregs)->regs[29] : (ptregs)->compat_fp)
|
|
|
|
static int
|
|
arm64_is_kernel_exception_frame(struct bt_info *bt, ulong stkptr)
|
|
{
|
|
struct arm64_pt_regs *regs;
|
|
struct machine_specific *ms = machdep->machspec;
|
|
|
|
if (stkptr > STACKSIZE() && !INSTACK(stkptr, bt)) {
|
|
if (CRASHDEBUG(1))
|
|
error(WARNING, "stkptr: %lx is outside the kernel stack range\n", stkptr);
|
|
return FALSE;
|
|
}
|
|
|
|
regs = (struct arm64_pt_regs *)&bt->stackbuf[(ulong)(STACK_OFFSET_TYPE(stkptr))];
|
|
|
|
if (INSTACK(regs->sp, bt) && INSTACK(regs->regs[29], bt) &&
|
|
!(regs->pstate & (0xffffffff00000000ULL | PSR_MODE32_BIT)) &&
|
|
is_kernel_text(regs->pc) &&
|
|
is_kernel_text(regs->regs[30] | ms->CONFIG_ARM64_KERNELPACMASK)) {
|
|
switch (regs->pstate & PSR_MODE_MASK)
|
|
{
|
|
case PSR_MODE_EL1t:
|
|
case PSR_MODE_EL1h:
|
|
case PSR_MODE_EL2t:
|
|
case PSR_MODE_EL2h:
|
|
return TRUE;
|
|
}
|
|
}
|
|
|
|
return FALSE;
|
|
}
|
|
|
|
static int
|
|
arm64_eframe_search(struct bt_info *bt)
|
|
{
|
|
int c;
|
|
ulong ptr, count;
|
|
struct machine_specific *ms;
|
|
|
|
if (bt->flags & BT_EFRAME_SEARCH2) {
|
|
if (!(machdep->flags & IRQ_STACKS))
|
|
error(FATAL, "IRQ stacks do not exist in this kernel\n");
|
|
|
|
ms = machdep->machspec;
|
|
|
|
for (c = 0; c < kt->cpus; c++) {
|
|
if ((bt->flags & BT_CPUMASK) &&
|
|
!(NUM_IN_BITMAP(bt->cpumask, c)))
|
|
continue;
|
|
|
|
fprintf(fp, "CPU %d IRQ STACK:", c);
|
|
bt->stackbase = ms->irq_stacks[c];
|
|
bt->stacktop = bt->stackbase + ms->irq_stack_size;
|
|
alter_stackbuf(bt);
|
|
count = 0;
|
|
|
|
for (ptr = bt->stackbase; ptr < bt->stacktop - SIZE(pt_regs); ptr++) {
|
|
if (arm64_is_kernel_exception_frame(bt, ptr)) {
|
|
fprintf(fp, "%s\nKERNEL-MODE EXCEPTION FRAME AT: %lx\n",
|
|
count ? "" : "\n", ptr);
|
|
arm64_print_exception_frame(bt, ptr, KERNEL_MODE, fp);
|
|
count++;
|
|
}
|
|
}
|
|
|
|
if (count)
|
|
fprintf(fp, "\n");
|
|
else
|
|
fprintf(fp, "(none found)\n\n");
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
|
|
count = 0;
|
|
for (ptr = bt->stackbase; ptr < bt->stacktop - SIZE(pt_regs); ptr++) {
|
|
if (arm64_is_kernel_exception_frame(bt, ptr)) {
|
|
fprintf(fp, "\nKERNEL-MODE EXCEPTION FRAME AT: %lx\n", ptr);
|
|
arm64_print_exception_frame(bt, ptr, KERNEL_MODE, fp);
|
|
count++;
|
|
}
|
|
}
|
|
|
|
if (is_kernel_thread(bt->tc->task))
|
|
return count;
|
|
|
|
ptr = bt->stacktop - USER_EFRAME_OFFSET;
|
|
fprintf(fp, "%sUSER-MODE EXCEPTION FRAME AT: %lx\n",
|
|
count++ ? "\n" : "", ptr);
|
|
arm64_print_exception_frame(bt, ptr, USER_MODE, fp);
|
|
|
|
return count;
|
|
}
|
|
|
|
static char *arm64_exception_functions[] = {
|
|
"do_undefinstr",
|
|
"do_sysinstr",
|
|
"do_debug_exception",
|
|
"do_mem_abort",
|
|
"do_el0_irq_bp_hardening",
|
|
"do_sp_pc_abort",
|
|
"handle_bad_stack",
|
|
NULL
|
|
};
|
|
|
|
static int
|
|
arm64_in_exception_text(ulong ptr)
|
|
{
|
|
struct machine_specific *ms = machdep->machspec;
|
|
char *name, **func;
|
|
|
|
if (ms->__irqentry_text_start && ms->__irqentry_text_end &&
|
|
((ptr >= ms->__irqentry_text_start) &&
|
|
(ptr < ms->__irqentry_text_end)))
|
|
return TRUE;
|
|
|
|
if (ms->__exception_text_start && ms->__exception_text_end) {
|
|
if ((ptr >= ms->__exception_text_start) &&
|
|
(ptr < ms->__exception_text_end))
|
|
return TRUE;
|
|
}
|
|
|
|
name = closest_symbol(ptr);
|
|
if (name != NULL) { /* Linux 5.5 and later */
|
|
for (func = &arm64_exception_functions[0]; *func; func++) {
|
|
if (STREQ(name, *func))
|
|
return TRUE;
|
|
}
|
|
}
|
|
|
|
return FALSE;
|
|
}
|
|
|
|
static int
|
|
arm64_in_exp_entry(ulong addr)
|
|
{
|
|
struct machine_specific *ms;
|
|
|
|
ms = machdep->machspec;
|
|
if ((ms->exp_entry1_start <= addr) && (addr < ms->exp_entry1_end))
|
|
return TRUE;
|
|
if ((ms->exp_entry2_start <= addr) && (addr < ms->exp_entry2_end))
|
|
return TRUE;
|
|
return FALSE;
|
|
}
|
|
|
|
#define BACKTRACE_CONTINUE (1)
|
|
#define BACKTRACE_COMPLETE_KERNEL (2)
|
|
#define BACKTRACE_COMPLETE_USER (3)
|
|
|
|
static int
|
|
arm64_print_stackframe_entry(struct bt_info *bt, int level, struct arm64_stackframe *frame, FILE *ofp)
|
|
{
|
|
char *name, *name_plus_offset;
|
|
ulong branch_pc, symbol_offset;
|
|
struct syment *sp;
|
|
struct load_module *lm;
|
|
char buf[BUFSIZE];
|
|
|
|
/*
|
|
* if pc comes from a saved lr, it actually points to an instruction
|
|
* after branch. To avoid any confusion, decrement pc by 4.
|
|
* See, for example, "bl schedule" before ret_to_user().
|
|
*/
|
|
branch_pc = frame->pc - 4;
|
|
|
|
name = closest_symbol(branch_pc);
|
|
name_plus_offset = NULL;
|
|
|
|
if (bt->flags & BT_SYMBOL_OFFSET) {
|
|
sp = value_search(branch_pc, &symbol_offset);
|
|
if (sp && symbol_offset)
|
|
name_plus_offset =
|
|
value_to_symstr(branch_pc, buf, bt->radix);
|
|
}
|
|
|
|
if (!INSTACK(frame->fp, bt) && IN_TASK_VMA(bt->task, frame->fp))
|
|
frame->fp = 0;
|
|
|
|
if (bt->flags & BT_FULL) {
|
|
if (level)
|
|
arm64_display_full_frame(bt, frame->fp);
|
|
bt->frameptr = frame->fp;
|
|
}
|
|
|
|
fprintf(ofp, "%s#%d [%8lx] %s at %lx", level < 10 ? " " : "", level,
|
|
frame->fp ? frame->fp : bt->stacktop - USER_EFRAME_OFFSET,
|
|
name_plus_offset ? name_plus_offset : name, branch_pc);
|
|
|
|
if (BT_REFERENCE_CHECK(bt)) {
|
|
arm64_do_bt_reference_check(bt, frame->pc, closest_symbol(frame->pc));
|
|
arm64_do_bt_reference_check(bt, branch_pc, name);
|
|
}
|
|
|
|
if (module_symbol(branch_pc, NULL, &lm, NULL, 0))
|
|
fprintf(ofp, " [%s]", lm->mod_name);
|
|
|
|
fprintf(ofp, "\n");
|
|
|
|
if (bt->flags & BT_LINE_NUMBERS) {
|
|
get_line_number(branch_pc, buf, FALSE);
|
|
if (strlen(buf))
|
|
fprintf(ofp, " %s\n", buf);
|
|
}
|
|
|
|
if (STREQ(name, "start_kernel") || STREQ(name, "secondary_start_kernel") ||
|
|
STREQ(name, "kthread") || STREQ(name, "kthreadd"))
|
|
return BACKTRACE_COMPLETE_KERNEL;
|
|
|
|
return BACKTRACE_CONTINUE;
|
|
}
|
|
|
|
static int
|
|
arm64_print_stackframe_entry_v2(struct bt_info *bt, int level, struct arm64_stackframe *frame, FILE *ofp)
|
|
{
|
|
char *name, *name_plus_offset;
|
|
ulong pc, symbol_offset;
|
|
struct syment *sp;
|
|
struct load_module *lm;
|
|
char buf[BUFSIZE];
|
|
|
|
/*
|
|
* if pc comes from a saved lr, it actually points to an instruction
|
|
* after branch. To avoid any confusion, decrement pc by 4.
|
|
* See, for example, "bl schedule" before ret_to_user().
|
|
*/
|
|
pc = frame->pc - 0x4;
|
|
name = closest_symbol(pc);
|
|
name_plus_offset = NULL;
|
|
|
|
if (bt->flags & BT_SYMBOL_OFFSET) {
|
|
sp = value_search(pc, &symbol_offset);
|
|
if (sp && symbol_offset)
|
|
name_plus_offset = value_to_symstr(pc, buf, bt->radix);
|
|
}
|
|
|
|
if (bt->flags & BT_USER_EFRAME)
|
|
frame->fp = 0;
|
|
|
|
fprintf(ofp, "%s#%d [%8lx] %s at %lx", level < 10 ? " " : "", level,
|
|
frame->fp ? frame->fp : bt->stacktop - USER_EFRAME_OFFSET,
|
|
name_plus_offset ? name_plus_offset : name, pc);
|
|
|
|
if (BT_REFERENCE_CHECK(bt))
|
|
arm64_do_bt_reference_check(bt, pc, name);
|
|
|
|
if (module_symbol(pc, NULL, &lm, NULL, 0))
|
|
fprintf(ofp, " [%s]", lm->mod_name);
|
|
|
|
fprintf(ofp, "\n");
|
|
|
|
if (bt->flags & BT_LINE_NUMBERS) {
|
|
get_line_number(pc, buf, FALSE);
|
|
if (strlen(buf))
|
|
fprintf(ofp, " %s\n", buf);
|
|
}
|
|
|
|
if (STREQ(name, "start_kernel") ||
|
|
STREQ(name, "secondary_start_kernel") ||
|
|
STREQ(name, "kthread") || STREQ(name, "kthreadd"))
|
|
return BACKTRACE_COMPLETE_KERNEL;
|
|
|
|
return BACKTRACE_CONTINUE;
|
|
}
|
|
|
|
static void
|
|
arm64_display_full_frame(struct bt_info *bt, ulong sp)
|
|
{
|
|
int i, u_idx;
|
|
ulong *up;
|
|
ulong words, addr;
|
|
char buf[BUFSIZE];
|
|
|
|
if (bt->frameptr == sp)
|
|
return;
|
|
|
|
if (INSTACK(bt->frameptr, bt)) {
|
|
if (INSTACK(sp, bt)) {
|
|
; /* normal case */
|
|
} else {
|
|
if (sp == 0)
|
|
/* interrupt in user mode */
|
|
sp = bt->stacktop - USER_EFRAME_OFFSET;
|
|
else
|
|
/* interrupt in kernel mode */
|
|
sp = bt->stacktop;
|
|
}
|
|
} else {
|
|
/* This is a transition case from irq to process stack. */
|
|
return;
|
|
}
|
|
|
|
words = (sp - bt->frameptr) / sizeof(ulong);
|
|
|
|
addr = bt->frameptr;
|
|
u_idx = (bt->frameptr - bt->stackbase)/sizeof(ulong);
|
|
for (i = 0; i < words; i++, u_idx++) {
|
|
if (!(i & 1))
|
|
fprintf(fp, "%s %lx: ", i ? "\n" : "", addr);
|
|
|
|
up = (ulong *)(&bt->stackbuf[u_idx*sizeof(ulong)]);
|
|
fprintf(fp, "%s ", format_stack_entry(bt, buf, *up, 0));
|
|
|
|
addr += sizeof(ulong);
|
|
}
|
|
fprintf(fp, "\n");
|
|
}
|
|
|
|
static void
|
|
arm64_display_full_frame_v2(struct bt_info *bt, struct arm64_stackframe *cur,
|
|
struct arm64_stackframe *next)
|
|
{
|
|
struct machine_specific *ms;
|
|
ulong next_fp, stackbase;
|
|
char *stackbuf;
|
|
int i, u_idx;
|
|
ulong *up;
|
|
ulong words, addr;
|
|
char buf[BUFSIZE];
|
|
|
|
stackbase = bt->stackbase;
|
|
stackbuf = bt->stackbuf;
|
|
ms = machdep->machspec;
|
|
|
|
/* Calc next fp for dump */
|
|
if (next->fp == 0)
|
|
/* last stackframe on kernel tack */
|
|
next_fp = bt->stacktop - 0x10;
|
|
else if (!INSTACK(cur->sp, bt)) {
|
|
/* We have just switched over stacks */
|
|
next_fp = ms->irq_stacks[bt->tc->processor]
|
|
+ ms->irq_stack_size - 0x10;
|
|
|
|
/*
|
|
* We are already buffering a process stack.
|
|
* So use an old buffer for IRQ stack.
|
|
*/
|
|
stackbase = ms->irq_stacks[bt->tc->processor];
|
|
stackbuf = ms->irq_stackbuf;
|
|
} else
|
|
next_fp = next->fp;
|
|
|
|
if (CRASHDEBUG(1))
|
|
fprintf(fp, " frame <%016lx:%016lx>\n", cur->fp, next_fp);
|
|
|
|
/* Check here because we want to see a debug message above. */
|
|
if (!(bt->flags & BT_FULL))
|
|
return;
|
|
if (next_fp <= cur->fp)
|
|
return;
|
|
|
|
/* Dump */
|
|
words = (next_fp - cur->fp) / sizeof(ulong);
|
|
addr = cur->fp;
|
|
u_idx = (cur->fp - stackbase)/sizeof(ulong);
|
|
for (i = 0; i < words; i++, u_idx++) {
|
|
if (!(i & 1))
|
|
fprintf(fp, "%s %lx: ", i ? "\n" : "", addr);
|
|
|
|
up = (ulong *)(&stackbuf[u_idx*sizeof(ulong)]);
|
|
fprintf(fp, "%s ", format_stack_entry(bt, buf, *up, 0));
|
|
|
|
addr += sizeof(ulong);
|
|
}
|
|
fprintf(fp, "\n");
|
|
|
|
if (stackbuf == ms->irq_stackbuf)
|
|
FREEBUF(stackbuf);
|
|
}
|
|
|
|
static int
|
|
arm64_unwind_frame(struct bt_info *bt, struct arm64_stackframe *frame)
|
|
{
|
|
unsigned long high, low, fp;
|
|
unsigned long stack_mask;
|
|
unsigned long irq_stack_ptr, orig_sp;
|
|
struct arm64_pt_regs *ptregs;
|
|
struct machine_specific *ms = machdep->machspec;
|
|
|
|
stack_mask = (unsigned long)(ARM64_STACK_SIZE) - 1;
|
|
fp = frame->fp;
|
|
|
|
low = frame->sp;
|
|
high = (low + stack_mask) & ~(stack_mask);
|
|
|
|
if (fp < low || fp > high || fp & 0xf || !INSTACK(fp, bt))
|
|
return FALSE;
|
|
|
|
frame->sp = fp + 0x10;
|
|
frame->fp = GET_STACK_ULONG(fp);
|
|
frame->pc = GET_STACK_ULONG(fp + 8);
|
|
if (is_kernel_text(frame->pc | ms->CONFIG_ARM64_KERNELPACMASK))
|
|
frame->pc |= ms->CONFIG_ARM64_KERNELPACMASK;
|
|
|
|
if ((frame->fp == 0) && (frame->pc == 0))
|
|
return FALSE;
|
|
|
|
if (!(machdep->flags & (IRQ_STACKS | OVERFLOW_STACKS)))
|
|
return TRUE;
|
|
|
|
if (machdep->flags & UNW_4_14) {
|
|
if (((bt->flags & BT_IRQSTACK) &&
|
|
!arm64_on_irq_stack(bt->tc->processor, frame->fp)) ||
|
|
((bt->flags & BT_OVERFLOW_STACK) &&
|
|
!arm64_on_overflow_stack(bt->tc->processor, frame->fp))) {
|
|
if (arm64_on_process_stack(bt, frame->fp)) {
|
|
arm64_set_process_stack(bt);
|
|
|
|
frame->sp = frame->fp - KERN_EFRAME_OFFSET;
|
|
/*
|
|
* for switch_stack
|
|
* fp still points to irq stack
|
|
*/
|
|
bt->bptr = fp;
|
|
/*
|
|
* for display_full_frame
|
|
* sp points to process stack
|
|
*
|
|
* If we want to see pt_regs,
|
|
* comment out the below.
|
|
* bt->frameptr = frame->sp;
|
|
*/
|
|
} else {
|
|
/* irq -> user */
|
|
return FALSE;
|
|
}
|
|
}
|
|
|
|
return TRUE;
|
|
}
|
|
|
|
/*
|
|
* The kernel's manner of determining the end of the IRQ stack:
|
|
*
|
|
* #define THREAD_SIZE 16384
|
|
* #define THREAD_START_SP (THREAD_SIZE - 16)
|
|
* #define IRQ_STACK_START_SP THREAD_START_SP
|
|
* #define IRQ_STACK_PTR(cpu) ((unsigned long)per_cpu(irq_stack, cpu) + IRQ_STACK_START_SP)
|
|
* #define IRQ_STACK_TO_TASK_STACK(ptr) (*((unsigned long *)((ptr) - 0x08)))
|
|
*
|
|
* irq_stack_ptr = IRQ_STACK_PTR(raw_smp_processor_id());
|
|
* orig_sp = IRQ_STACK_TO_TASK_STACK(irq_stack_ptr); (pt_regs pointer on process stack)
|
|
*/
|
|
irq_stack_ptr = ms->irq_stacks[bt->tc->processor] + ms->irq_stack_size - 16;
|
|
|
|
if (frame->sp == irq_stack_ptr) {
|
|
orig_sp = GET_STACK_ULONG(irq_stack_ptr - 8);
|
|
arm64_set_process_stack(bt);
|
|
if (INSTACK(orig_sp, bt) && (INSTACK(frame->fp, bt) || (frame->fp == 0))) {
|
|
ptregs = (struct arm64_pt_regs *)&bt->stackbuf[(ulong)(STACK_OFFSET_TYPE(orig_sp))];
|
|
frame->sp = orig_sp;
|
|
frame->pc = ptregs->pc;
|
|
bt->bptr = fp;
|
|
if (CRASHDEBUG(1))
|
|
error(INFO,
|
|
"arm64_unwind_frame: switch stacks: fp: %lx sp: %lx pc: %lx\n",
|
|
frame->fp, frame->sp, frame->pc);
|
|
} else {
|
|
error(WARNING,
|
|
"arm64_unwind_frame: on IRQ stack: oriq_sp: %lx%s fp: %lx%s\n",
|
|
orig_sp, INSTACK(orig_sp, bt) ? "" : " (?)",
|
|
frame->fp, INSTACK(frame->fp, bt) ? "" : " (?)");
|
|
return FALSE;
|
|
}
|
|
}
|
|
|
|
return TRUE;
|
|
}
|
|
|
|
/*
|
|
* The following figure shows how unwinding can be done.
|
|
* Here we assume that the callstack order is:
|
|
* #(X-1) ppc (previous PC)
|
|
* #X cpc (current PC)
|
|
* < #(X+ 1) epc (Exception entry) >
|
|
* #(X+1/2) npc (Next PC)
|
|
* #(X+2/3) Npc (One before Next)
|
|
* #(X+3/4) NNpc (One before 'Npc')
|
|
* and unwind frames from #X to #(X+1).
|
|
* When we add a faked frame for exception entry (exception frame)
|
|
* as #(X+1), the next frame for npc will be recognized as #(x+2).
|
|
*
|
|
* (1)Normal stackframe:
|
|
* +------+
|
|
* | pfp |
|
|
* | cpc |
|
|
* psp + +
|
|
* | |
|
|
* | |
|
|
* pfp +------+ <--- :prev stackframe = <pfp, psp, ppc>
|
|
* | cfp |
|
|
* | npc |
|
|
* csp + +
|
|
* | |
|
|
* | |
|
|
* cfp +------+ <--- :curr stackframe = <cfp, csp, cpc>
|
|
* | nfp | cfp = *pfp
|
|
* | Npc | csp = pfp + 0x10
|
|
* nsp + +
|
|
* | |
|
|
* | |
|
|
* nfp +------+ <--- :next stackframe = <nfp, nsp, npc>
|
|
* | |
|
|
*
|
|
* (2)Exception on the same (IRQ or process) stack:
|
|
* +------+
|
|
* | pfp |
|
|
* | cpc |
|
|
* psp + +
|
|
* | |
|
|
* | |
|
|
* pfp +------+ <--- :prev stackframe = <pfp, psp, ppc>
|
|
* | cfp |
|
|
* | npc |
|
|
* csp + +
|
|
* | |
|
|
* | |
|
|
* cfp +------+ <--- :curr stackframe = <cfp, csp, cpc>
|
|
* | nfp |
|
|
* | epc |
|
|
* + +
|
|
* | |
|
|
* | | faked(*)
|
|
* esp +------+ <--- :excp stackframe = <---, esp, epc
|
|
* | | esp = nsp - sizeof(pt_regs)
|
|
* | |
|
|
* | Npc | (*) If we didn't add this frame, the next frame
|
|
* | nfp | would be
|
|
* | nsp | <nfp, nfp + 0x10, epc>
|
|
* | npc | and the frame below for npc would be lost.
|
|
* nsp + +
|
|
* | |
|
|
* nfp +------+ <--- :task stackframe = <nfp, nsp, npc>
|
|
* | Nfp |
|
|
* | NNpc |
|
|
* Nsp + +
|
|
* | |
|
|
* Nfp +------+ <--- :task stackframe = <Nfp, Nsp, Npc>
|
|
* | NNfp |
|
|
*
|
|
* (3)Interrupt:
|
|
* +------+
|
|
* | cfp |
|
|
* | ipc |
|
|
* csp + +
|
|
* | |
|
|
* | |
|
|
* cfp +------+ <--- :curr stackframe = <cfp, csp, cpc>
|
|
* | ifp |
|
|
* | epc |
|
|
* isp + +
|
|
* | |
|
|
* | | (*)
|
|
* ifp +------+ <--- :irq stackframe = <ifp, isp, epc>
|
|
* | nfp | ifp == IRQ_STACK_PTR
|
|
* | esp | (*) Before the kernel enters an irq handler, frame
|
|
* top +------+ pointer moves to the top of IRQ stack.
|
|
* IRQ stack So we have to skip this frame in unwinding.
|
|
*
|
|
* faked
|
|
* esp +------+ <--- :excp stackframe = <---, esp, epc>
|
|
* | | esp = nsp - sizeof(pt_regs)
|
|
* | |
|
|
* | Npc |
|
|
* | nfp |
|
|
* | nsp |
|
|
* | npc |
|
|
* nsp + +
|
|
* | |
|
|
* nfp +------+ <--- :task stackframe = <nfp, nsp, npc>
|
|
* | Nfp |
|
|
* | NNpc |
|
|
* Nsp + +
|
|
* | |
|
|
* Nfp +------+ <--- :task stackframe = <Nfp, Nsp, Npc>
|
|
* | NNfp |
|
|
*/
|
|
|
|
static struct arm64_stackframe ext_frame;
|
|
|
|
static int
|
|
arm64_unwind_frame_v2(struct bt_info *bt, struct arm64_stackframe *frame,
|
|
FILE *ofp)
|
|
{
|
|
unsigned long high, low, fp;
|
|
unsigned long stack_mask;
|
|
unsigned long irq_stack_ptr;
|
|
struct machine_specific *ms;
|
|
|
|
stack_mask = (unsigned long)(ARM64_STACK_SIZE) - 1;
|
|
fp = frame->fp;
|
|
|
|
low = frame->sp;
|
|
high = (low + stack_mask) & ~(stack_mask);
|
|
|
|
if (fp < low || fp > high || fp & 0xf || !INSTACK(fp, bt))
|
|
return FALSE;
|
|
|
|
if (CRASHDEBUG(1))
|
|
fprintf(ofp, " cur fp:%016lx sp:%016lx pc:%016lx\n",
|
|
frame->fp, frame->sp, frame->pc);
|
|
|
|
if (ext_frame.pc) {
|
|
/*
|
|
* The previous frame was a dummy for exception entry.
|
|
* So complement a missing (task) stackframe now.
|
|
*/
|
|
frame->fp = ext_frame.fp;
|
|
frame->sp = ext_frame.sp;
|
|
frame->pc = ext_frame.pc;
|
|
|
|
ext_frame.pc = 0; /* back to normal unwinding */
|
|
|
|
goto unwind_done;
|
|
}
|
|
|
|
frame->pc = GET_STACK_ULONG(fp + 8);
|
|
if (!arm64_in_exp_entry(frame->pc)) {
|
|
/* (1) Normal stack frame */
|
|
|
|
frame->sp = fp + 0x10;
|
|
frame->fp = GET_STACK_ULONG(fp);
|
|
} else {
|
|
/*
|
|
* We are in exception entry code, and so
|
|
* - add a faked frame for exception entry, and
|
|
* - prepare for a stackframe hidden by exception
|
|
*/
|
|
|
|
ext_frame.fp = GET_STACK_ULONG(fp);
|
|
/*
|
|
* Note:
|
|
* In the following code, we determine a stack pointer for
|
|
* exception entry based on ext_frame.fp because we have
|
|
* no way to know a ext_frame.sp.
|
|
* Fortunately, this will work fine for most functions
|
|
* in the kernel.
|
|
*/
|
|
if (ext_frame.fp == 0) {
|
|
/*
|
|
* (2)
|
|
* Either on process stack or on IRQ stack,
|
|
* the next frame is the last one on process stack.
|
|
*/
|
|
|
|
frame->sp = bt->stacktop
|
|
- sizeof(struct arm64_pt_regs) - 0x10;
|
|
frame->fp = frame->sp;
|
|
} else if (!arm64_on_irq_stack(bt->tc->processor, frame->sp)) {
|
|
/*
|
|
* (2)
|
|
* We are on process stack. Just add a faked frame
|
|
*/
|
|
|
|
if (!arm64_on_irq_stack(bt->tc->processor, ext_frame.fp))
|
|
frame->sp = ext_frame.fp
|
|
- sizeof(struct arm64_pt_regs);
|
|
else {
|
|
/*
|
|
* FIXME: very exceptional case
|
|
* We are already back on process stack, but
|
|
* a saved frame pointer indicates that we are
|
|
* on IRQ stack. Unfortunately this can happen
|
|
* when some functions are called after
|
|
* an irq handler is done because irq_exit()
|
|
* doesn't restore a frame pointer (x29).
|
|
* Those functions include
|
|
* - do_notify_resume()
|
|
* - trace_hardirqs_off()
|
|
* - schedule()
|
|
*
|
|
* We have no perfect way to determine a true
|
|
* stack pointer value here.
|
|
* 0x20 is a stackframe size of schedule().
|
|
* Really ugly
|
|
*/
|
|
frame->sp = frame->fp + 0x20;
|
|
fprintf(ofp, " (Next exception frame might be wrong)\n");
|
|
}
|
|
|
|
frame->fp = frame->sp;
|
|
} else {
|
|
/* We are on IRQ stack */
|
|
|
|
ms = machdep->machspec;
|
|
irq_stack_ptr = ms->irq_stacks[bt->tc->processor]
|
|
+ ms->irq_stack_size - 0x20;
|
|
if (ext_frame.fp != irq_stack_ptr) {
|
|
/* (2) Just add a faked frame */
|
|
|
|
frame->sp = ext_frame.fp
|
|
- sizeof(struct arm64_pt_regs);
|
|
frame->fp = frame->sp;
|
|
} else {
|
|
/*
|
|
* (3)
|
|
* Switch from IRQ stack to process stack
|
|
*/
|
|
|
|
frame->sp = GET_STACK_ULONG(irq_stack_ptr + 8);
|
|
frame->fp = frame->sp;
|
|
|
|
/*
|
|
* Keep a buffer for a while until
|
|
* displaying the last frame on IRQ stack
|
|
* at next arm64_print_stackframe_entry_v2()
|
|
*/
|
|
if (bt->flags & BT_FULL)
|
|
ms->irq_stackbuf = bt->stackbuf;
|
|
|
|
arm64_set_process_stack(bt);
|
|
}
|
|
}
|
|
|
|
/* prepare for a stackframe hidden by exception */
|
|
arm64_gen_hidden_frame(bt, frame->sp, &ext_frame);
|
|
}
|
|
|
|
unwind_done:
|
|
if (CRASHDEBUG(1))
|
|
fprintf(ofp, " nxt fp:%016lx sp:%016lx pc:%016lx\n",
|
|
frame->fp, frame->sp, frame->pc);
|
|
|
|
return TRUE;
|
|
}
|
|
|
|
/*
|
|
* A layout of a stack frame in a function looks like:
|
|
*
|
|
* stack grows to lower addresses.
|
|
* /|\
|
|
* |
|
|
* | |
|
|
* new sp +------+ <---
|
|
* |dyn | |
|
|
* | vars | |
|
|
* new fp +- - - + |
|
|
* |old fp| | a function's stack frame
|
|
* |old lr| |
|
|
* |static| |
|
|
* | vars| |
|
|
* old sp +------+ <---
|
|
* |dyn |
|
|
* | vars |
|
|
* old fp +------+
|
|
* | |
|
|
*
|
|
* - On function entry, sp is decremented down to new fp.
|
|
*
|
|
* - and old fp and sp are saved into this stack frame.
|
|
* "Static" local variables are allocated at the same time.
|
|
*
|
|
* - Later on, "dynamic" local variables may be allocated on a stack.
|
|
* But those dynamic variables are rarely used in the kernel image,
|
|
* and, as a matter of fact, sp is equal to fp in almost all functions.
|
|
* (not 100% though.)
|
|
*
|
|
* - Currently, sp is determined in arm64_unwind_frame() by
|
|
* sp = a callee's fp + 0x10
|
|
* where 0x10 stands for a saved area for fp and sp
|
|
*
|
|
* - As you can see, however, this calculated sp still points to the top of
|
|
* callee's static local variables and doesn't match with a *real* sp.
|
|
*
|
|
* - So, generally, dumping a stack from this calculated sp to the next frame's
|
|
* sp shows "callee's static local variables", old fp and sp.
|
|
*
|
|
* Diagram and explanation courtesy of Takahiro Akashi
|
|
*/
|
|
|
|
static void
|
|
arm64_back_trace_cmd(struct bt_info *bt)
|
|
{
|
|
struct arm64_stackframe stackframe;
|
|
int level;
|
|
ulong exception_frame;
|
|
FILE *ofp;
|
|
|
|
if (bt->flags & BT_OPT_BACK_TRACE) {
|
|
if (machdep->flags & UNW_4_14) {
|
|
option_not_supported('o');
|
|
return;
|
|
}
|
|
|
|
arm64_back_trace_cmd_v2(bt);
|
|
return;
|
|
}
|
|
|
|
ofp = BT_REFERENCE_CHECK(bt) ? pc->nullfp : fp;
|
|
|
|
/*
|
|
* stackframes are created from 3 contiguous stack addresses:
|
|
*
|
|
* x: contains stackframe.fp -- points to next triplet
|
|
* x+8: contains stackframe.pc -- text return address
|
|
* x+16: is the stackframe.sp address
|
|
*/
|
|
|
|
if (bt->flags & BT_KDUMP_ADJUST) {
|
|
if (arm64_on_irq_stack(bt->tc->processor, bt->bptr)) {
|
|
arm64_set_irq_stack(bt);
|
|
bt->flags |= BT_IRQSTACK;
|
|
}
|
|
stackframe.fp = GET_STACK_ULONG(bt->bptr - 8);
|
|
stackframe.pc = GET_STACK_ULONG(bt->bptr);
|
|
stackframe.sp = bt->bptr + 8;
|
|
bt->frameptr = stackframe.sp;
|
|
} else if (bt->hp && bt->hp->esp) {
|
|
if (arm64_on_irq_stack(bt->tc->processor, bt->hp->esp)) {
|
|
arm64_set_irq_stack(bt);
|
|
bt->flags |= BT_IRQSTACK;
|
|
}
|
|
stackframe.fp = GET_STACK_ULONG(bt->hp->esp - 8);
|
|
stackframe.pc = bt->hp->eip ?
|
|
bt->hp->eip : GET_STACK_ULONG(bt->hp->esp);
|
|
stackframe.sp = bt->hp->esp + 8;
|
|
bt->flags &= ~BT_REGS_NOT_FOUND;
|
|
} else {
|
|
if (arm64_on_irq_stack(bt->tc->processor, bt->frameptr)) {
|
|
arm64_set_irq_stack(bt);
|
|
bt->flags |= BT_IRQSTACK;
|
|
} else if (arm64_on_overflow_stack(bt->tc->processor, bt->frameptr)) {
|
|
arm64_set_overflow_stack(bt);
|
|
bt->flags |= BT_OVERFLOW_STACK;
|
|
}
|
|
stackframe.sp = bt->stkptr;
|
|
stackframe.pc = bt->instptr;
|
|
stackframe.fp = bt->frameptr;
|
|
}
|
|
|
|
if (bt->flags & BT_TEXT_SYMBOLS) {
|
|
arm64_print_text_symbols(bt, &stackframe, ofp);
|
|
if (BT_REFERENCE_FOUND(bt)) {
|
|
print_task_header(fp, task_to_context(bt->task), 0);
|
|
arm64_print_text_symbols(bt, &stackframe, fp);
|
|
fprintf(fp, "\n");
|
|
}
|
|
return;
|
|
}
|
|
|
|
if (bt->flags & BT_REGS_NOT_FOUND)
|
|
return;
|
|
|
|
if (!(bt->flags & BT_KDUMP_ADJUST)) {
|
|
if (bt->flags & BT_USER_SPACE)
|
|
goto complete_user;
|
|
|
|
if (DUMPFILE() && is_task_active(bt->task)) {
|
|
exception_frame = stackframe.fp - KERN_EFRAME_OFFSET;
|
|
if (arm64_is_kernel_exception_frame(bt, exception_frame))
|
|
arm64_print_exception_frame(bt, exception_frame,
|
|
KERNEL_MODE, ofp);
|
|
}
|
|
}
|
|
|
|
level = exception_frame = 0;
|
|
while (1) {
|
|
bt->instptr = stackframe.pc;
|
|
|
|
switch (arm64_print_stackframe_entry(bt, level, &stackframe, ofp))
|
|
{
|
|
case BACKTRACE_COMPLETE_KERNEL:
|
|
return;
|
|
case BACKTRACE_COMPLETE_USER:
|
|
goto complete_user;
|
|
case BACKTRACE_CONTINUE:
|
|
break;
|
|
}
|
|
|
|
if (exception_frame) {
|
|
arm64_print_exception_frame(bt, exception_frame, KERNEL_MODE, ofp);
|
|
exception_frame = 0;
|
|
}
|
|
|
|
if (!arm64_unwind_frame(bt, &stackframe))
|
|
break;
|
|
|
|
if (arm64_in_exception_text(bt->instptr) && INSTACK(stackframe.fp, bt)) {
|
|
if (bt->flags & BT_OVERFLOW_STACK) {
|
|
exception_frame = stackframe.fp - KERN_EFRAME_OFFSET;
|
|
} else if (!(bt->flags & BT_IRQSTACK) ||
|
|
((stackframe.sp + SIZE(pt_regs)) < bt->stacktop)) {
|
|
if (arm64_is_kernel_exception_frame(bt, stackframe.fp - KERN_EFRAME_OFFSET))
|
|
exception_frame = stackframe.fp - KERN_EFRAME_OFFSET;
|
|
}
|
|
}
|
|
|
|
if ((bt->flags & BT_IRQSTACK) &&
|
|
!arm64_on_irq_stack(bt->tc->processor, stackframe.fp)) {
|
|
bt->flags &= ~BT_IRQSTACK;
|
|
if (arm64_switch_stack(bt, &stackframe, ofp) == USER_MODE)
|
|
break;
|
|
}
|
|
|
|
if ((bt->flags & BT_OVERFLOW_STACK) &&
|
|
!arm64_on_overflow_stack(bt->tc->processor, stackframe.fp)) {
|
|
bt->flags &= ~BT_OVERFLOW_STACK;
|
|
if (arm64_switch_stack_from_overflow(bt, &stackframe, ofp) == USER_MODE)
|
|
break;
|
|
}
|
|
|
|
level++;
|
|
}
|
|
|
|
if (is_kernel_thread(bt->tc->task))
|
|
return;
|
|
|
|
complete_user:
|
|
exception_frame = bt->stacktop - USER_EFRAME_OFFSET;
|
|
arm64_print_exception_frame(bt, exception_frame, USER_MODE, ofp);
|
|
if ((bt->flags & (BT_USER_SPACE|BT_KDUMP_ADJUST)) == BT_USER_SPACE)
|
|
fprintf(ofp, " #0 [user space]\n");
|
|
}
|
|
|
|
static void
|
|
arm64_back_trace_cmd_v2(struct bt_info *bt)
|
|
{
|
|
struct arm64_stackframe stackframe, cur_frame;
|
|
int level, mode;
|
|
ulong exception_frame;
|
|
FILE *ofp;
|
|
|
|
ofp = BT_REFERENCE_CHECK(bt) ? pc->nullfp : fp;
|
|
|
|
/*
|
|
* stackframes are created from 3 contiguous stack addresses:
|
|
*
|
|
* x: contains stackframe.fp -- points to next triplet
|
|
* x+8: contains stackframe.pc -- text return address
|
|
* x+16: is the stackframe.sp address
|
|
*/
|
|
|
|
if (bt->flags & BT_KDUMP_ADJUST) {
|
|
if (arm64_on_irq_stack(bt->tc->processor, bt->bptr)) {
|
|
arm64_set_irq_stack(bt);
|
|
bt->flags |= BT_IRQSTACK;
|
|
}
|
|
stackframe.fp = GET_STACK_ULONG(bt->bptr);
|
|
stackframe.pc = GET_STACK_ULONG(bt->bptr + 8);
|
|
stackframe.sp = bt->bptr + 16;
|
|
bt->frameptr = stackframe.fp;
|
|
} else {
|
|
if (arm64_on_irq_stack(bt->tc->processor, bt->frameptr)) {
|
|
arm64_set_irq_stack(bt);
|
|
bt->flags |= BT_IRQSTACK;
|
|
}
|
|
stackframe.sp = bt->stkptr;
|
|
stackframe.pc = bt->instptr;
|
|
stackframe.fp = bt->frameptr;
|
|
}
|
|
|
|
if (bt->flags & BT_TEXT_SYMBOLS) {
|
|
arm64_print_text_symbols(bt, &stackframe, ofp);
|
|
if (BT_REFERENCE_FOUND(bt)) {
|
|
print_task_header(fp, task_to_context(bt->task), 0);
|
|
arm64_print_text_symbols(bt, &stackframe, fp);
|
|
fprintf(fp, "\n");
|
|
}
|
|
return;
|
|
}
|
|
|
|
if (bt->flags & BT_REGS_NOT_FOUND)
|
|
return;
|
|
|
|
if (!(bt->flags & BT_KDUMP_ADJUST)) {
|
|
if (bt->flags & BT_USER_SPACE) {
|
|
user_space:
|
|
exception_frame = bt->stacktop - USER_EFRAME_OFFSET;
|
|
arm64_print_exception_frame(bt, exception_frame,
|
|
USER_MODE, ofp);
|
|
// fprintf(ofp, " #0 [user space]\n");
|
|
|
|
return;
|
|
}
|
|
|
|
if (DUMPFILE() && is_task_active(bt->task)) {
|
|
exception_frame = stackframe.fp - SIZE(pt_regs);
|
|
if (arm64_is_kernel_exception_frame(bt, exception_frame))
|
|
arm64_print_exception_frame(bt, exception_frame,
|
|
KERNEL_MODE, ofp);
|
|
}
|
|
}
|
|
|
|
for (level = 0;; level++) {
|
|
bt->instptr = stackframe.pc;
|
|
|
|
/*
|
|
* Show one-line stackframe info
|
|
*/
|
|
if (arm64_print_stackframe_entry_v2(bt, level, &stackframe, ofp)
|
|
== BACKTRACE_COMPLETE_KERNEL)
|
|
break;
|
|
|
|
cur_frame = stackframe;
|
|
if (!arm64_unwind_frame_v2(bt, &stackframe, ofp))
|
|
break;
|
|
|
|
/*
|
|
* Dump the contents of the current stackframe.
|
|
* We need to know the next stackframe to determine
|
|
* the dump range:
|
|
* <cur_frame.fp:stackframe.fp>
|
|
*/
|
|
arm64_display_full_frame_v2(bt, &cur_frame, &stackframe);
|
|
|
|
/*
|
|
* If we are in a normal stackframe, just continue,
|
|
* otherwise show an exception frame.
|
|
* Since exception entry code doesn't have a real
|
|
* stackframe, we fake a dummy frame here.
|
|
*/
|
|
if (!arm64_in_exp_entry(stackframe.pc))
|
|
continue;
|
|
|
|
if (!INSTACK(cur_frame.sp, bt))
|
|
fprintf(ofp, "--- <IRQ stack> ---\n");
|
|
|
|
arm64_print_stackframe_entry_v2(bt, ++level, &stackframe, ofp);
|
|
if (bt->flags & BT_USER_EFRAME)
|
|
goto user_space;
|
|
cur_frame = stackframe;
|
|
arm64_unwind_frame_v2(bt, &stackframe, ofp);
|
|
|
|
/*
|
|
* and don't show the contenxts. Instead,
|
|
* show an exception frame below
|
|
*/
|
|
|
|
if (!INSTACK(cur_frame.sp, bt)) {
|
|
/* This check is a safeguard. See unwind_frame(). */
|
|
error(WARNING,
|
|
"stack pointer for exception frame is wrong\n");
|
|
return;
|
|
}
|
|
mode = (stackframe.pc < machdep->machspec->userspace_top) ?
|
|
USER_MODE : KERNEL_MODE;
|
|
// fprintf(ofp, "--- <Exception in %s> ---\n",
|
|
// mode == KERNEL_MODE ? "kernel" : "user");
|
|
arm64_print_exception_frame(bt, cur_frame.sp, mode, ofp);
|
|
|
|
if (mode == USER_MODE)
|
|
break;
|
|
}
|
|
}
|
|
|
|
static void
|
|
arm64_print_text_symbols(struct bt_info *bt, struct arm64_stackframe *frame, FILE *ofp)
|
|
{
|
|
int i;
|
|
ulong *up;
|
|
struct load_module *lm;
|
|
char buf1[BUFSIZE];
|
|
char buf2[BUFSIZE];
|
|
char *name;
|
|
ulong start;
|
|
ulong val;
|
|
struct machine_specific *ms = machdep->machspec;
|
|
|
|
if (bt->flags & BT_TEXT_SYMBOLS_ALL)
|
|
start = bt->stackbase;
|
|
else {
|
|
start = frame->sp - 8;
|
|
fprintf(ofp, "%sSTART: %s at %lx\n",
|
|
space(VADDR_PRLEN > 8 ? 14 : 6),
|
|
bt->flags & BT_SYMBOL_OFFSET ?
|
|
value_to_symstr(frame->pc, buf2, bt->radix) :
|
|
closest_symbol(frame->pc), frame->pc);
|
|
}
|
|
|
|
for (i = (start - bt->stackbase)/sizeof(ulong); i < LONGS_PER_STACK; i++) {
|
|
up = (ulong *)(&bt->stackbuf[i*sizeof(ulong)]);
|
|
val = *up;
|
|
if (is_kernel_text(val | ms->CONFIG_ARM64_KERNELPACMASK)) {
|
|
val |= ms->CONFIG_ARM64_KERNELPACMASK;
|
|
name = closest_symbol(val);
|
|
fprintf(ofp, " %s[%s] %s at %lx",
|
|
bt->flags & BT_ERROR_MASK ?
|
|
" " : "",
|
|
mkstring(buf1, VADDR_PRLEN,
|
|
RJUST|LONG_HEX,
|
|
MKSTR(bt->stackbase +
|
|
(i * sizeof(long)))),
|
|
bt->flags & BT_SYMBOL_OFFSET ?
|
|
value_to_symstr(val, buf2, bt->radix) :
|
|
name, val);
|
|
if (module_symbol(val, NULL, &lm, NULL, 0))
|
|
fprintf(ofp, " [%s]", lm->mod_name);
|
|
fprintf(ofp, "\n");
|
|
if (BT_REFERENCE_CHECK(bt))
|
|
arm64_do_bt_reference_check(bt, val, name);
|
|
}
|
|
}
|
|
}
|
|
|
|
static int
|
|
arm64_in_kdump_text(struct bt_info *bt, struct arm64_stackframe *frame)
|
|
{
|
|
ulong *ptr, *start, *base;
|
|
struct machine_specific *ms;
|
|
ulong crash_kexec_frame;
|
|
|
|
if (!(machdep->flags & KDUMP_ENABLED))
|
|
return FALSE;
|
|
|
|
base = (ulong *)&bt->stackbuf[(ulong)(STACK_OFFSET_TYPE(bt->stackbase))];
|
|
if (bt->flags & BT_USER_SPACE)
|
|
start = (ulong *)&bt->stackbuf[(ulong)(STACK_OFFSET_TYPE(bt->stacktop))];
|
|
else {
|
|
if (INSTACK(frame->fp, bt))
|
|
start = (ulong *)&bt->stackbuf[(ulong)(STACK_OFFSET_TYPE(frame->fp))];
|
|
else
|
|
start = (ulong *)&bt->stackbuf[(ulong)(STACK_OFFSET_TYPE(bt->stacktop))];
|
|
}
|
|
|
|
crash_kexec_frame = 0;
|
|
ms = machdep->machspec;
|
|
for (ptr = start - 8; ptr >= base; ptr--) {
|
|
if (bt->flags & BT_OPT_BACK_TRACE) {
|
|
if ((*ptr > ms->crash_kexec_start) &&
|
|
(*ptr < ms->crash_kexec_end) &&
|
|
INSTACK(*(ptr - 1), bt)) {
|
|
bt->bptr = ((ulong)(ptr - 1) - (ulong)base)
|
|
+ task_to_stackbase(bt->tc->task);
|
|
if (CRASHDEBUG(1))
|
|
fprintf(fp, "%lx: %lx (crash_kexec)\n", bt->bptr, *ptr);
|
|
return TRUE;
|
|
}
|
|
if ((*ptr > ms->crash_save_cpu_start) &&
|
|
(*ptr < ms->crash_save_cpu_end) &&
|
|
INSTACK(*(ptr - 1), bt)) {
|
|
bt->bptr = ((ulong)(ptr - 1) - (ulong)base)
|
|
+ task_to_stackbase(bt->tc->task);
|
|
if (CRASHDEBUG(1))
|
|
fprintf(fp, "%lx: %lx (crash_save_cpu)\n", bt->bptr, *ptr);
|
|
return TRUE;
|
|
}
|
|
} else {
|
|
if ((*ptr > ms->machine_kexec_start) && (*ptr < ms->machine_kexec_end)) {
|
|
bt->bptr = ((ulong)ptr - (ulong)base)
|
|
+ task_to_stackbase(bt->tc->task);
|
|
if (CRASHDEBUG(1))
|
|
fprintf(fp, "%lx: %lx (machine_kexec)\n", bt->bptr, *ptr);
|
|
return TRUE;
|
|
}
|
|
if ((*ptr > ms->crash_kexec_start) && (*ptr < ms->crash_kexec_end)) {
|
|
/*
|
|
* Stash the first crash_kexec frame in case the machine_kexec
|
|
* frame is not found.
|
|
*/
|
|
if (!crash_kexec_frame) {
|
|
crash_kexec_frame = ((ulong)ptr - (ulong)base)
|
|
+ task_to_stackbase(bt->tc->task);
|
|
if (CRASHDEBUG(1))
|
|
fprintf(fp, "%lx: %lx (crash_kexec)\n",
|
|
bt->bptr, *ptr);
|
|
}
|
|
continue;
|
|
}
|
|
if ((*ptr > ms->crash_save_cpu_start) && (*ptr < ms->crash_save_cpu_end)) {
|
|
bt->bptr = ((ulong)ptr - (ulong)base)
|
|
+ task_to_stackbase(bt->tc->task);
|
|
if (CRASHDEBUG(1))
|
|
fprintf(fp, "%lx: %lx (crash_save_cpu)\n", bt->bptr, *ptr);
|
|
return TRUE;
|
|
}
|
|
}
|
|
}
|
|
|
|
if (crash_kexec_frame) {
|
|
bt->bptr = crash_kexec_frame;
|
|
return TRUE;
|
|
}
|
|
|
|
return FALSE;
|
|
}
|
|
|
|
static int
|
|
arm64_in_kdump_text_on_irq_stack(struct bt_info *bt)
|
|
{
|
|
int cpu;
|
|
ulong stackbase;
|
|
char *stackbuf;
|
|
ulong *ptr, *start, *base;
|
|
struct machine_specific *ms;
|
|
|
|
if ((machdep->flags & (IRQ_STACKS|KDUMP_ENABLED)) != (IRQ_STACKS|KDUMP_ENABLED))
|
|
return FALSE;
|
|
|
|
ms = machdep->machspec;
|
|
cpu = bt->tc->processor;
|
|
stackbase = ms->irq_stacks[cpu];
|
|
stackbuf = GETBUF(ms->irq_stack_size);
|
|
|
|
if (!readmem(stackbase, KVADDR, stackbuf,
|
|
ms->irq_stack_size, "IRQ stack contents", RETURN_ON_ERROR)) {
|
|
error(INFO, "read of IRQ stack at %lx failed\n", stackbase);
|
|
FREEBUF(stackbuf);
|
|
return FALSE;
|
|
}
|
|
|
|
base = (ulong *)stackbuf;
|
|
start = (ulong *)(stackbuf + ms->irq_stack_size);
|
|
|
|
for (ptr = start - 8; ptr >= base; ptr--) {
|
|
if (bt->flags & BT_OPT_BACK_TRACE) {
|
|
if ((*ptr > ms->crash_kexec_start) &&
|
|
(*ptr < ms->crash_kexec_end) &&
|
|
INSTACK(*(ptr - 1), bt)) {
|
|
bt->bptr = ((ulong)(ptr - 1) - (ulong)base) + stackbase;
|
|
if (CRASHDEBUG(1))
|
|
fprintf(fp, "%lx: %lx (crash_kexec on IRQ stack)\n",
|
|
bt->bptr, *ptr);
|
|
FREEBUF(stackbuf);
|
|
return TRUE;
|
|
}
|
|
if ((*ptr > ms->crash_save_cpu_start) &&
|
|
(*ptr < ms->crash_save_cpu_end) &&
|
|
INSTACK(*(ptr - 1), bt)) {
|
|
bt->bptr = ((ulong)(ptr - 1) - (ulong)base) + stackbase;
|
|
if (CRASHDEBUG(1))
|
|
fprintf(fp, "%lx: %lx (crash_save_cpu on IRQ stack)\n",
|
|
bt->bptr, *ptr);
|
|
FREEBUF(stackbuf);
|
|
return TRUE;
|
|
}
|
|
} else {
|
|
if ((*ptr > ms->crash_kexec_start) && (*ptr < ms->crash_kexec_end)) {
|
|
bt->bptr = ((ulong)ptr - (ulong)base) + stackbase;
|
|
if (CRASHDEBUG(1))
|
|
fprintf(fp, "%lx: %lx (crash_kexec on IRQ stack)\n",
|
|
bt->bptr, *ptr);
|
|
FREEBUF(stackbuf);
|
|
return TRUE;
|
|
}
|
|
if ((*ptr > ms->crash_save_cpu_start) && (*ptr < ms->crash_save_cpu_end)) {
|
|
bt->bptr = ((ulong)ptr - (ulong)base) + stackbase;
|
|
if (CRASHDEBUG(1))
|
|
fprintf(fp, "%lx: %lx (crash_save_cpu on IRQ stack)\n",
|
|
bt->bptr, *ptr);
|
|
FREEBUF(stackbuf);
|
|
return TRUE;
|
|
}
|
|
}
|
|
}
|
|
|
|
FREEBUF(stackbuf);
|
|
return FALSE;
|
|
}
|
|
|
|
static int
|
|
arm64_switch_stack(struct bt_info *bt, struct arm64_stackframe *frame, FILE *ofp)
|
|
{
|
|
int i;
|
|
ulong stacktop, words, addr;
|
|
ulong *stackbuf;
|
|
char buf[BUFSIZE];
|
|
struct machine_specific *ms = machdep->machspec;
|
|
|
|
if (bt->flags & BT_FULL) {
|
|
stacktop = ms->irq_stacks[bt->tc->processor] + ms->irq_stack_size;
|
|
words = (stacktop - bt->bptr) / sizeof(ulong);
|
|
stackbuf = (ulong *)GETBUF(words * sizeof(ulong));
|
|
readmem(bt->bptr, KVADDR, stackbuf, words * sizeof(long),
|
|
"top of IRQ stack", FAULT_ON_ERROR);
|
|
|
|
addr = bt->bptr;
|
|
for (i = 0; i < words; i++) {
|
|
if (!(i & 1))
|
|
fprintf(ofp, "%s %lx: ", i ? "\n" : "", addr);
|
|
fprintf(ofp, "%s ", format_stack_entry(bt, buf, stackbuf[i], 0));
|
|
addr += sizeof(ulong);
|
|
}
|
|
fprintf(ofp, "\n");
|
|
FREEBUF(stackbuf);
|
|
}
|
|
fprintf(ofp, "--- <IRQ stack> ---\n");
|
|
|
|
if (frame->fp == 0)
|
|
return USER_MODE;
|
|
|
|
if (!(machdep->flags & UNW_4_14))
|
|
arm64_print_exception_frame(bt, frame->sp, KERNEL_MODE, ofp);
|
|
|
|
return KERNEL_MODE;
|
|
}
|
|
|
|
static int
|
|
arm64_switch_stack_from_overflow(struct bt_info *bt, struct arm64_stackframe *frame, FILE *ofp)
|
|
{
|
|
int i;
|
|
ulong stacktop, words, addr;
|
|
ulong *stackbuf;
|
|
char buf[BUFSIZE];
|
|
struct machine_specific *ms = machdep->machspec;
|
|
|
|
if (bt->flags & BT_FULL) {
|
|
stacktop = ms->overflow_stacks[bt->tc->processor] + ms->overflow_stack_size;
|
|
words = (stacktop - bt->bptr) / sizeof(ulong);
|
|
stackbuf = (ulong *)GETBUF(words * sizeof(ulong));
|
|
readmem(bt->bptr, KVADDR, stackbuf, words * sizeof(long),
|
|
"top of overflow stack", FAULT_ON_ERROR);
|
|
|
|
addr = bt->bptr;
|
|
for (i = 0; i < words; i++) {
|
|
if (!(i & 1))
|
|
fprintf(ofp, "%s %lx: ", i ? "\n" : "", addr);
|
|
fprintf(ofp, "%s ", format_stack_entry(bt, buf, stackbuf[i], 0));
|
|
addr += sizeof(ulong);
|
|
}
|
|
fprintf(ofp, "\n");
|
|
FREEBUF(stackbuf);
|
|
}
|
|
fprintf(ofp, "--- <Overflow stack> ---\n");
|
|
|
|
if (frame->fp == 0)
|
|
return USER_MODE;
|
|
|
|
if (!(machdep->flags & UNW_4_14))
|
|
arm64_print_exception_frame(bt, frame->sp, KERNEL_MODE, ofp);
|
|
|
|
return KERNEL_MODE;
|
|
}
|
|
|
|
static int
|
|
arm64_get_dumpfile_stackframe(struct bt_info *bt, struct arm64_stackframe *frame)
|
|
{
|
|
struct machine_specific *ms = machdep->machspec;
|
|
struct arm64_pt_regs *ptregs;
|
|
bool skip = false;
|
|
|
|
if (bt->flags & BT_SKIP_IDLE) {
|
|
skip = true;
|
|
bt->flags &= ~BT_SKIP_IDLE;
|
|
}
|
|
|
|
if (!ms->panic_task_regs ||
|
|
(!ms->panic_task_regs[bt->tc->processor].sp &&
|
|
!ms->panic_task_regs[bt->tc->processor].pc)) {
|
|
bt->flags |= BT_REGS_NOT_FOUND;
|
|
return FALSE;
|
|
}
|
|
|
|
ptregs = &ms->panic_task_regs[bt->tc->processor];
|
|
frame->pc = ptregs->pc;
|
|
if (user_mode(ptregs)) {
|
|
frame->sp = user_stack_pointer(ptregs);
|
|
frame->fp = user_frame_pointer(ptregs);
|
|
if (is_kernel_text(frame->pc) ||
|
|
!in_user_stack(bt->tc->task, frame->sp)) {
|
|
error(WARNING,
|
|
"corrupt NT_PRSTATUS? pstate: 0x%lx, but no user frame found\n",
|
|
ptregs->pstate);
|
|
if (is_kernel_text(frame->pc) &&
|
|
INSTACK(frame->sp, bt) && INSTACK(frame->fp, bt))
|
|
goto try_kernel;
|
|
bt->flags |= BT_REGS_NOT_FOUND;
|
|
return FALSE;
|
|
}
|
|
bt->flags |= BT_USER_SPACE;
|
|
} else {
|
|
try_kernel:
|
|
frame->sp = ptregs->sp;
|
|
frame->fp = ptregs->regs[29];
|
|
bt->machdep = ptregs;
|
|
}
|
|
|
|
if (arm64_in_kdump_text(bt, frame) ||
|
|
arm64_in_kdump_text_on_irq_stack(bt)) {
|
|
bt->flags |= BT_KDUMP_ADJUST;
|
|
if (skip && is_idle_thread(bt->task))
|
|
bt->flags |= BT_SKIP_IDLE;
|
|
}
|
|
|
|
return TRUE;
|
|
}
|
|
|
|
static int
|
|
arm64_get_stackframe(struct bt_info *bt, struct arm64_stackframe *frame)
|
|
{
|
|
if (!fill_task_struct(bt->task))
|
|
return FALSE;
|
|
|
|
frame->sp = ULONG(tt->task_struct + OFFSET(task_struct_thread_context_sp));
|
|
frame->pc = ULONG(tt->task_struct + OFFSET(task_struct_thread_context_pc));
|
|
frame->fp = ULONG(tt->task_struct + OFFSET(task_struct_thread_context_fp));
|
|
|
|
return TRUE;
|
|
}
|
|
|
|
static void
|
|
arm64_get_stack_frame(struct bt_info *bt, ulong *pcp, ulong *spp)
|
|
{
|
|
struct user_regs_bitmap_struct *ur_bitmap;
|
|
struct arm64_stackframe stackframe = { 0 };
|
|
|
|
if (DUMPFILE() && is_task_active(bt->task)) {
|
|
arm64_get_dumpfile_stackframe(bt, &stackframe);
|
|
bt->need_free = FALSE;
|
|
} else {
|
|
if (bt->flags & BT_SKIP_IDLE)
|
|
bt->flags &= ~BT_SKIP_IDLE;
|
|
|
|
arm64_get_stackframe(bt, &stackframe);
|
|
|
|
ur_bitmap = (struct user_regs_bitmap_struct *)GETBUF(sizeof(*ur_bitmap));
|
|
memset(ur_bitmap, 0, sizeof(*ur_bitmap));
|
|
ur_bitmap->ur.pc = stackframe.pc;
|
|
ur_bitmap->ur.sp = stackframe.sp;
|
|
ur_bitmap->ur.regs[29] = stackframe.fp;
|
|
SET_BIT(ur_bitmap->bitmap, REG_SEQ(arm64_pt_regs, pc));
|
|
SET_BIT(ur_bitmap->bitmap, REG_SEQ(arm64_pt_regs, sp));
|
|
SET_BIT(ur_bitmap->bitmap, REG_SEQ(arm64_pt_regs, regs[0])
|
|
+ X29_REGNUM - X0_REGNUM);
|
|
bt->machdep = ur_bitmap;
|
|
bt->need_free = TRUE;
|
|
}
|
|
|
|
bt->frameptr = stackframe.fp;
|
|
if (pcp)
|
|
*pcp = stackframe.pc;
|
|
if (spp)
|
|
*spp = stackframe.sp;
|
|
}
|
|
|
|
static void
|
|
arm64_gen_hidden_frame(struct bt_info *bt, ulong sp,
|
|
struct arm64_stackframe *frame)
|
|
{
|
|
struct arm64_pt_regs *ptregs;
|
|
|
|
if (IN_TASK_VMA(bt->task, sp)) {
|
|
bt->flags |= BT_USER_EFRAME;
|
|
return;
|
|
}
|
|
|
|
ptregs = (struct arm64_pt_regs *)
|
|
&bt->stackbuf[(ulong)(STACK_OFFSET_TYPE(sp))];
|
|
|
|
frame->pc = ptregs->pc;
|
|
frame->fp = ptregs->regs[29];
|
|
frame->sp = ptregs->sp;
|
|
}
|
|
|
|
static void
|
|
arm64_print_exception_frame(struct bt_info *bt, ulong pt_regs, int mode, FILE *ofp)
|
|
{
|
|
int i, r, rows, top_reg, is_64_bit;
|
|
struct arm64_pt_regs *regs;
|
|
struct syment *sp;
|
|
ulong LR, SP, offset;
|
|
char buf[BUFSIZE];
|
|
struct machine_specific *ms = machdep->machspec;
|
|
|
|
if (CRASHDEBUG(1))
|
|
fprintf(ofp, "pt_regs: %lx\n", pt_regs);
|
|
|
|
regs = (struct arm64_pt_regs *)
|
|
&bt->stackbuf[(ulong)(STACK_OFFSET_TYPE(pt_regs))];
|
|
|
|
if ((mode == USER_MODE) && (regs->pstate & PSR_MODE32_BIT)) {
|
|
LR = regs->regs[14];
|
|
SP = regs->regs[13];
|
|
top_reg = 12;
|
|
is_64_bit = FALSE;
|
|
rows = 4;
|
|
} else {
|
|
LR = regs->regs[30];
|
|
if (is_kernel_text (LR | ms->CONFIG_ARM64_KERNELPACMASK))
|
|
LR |= ms->CONFIG_ARM64_KERNELPACMASK;
|
|
SP = regs->sp;
|
|
top_reg = 29;
|
|
is_64_bit = TRUE;
|
|
rows = 3;
|
|
}
|
|
|
|
switch (mode) {
|
|
case USER_MODE:
|
|
if (is_64_bit)
|
|
fprintf(ofp,
|
|
" PC: %016lx LR: %016lx SP: %016lx\n ",
|
|
(ulong)regs->pc, LR, SP);
|
|
else
|
|
fprintf(ofp,
|
|
" PC: %08lx LR: %08lx SP: %08lx PSTATE: %08lx\n ",
|
|
(ulong)regs->pc, LR, SP, (ulong)regs->pstate);
|
|
break;
|
|
|
|
case KERNEL_MODE:
|
|
fprintf(ofp, " PC: %016lx ", (ulong)regs->pc);
|
|
if (is_kernel_text(regs->pc) &&
|
|
(sp = value_search(regs->pc, &offset))) {
|
|
fprintf(ofp, "[%s", sp->name);
|
|
if (offset)
|
|
fprintf(ofp, (*gdb_output_radix == 16) ?
|
|
"+0x%lx" : "+%ld",
|
|
offset);
|
|
fprintf(ofp, "]\n");
|
|
} else
|
|
fprintf(ofp, "[unknown or invalid address]\n");
|
|
|
|
fprintf(ofp, " LR: %016lx ", LR);
|
|
if (is_kernel_text(LR) &&
|
|
(sp = value_search(LR, &offset))) {
|
|
fprintf(ofp, "[%s", sp->name);
|
|
if (offset)
|
|
fprintf(ofp, (*gdb_output_radix == 16) ?
|
|
"+0x%lx" : "+%ld",
|
|
offset);
|
|
fprintf(ofp, "]\n");
|
|
} else
|
|
fprintf(ofp, "[unknown or invalid address]\n");
|
|
|
|
fprintf(ofp, " SP: %016lx PSTATE: %08lx\n ",
|
|
SP, (ulong)regs->pstate);
|
|
break;
|
|
}
|
|
|
|
for (i = top_reg, r = 1; i >= 0; r++, i--) {
|
|
fprintf(ofp, "%sX%d: ",
|
|
i < 10 ? " " : "", i);
|
|
fprintf(ofp, is_64_bit ? "%016lx" : "%08lx",
|
|
(ulong)regs->regs[i]);
|
|
if ((i == 0) && !is_64_bit)
|
|
fprintf(ofp, "\n");
|
|
else if ((i == 0) || ((r % rows) == 0))
|
|
fprintf(ofp, "\n%s",
|
|
(i == 0) && (mode == KERNEL_MODE) ? "" : " ");
|
|
else
|
|
fprintf(ofp, "%s", is_64_bit ? " " : " ");
|
|
}
|
|
|
|
if (is_64_bit) {
|
|
if (mode == USER_MODE) {
|
|
fprintf(ofp, "ORIG_X0: %016lx SYSCALLNO: %lx",
|
|
(ulong)regs->orig_x0, (ulong)regs->syscallno);
|
|
fprintf(ofp, " PSTATE: %08lx\n", (ulong)regs->pstate);
|
|
}
|
|
}
|
|
|
|
if (is_kernel_text(regs->pc) && (bt->flags & BT_LINE_NUMBERS)) {
|
|
get_line_number(regs->pc, buf, FALSE);
|
|
if (strlen(buf))
|
|
fprintf(ofp, " %s\n", buf);
|
|
}
|
|
|
|
if (BT_REFERENCE_CHECK(bt)) {
|
|
arm64_do_bt_reference_check(bt, regs->pc, NULL);
|
|
if ((sp = value_search(regs->pc, &offset)))
|
|
arm64_do_bt_reference_check(bt, 0, sp->name);
|
|
arm64_do_bt_reference_check(bt, LR, NULL);
|
|
arm64_do_bt_reference_check(bt, SP, NULL);
|
|
arm64_do_bt_reference_check(bt, regs->pstate, NULL);
|
|
for (i = 0; i <= top_reg; i++)
|
|
arm64_do_bt_reference_check(bt, regs->regs[i], NULL);
|
|
if (is_64_bit) {
|
|
arm64_do_bt_reference_check(bt, regs->orig_x0, NULL);
|
|
arm64_do_bt_reference_check(bt, regs->syscallno, NULL);
|
|
}
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Check a frame for a requested reference.
|
|
*/
|
|
static void
|
|
arm64_do_bt_reference_check(struct bt_info *bt, ulong text, char *name)
|
|
{
|
|
ulong offset;
|
|
struct syment *sp = NULL;
|
|
|
|
if (!name)
|
|
sp = value_search(text, &offset);
|
|
else if (!text)
|
|
sp = symbol_search(name);
|
|
|
|
switch (bt->ref->cmdflags & (BT_REF_SYMBOL|BT_REF_HEXVAL))
|
|
{
|
|
case BT_REF_SYMBOL:
|
|
if (name) {
|
|
if (STREQ(name, bt->ref->str))
|
|
bt->ref->cmdflags |= BT_REF_FOUND;
|
|
} else {
|
|
if (sp && !offset && STREQ(sp->name, bt->ref->str))
|
|
bt->ref->cmdflags |= BT_REF_FOUND;
|
|
}
|
|
break;
|
|
|
|
case BT_REF_HEXVAL:
|
|
if (text) {
|
|
if (bt->ref->hexval == text)
|
|
bt->ref->cmdflags |= BT_REF_FOUND;
|
|
} else if (sp && (bt->ref->hexval == sp->value))
|
|
bt->ref->cmdflags |= BT_REF_FOUND;
|
|
else if (!name && !text && (bt->ref->hexval == 0))
|
|
bt->ref->cmdflags |= BT_REF_FOUND;
|
|
break;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Translate a PTE, returning TRUE if the page is present.
|
|
* If a physaddr pointer is passed in, don't print anything.
|
|
*/
|
|
static int
|
|
arm64_translate_pte(ulong pte, void *physaddr, ulonglong unused)
|
|
{
|
|
int c, others, len1, len2, len3;
|
|
ulong paddr;
|
|
char buf1[BUFSIZE];
|
|
char buf2[BUFSIZE];
|
|
char buf3[BUFSIZE];
|
|
char ptebuf[BUFSIZE];
|
|
char physbuf[BUFSIZE];
|
|
char *arglist[MAXARGS];
|
|
int page_present;
|
|
|
|
paddr = PTE_TO_PHYS(pte);
|
|
page_present = pte & (PTE_VALID | machdep->machspec->PTE_PROT_NONE);
|
|
|
|
if (physaddr) {
|
|
*((ulong *)physaddr) = paddr;
|
|
return page_present;
|
|
}
|
|
|
|
sprintf(ptebuf, "%lx", pte);
|
|
len1 = MAX(strlen(ptebuf), strlen("PTE"));
|
|
fprintf(fp, "%s ", mkstring(buf1, len1, CENTER|LJUST, "PTE"));
|
|
|
|
if (!page_present) {
|
|
swap_location(pte, buf1);
|
|
if ((c = parse_line(buf1, arglist)) != 3)
|
|
error(FATAL, "cannot determine swap location\n");
|
|
|
|
len2 = MAX(strlen(arglist[0]), strlen("SWAP"));
|
|
len3 = MAX(strlen(arglist[2]), strlen("OFFSET"));
|
|
|
|
fprintf(fp, "%s %s\n",
|
|
mkstring(buf2, len2, CENTER|LJUST, "SWAP"),
|
|
mkstring(buf3, len3, CENTER|LJUST, "OFFSET"));
|
|
|
|
strcpy(buf2, arglist[0]);
|
|
strcpy(buf3, arglist[2]);
|
|
fprintf(fp, "%s %s %s\n",
|
|
mkstring(ptebuf, len1, CENTER|RJUST, NULL),
|
|
mkstring(buf2, len2, CENTER|RJUST, NULL),
|
|
mkstring(buf3, len3, CENTER|RJUST, NULL));
|
|
return page_present;
|
|
}
|
|
|
|
sprintf(physbuf, "%lx", paddr);
|
|
len2 = MAX(strlen(physbuf), strlen("PHYSICAL"));
|
|
fprintf(fp, "%s ", mkstring(buf1, len2, CENTER|LJUST, "PHYSICAL"));
|
|
|
|
fprintf(fp, "FLAGS\n");
|
|
|
|
fprintf(fp, "%s %s ",
|
|
mkstring(ptebuf, len1, CENTER|RJUST, NULL),
|
|
mkstring(physbuf, len2, CENTER|RJUST, NULL));
|
|
fprintf(fp, "(");
|
|
others = 0;
|
|
|
|
if (pte) {
|
|
if (pte & PTE_VALID)
|
|
fprintf(fp, "%sVALID", others++ ? "|" : "");
|
|
if (pte & machdep->machspec->PTE_FILE)
|
|
fprintf(fp, "%sFILE", others++ ? "|" : "");
|
|
if (pte & machdep->machspec->PTE_PROT_NONE)
|
|
fprintf(fp, "%sPROT_NONE", others++ ? "|" : "");
|
|
if (pte & PTE_USER)
|
|
fprintf(fp, "%sUSER", others++ ? "|" : "");
|
|
if (pte & PTE_RDONLY)
|
|
fprintf(fp, "%sRDONLY", others++ ? "|" : "");
|
|
if (pte & PTE_SHARED)
|
|
fprintf(fp, "%sSHARED", others++ ? "|" : "");
|
|
if (pte & PTE_AF)
|
|
fprintf(fp, "%sAF", others++ ? "|" : "");
|
|
if (pte & PTE_NG)
|
|
fprintf(fp, "%sNG", others++ ? "|" : "");
|
|
if (pte & PTE_PXN)
|
|
fprintf(fp, "%sPXN", others++ ? "|" : "");
|
|
if (pte & PTE_UXN)
|
|
fprintf(fp, "%sUXN", others++ ? "|" : "");
|
|
if (pte & PTE_DIRTY)
|
|
fprintf(fp, "%sDIRTY", others++ ? "|" : "");
|
|
if (pte & PTE_SPECIAL)
|
|
fprintf(fp, "%sSPECIAL", others++ ? "|" : "");
|
|
} else {
|
|
fprintf(fp, "no mapping");
|
|
}
|
|
|
|
fprintf(fp, ")\n");
|
|
|
|
return (page_present);
|
|
}
|
|
|
|
static ulong
|
|
arm64_vmalloc_start(void)
|
|
{
|
|
return machdep->machspec->vmalloc_start_addr;
|
|
}
|
|
|
|
/*
|
|
* Not so accurate since thread_info introduction.
|
|
*/
|
|
static int
|
|
arm64_is_task_addr(ulong task)
|
|
{
|
|
if (tt->flags & THREAD_INFO)
|
|
return IS_KVADDR(task);
|
|
else
|
|
return (IS_KVADDR(task) && (ALIGNED_STACK_OFFSET(task) == 0));
|
|
}
|
|
|
|
static ulong
|
|
PLT_veneer_to_kvaddr(ulong value)
|
|
{
|
|
uint32_t insn;
|
|
ulong addr = 0;
|
|
int i;
|
|
|
|
/*
|
|
* PLT veneer always looks:
|
|
* movn x16, #0x....
|
|
* movk x16, #0x...., lsl #16
|
|
* movk x16, #0x...., lsl #32
|
|
* br x16
|
|
*/
|
|
for (i = 0; i < 4; i++) {
|
|
if (!readmem(value + i * sizeof(insn), KVADDR, &insn,
|
|
sizeof(insn), "PLT veneer", RETURN_ON_ERROR)) {
|
|
error(WARNING, "cannot read PLT veneer instruction at %lx\n",
|
|
value + i * sizeof(insn));
|
|
return value;
|
|
}
|
|
switch (i) {
|
|
case 0:
|
|
if ((insn & 0xffe0001f) != 0x92800010)
|
|
goto not_plt;
|
|
addr = ~((ulong)(insn & 0x1fffe0) >> 5);
|
|
break;
|
|
case 1:
|
|
if ((insn & 0xffe0001f) != 0xf2a00010)
|
|
goto not_plt;
|
|
addr &= 0xffffffff0000ffff;
|
|
addr |= (ulong)(insn & 0x1fffe0) << (16 - 5);
|
|
break;
|
|
case 2:
|
|
if ((insn & 0xffe0001f) != 0xf2c00010)
|
|
goto not_plt;
|
|
addr &= 0xffff0000ffffffff;
|
|
addr |= (ulong)(insn & 0x1fffe0) << (32 - 5);
|
|
break;
|
|
case 3:
|
|
if (insn != 0xd61f0200)
|
|
goto not_plt;
|
|
break;
|
|
default:
|
|
return value; /* to avoid any warnings */
|
|
}
|
|
}
|
|
|
|
return addr;
|
|
|
|
not_plt:
|
|
return value;
|
|
}
|
|
|
|
/*
|
|
* Filter dissassembly output if the output radix is not gdb's default 10
|
|
*/
|
|
static int
|
|
arm64_dis_filter(ulong vaddr, char *inbuf, unsigned int output_radix)
|
|
{
|
|
char buf1[BUFSIZE];
|
|
char buf2[BUFSIZE];
|
|
char *colon, *p1;
|
|
int argc;
|
|
char *argv[MAXARGS];
|
|
ulong value;
|
|
|
|
if (!inbuf)
|
|
return TRUE;
|
|
|
|
console("IN: %s", inbuf);
|
|
|
|
colon = strstr(inbuf, ":");
|
|
|
|
if (colon) {
|
|
sprintf(buf1, "0x%lx <%s>", vaddr,
|
|
value_to_symstr(vaddr, buf2, output_radix));
|
|
sprintf(buf2, "%s%s", buf1, colon);
|
|
strcpy(inbuf, buf2);
|
|
}
|
|
|
|
strcpy(buf1, inbuf);
|
|
argc = parse_line(buf1, argv);
|
|
|
|
if ((FIRSTCHAR(argv[argc-1]) == '<') &&
|
|
(LASTCHAR(argv[argc-1]) == '>')) {
|
|
p1 = rindex(inbuf, '<');
|
|
while ((p1 > inbuf) && !(STRNEQ(p1, " 0x") || STRNEQ(p1, "\t0x")))
|
|
p1--;
|
|
|
|
if (!(STRNEQ(p1, " 0x") || STRNEQ(p1, "\t0x")))
|
|
return FALSE;
|
|
p1++;
|
|
|
|
if (!extract_hex(p1, &value, NULLCHAR, TRUE))
|
|
return FALSE;
|
|
|
|
sprintf(buf1, "0x%lx <%s>\n", value,
|
|
value_to_symstr(value, buf2, output_radix));
|
|
|
|
sprintf(p1, "%s", buf1);
|
|
}
|
|
|
|
if (IS_MODULE_VADDR(vaddr)) {
|
|
ulong orig_value;
|
|
|
|
p1 = &inbuf[strlen(inbuf)-1];
|
|
strcpy(buf1, inbuf);
|
|
argc = parse_line(buf1, argv);
|
|
|
|
if ((STREQ(argv[argc-2], "b") || STREQ(argv[argc-2], "bl")) &&
|
|
extract_hex(argv[argc-1], &orig_value, NULLCHAR, TRUE)) {
|
|
value = PLT_veneer_to_kvaddr(orig_value);
|
|
sprintf(p1, " <%s%s>\n",
|
|
value == orig_value ? "" : "plt:",
|
|
value_to_symstr(value, buf2, output_radix));
|
|
}
|
|
}
|
|
|
|
console(" %s", inbuf);
|
|
|
|
return TRUE;
|
|
}
|
|
|
|
/*
|
|
* Machine dependent command.
|
|
*/
|
|
static void
|
|
arm64_cmd_mach(void)
|
|
{
|
|
int c;
|
|
|
|
while ((c = getopt(argcnt, args, "cm")) != -1) {
|
|
switch (c) {
|
|
case 'c':
|
|
case 'm':
|
|
option_not_supported(c);
|
|
break;
|
|
|
|
default:
|
|
argerrs++;
|
|
break;
|
|
}
|
|
}
|
|
|
|
if (argerrs)
|
|
cmd_usage(pc->curcmd, SYNOPSIS);
|
|
|
|
arm64_display_machine_stats();
|
|
}
|
|
|
|
static void
|
|
arm64_display_machine_stats(void)
|
|
{
|
|
int i, pad;
|
|
struct new_utsname *uts;
|
|
char buf[BUFSIZE];
|
|
ulong mhz;
|
|
|
|
uts = &kt->utsname;
|
|
|
|
fprintf(fp, " MACHINE TYPE: %s\n", uts->machine);
|
|
fprintf(fp, " MEMORY SIZE: %s\n", get_memory_size(buf));
|
|
fprintf(fp, " CPUS: %d\n", get_cpus_to_display());
|
|
if ((mhz = machdep->processor_speed()))
|
|
fprintf(fp, " PROCESSOR SPEED: %ld Mhz\n", mhz);
|
|
fprintf(fp, " HZ: %d\n", machdep->hz);
|
|
fprintf(fp, " PAGE SIZE: %d\n", PAGESIZE());
|
|
fprintf(fp, "KERNEL VIRTUAL BASE: %lx\n", machdep->machspec->page_offset);
|
|
fprintf(fp, "KERNEL MODULES BASE: %lx\n", machdep->machspec->modules_vaddr);
|
|
fprintf(fp, "KERNEL VMALLOC BASE: %lx\n", machdep->machspec->vmalloc_start_addr);
|
|
fprintf(fp, "KERNEL VMEMMAP BASE: %lx\n", machdep->machspec->vmemmap_vaddr);
|
|
fprintf(fp, " KERNEL STACK SIZE: %ld\n", STACKSIZE());
|
|
if (machdep->machspec->irq_stack_size) {
|
|
fprintf(fp, " IRQ STACK SIZE: %ld\n",
|
|
machdep->machspec->irq_stack_size);
|
|
fprintf(fp, " IRQ STACKS:\n");
|
|
for (i = 0; i < kt->cpus; i++) {
|
|
pad = (i < 10) ? 3 : (i < 100) ? 2 : (i < 1000) ? 1 : 0;
|
|
fprintf(fp, "%s CPU %d: %lx\n", space(pad), i,
|
|
machdep->machspec->irq_stacks[i]);
|
|
}
|
|
}
|
|
if (machdep->machspec->overflow_stack_size) {
|
|
fprintf(fp, "OVERFLOW STACK SIZE: %ld\n",
|
|
machdep->machspec->overflow_stack_size);
|
|
fprintf(fp, " OVERFLOW STACKS:\n");
|
|
for (i = 0; i < kt->cpus; i++) {
|
|
pad = (i < 10) ? 3 : (i < 100) ? 2 : (i < 1000) ? 1 : 0;
|
|
fprintf(fp, "%s CPU %d: %lx\n", space(pad), i,
|
|
machdep->machspec->overflow_stacks[i]);
|
|
}
|
|
}
|
|
}
|
|
|
|
static int
|
|
arm64_get_smp_cpus(void)
|
|
{
|
|
int cpus;
|
|
|
|
if ((cpus = get_cpus_present()))
|
|
return cpus;
|
|
else
|
|
return MAX(get_cpus_online(), get_highest_cpu_online()+1);
|
|
}
|
|
|
|
|
|
/*
|
|
* Retrieve task registers for the time of the crash.
|
|
*/
|
|
static void
|
|
arm64_get_crash_notes(void)
|
|
{
|
|
struct machine_specific *ms = machdep->machspec;
|
|
ulong crash_notes;
|
|
Elf64_Nhdr *note = NULL;
|
|
ulong offset;
|
|
char *buf, *p;
|
|
ulong *notes_ptrs;
|
|
ulong i, found;
|
|
|
|
if (!symbol_exists("crash_notes")) {
|
|
if (DISKDUMP_DUMPFILE() || KDUMP_DUMPFILE()) {
|
|
if (!(ms->panic_task_regs = calloc((size_t)kt->cpus, sizeof(struct arm64_pt_regs))))
|
|
error(FATAL, "cannot calloc panic_task_regs space\n");
|
|
|
|
for (i = found = 0; i < kt->cpus; i++) {
|
|
if (DISKDUMP_DUMPFILE())
|
|
note = diskdump_get_prstatus_percpu(i);
|
|
else if (KDUMP_DUMPFILE())
|
|
note = netdump_get_prstatus_percpu(i);
|
|
|
|
if (!note) {
|
|
error(WARNING, "cpu %d: cannot find NT_PRSTATUS note\n", i);
|
|
continue;
|
|
}
|
|
|
|
/*
|
|
* Find correct location of note data. This contains elf_prstatus
|
|
* structure which has registers etc. for the crashed task.
|
|
*/
|
|
offset = sizeof(Elf64_Nhdr);
|
|
offset = roundup(offset + note->n_namesz, 4);
|
|
p = (char *)note + offset; /* start of elf_prstatus */
|
|
|
|
BCOPY(p + OFFSET(elf_prstatus_pr_reg), &ms->panic_task_regs[i],
|
|
sizeof(struct arm64_pt_regs));
|
|
|
|
found++;
|
|
}
|
|
if (!found) {
|
|
free(ms->panic_task_regs);
|
|
ms->panic_task_regs = NULL;
|
|
}
|
|
}
|
|
return;
|
|
}
|
|
|
|
crash_notes = symbol_value("crash_notes");
|
|
|
|
notes_ptrs = (ulong *)GETBUF(kt->cpus*sizeof(notes_ptrs[0]));
|
|
|
|
/*
|
|
* Read crash_notes for the first CPU. crash_notes are in standard ELF
|
|
* note format.
|
|
*/
|
|
if (!readmem(crash_notes, KVADDR, ¬es_ptrs[kt->cpus-1],
|
|
sizeof(notes_ptrs[kt->cpus-1]), "crash_notes", RETURN_ON_ERROR)) {
|
|
error(WARNING, "cannot read \"crash_notes\"\n");
|
|
FREEBUF(notes_ptrs);
|
|
return;
|
|
}
|
|
|
|
if (symbol_exists("__per_cpu_offset")) {
|
|
/*
|
|
* Add __per_cpu_offset for each cpu to form the notes pointer.
|
|
*/
|
|
for (i = 0; i<kt->cpus; i++)
|
|
notes_ptrs[i] = notes_ptrs[kt->cpus-1] + kt->__per_cpu_offset[i];
|
|
}
|
|
|
|
buf = GETBUF(SIZE(note_buf));
|
|
|
|
if (!(ms->panic_task_regs = calloc((size_t)kt->cpus, sizeof(struct arm64_pt_regs))))
|
|
error(FATAL, "cannot calloc panic_task_regs space\n");
|
|
|
|
for (i = found = 0; i < kt->cpus; i++) {
|
|
if (!readmem(notes_ptrs[i], KVADDR, buf, SIZE(note_buf),
|
|
"note_buf_t", RETURN_ON_ERROR)) {
|
|
error(WARNING, "cpu %d: cannot read NT_PRSTATUS note\n", i);
|
|
continue;
|
|
}
|
|
|
|
/*
|
|
* Do some sanity checks for this note before reading registers from it.
|
|
*/
|
|
note = (Elf64_Nhdr *)buf;
|
|
p = buf + sizeof(Elf64_Nhdr);
|
|
|
|
/*
|
|
* dumpfiles created with qemu won't have crash_notes, but there will
|
|
* be elf notes; dumpfiles created by kdump do not create notes for
|
|
* offline cpus.
|
|
*/
|
|
if (note->n_namesz == 0 && (DISKDUMP_DUMPFILE() || KDUMP_DUMPFILE())) {
|
|
if (DISKDUMP_DUMPFILE())
|
|
note = diskdump_get_prstatus_percpu(i);
|
|
else if (KDUMP_DUMPFILE())
|
|
note = netdump_get_prstatus_percpu(i);
|
|
if (note) {
|
|
/*
|
|
* SIZE(note_buf) accounts for a "final note", which is a
|
|
* trailing empty elf note header.
|
|
*/
|
|
long notesz = SIZE(note_buf) - sizeof(Elf64_Nhdr);
|
|
|
|
if (sizeof(Elf64_Nhdr) + roundup(note->n_namesz, 4) +
|
|
note->n_descsz == notesz)
|
|
BCOPY((char *)note, buf, notesz);
|
|
} else {
|
|
error(WARNING, "cpu %d: cannot find NT_PRSTATUS note\n", i);
|
|
continue;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Check the sanity of NT_PRSTATUS note only for each online cpu.
|
|
* If this cpu has invalid note, continue to find the crash notes
|
|
* for other online cpus.
|
|
*/
|
|
if (note->n_type != NT_PRSTATUS) {
|
|
error(WARNING, "cpu %d: invalid NT_PRSTATUS note (n_type != NT_PRSTATUS)\n", i);
|
|
continue;
|
|
}
|
|
|
|
if (!STRNEQ(p, "CORE")) {
|
|
error(WARNING, "cpu %d: invalid NT_PRSTATUS note (name != \"CORE\")\n", i);
|
|
continue;
|
|
}
|
|
|
|
/*
|
|
* Find correct location of note data. This contains elf_prstatus
|
|
* structure which has registers etc. for the crashed task.
|
|
*/
|
|
offset = sizeof(Elf64_Nhdr);
|
|
offset = roundup(offset + note->n_namesz, 4);
|
|
p = buf + offset; /* start of elf_prstatus */
|
|
|
|
BCOPY(p + OFFSET(elf_prstatus_pr_reg), &ms->panic_task_regs[i],
|
|
sizeof(struct arm64_pt_regs));
|
|
|
|
found++;
|
|
}
|
|
|
|
FREEBUF(buf);
|
|
FREEBUF(notes_ptrs);
|
|
|
|
if (!found) {
|
|
free(ms->panic_task_regs);
|
|
ms->panic_task_regs = NULL;
|
|
}
|
|
}
|
|
|
|
static void
|
|
arm64_clear_machdep_cache(void) {
|
|
/*
|
|
* TBD: probably not necessary...
|
|
*/
|
|
return;
|
|
}
|
|
|
|
static int
|
|
arm64_on_process_stack(struct bt_info *bt, ulong stkptr)
|
|
{
|
|
ulong stackbase, stacktop;
|
|
|
|
stackbase = GET_STACKBASE(bt->task);
|
|
stacktop = GET_STACKTOP(bt->task);
|
|
|
|
if ((stkptr >= stackbase) && (stkptr < stacktop))
|
|
return TRUE;
|
|
|
|
return FALSE;
|
|
}
|
|
|
|
static int
|
|
arm64_in_alternate_stackv(int cpu, ulong stkptr, ulong *stacks, ulong stack_size)
|
|
{
|
|
if ((cpu >= kt->cpus) || (stacks == NULL) || !stack_size)
|
|
return FALSE;
|
|
|
|
if ((stkptr >= stacks[cpu]) &&
|
|
(stkptr < (stacks[cpu] + stack_size)))
|
|
return TRUE;
|
|
|
|
return FALSE;
|
|
}
|
|
|
|
static int
|
|
arm64_in_alternate_stack(int cpu, ulong stkptr)
|
|
{
|
|
return (arm64_on_irq_stack(cpu, stkptr) ||
|
|
arm64_on_overflow_stack(cpu, stkptr));
|
|
}
|
|
|
|
static int
|
|
arm64_on_irq_stack(int cpu, ulong stkptr)
|
|
{
|
|
struct machine_specific *ms = machdep->machspec;
|
|
|
|
return arm64_in_alternate_stackv(cpu, stkptr,
|
|
ms->irq_stacks, ms->irq_stack_size);
|
|
}
|
|
|
|
static int
|
|
arm64_on_overflow_stack(int cpu, ulong stkptr)
|
|
{
|
|
struct machine_specific *ms = machdep->machspec;
|
|
|
|
return arm64_in_alternate_stackv(cpu, stkptr,
|
|
ms->overflow_stacks, ms->overflow_stack_size);
|
|
}
|
|
|
|
static void
|
|
arm64_set_irq_stack(struct bt_info *bt)
|
|
{
|
|
struct machine_specific *ms = machdep->machspec;
|
|
|
|
bt->stackbase = ms->irq_stacks[bt->tc->processor];
|
|
bt->stacktop = bt->stackbase + ms->irq_stack_size;
|
|
alter_stackbuf(bt);
|
|
}
|
|
|
|
static void
|
|
arm64_set_overflow_stack(struct bt_info *bt)
|
|
{
|
|
struct machine_specific *ms = machdep->machspec;
|
|
|
|
bt->stackbase = ms->overflow_stacks[bt->tc->processor];
|
|
bt->stacktop = bt->stackbase + ms->overflow_stack_size;
|
|
alter_stackbuf(bt);
|
|
}
|
|
|
|
static void
|
|
arm64_set_process_stack(struct bt_info *bt)
|
|
{
|
|
bt->stackbase = GET_STACKBASE(bt->task);
|
|
bt->stacktop = GET_STACKTOP(bt->task);
|
|
alter_stackbuf(bt);
|
|
}
|
|
|
|
|
|
static int
|
|
compare_kvaddr(const void *v1, const void *v2)
|
|
{
|
|
struct vaddr_range *r1, *r2;
|
|
|
|
r1 = (struct vaddr_range *)v1;
|
|
r2 = (struct vaddr_range *)v2;
|
|
|
|
return (r1->start < r2->start ? -1 :
|
|
r1->start == r2->start ? 0 : 1);
|
|
}
|
|
|
|
static int
|
|
arm64_get_kvaddr_ranges(struct vaddr_range *vrp)
|
|
{
|
|
int cnt;
|
|
|
|
cnt = 0;
|
|
|
|
vrp[cnt].type = KVADDR_UNITY_MAP;
|
|
vrp[cnt].start = machdep->machspec->page_offset;
|
|
vrp[cnt++].end = vt->high_memory;
|
|
|
|
vrp[cnt].type = KVADDR_VMALLOC;
|
|
vrp[cnt].start = machdep->machspec->vmalloc_start_addr;
|
|
vrp[cnt++].end = last_vmalloc_address();
|
|
|
|
if (st->mods_installed) {
|
|
vrp[cnt].type = KVADDR_MODULES;
|
|
vrp[cnt].start = lowest_module_address();
|
|
vrp[cnt++].end = roundup(highest_module_address(),
|
|
PAGESIZE());
|
|
}
|
|
|
|
if (machdep->flags & VMEMMAP) {
|
|
vrp[cnt].type = KVADDR_VMEMMAP;
|
|
vrp[cnt].start = machdep->machspec->vmemmap_vaddr;
|
|
vrp[cnt++].end = vt->node_table[vt->numnodes-1].mem_map +
|
|
(vt->node_table[vt->numnodes-1].size * SIZE(page));
|
|
}
|
|
|
|
qsort(vrp, cnt, sizeof(struct vaddr_range), compare_kvaddr);
|
|
|
|
return cnt;
|
|
}
|
|
|
|
/*
|
|
* Include both vmalloc'd, module and vmemmap address space as VMALLOC space.
|
|
*/
|
|
int
|
|
arm64_IS_VMALLOC_ADDR(ulong vaddr)
|
|
{
|
|
struct machine_specific *ms = machdep->machspec;
|
|
|
|
if (is_mte_kvaddr(vaddr))
|
|
vaddr = mte_tag_reset(vaddr);
|
|
|
|
if ((machdep->flags & NEW_VMEMMAP) &&
|
|
(vaddr >= machdep->machspec->kimage_text) &&
|
|
(vaddr <= machdep->machspec->kimage_end))
|
|
return FALSE;
|
|
|
|
if (ms->VA_START && (vaddr >= ms->VA_START))
|
|
return TRUE;
|
|
|
|
return ((vaddr >= ms->vmalloc_start_addr && vaddr <= ms->vmalloc_end) ||
|
|
((machdep->flags & VMEMMAP) &&
|
|
((vaddr >= ms->vmemmap_vaddr && vaddr <= ms->vmemmap_end) ||
|
|
(vaddr >= ms->vmalloc_end && vaddr <= ms->vmemmap_vaddr))) ||
|
|
(vaddr >= ms->modules_vaddr && vaddr <= ms->modules_end));
|
|
}
|
|
|
|
/* Return TRUE if we succeed, return FALSE on failure. */
|
|
static int
|
|
arm64_set_va_bits_by_tcr(void)
|
|
{
|
|
ulong value;
|
|
|
|
if (arm64_get_vmcoreinfo(&value, "NUMBER(TCR_EL1_T1SZ)", NUM_HEX) ||
|
|
arm64_get_vmcoreinfo(&value, "NUMBER(tcr_el1_t1sz)", NUM_HEX)) {
|
|
/* See ARMv8 ARM for the description of
|
|
* TCR_EL1.T1SZ and how it can be used
|
|
* to calculate the vabits_actual
|
|
* supported by underlying kernel.
|
|
*
|
|
* Basically:
|
|
* vabits_actual = 64 - T1SZ;
|
|
*/
|
|
value = 64 - value;
|
|
if (CRASHDEBUG(1))
|
|
fprintf(fp, "vmcoreinfo : vabits_actual: %ld\n", value);
|
|
machdep->machspec->VA_BITS_ACTUAL = value;
|
|
machdep->machspec->VA_BITS = value;
|
|
machdep->machspec->VA_START = _VA_START(machdep->machspec->VA_BITS_ACTUAL);
|
|
return TRUE;
|
|
}
|
|
|
|
return FALSE;
|
|
}
|
|
|
|
static void
|
|
arm64_calc_VA_BITS(void)
|
|
{
|
|
int bitval;
|
|
struct syment *sp;
|
|
ulong vabits_actual, value;
|
|
|
|
arm64_get_vmcoreinfo(&machdep->machspec->CONFIG_ARM64_VA_BITS, "NUMBER(VA_BITS)", NUM_DEC);
|
|
|
|
if (kernel_symbol_exists("vabits_actual")) {
|
|
if (pc->flags & PROC_KCORE) {
|
|
vabits_actual = symbol_value_from_proc_kallsyms("vabits_actual");
|
|
if ((vabits_actual != BADVAL) && (READMEM(pc->mfd, &value, sizeof(ulong),
|
|
vabits_actual, KCORE_USE_VADDR) > 0)) {
|
|
if (CRASHDEBUG(1))
|
|
fprintf(fp,
|
|
"/proc/kcore: vabits_actual: %ld\n", value);
|
|
machdep->machspec->VA_BITS_ACTUAL = value;
|
|
machdep->machspec->VA_BITS = value;
|
|
machdep->machspec->VA_START = _VA_START(machdep->machspec->VA_BITS_ACTUAL);
|
|
} else
|
|
error(FATAL, "/proc/kcore: cannot read vabits_actual\n");
|
|
} else if (ACTIVE())
|
|
error(FATAL, "cannot determine VA_BITS_ACTUAL: please use /proc/kcore\n");
|
|
else {
|
|
if (arm64_set_va_bits_by_tcr()) {
|
|
/* nothing */
|
|
} else if (machdep->machspec->VA_BITS_ACTUAL) {
|
|
machdep->machspec->VA_BITS = machdep->machspec->VA_BITS_ACTUAL;
|
|
machdep->machspec->VA_START = _VA_START(machdep->machspec->VA_BITS_ACTUAL);
|
|
} else if (machdep->machspec->CONFIG_ARM64_VA_BITS) {
|
|
/* guess */
|
|
machdep->machspec->VA_BITS_ACTUAL = machdep->machspec->CONFIG_ARM64_VA_BITS;
|
|
machdep->machspec->VA_BITS = machdep->machspec->CONFIG_ARM64_VA_BITS;
|
|
machdep->machspec->VA_START = _VA_START(machdep->machspec->VA_BITS_ACTUAL);
|
|
} else
|
|
error(FATAL, "cannot determine VA_BITS_ACTUAL\n");
|
|
}
|
|
|
|
if (machdep->machspec->CONFIG_ARM64_VA_BITS)
|
|
machdep->machspec->VA_BITS = machdep->machspec->CONFIG_ARM64_VA_BITS;
|
|
|
|
/*
|
|
* The mm flip commit is introduced before 52-bits VA, which is before the
|
|
* commit to export NUMBER(TCR_EL1_T1SZ)
|
|
*/
|
|
machdep->flags |= FLIPPED_VM;
|
|
return;
|
|
} else if (arm64_set_va_bits_by_tcr()) {
|
|
return;
|
|
} else if (machdep->machspec->VA_BITS_ACTUAL) {
|
|
machdep->machspec->VA_BITS = machdep->machspec->VA_BITS_ACTUAL;
|
|
machdep->machspec->VA_START = _VA_START(machdep->machspec->VA_BITS_ACTUAL);
|
|
return;
|
|
}
|
|
|
|
if (!(sp = symbol_search("swapper_pg_dir")) &&
|
|
!(sp = symbol_search("idmap_pg_dir")) &&
|
|
!(sp = symbol_search("_text")) &&
|
|
!(sp = symbol_search("stext"))) {
|
|
for (sp = st->symtable; sp < st->symend; sp++) {
|
|
if (highest_bit_long(sp->value) == 63)
|
|
break;
|
|
}
|
|
}
|
|
|
|
if (sp)
|
|
value = sp->value;
|
|
else
|
|
value = kt->vmcoreinfo.log_buf_SYMBOL; /* crash --log */
|
|
|
|
for (bitval = highest_bit_long(value); bitval; bitval--) {
|
|
if ((value & (1UL << bitval)) == 0) {
|
|
if (machdep->flags & NEW_VMEMMAP)
|
|
machdep->machspec->VA_BITS = bitval + 1;
|
|
else
|
|
machdep->machspec->VA_BITS = bitval + 2;
|
|
break;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Verify against dumpfiles that export VA_BITS in vmcoreinfo
|
|
*/
|
|
if (machdep->machspec->CONFIG_ARM64_VA_BITS &&
|
|
(machdep->machspec->VA_BITS != machdep->machspec->CONFIG_ARM64_VA_BITS)) {
|
|
error(WARNING, "VA_BITS: calculated: %ld vmcoreinfo: %ld\n",
|
|
machdep->machspec->VA_BITS, machdep->machspec->CONFIG_ARM64_VA_BITS);
|
|
machdep->machspec->VA_BITS = machdep->machspec->CONFIG_ARM64_VA_BITS;
|
|
}
|
|
|
|
if (CRASHDEBUG(1))
|
|
fprintf(fp, "VA_BITS: %ld\n", machdep->machspec->VA_BITS);
|
|
|
|
}
|
|
|
|
/*
|
|
* The size and end of the vmalloc range is dependent upon the kernel's
|
|
* VMEMMAP_SIZE value, and the vmemmap range is dependent upon the end
|
|
* of the vmalloc range as well as the VMEMMAP_SIZE:
|
|
*
|
|
* #define VMEMMAP_SIZE ALIGN((1UL << (VA_BITS - PAGE_SHIFT)) * sizeof(struct page), PUD_SIZE)
|
|
* #define VMALLOC_START (UL(0xffffffffffffffff) << VA_BITS)
|
|
* #define VMALLOC_END (PAGE_OFFSET - PUD_SIZE - VMEMMAP_SIZE - SZ_64K)
|
|
*
|
|
* Since VMEMMAP_SIZE is dependent upon the size of a struct page,
|
|
* the two ranges cannot be determined until POST_GDB.
|
|
*
|
|
* Since 52-bit VA was introduced:
|
|
*
|
|
* #define STRUCT_PAGE_MAX_SHIFT 6
|
|
* #define VMEMMAP_SIZE (UL(1) << (VA_BITS - PAGE_SHIFT - 1 + STRUCT_PAGE_MAX_SHIFT))
|
|
* #define VMEMMAP_START (-VMEMMAP_SIZE)
|
|
* #define VMALLOC_START (MODULES_END)
|
|
* #define VMALLOC_END (- PUD_SIZE - VMEMMAP_SIZE - SZ_64K)
|
|
* #define vmemmap ((struct page *)VMEMMAP_START - (memstart_addr >> PAGE_SHIFT))
|
|
*/
|
|
|
|
#define ALIGN(x, a) __ALIGN_KERNEL((x), (a))
|
|
#define __ALIGN_KERNEL(x, a) __ALIGN_KERNEL_MASK(x, (typeof(x))(a) - 1)
|
|
#define __ALIGN_KERNEL_MASK(x, mask) (((x) + (mask)) & ~(mask))
|
|
|
|
static void
|
|
arm64_calc_virtual_memory_ranges(void)
|
|
{
|
|
struct machine_specific *ms = machdep->machspec;
|
|
ulong value, vmemmap_start, vmemmap_end, vmemmap_size, vmalloc_end;
|
|
char *string;
|
|
int ret;
|
|
ulong PUD_SIZE = UNINITIALIZED;
|
|
|
|
if (!machdep->machspec->CONFIG_ARM64_VA_BITS) {
|
|
if (arm64_get_vmcoreinfo(&value, "NUMBER(VA_BITS)", NUM_DEC)) {
|
|
machdep->machspec->CONFIG_ARM64_VA_BITS = value;
|
|
} else if (kt->ikconfig_flags & IKCONFIG_AVAIL) {
|
|
if ((ret = get_kernel_config("CONFIG_ARM64_VA_BITS",
|
|
&string)) == IKCONFIG_STR)
|
|
machdep->machspec->CONFIG_ARM64_VA_BITS = atol(string);
|
|
}
|
|
}
|
|
|
|
if (THIS_KERNEL_VERSION < LINUX(3,17,0)) /* use original hardwired values */
|
|
return;
|
|
|
|
STRUCT_SIZE_INIT(page, "page");
|
|
ms->struct_page_size = SIZE(page);
|
|
|
|
switch (machdep->flags & (VM_L2_64K|VM_L3_64K|VM_L3_4K|VM_L4_4K))
|
|
{
|
|
case VM_L2_64K:
|
|
case VM_L3_64K:
|
|
PUD_SIZE = PGDIR_SIZE_L2_64K;
|
|
break;
|
|
case VM_L3_4K:
|
|
PUD_SIZE = PGDIR_SIZE_L3_4K;
|
|
case VM_L4_4K:
|
|
PUD_SIZE = PUD_SIZE_L4_4K;
|
|
break;
|
|
}
|
|
|
|
#define STRUCT_PAGE_MAX_SHIFT 6
|
|
|
|
if (ms->VA_BITS_ACTUAL) {
|
|
ulong va_bits_min = 48;
|
|
|
|
if (machdep->machspec->CONFIG_ARM64_VA_BITS < 48)
|
|
va_bits_min = ms->CONFIG_ARM64_VA_BITS;
|
|
|
|
vmemmap_size = (1UL) << (va_bits_min - machdep->pageshift - 1 + STRUCT_PAGE_MAX_SHIFT);
|
|
vmalloc_end = (- PUD_SIZE - vmemmap_size - KILOBYTES(64));
|
|
vmemmap_start = (-vmemmap_size - MEGABYTES(2));
|
|
ms->vmalloc_end = vmalloc_end - 1;
|
|
ms->vmemmap_vaddr = vmemmap_start;
|
|
ms->vmemmap_end = vmemmap_start + vmemmap_size;
|
|
|
|
return;
|
|
}
|
|
|
|
if (machdep->flags & NEW_VMEMMAP)
|
|
vmemmap_size = 1UL << (ms->VA_BITS - machdep->pageshift - 1
|
|
+ STRUCT_PAGE_MAX_SHIFT);
|
|
else
|
|
vmemmap_size = ALIGN((1UL << (ms->VA_BITS - machdep->pageshift)) * SIZE(page), PUD_SIZE);
|
|
|
|
vmalloc_end = (ms->page_offset - PUD_SIZE - vmemmap_size - SZ_64K);
|
|
|
|
if (machdep->flags & NEW_VMEMMAP) {
|
|
vmemmap_start = ms->page_offset - vmemmap_size;
|
|
vmemmap_end = ms->page_offset;
|
|
} else {
|
|
vmemmap_start = vmalloc_end + SZ_64K;
|
|
vmemmap_end = vmemmap_start + vmemmap_size;
|
|
}
|
|
|
|
ms->vmalloc_end = vmalloc_end - 1;
|
|
ms->vmemmap_vaddr = vmemmap_start;
|
|
ms->vmemmap_end = vmemmap_end - 1;
|
|
}
|
|
|
|
static int
|
|
arm64_is_uvaddr(ulong addr, struct task_context *tc)
|
|
{
|
|
return (addr < machdep->machspec->userspace_top);
|
|
}
|
|
|
|
|
|
ulong
|
|
arm64_swp_type(ulong pte)
|
|
{
|
|
struct machine_specific *ms = machdep->machspec;
|
|
|
|
pte >>= ms->__SWP_TYPE_SHIFT;
|
|
pte &= ms->__SWP_TYPE_MASK;
|
|
return pte;
|
|
}
|
|
|
|
ulong
|
|
arm64_swp_offset(ulong pte)
|
|
{
|
|
struct machine_specific *ms = machdep->machspec;
|
|
|
|
pte >>= ms->__SWP_OFFSET_SHIFT;
|
|
if (ms->__SWP_OFFSET_MASK)
|
|
pte &= ms->__SWP_OFFSET_MASK;
|
|
return pte;
|
|
}
|
|
|
|
static void arm64_calc_KERNELPACMASK(void)
|
|
{
|
|
ulong value;
|
|
|
|
if (arm64_get_vmcoreinfo(&value, "NUMBER(KERNELPACMASK)", NUM_HEX)) {
|
|
machdep->machspec->CONFIG_ARM64_KERNELPACMASK = value;
|
|
if (CRASHDEBUG(1))
|
|
fprintf(fp, "CONFIG_ARM64_KERNELPACMASK: %lx\n", value);
|
|
}
|
|
}
|
|
|
|
#define GENMASK_UL(h, l) \
|
|
(((~0UL) << (l)) & (~0UL >> (BITS_PER_LONG - 1 - (h))))
|
|
|
|
static void arm64_recalc_KERNELPACMASK(void){
|
|
/*
|
|
* Check if PAC is enabled according to the existence of
|
|
* kernel symbol 'ptrauth_keys_kernel'.
|
|
*/
|
|
if (STRUCT_EXISTS("ptrauth_keys_kernel") &&
|
|
machdep->machspec->VA_BITS_ACTUAL){
|
|
machdep->machspec->CONFIG_ARM64_KERNELPACMASK =
|
|
GENMASK_UL(63, machdep->machspec->VA_BITS_ACTUAL);
|
|
if (CRASHDEBUG(1))
|
|
fprintf(fp, "CONFIG_ARM64_KERNELPACMASK: %lx\n",
|
|
machdep->machspec->CONFIG_ARM64_KERNELPACMASK);
|
|
}
|
|
}
|
|
|
|
static ulong arm64_set_irq_stack_size(void)
|
|
{
|
|
int min_thread_shift = 14;
|
|
ulong thread_shift = 0;
|
|
char buf1[BUFSIZE];
|
|
char *pos1, *pos2;
|
|
int errflag = 0;
|
|
|
|
if (kernel_symbol_exists("vmcoreinfo_data") &&
|
|
kernel_symbol_exists("vmcoreinfo_size")) {
|
|
/*
|
|
* Referring to arch/arm64/include/asm/memory.h
|
|
*/
|
|
if (kernel_symbol_exists("kasan_enable_current"))
|
|
min_thread_shift += 1;
|
|
|
|
if (MEMBER_EXISTS("task_struct", "stack_vm_area") &&
|
|
(min_thread_shift < machdep->pageshift))
|
|
thread_shift = machdep->pageshift;
|
|
else
|
|
thread_shift = min_thread_shift;
|
|
} else {
|
|
sprintf(buf1, "x/32i vectors");
|
|
open_tmpfile();
|
|
if (!gdb_pass_through(buf1, pc->tmpfile, GNU_RETURN_ON_ERROR))
|
|
goto out;
|
|
|
|
rewind(pc->tmpfile);
|
|
while (fgets(buf1, BUFSIZE, pc->tmpfile)) {
|
|
if ((pos1 = strstr(buf1, "tbnz"))) {
|
|
if ((pos2 = strchr(pos1, '#'))) {
|
|
pos2 += 1;
|
|
for (pos1 = pos2; *pos2 != '\0' && *pos2 != ','; pos2++);
|
|
*pos2 = '\0';
|
|
thread_shift = stol(pos1, RETURN_ON_ERROR|QUIET, &errflag);
|
|
if (errflag)
|
|
thread_shift = 0;
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
out:
|
|
close_tmpfile();
|
|
}
|
|
|
|
if (thread_shift)
|
|
return ((1UL) << thread_shift);
|
|
|
|
return 0;
|
|
}
|
|
|
|
#endif /* ARM64 */
|
|
|
|
|