crash/arm64.c

5444 lines
157 KiB
C

/*
* arm64.c - core analysis suite
*
* Copyright (C) 2012-2020 David Anderson
* Copyright (C) 2012-2020 Red Hat, Inc. All rights reserved.
*
* This program is free software; you can redistribute it and/or modify
* it under the terms of the GNU General Public License as published by
* the Free Software Foundation; either version 2 of the License, or
* (at your option) any later version.
*
* This program is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
* GNU General Public License for more details.
*/
#ifdef ARM64
#include "defs.h"
#include <elf.h>
#include <endian.h>
#include <math.h>
#include <sys/ioctl.h>
#define NOT_IMPLEMENTED(X) error((X), "%s: function not implemented\n", __func__)
static struct machine_specific arm64_machine_specific = { 0 };
static int arm64_verify_symbol(const char *, ulong, char);
static void arm64_parse_cmdline_args(void);
static int arm64_search_for_kimage_voffset(ulong);
static int verify_kimage_voffset(void);
static void arm64_calc_kimage_voffset(void);
static void arm64_calc_phys_offset(void);
static void arm64_calc_physvirt_offset(void);
static void arm64_calc_virtual_memory_ranges(void);
static void arm64_get_section_size_bits(void);
static int arm64_kdump_phys_base(ulong *);
static ulong arm64_processor_speed(void);
static void arm64_init_kernel_pgd(void);
static int arm64_kvtop(struct task_context *, ulong, physaddr_t *, int);
static int arm64_uvtop(struct task_context *, ulong, physaddr_t *, int);
static int arm64_vtop_2level_64k(ulong, ulong, physaddr_t *, int);
static int arm64_vtop_3level_64k(ulong, ulong, physaddr_t *, int);
static int arm64_vtop_2level_16k(ulong, ulong, physaddr_t *, int);
static int arm64_vtop_3level_16k(ulong, ulong, physaddr_t *, int);
static int arm64_vtop_4level_16k(ulong, ulong, physaddr_t *, int);
static int arm64_vtop_3level_4k(ulong, ulong, physaddr_t *, int);
static int arm64_vtop_4level_4k(ulong, ulong, physaddr_t *, int);
static ulong arm64_get_task_pgd(ulong);
static void arm64_irq_stack_init(void);
static void arm64_overflow_stack_init(void);
static void arm64_stackframe_init(void);
static int arm64_eframe_search(struct bt_info *);
static int arm64_is_kernel_exception_frame(struct bt_info *, ulong);
static int arm64_in_exception_text(ulong);
static int arm64_in_exp_entry(ulong);
static void arm64_back_trace_cmd(struct bt_info *);
static void arm64_back_trace_cmd_v2(struct bt_info *);
static void arm64_print_text_symbols(struct bt_info *, struct arm64_stackframe *, FILE *);
static int arm64_print_stackframe_entry(struct bt_info *, int, struct arm64_stackframe *, FILE *);
static int arm64_print_stackframe_entry_v2(struct bt_info *, int, struct arm64_stackframe *, FILE *);
static void arm64_display_full_frame(struct bt_info *, ulong);
static void arm64_display_full_frame_v2(struct bt_info *, struct arm64_stackframe *, struct arm64_stackframe *);
static int arm64_unwind_frame(struct bt_info *, struct arm64_stackframe *);
static int arm64_unwind_frame_v2(struct bt_info *, struct arm64_stackframe *, FILE *);
static int arm64_get_dumpfile_stackframe(struct bt_info *, struct arm64_stackframe *);
static int arm64_in_kdump_text(struct bt_info *, struct arm64_stackframe *);
static int arm64_in_kdump_text_on_irq_stack(struct bt_info *);
static int arm64_switch_stack(struct bt_info *, struct arm64_stackframe *, FILE *);
static int arm64_switch_stack_from_overflow(struct bt_info *, struct arm64_stackframe *, FILE *);
static int arm64_get_stackframe(struct bt_info *, struct arm64_stackframe *);
static void arm64_get_stack_frame(struct bt_info *, ulong *, ulong *);
static void arm64_gen_hidden_frame(struct bt_info *bt, ulong, struct arm64_stackframe *);
static void arm64_print_exception_frame(struct bt_info *, ulong, int, FILE *);
static void arm64_do_bt_reference_check(struct bt_info *, ulong, char *);
static int arm64_translate_pte(ulong, void *, ulonglong);
static ulong arm64_vmalloc_start(void);
static int arm64_is_task_addr(ulong);
static int arm64_dis_filter(ulong, char *, unsigned int);
static void arm64_cmd_mach(void);
static void arm64_display_machine_stats(void);
static int arm64_get_smp_cpus(void);
static void arm64_clear_machdep_cache(void);
static int arm64_on_process_stack(struct bt_info *, ulong);
static int arm64_in_alternate_stack(int, ulong);
static int arm64_in_alternate_stackv(int cpu, ulong stkptr, ulong *stacks, ulong stack_size);
static int arm64_on_irq_stack(int, ulong);
static int arm64_on_overflow_stack(int, ulong);
static void arm64_set_irq_stack(struct bt_info *);
static void arm64_set_overflow_stack(struct bt_info *);
static void arm64_set_process_stack(struct bt_info *);
static int arm64_get_kvaddr_ranges(struct vaddr_range *);
static void arm64_get_crash_notes(void);
static void arm64_calc_VA_BITS(void);
static int arm64_is_uvaddr(ulong, struct task_context *);
static void arm64_calc_KERNELPACMASK(void);
static void arm64_recalc_KERNELPACMASK(void);
static int arm64_get_vmcoreinfo(unsigned long *vaddr, const char *label, int base);
static ulong arm64_set_irq_stack_size(void);
struct kernel_range {
unsigned long modules_vaddr, modules_end;
unsigned long vmalloc_start_addr, vmalloc_end;
unsigned long vmemmap_vaddr, vmemmap_end;
};
static struct kernel_range *arm64_get_va_range(struct machine_specific *ms);
static void arm64_get_struct_page_size(struct machine_specific *ms);
/* mte tag shift bit */
#define MTE_TAG_SHIFT 56
/* native kernel pointers tag */
#define KASAN_TAG_KERNEL 0xFF
/* minimum value for random tags */
#define KASAN_TAG_MIN 0xF0
/* right shift the tag to MTE_TAG_SHIFT bit */
#define mte_tag_shifted(tag) ((ulong)(tag) << MTE_TAG_SHIFT)
/* get the top byte value of the original kvaddr */
#define mte_tag_get(addr) (unsigned char)((ulong)(addr) >> MTE_TAG_SHIFT)
/* reset the top byte to get an untaggged kvaddr */
#define mte_tag_reset(addr) (((ulong)addr & ~mte_tag_shifted(KASAN_TAG_KERNEL)) | \
mte_tag_shifted(KASAN_TAG_KERNEL))
struct user_regs_bitmap_struct {
struct arm64_pt_regs ur;
ulong bitmap[32];
};
static inline bool is_mte_kvaddr(ulong addr)
{
/* check for ARM64_MTE enabled */
if (!(machdep->flags & ARM64_MTE))
return false;
/* check the validity of HW Tag-Based kvaddr */
if (mte_tag_get(addr) >= KASAN_TAG_MIN && mte_tag_get(addr) < KASAN_TAG_KERNEL)
return true;
return false;
}
static int arm64_is_kvaddr(ulong addr)
{
if (is_mte_kvaddr(addr))
return (mte_tag_reset(addr) >= (ulong)(machdep->kvbase));
return (addr >= (ulong)(machdep->kvbase));
}
static void arm64_calc_kernel_start(void)
{
struct machine_specific *ms = machdep->machspec;
struct syment *sp;
if (THIS_KERNEL_VERSION >= LINUX(5,11,0))
sp = kernel_symbol_search("_stext");
else
sp = kernel_symbol_search("_text");
ms->kimage_text = (sp ? sp->value : 0);
sp = kernel_symbol_search("_end");
ms->kimage_end = (sp ? sp->value : 0);
}
static int
arm64_vmemmap_is_page_ptr(ulong addr, physaddr_t *phys)
{
ulong size = SIZE(page);
ulong pfn, nr;
if (IS_SPARSEMEM() && (machdep->flags & VMEMMAP) &&
(addr >= VMEMMAP_VADDR && addr <= VMEMMAP_END) &&
!((addr - VMEMMAP_VADDR) % size)) {
pfn = (addr - machdep->machspec->vmemmap) / size;
nr = pfn_to_section_nr(pfn);
if (valid_section_nr(nr)) {
if (phys)
*phys = PTOB(pfn);
return TRUE;
}
}
return FALSE;
}
static void arm64_get_vmemmap_page_ptr(void)
{
struct machine_specific *ms = machdep->machspec;
/* If vmemmap exists, it means kernel enabled CONFIG_SPARSEMEM_VMEMMAP */
if (arm64_get_vmcoreinfo(&ms->vmemmap, "SYMBOL(vmemmap)", NUM_HEX))
goto out;
/* The global symbol of vmemmap is removed since kernel commit 7bc1a0f9e1765 */
if (kernel_symbol_exists("vmemmap"))
ms->vmemmap = symbol_value("vmemmap");
else
ms->vmemmap = ms->vmemmap_vaddr - ((ms->phys_offset >> machdep->pageshift) * ms->struct_page_size);
out:
if (ms->vmemmap)
machdep->is_page_ptr = arm64_vmemmap_is_page_ptr;
}
static int
arm64_get_current_task_reg(int regno, const char *name,
int size, void *value)
{
struct bt_info bt_info, bt_setup;
struct task_context *tc;
struct user_regs_bitmap_struct *ur_bitmap;
ulong ip, sp;
bool ret = FALSE;
switch (regno) {
case X0_REGNUM ... PC_REGNUM:
break;
default:
return FALSE;
}
tc = CURRENT_CONTEXT();
if (!tc)
return FALSE;
BZERO(&bt_setup, sizeof(struct bt_info));
clone_bt_info(&bt_setup, &bt_info, tc);
fill_stackbuf(&bt_info);
get_dumpfile_regs(&bt_info, &sp, &ip);
if (bt_info.stackbuf)
FREEBUF(bt_info.stackbuf);
ur_bitmap = (struct user_regs_bitmap_struct *)bt_info.machdep;
if (!ur_bitmap)
return FALSE;
if (!bt_info.need_free) {
goto get_all;
}
switch (regno) {
case X0_REGNUM ... X30_REGNUM:
if (!NUM_IN_BITMAP(ur_bitmap->bitmap,
REG_SEQ(arm64_pt_regs, regs[0]) + regno - X0_REGNUM)) {
FREEBUF(ur_bitmap);
return FALSE;
}
break;
case SP_REGNUM:
if (!NUM_IN_BITMAP(ur_bitmap->bitmap,
REG_SEQ(arm64_pt_regs, sp))) {
FREEBUF(ur_bitmap);
return FALSE;
}
break;
case PC_REGNUM:
if (!NUM_IN_BITMAP(ur_bitmap->bitmap,
REG_SEQ(arm64_pt_regs, pc))) {
FREEBUF(ur_bitmap);
return FALSE;
}
break;
}
get_all:
switch (regno) {
case X0_REGNUM ... X30_REGNUM:
if (size != sizeof(ur_bitmap->ur.regs[regno]))
break;
memcpy(value, &ur_bitmap->ur.regs[regno], size);
ret = TRUE;
break;
case SP_REGNUM:
if (size != sizeof(ur_bitmap->ur.sp))
break;
memcpy(value, &ur_bitmap->ur.sp, size);
ret = TRUE;
break;
case PC_REGNUM:
if (size != sizeof(ur_bitmap->ur.pc))
break;
memcpy(value, &ur_bitmap->ur.pc, size);
ret = TRUE;
break;
}
if (bt_info.need_free) {
FREEBUF(ur_bitmap);
bt_info.need_free = FALSE;
}
return ret;
}
/*
* Do all necessary machine-specific setup here. This is called several times
* during initialization.
*/
void
arm64_init(int when)
{
ulong value;
struct machine_specific *ms;
#if defined(__x86_64__)
if (ACTIVE())
error(FATAL, "compiled for the ARM64 architecture\n");
#endif
switch (when) {
case SETUP_ENV:
machdep->process_elf_notes = process_elf64_notes;
break;
case PRE_SYMTAB:
machdep->machspec = &arm64_machine_specific;
machdep->verify_symbol = arm64_verify_symbol;
if (pc->flags & KERNEL_DEBUG_QUERY)
return;
machdep->verify_paddr = generic_verify_paddr;
if (machdep->cmdline_args[0])
arm64_parse_cmdline_args();
machdep->flags |= MACHDEP_BT_TEXT;
ms = machdep->machspec;
/*
* The st->_stext_vmlinux is needed in arm64_init(PRE_GDB) when a
* dumpfile does not have vmcoreinfo and we use -m vabits_actual
* option, e.g. a raw RAM dumpfile.
*/
if (ms->VA_BITS_ACTUAL)
st->_stext_vmlinux = UNINITIALIZED;
if (!ms->kimage_voffset && STREQ(pc->live_memsrc, "/dev/crash"))
ioctl(pc->mfd, DEV_CRASH_ARCH_DATA, &ms->kimage_voffset);
if (!ms->kimage_voffset)
arm64_get_vmcoreinfo(&ms->kimage_voffset, "NUMBER(kimage_voffset)", NUM_HEX);
if (ms->kimage_voffset ||
(ACTIVE() && (symbol_value_from_proc_kallsyms("kimage_voffset") != BADVAL))) {
machdep->flags |= NEW_VMEMMAP;
/*
* Even if CONFIG_RANDOMIZE_BASE is not configured,
* derive_kaslr_offset() should work and set
* kt->relocate to 0
*/
if (!kt->relocate && !(kt->flags2 & (RELOC_AUTO|KASLR)))
kt->flags2 |= (RELOC_AUTO|KASLR);
}
break;
case PRE_GDB:
if (kernel_symbol_exists("kimage_voffset"))
machdep->flags |= NEW_VMEMMAP;
if (kernel_symbol_exists("cpu_enable_mte"))
machdep->flags |= ARM64_MTE;
if (!machdep->pagesize && arm64_get_vmcoreinfo(&value, "PAGESIZE", NUM_DEC))
machdep->pagesize = (unsigned int)value;
if (!machdep->pagesize) {
/*
* Kerneldoc Documentation/arm64/booting.txt describes
* the kernel image header flags field.
*/
value = machdep->machspec->kernel_flags;
value = (value >> 1) & 3;
switch(value)
{
case 0:
break;
case 1:
machdep->pagesize = 4096;
break;
case 2:
machdep->pagesize = 16384;
break;
case 3:
machdep->pagesize = 65536;
break;
}
}
/*
* This code section will only be executed if the kernel is
* earlier than Linux 4.4 (if there is no vmcoreinfo)
*/
if (!machdep->pagesize &&
kernel_symbol_exists("swapper_pg_dir") &&
kernel_symbol_exists("idmap_pg_dir")) {
value = symbol_value("swapper_pg_dir") -
symbol_value("idmap_pg_dir");
/*
* idmap_pg_dir is 2 pages prior to 4.1,
* and 3 pages thereafter. Only 4K and 64K
* page sizes are supported.
*/
switch (value)
{
case (4096 * 2):
case (4096 * 3):
machdep->pagesize = 4096;
break;
case (65536 * 2):
case (65536 * 3):
machdep->pagesize = 65536;
break;
}
} else if (ACTIVE())
machdep->pagesize = memory_page_size(); /* host */
machdep->pageshift = ffs(machdep->pagesize) - 1;
machdep->pageoffset = machdep->pagesize - 1;
machdep->pagemask = ~((ulonglong)machdep->pageoffset);
ms = machdep->machspec;
arm64_get_struct_page_size(ms);
arm64_calc_VA_BITS();
arm64_calc_KERNELPACMASK();
/* vabits_actual introduced after mm flip, so it should be flipped layout */
if (ms->VA_BITS_ACTUAL) {
ms->page_offset = ARM64_FLIP_PAGE_OFFSET;
/* useless on arm64 */
machdep->identity_map_base = ARM64_FLIP_PAGE_OFFSET;
machdep->kvbase = ARM64_FLIP_PAGE_OFFSET;
ms->userspace_top = ARM64_USERSPACE_TOP_ACTUAL;
} else {
ms->page_offset = ARM64_PAGE_OFFSET;
machdep->identity_map_base = ARM64_PAGE_OFFSET;
machdep->kvbase = ARM64_VA_START;
ms->userspace_top = ARM64_USERSPACE_TOP;
}
machdep->is_kvaddr = arm64_is_kvaddr;
machdep->kvtop = arm64_kvtop;
/* The defaults */
ms->vmalloc_end = ARM64_VMALLOC_END;
ms->vmemmap_vaddr = ARM64_VMEMMAP_VADDR;
ms->vmemmap_end = ARM64_VMEMMAP_END;
if (machdep->flags & NEW_VMEMMAP) {
struct syment *sp;
struct kernel_range *r;
/* It is finally decided in arm64_calc_kernel_start() */
sp = kernel_symbol_search("_text");
ms->kimage_text = (sp ? sp->value : 0);
sp = kernel_symbol_search("_end");
ms->kimage_end = (sp ? sp->value : 0);
if (ms->struct_page_size && (r = arm64_get_va_range(ms))) {
/* We can get all the MODULES/VMALLOC/VMEMMAP ranges now.*/
ms->modules_vaddr = r->modules_vaddr;
ms->modules_end = r->modules_end - 1;
ms->vmalloc_start_addr = r->vmalloc_start_addr;
ms->vmalloc_end = r->vmalloc_end - 1;
ms->vmemmap_vaddr = r->vmemmap_vaddr;
ms->vmemmap_end = r->vmemmap_end - 1;
} else if (ms->VA_BITS_ACTUAL) {
ms->modules_vaddr = (st->_stext_vmlinux & TEXT_OFFSET_MASK) - ARM64_MODULES_VSIZE;
ms->modules_end = ms->modules_vaddr + ARM64_MODULES_VSIZE -1;
ms->vmalloc_start_addr = ms->modules_end + 1;
} else {
ms->modules_vaddr = ARM64_VA_START;
if (kernel_symbol_exists("kasan_init"))
ms->modules_vaddr += ARM64_KASAN_SHADOW_SIZE;
ms->modules_end = ms->modules_vaddr + ARM64_MODULES_VSIZE -1;
ms->vmalloc_start_addr = ms->modules_end + 1;
}
arm64_calc_kimage_voffset();
} else {
ms->modules_vaddr = ARM64_PAGE_OFFSET - MEGABYTES(64);
ms->modules_end = ARM64_PAGE_OFFSET - 1;
ms->vmalloc_start_addr = ARM64_VA_START;
}
switch (machdep->pagesize)
{
case 4096:
machdep->ptrs_per_pgd = PTRS_PER_PGD_L3_4K;
if ((machdep->pgd =
(char *)malloc(PTRS_PER_PGD_L3_4K * 8)) == NULL)
error(FATAL, "cannot malloc pgd space.");
if (machdep->machspec->VA_BITS > PGDIR_SHIFT_L4_4K) {
machdep->flags |= VM_L4_4K;
if ((machdep->pud =
(char *)malloc(PTRS_PER_PUD_L4_4K * 8))
== NULL)
error(FATAL, "cannot malloc pud space.");
} else {
machdep->flags |= VM_L3_4K;
machdep->pud = NULL; /* not used */
}
if ((machdep->pmd =
(char *)malloc(PTRS_PER_PMD_L3_4K * 8)) == NULL)
error(FATAL, "cannot malloc pmd space.");
if ((machdep->ptbl =
(char *)malloc(PTRS_PER_PTE_L3_4K * 8)) == NULL)
error(FATAL, "cannot malloc ptbl space.");
break;
case 16384:
if (machdep->machspec->VA_BITS == 48) {
machdep->flags |= VM_L4_16K;
if (!machdep->ptrs_per_pgd)
machdep->ptrs_per_pgd = PTRS_PER_PGD_L4_16K;
if ((machdep->pgd =
(char *)malloc(machdep->ptrs_per_pgd * 8)) == NULL)
error(FATAL, "cannot malloc pgd space.");
if ((machdep->pud =
(char *)malloc(PTRS_PER_PUD_L4_16K * 8)) == NULL)
error(FATAL, "cannot malloc pud space.");
if ((machdep->pmd =
(char *)malloc(PTRS_PER_PMD_L4_16K * 8)) == NULL)
error(FATAL, "cannot malloc pmd space.");
if ((machdep->ptbl =
(char *)malloc(PTRS_PER_PTE_L4_16K * 8)) == NULL)
error(FATAL, "cannot malloc ptbl space.");
}
else if (machdep->machspec->VA_BITS == 47) {
machdep->flags |= VM_L3_16K;
if (!machdep->ptrs_per_pgd)
machdep->ptrs_per_pgd = PTRS_PER_PGD_L3_16K;
if ((machdep->pgd =
(char *)malloc(machdep->ptrs_per_pgd * 8)) == NULL)
error(FATAL, "cannot malloc pgd space.");
if ((machdep->pmd =
(char *)malloc(PTRS_PER_PMD_L3_16K * 8)) == NULL)
error(FATAL, "cannot malloc pmd space.");
if ((machdep->ptbl =
(char *)malloc(PTRS_PER_PTE_L3_16K * 8)) == NULL)
error(FATAL, "cannot malloc ptbl space.");
machdep->pud = NULL; /* not used */
} else if (machdep->machspec->VA_BITS == 36) {
machdep->flags |= VM_L2_16K;
if (!machdep->ptrs_per_pgd)
machdep->ptrs_per_pgd = PTRS_PER_PGD_L2_16K;
if ((machdep->pgd =
(char *)malloc(machdep->ptrs_per_pgd * 8)) == NULL)
error(FATAL, "cannot malloc pgd space.");
if ((machdep->ptbl =
(char *)malloc(PTRS_PER_PTE_L2_16K * 8)) == NULL)
error(FATAL, "cannot malloc ptbl space.");
machdep->pmd = NULL; /* not used */
machdep->pud = NULL; /* not used */
} else {
error(FATAL, "Do not support 52 bits, 4-level for 16K page now.");
}
break;
case 65536:
/*
* idmap_ptrs_per_pgd has been removed since Linux-v6.0-rc1, see:
* commit ebd9aea1f27e ("arm64: head: drop idmap_ptrs_per_pgd")
*/
if (kernel_symbol_exists("idmap_ptrs_per_pgd") &&
readmem(symbol_value("idmap_ptrs_per_pgd"), KVADDR,
&value, sizeof(ulong), "idmap_ptrs_per_pgd", QUIET|RETURN_ON_ERROR))
machdep->ptrs_per_pgd = value;
if (machdep->machspec->VA_BITS > PGDIR_SHIFT_L3_64K) {
machdep->flags |= VM_L3_64K;
if (!machdep->ptrs_per_pgd) {
if (machdep->machspec->VA_BITS == 52)
machdep->ptrs_per_pgd = PTRS_PER_PGD_L3_64K_52;
else if (machdep->machspec->VA_BITS == 48)
machdep->ptrs_per_pgd = PTRS_PER_PGD_L3_64K_48;
else
error(FATAL, "wrong VA_BITS for 64K page.");
}
if ((machdep->pgd =
(char *)malloc(machdep->ptrs_per_pgd * 8)) == NULL)
error(FATAL, "cannot malloc pgd space.");
if ((machdep->pmd =
(char *)malloc(PTRS_PER_PMD_L3_64K * 8)) == NULL)
error(FATAL, "cannot malloc pmd space.");
if ((machdep->ptbl =
(char *)malloc(PTRS_PER_PTE_L3_64K * 8)) == NULL)
error(FATAL, "cannot malloc ptbl space.");
} else {
machdep->flags |= VM_L2_64K;
if (!machdep->ptrs_per_pgd)
machdep->ptrs_per_pgd = PTRS_PER_PGD_L2_64K;
if ((machdep->pgd =
(char *)malloc(machdep->ptrs_per_pgd * 8)) == NULL)
error(FATAL, "cannot malloc pgd space.");
if ((machdep->ptbl =
(char *)malloc(PTRS_PER_PTE_L2_64K * 8)) == NULL)
error(FATAL, "cannot malloc ptbl space.");
machdep->pmd = NULL; /* not used */
}
machdep->pud = NULL; /* not used */
break;
default:
if (machdep->pagesize)
error(FATAL, "invalid/unsupported page size: %d\n",
machdep->pagesize);
else
error(FATAL, "cannot determine page size\n");
}
machdep->last_pgd_read = 0;
machdep->last_pud_read = 0;
machdep->last_pmd_read = 0;
machdep->last_ptbl_read = 0;
machdep->clear_machdep_cache = arm64_clear_machdep_cache;
machdep->stacksize = ARM64_STACK_SIZE;
machdep->flags |= VMEMMAP;
machdep->uvtop = arm64_uvtop;
machdep->is_uvaddr = arm64_is_uvaddr;
machdep->eframe_search = arm64_eframe_search;
machdep->back_trace = arm64_back_trace_cmd;
machdep->in_alternate_stack = arm64_in_alternate_stack;
machdep->processor_speed = arm64_processor_speed;
machdep->get_task_pgd = arm64_get_task_pgd;
machdep->get_stack_frame = arm64_get_stack_frame;
machdep->get_stackbase = generic_get_stackbase;
machdep->get_stacktop = generic_get_stacktop;
machdep->translate_pte = arm64_translate_pte;
machdep->memory_size = generic_memory_size;
machdep->vmalloc_start = arm64_vmalloc_start;
machdep->get_kvaddr_ranges = arm64_get_kvaddr_ranges;
machdep->is_task_addr = arm64_is_task_addr;
machdep->dis_filter = arm64_dis_filter;
machdep->cmd_mach = arm64_cmd_mach;
machdep->get_smp_cpus = arm64_get_smp_cpus;
machdep->line_number_hooks = NULL;
machdep->value_to_symbol = generic_machdep_value_to_symbol;
machdep->dump_irq = generic_dump_irq;
machdep->show_interrupts = generic_show_interrupts;
machdep->get_irq_affinity = generic_get_irq_affinity;
machdep->dumpfile_init = NULL;
machdep->verify_line_number = NULL;
machdep->init_kernel_pgd = arm64_init_kernel_pgd;
machdep->get_current_task_reg = arm64_get_current_task_reg;
/* use machdep parameters */
arm64_calc_phys_offset();
arm64_calc_physvirt_offset();
if (CRASHDEBUG(1)) {
if (machdep->flags & NEW_VMEMMAP)
fprintf(fp, "kimage_voffset: %lx\n",
machdep->machspec->kimage_voffset);
fprintf(fp, "phys_offset: %lx\n",
machdep->machspec->phys_offset);
fprintf(fp, "physvirt_offset: %lx\n", machdep->machspec->physvirt_offset);
}
break;
case POST_GDB:
/* Rely on kernel version to decide the kernel start address */
arm64_calc_kernel_start();
/* Can we get the size of struct page before POST_GDB */
ms = machdep->machspec;
if (!ms->struct_page_size)
arm64_calc_virtual_memory_ranges();
arm64_get_vmemmap_page_ptr();
arm64_get_section_size_bits();
if (!machdep->max_physmem_bits) {
if (arm64_get_vmcoreinfo(&machdep->max_physmem_bits, "NUMBER(MAX_PHYSMEM_BITS)", NUM_DEC)) {
/* nothing */
} else if (machdep->machspec->VA_BITS == 52) /* guess */
machdep->max_physmem_bits = _MAX_PHYSMEM_BITS_52;
else if (THIS_KERNEL_VERSION >= LINUX(3,17,0))
machdep->max_physmem_bits = _MAX_PHYSMEM_BITS_3_17;
else
machdep->max_physmem_bits = _MAX_PHYSMEM_BITS;
}
if (CRASHDEBUG(1)) {
if (ms->VA_BITS_ACTUAL) {
fprintf(fp, "CONFIG_ARM64_VA_BITS: %ld\n", ms->CONFIG_ARM64_VA_BITS);
fprintf(fp, " VA_BITS_ACTUAL: %ld\n", ms->VA_BITS_ACTUAL);
fprintf(fp, "(calculated) VA_BITS: %ld\n", ms->VA_BITS);
fprintf(fp, " PAGE_OFFSET: %lx\n", ARM64_FLIP_PAGE_OFFSET);
fprintf(fp, " VA_START: %lx\n", ms->VA_START);
fprintf(fp, " modules: %lx - %lx\n", ms->modules_vaddr, ms->modules_end);
fprintf(fp, " vmalloc: %lx - %lx\n", ms->vmalloc_start_addr, ms->vmalloc_end);
fprintf(fp, "kernel image: %lx - %lx\n", ms->kimage_text, ms->kimage_end);
fprintf(fp, " vmemmap: %lx - %lx\n\n", ms->vmemmap_vaddr, ms->vmemmap_end);
}
}
if (THIS_KERNEL_VERSION >= LINUX(5,19,0)) {
ms->__SWP_TYPE_BITS = 5;
ms->__SWP_TYPE_SHIFT = 3;
ms->__SWP_TYPE_MASK = ((1UL << ms->__SWP_TYPE_BITS) - 1);
ms->__SWP_OFFSET_SHIFT = (ms->__SWP_TYPE_BITS + ms->__SWP_TYPE_SHIFT);
ms->__SWP_OFFSET_BITS = 50;
ms->__SWP_OFFSET_MASK = ((1UL << ms->__SWP_OFFSET_BITS) - 1);
ms->PTE_PROT_NONE = (1UL << 58);
ms->PTE_FILE = 0; /* unused */
} else if (THIS_KERNEL_VERSION >= LINUX(4,0,0)) {
ms->__SWP_TYPE_BITS = 6;
ms->__SWP_TYPE_SHIFT = 2;
ms->__SWP_TYPE_MASK = ((1UL << ms->__SWP_TYPE_BITS) - 1);
ms->__SWP_OFFSET_SHIFT = (ms->__SWP_TYPE_BITS + ms->__SWP_TYPE_SHIFT);
ms->__SWP_OFFSET_BITS = 50;
ms->__SWP_OFFSET_MASK = ((1UL << ms->__SWP_OFFSET_BITS) - 1);
ms->PTE_PROT_NONE = (1UL << 58);
ms->PTE_FILE = 0; /* unused */
} else if (THIS_KERNEL_VERSION >= LINUX(3,13,0)) {
ms->__SWP_TYPE_BITS = 6;
ms->__SWP_TYPE_SHIFT = 3;
ms->__SWP_TYPE_MASK = ((1UL << ms->__SWP_TYPE_BITS) - 1);
ms->__SWP_OFFSET_SHIFT = (ms->__SWP_TYPE_BITS + ms->__SWP_TYPE_SHIFT);
ms->__SWP_OFFSET_BITS = 49;
ms->__SWP_OFFSET_MASK = ((1UL << ms->__SWP_OFFSET_BITS) - 1);
ms->PTE_PROT_NONE = (1UL << 58);
ms->PTE_FILE = (1UL << 2);
} else if (THIS_KERNEL_VERSION >= LINUX(3,11,0)) {
ms->__SWP_TYPE_BITS = 6;
ms->__SWP_TYPE_SHIFT = 4;
ms->__SWP_TYPE_MASK = ((1UL << ms->__SWP_TYPE_BITS) - 1);
ms->__SWP_OFFSET_SHIFT = (ms->__SWP_TYPE_BITS + ms->__SWP_TYPE_SHIFT);
ms->__SWP_OFFSET_BITS = 0; /* unused */
ms->__SWP_OFFSET_MASK = 0; /* unused */
ms->PTE_PROT_NONE = (1UL << 2);
ms->PTE_FILE = (1UL << 3);
} else {
ms->__SWP_TYPE_BITS = 6;
ms->__SWP_TYPE_SHIFT = 3;
ms->__SWP_TYPE_MASK = ((1UL << ms->__SWP_TYPE_BITS) - 1);
ms->__SWP_OFFSET_SHIFT = (ms->__SWP_TYPE_BITS + ms->__SWP_TYPE_SHIFT);
ms->__SWP_OFFSET_BITS = 0; /* unused */
ms->__SWP_OFFSET_MASK = 0; /* unused */
ms->PTE_PROT_NONE = (1UL << 1);
ms->PTE_FILE = (1UL << 2);
}
if (symbol_exists("irq_desc"))
ARRAY_LENGTH_INIT(machdep->nr_irqs, irq_desc,
"irq_desc", NULL, 0);
else if (kernel_symbol_exists("nr_irqs"))
get_symbol_data("nr_irqs", sizeof(unsigned int),
&machdep->nr_irqs);
if (!machdep->hz)
machdep->hz = 100;
/*
* Let's calculate the KERNELPACMASK value based on the
* vabits, see:
* arch/arm64/kernel/vmcore_info.c
* arch/arm64/include/asm/pointer_auth.h
*/
if(!machdep->machspec->CONFIG_ARM64_KERNELPACMASK)
arm64_recalc_KERNELPACMASK();
arm64_irq_stack_init();
arm64_overflow_stack_init();
arm64_stackframe_init();
break;
case POST_INIT:
/*
* crash_notes contains machine specific information about the
* crash. In particular, it contains CPU registers at the time
* of the crash. We need this information to extract correct
* backtraces from the panic task.
*/
if (!LIVE())
arm64_get_crash_notes();
gdb_change_thread_context();
break;
case LOG_ONLY:
machdep->machspec = &arm64_machine_specific;
arm64_calc_VA_BITS();
arm64_calc_KERNELPACMASK();
arm64_calc_phys_offset();
machdep->machspec->page_offset = ARM64_PAGE_OFFSET;
arm64_calc_physvirt_offset();
break;
}
}
struct kernel_va_range_handler {
unsigned long kernel_versions_start; /* include */
unsigned long kernel_versions_end; /* exclude */
struct kernel_range *(*get_range)(struct machine_specific *);
};
static struct kernel_range tmp_range;
#define _PAGE_END(va) (-(1UL << ((va) - 1)))
#define SZ_64K 0x00010000
#define SZ_2M 0x00200000
/*
* Get the max shift of the size of struct page.
* Most of the time, it is 64 bytes, but not sure.
*/
static int arm64_get_struct_page_max_shift(struct machine_specific *ms)
{
return (int)ceil(log2(ms->struct_page_size));
}
/* Return TRUE if we succeed, return FALSE on failure. */
static int
arm64_get_vmcoreinfo(unsigned long *vaddr, const char *label, int base)
{
int err = 0;
char *string = pc->read_vmcoreinfo(label);
if (!string)
return FALSE;
switch (base) {
case NUM_HEX:
*vaddr = strtoul(string, NULL, 16);
break;
case NUM_DEC:
*vaddr = strtoul(string, NULL, 10);
break;
default:
err++;
error(INFO, "Unknown type:%#x, (NUM_HEX|NUM_DEC)\n", base);
}
free(string);
return err ? FALSE: TRUE;
}
/*
* The change is caused by the kernel patch since v5.18-rc1:
* "arm64: crash_core: Export MODULES, VMALLOC, and VMEMMAP ranges"
*/
static struct kernel_range *arm64_get_range_v5_18(struct machine_specific *ms)
{
struct kernel_range *r = &tmp_range;
/* Get the MODULES_VADDR ~ MODULES_END */
if (!arm64_get_vmcoreinfo(&r->modules_vaddr, "NUMBER(MODULES_VADDR)", NUM_HEX))
return NULL;
if (!arm64_get_vmcoreinfo(&r->modules_end, "NUMBER(MODULES_END)", NUM_HEX))
return NULL;
/* Get the VMEMMAP_START ~ VMEMMAP_END */
if (!arm64_get_vmcoreinfo(&r->vmemmap_vaddr, "NUMBER(VMEMMAP_START)", NUM_HEX))
return NULL;
if (!arm64_get_vmcoreinfo(&r->vmemmap_end, "NUMBER(VMEMMAP_END)", NUM_HEX))
return NULL;
/* Get the VMALLOC_START ~ VMALLOC_END */
if (!arm64_get_vmcoreinfo(&r->vmalloc_start_addr, "NUMBER(VMALLOC_START)", NUM_HEX))
return NULL;
if (!arm64_get_vmcoreinfo(&r->vmalloc_end, "NUMBER(VMALLOC_END)", NUM_HEX))
return NULL;
return r;
}
/*
* The change is caused by the kernel patch since v5.17-rc1:
* "b89ddf4cca43 arm64/bpf: Remove 128MB limit for BPF JIT programs"
*/
static struct kernel_range *arm64_get_range_v5_17(struct machine_specific *ms)
{
struct kernel_range *r = &tmp_range;
unsigned long v = ms->CONFIG_ARM64_VA_BITS;
unsigned long vmem_shift, vmemmap_size;
/* Not initialized yet */
if (v == 0)
return NULL;
if (v > 48)
v = 48;
/* Get the MODULES_VADDR ~ MODULES_END */
r->modules_vaddr = _PAGE_END(v);
r->modules_end = r->modules_vaddr + MEGABYTES(128);
/* Get the VMEMMAP_START ~ VMEMMAP_END */
vmem_shift = machdep->pageshift - arm64_get_struct_page_max_shift(ms);
vmemmap_size = (_PAGE_END(v) - PAGE_OFFSET) >> vmem_shift;
r->vmemmap_vaddr = (-(1UL << (ms->CONFIG_ARM64_VA_BITS - vmem_shift)));
r->vmemmap_end = r->vmemmap_vaddr + vmemmap_size;
/* Get the VMALLOC_START ~ VMALLOC_END */
r->vmalloc_start_addr = r->modules_end;
r->vmalloc_end = r->vmemmap_vaddr - MEGABYTES(256);
return r;
}
/*
* The change is caused by the kernel patch since v5.11:
* "9ad7c6d5e75b arm64: mm: tidy up top of kernel VA space"
*/
static struct kernel_range *arm64_get_range_v5_11(struct machine_specific *ms)
{
struct kernel_range *r = &tmp_range;
unsigned long v = ms->CONFIG_ARM64_VA_BITS;
unsigned long vmem_shift, vmemmap_size, bpf_jit_size = MEGABYTES(128);
/* Not initialized yet */
if (v == 0)
return NULL;
if (v > 48)
v = 48;
/* Get the MODULES_VADDR ~ MODULES_END */
r->modules_vaddr = _PAGE_END(v) + bpf_jit_size;
r->modules_end = r->modules_vaddr + MEGABYTES(128);
/* Get the VMEMMAP_START ~ VMEMMAP_END */
vmem_shift = machdep->pageshift - arm64_get_struct_page_max_shift(ms);
vmemmap_size = (_PAGE_END(v) - PAGE_OFFSET) >> vmem_shift;
r->vmemmap_vaddr = (-(1UL << (ms->CONFIG_ARM64_VA_BITS - vmem_shift)));
r->vmemmap_end = r->vmemmap_vaddr + vmemmap_size;
/* Get the VMALLOC_START ~ VMALLOC_END */
r->vmalloc_start_addr = r->modules_end;
r->vmalloc_end = r->vmemmap_vaddr - MEGABYTES(256);
return r;
}
static unsigned long arm64_get_pud_size(void)
{
unsigned long PUD_SIZE = 0;
switch (machdep->pagesize) {
case 4096:
if (machdep->machspec->VA_BITS > PGDIR_SHIFT_L4_4K) {
PUD_SIZE = PUD_SIZE_L4_4K;
} else {
PUD_SIZE = PGDIR_SIZE_L3_4K;
}
break;
case 65536:
PUD_SIZE = PGDIR_SIZE_L2_64K;
default:
break;
}
return PUD_SIZE;
}
/*
* The change is caused by the kernel patches since v5.4, such as:
* "ce3aaed87344 arm64: mm: Modify calculation of VMEMMAP_SIZE"
* "14c127c957c1 arm64: mm: Flip kernel VA space"
*/
static struct kernel_range *arm64_get_range_v5_4(struct machine_specific *ms)
{
struct kernel_range *r = &tmp_range;
unsigned long v = ms->CONFIG_ARM64_VA_BITS;
unsigned long kasan_shadow_shift, kasan_shadow_offset, PUD_SIZE;
unsigned long vmem_shift, vmemmap_size, bpf_jit_size = MEGABYTES(128);
char *string;
int ret;
/* Not initialized yet */
if (v == 0)
return NULL;
if (v > 48)
v = 48;
/* Get the MODULES_VADDR ~ MODULES_END */
if (kernel_symbol_exists("kasan_init")) {
/* See the arch/arm64/Makefile */
ret = get_kernel_config("CONFIG_KASAN_SW_TAGS", NULL);
if (ret == IKCONFIG_N)
return NULL;
kasan_shadow_shift = (ret == IKCONFIG_Y) ? 4: 3;
/* See the arch/arm64/Kconfig*/
ret = get_kernel_config("CONFIG_KASAN_SHADOW_OFFSET", &string);
if (ret != IKCONFIG_STR)
return NULL;
kasan_shadow_offset = atol(string);
r->modules_vaddr = (1UL << (64 - kasan_shadow_shift)) + kasan_shadow_offset
+ bpf_jit_size;
} else {
r->modules_vaddr = _PAGE_END(v) + bpf_jit_size;
}
r->modules_end = r->modules_vaddr + MEGABYTES(128);
/* Get the VMEMMAP_START ~ VMEMMAP_END */
vmem_shift = machdep->pageshift - arm64_get_struct_page_max_shift(ms);
vmemmap_size = (_PAGE_END(v) - PAGE_OFFSET) >> vmem_shift;
r->vmemmap_vaddr = (-vmemmap_size - SZ_2M);
/*
* In the v5.7, the patch: "bbd6ec605c arm64/mm: Enable memory hot remove"
* adds the VMEMMAP_END.
*
* But before the VMEMMAP_END was added to kernel, we can also see
* the following in arch/arm64/mm/dump.c:
* { VMEMMAP_START + VMEMMAP_SIZE, "vmemmap end" },
*/
r->vmemmap_end = r->vmemmap_vaddr + vmemmap_size;
/* Get the VMALLOC_START ~ VMALLOC_END */
PUD_SIZE = arm64_get_pud_size();
r->vmalloc_start_addr = r->modules_end;
r->vmalloc_end = (-PUD_SIZE - vmemmap_size - SZ_64K);
return r;
}
/*
* The change is caused by the kernel patches since v5.0, such as:
* "91fc957c9b1d arm64/bpf: don't allocate BPF JIT programs in module memory"
*/
static struct kernel_range *arm64_get_range_v5_0(struct machine_specific *ms)
{
struct kernel_range *r = &tmp_range;
unsigned long v = ms->CONFIG_ARM64_VA_BITS;
unsigned long kasan_shadow_shift, PUD_SIZE;
unsigned long vmemmap_size, bpf_jit_size = MEGABYTES(128);
unsigned long va_start, page_offset;
int ret;
/* Not initialized yet */
if (v == 0)
return NULL;
va_start = (0xffffffffffffffffUL - (1UL << v) + 1);
page_offset = (0xffffffffffffffffUL - (1UL << (v - 1)) + 1);
/* Get the MODULES_VADDR ~ MODULES_END */
if (kernel_symbol_exists("kasan_init")) {
/* See the arch/arm64/Makefile */
ret = get_kernel_config("CONFIG_KASAN_SW_TAGS", NULL);
if (ret == IKCONFIG_N)
return NULL;
kasan_shadow_shift = (ret == IKCONFIG_Y) ? 4: 3;
r->modules_vaddr = va_start + (1UL << (v - kasan_shadow_shift)) + bpf_jit_size;
} else {
r->modules_vaddr = va_start + bpf_jit_size;
}
r->modules_end = r->modules_vaddr + MEGABYTES(128);
/* Get the VMEMMAP_START ~ VMEMMAP_END */
vmemmap_size = (1UL << (v - machdep->pageshift - 1 + arm64_get_struct_page_max_shift(ms)));
r->vmemmap_vaddr = page_offset - vmemmap_size;
r->vmemmap_end = r->vmemmap_vaddr + vmemmap_size; /* See the arch/arm64/mm/dump.c */
/* Get the VMALLOC_START ~ VMALLOC_END */
PUD_SIZE = arm64_get_pud_size();
r->vmalloc_start_addr = r->modules_end;
r->vmalloc_end = page_offset - PUD_SIZE - vmemmap_size - SZ_64K;
return r;
}
static struct kernel_va_range_handler kernel_va_range_handlers[] = {
{
LINUX(5,18,0),
LINUX(999,0,0), /* Just a boundary */
get_range: arm64_get_range_v5_18,
}, {
LINUX(5,17,0), LINUX(5,18,0),
get_range: arm64_get_range_v5_17,
}, {
LINUX(5,11,0), LINUX(5,17,0),
get_range: arm64_get_range_v5_11,
}, {
LINUX(5,4,0), LINUX(5,11,0),
get_range: arm64_get_range_v5_4,
}, {
LINUX(5,0,0), LINUX(5,4,0),
get_range: arm64_get_range_v5_0,
},
};
#define ARRAY_SIZE(a) (sizeof (a) / sizeof ((a)[0]))
static unsigned long arm64_get_kernel_version(void)
{
char *string;
if (THIS_KERNEL_VERSION)
return THIS_KERNEL_VERSION;
if ((string = pc->read_vmcoreinfo("OSRELEASE"))) {
parse_kernel_version(string);
free(string);
}
return THIS_KERNEL_VERSION;
}
/* Return NULL if we fail. */
static struct kernel_range *arm64_get_va_range(struct machine_specific *ms)
{
struct kernel_va_range_handler *h;
unsigned long kernel_version = arm64_get_kernel_version();
struct kernel_range *r = NULL;
int i;
if (!kernel_version)
goto range_failed;
for (i = 0; i < ARRAY_SIZE(kernel_va_range_handlers); i++) {
h = kernel_va_range_handlers + i;
/* Get the right hook for this kernel version */
if (h->kernel_versions_start <= kernel_version &&
kernel_version < h->kernel_versions_end) {
/* Get the correct virtual address ranges */
r = h->get_range(ms);
if (!r)
goto range_failed;
return r;
}
}
range_failed:
/* Reset ms->struct_page_size to 0 for arm64_calc_virtual_memory_ranges() */
ms->struct_page_size = 0;
return NULL;
}
/* Get the size of struct page {} */
static void arm64_get_struct_page_size(struct machine_specific *ms)
{
arm64_get_vmcoreinfo(&ms->struct_page_size, "SIZE(page)", NUM_DEC);
}
/*
* Accept or reject a symbol from the kernel namelist.
*/
static int
arm64_verify_symbol(const char *name, ulong value, char type)
{
if (!name || !strlen(name))
return FALSE;
if ((type == 'A') && STREQ(name, "_kernel_flags_le"))
machdep->machspec->kernel_flags = le64toh(value);
if ((type == 'A') && STREQ(name, "_kernel_flags_le_hi32"))
machdep->machspec->kernel_flags |= ((ulong)le32toh(value) << 32);
if ((type == 'A') && STREQ(name, "_kernel_flags_le_lo32"))
machdep->machspec->kernel_flags |= le32toh(value);
if (((type == 'A') || (type == 'a')) && (highest_bit_long(value) != 63))
return FALSE;
if ((value == 0) &&
((type == 'a') || (type == 'n') || (type == 'N') || (type == 'U')))
return FALSE;
if (STREQ(name, "$d") || STRNEQ(name, "$d.") ||
STREQ(name, "$x") || STRNEQ(name, "$x.") ||
STREQ(name, "$c") || STRNEQ(name, "$c."))
return FALSE;
if ((type == 'A') && STRNEQ(name, "__crc_"))
return FALSE;
if ((type == 'N') && strstr(name, "$d"))
return FALSE;
if (!(machdep->flags & KSYMS_START) && STREQ(name, "idmap_pg_dir"))
machdep->flags |= KSYMS_START;
return TRUE;
}
void
arm64_dump_machdep_table(ulong arg)
{
const struct machine_specific *ms = machdep->machspec;
int others, i;
others = 0;
fprintf(fp, " flags: %lx (", machdep->flags);
if (machdep->flags & KSYMS_START)
fprintf(fp, "%sKSYMS_START", others++ ? "|" : "");
if (machdep->flags & PHYS_OFFSET)
fprintf(fp, "%sPHYS_OFFSET", others++ ? "|" : "");
if (machdep->flags & VM_L2_64K)
fprintf(fp, "%sVM_L2_64K", others++ ? "|" : "");
if (machdep->flags & VM_L3_64K)
fprintf(fp, "%sVM_L3_64K", others++ ? "|" : "");
if (machdep->flags & VM_L2_16K)
fprintf(fp, "%sVM_L2_16K", others++ ? "|" : "");
if (machdep->flags & VM_L3_16K)
fprintf(fp, "%sVM_L3_16K", others++ ? "|" : "");
if (machdep->flags & VM_L4_16K)
fprintf(fp, "%sVM_L4_16K", others++ ? "|" : "");
if (machdep->flags & VM_L3_4K)
fprintf(fp, "%sVM_L3_4K", others++ ? "|" : "");
if (machdep->flags & VM_L4_4K)
fprintf(fp, "%sVM_L4_4K", others++ ? "|" : "");
if (machdep->flags & VMEMMAP)
fprintf(fp, "%sVMEMMAP", others++ ? "|" : "");
if (machdep->flags & KDUMP_ENABLED)
fprintf(fp, "%sKDUMP_ENABLED", others++ ? "|" : "");
if (machdep->flags & IRQ_STACKS)
fprintf(fp, "%sIRQ_STACKS", others++ ? "|" : "");
if (machdep->flags & UNW_4_14)
fprintf(fp, "%sUNW_4_14", others++ ? "|" : "");
if (machdep->flags & MACHDEP_BT_TEXT)
fprintf(fp, "%sMACHDEP_BT_TEXT", others++ ? "|" : "");
if (machdep->flags & NEW_VMEMMAP)
fprintf(fp, "%sNEW_VMEMMAP", others++ ? "|" : "");
if (machdep->flags & FLIPPED_VM)
fprintf(fp, "%sFLIPPED_VM", others++ ? "|" : "");
if (machdep->flags & HAS_PHYSVIRT_OFFSET)
fprintf(fp, "%sHAS_PHYSVIRT_OFFSET", others++ ? "|" : "");
if (machdep->flags & ARM64_MTE)
fprintf(fp, "%sARM64_MTE", others++ ? "|" : "");
fprintf(fp, ")\n");
fprintf(fp, " kvbase: %lx\n", machdep->kvbase);
fprintf(fp, " identity_map_base: %lx\n", machdep->identity_map_base);
fprintf(fp, " pagesize: %d\n", machdep->pagesize);
fprintf(fp, " pageshift: %d\n", machdep->pageshift);
fprintf(fp, " pagemask: %lx\n", (ulong)machdep->pagemask);
fprintf(fp, " pageoffset: %lx\n", machdep->pageoffset);
fprintf(fp, " stacksize: %ld\n", machdep->stacksize);
fprintf(fp, " hz: %d\n", machdep->hz);
fprintf(fp, " mhz: %ld\n", machdep->mhz);
fprintf(fp, " memsize: %lld (0x%llx)\n",
(ulonglong)machdep->memsize, (ulonglong)machdep->memsize);
fprintf(fp, " bits: %d\n", machdep->bits);
fprintf(fp, " nr_irqs: %d\n", machdep->nr_irqs);
fprintf(fp, " eframe_search: arm64_eframe_search()\n");
fprintf(fp, " back_trace: arm64_back_trace_cmd() (default: %s method)\n",
kt->flags & USE_OPT_BT ? "optional" : "original");
fprintf(fp, " in_alternate_stack: arm64_in_alternate_stack()\n");
fprintf(fp, " processor_speed: arm64_processor_speed()\n");
fprintf(fp, " uvtop: arm64_uvtop()->%s()\n",
machdep->flags & VM_L3_4K ?
"arm64_vtop_3level_4k" :
machdep->flags & VM_L4_4K ?
"arm64_vtop_4level_4k" :
machdep->flags & VM_L2_16K ?
"arm64_vtop_2level_16k" :
machdep->flags & VM_L3_16K ?
"arm64_vtop_3level_16k" :
machdep->flags & VM_L4_16K ?
"arm64_vtop_4level_16k" :
machdep->flags & VM_L3_64K ?
"arm64_vtop_3level_64k" : "arm64_vtop_2level_64k");
fprintf(fp, " kvtop: arm64_kvtop()->%s()\n",
machdep->flags & VM_L3_4K ?
"arm64_vtop_3level_4k" :
machdep->flags & VM_L4_4K ?
"arm64_vtop_4level_4k" :
machdep->flags & VM_L2_16K ?
"arm64_vtop_2level_16k" :
machdep->flags & VM_L3_16K ?
"arm64_vtop_3level_16k" :
machdep->flags & VM_L4_16K ?
"arm64_vtop_4level_16k" :
machdep->flags & VM_L3_64K ?
"arm64_vtop_3level_64k" : "arm64_vtop_2level_64k");
fprintf(fp, " get_task_pgd: arm64_get_task_pgd()\n");
fprintf(fp, " dump_irq: generic_dump_irq()\n");
fprintf(fp, " get_stack_frame: arm64_get_stack_frame()\n");
fprintf(fp, " get_stackbase: generic_get_stackbase()\n");
fprintf(fp, " get_stacktop: generic_get_stacktop()\n");
fprintf(fp, " translate_pte: arm64_translate_pte()\n");
fprintf(fp, " memory_size: generic_memory_size()\n");
fprintf(fp, " vmalloc_start: arm64_vmalloc_start()\n");
fprintf(fp, " get_kvaddr_ranges: arm64_get_kvaddr_ranges()\n");
fprintf(fp, " is_task_addr: arm64_is_task_addr()\n");
fprintf(fp, " verify_symbol: arm64_verify_symbol()\n");
fprintf(fp, " dis_filter: arm64_dis_filter()\n");
fprintf(fp, " cmd_mach: arm64_cmd_mach()\n");
fprintf(fp, " get_smp_cpus: arm64_get_smp_cpus()\n");
fprintf(fp, " is_kvaddr: arm64_is_kvaddr()\n");
fprintf(fp, " is_uvaddr: arm64_is_uvaddr()\n");
fprintf(fp, " value_to_symbol: generic_machdep_value_to_symbol()\n");
fprintf(fp, " init_kernel_pgd: arm64_init_kernel_pgd\n");
fprintf(fp, " verify_paddr: generic_verify_paddr()\n");
fprintf(fp, " show_interrupts: generic_show_interrupts()\n");
fprintf(fp, " get_irq_affinity: generic_get_irq_affinity()\n");
fprintf(fp, " dumpfile_init: (not used)\n");
fprintf(fp, " process_elf_notes: process_elf64_notes()\n");
fprintf(fp, " verify_line_number: (not used)\n");
fprintf(fp, " xendump_p2m_create: (n/a)\n");
fprintf(fp, "xen_kdump_p2m_create: (n/a)\n");
fprintf(fp, " xendump_panic_task: (n/a)\n");
fprintf(fp, " get_xendump_regs: (n/a)\n");
fprintf(fp, " line_number_hooks: (not used)\n");
fprintf(fp, " last_pgd_read: %lx\n", machdep->last_pgd_read);
fprintf(fp, " last_pud_read: ");
if ((!(machdep->flags & VM_L4_4K)) && (!(machdep->flags & VM_L4_16K)))
fprintf(fp, "(not used)\n");
else
fprintf(fp, "%lx\n", machdep->last_pud_read);
fprintf(fp, " last_pmd_read: ");
if ((machdep->flags & VM_L2_64K) || (machdep->flags & VM_L2_16K))
fprintf(fp, "(not used)\n");
else
fprintf(fp, "%lx\n", machdep->last_pmd_read);
fprintf(fp, " last_ptbl_read: %lx\n", machdep->last_ptbl_read);
fprintf(fp, " clear_machdep_cache: arm64_clear_machdep_cache()\n");
fprintf(fp, " pgd: %lx\n", (ulong)machdep->pgd);
fprintf(fp, " pud: %lx\n", (ulong)machdep->pud);
fprintf(fp, " pmd: %lx\n", (ulong)machdep->pmd);
fprintf(fp, " ptbl: %lx\n", (ulong)machdep->ptbl);
fprintf(fp, " ptrs_per_pgd: %d\n", machdep->ptrs_per_pgd);
fprintf(fp, " section_size_bits: %ld\n", machdep->section_size_bits);
fprintf(fp, " max_physmem_bits: %ld\n", machdep->max_physmem_bits);
fprintf(fp, " sections_per_root: %ld\n", machdep->sections_per_root);
for (i = 0; i < MAX_MACHDEP_ARGS; i++) {
fprintf(fp, " cmdline_args[%d]: %s\n",
i, machdep->cmdline_args[i] ?
machdep->cmdline_args[i] : "(unused)");
}
fprintf(fp, " machspec: %lx\n", (ulong)ms);
fprintf(fp, " struct_page_size: %ld\n", ms->struct_page_size);
fprintf(fp, " VA_BITS: %ld\n", ms->VA_BITS);
fprintf(fp, " CONFIG_ARM64_VA_BITS: %ld\n", ms->CONFIG_ARM64_VA_BITS);
fprintf(fp, " VA_START: ");
if (ms->VA_START)
fprintf(fp, "%lx\n", ms->VA_START);
else
fprintf(fp, "(unused)\n");
fprintf(fp, " VA_BITS_ACTUAL: ");
if (ms->VA_BITS_ACTUAL)
fprintf(fp, "%ld\n", ms->VA_BITS_ACTUAL);
else
fprintf(fp, "(unused)\n");
fprintf(fp, "CONFIG_ARM64_KERNELPACMASK: ");
if (ms->CONFIG_ARM64_KERNELPACMASK)
fprintf(fp, "%lx\n", ms->CONFIG_ARM64_KERNELPACMASK);
else
fprintf(fp, "(unused)\n");
fprintf(fp, " userspace_top: %016lx\n", ms->userspace_top);
fprintf(fp, " page_offset: %016lx\n", ms->page_offset);
fprintf(fp, " vmalloc_start_addr: %016lx\n", ms->vmalloc_start_addr);
fprintf(fp, " vmalloc_end: %016lx\n", ms->vmalloc_end);
fprintf(fp, " modules_vaddr: %016lx\n", ms->modules_vaddr);
fprintf(fp, " modules_end: %016lx\n", ms->modules_end);
fprintf(fp, " vmemmap_vaddr: %016lx\n", ms->vmemmap_vaddr);
fprintf(fp, " vmemmap_end: %016lx\n", ms->vmemmap_end);
if (machdep->flags & NEW_VMEMMAP) {
fprintf(fp, " vmemmap: %016lx\n", ms->vmemmap);
fprintf(fp, " kimage_text: %016lx\n", ms->kimage_text);
fprintf(fp, " kimage_end: %016lx\n", ms->kimage_end);
fprintf(fp, " kimage_voffset: %016lx\n", ms->kimage_voffset);
}
fprintf(fp, " phys_offset: %lx\n", ms->phys_offset);
fprintf(fp, " physvirt_offset: %lx\n", ms->physvirt_offset);
fprintf(fp, "__exception_text_start: %lx\n", ms->__exception_text_start);
fprintf(fp, " __exception_text_end: %lx\n", ms->__exception_text_end);
fprintf(fp, " __irqentry_text_start: %lx\n", ms->__irqentry_text_start);
fprintf(fp, " __irqentry_text_end: %lx\n", ms->__irqentry_text_end);
fprintf(fp, " exp_entry1_start: %lx\n", ms->exp_entry1_start);
fprintf(fp, " exp_entry1_end: %lx\n", ms->exp_entry1_end);
fprintf(fp, " exp_entry2_start: %lx\n", ms->exp_entry2_start);
fprintf(fp, " exp_entry2_end: %lx\n", ms->exp_entry2_end);
fprintf(fp, " panic_task_regs: %lx\n", (ulong)ms->panic_task_regs);
fprintf(fp, " user_eframe_offset: %ld\n", ms->user_eframe_offset);
fprintf(fp, " kern_eframe_offset: %ld\n", ms->kern_eframe_offset);
fprintf(fp, " PTE_PROT_NONE: %lx\n", ms->PTE_PROT_NONE);
fprintf(fp, " PTE_FILE: ");
if (ms->PTE_FILE)
fprintf(fp, "%lx\n", ms->PTE_FILE);
else
fprintf(fp, "(unused)\n");
fprintf(fp, " __SWP_TYPE_BITS: %ld\n", ms->__SWP_TYPE_BITS);
fprintf(fp, " __SWP_TYPE_SHIFT: %ld\n", ms->__SWP_TYPE_SHIFT);
fprintf(fp, " __SWP_TYPE_MASK: %lx\n", ms->__SWP_TYPE_MASK);
fprintf(fp, " __SWP_OFFSET_BITS: ");
if (ms->__SWP_OFFSET_BITS)
fprintf(fp, "%ld\n", ms->__SWP_OFFSET_BITS);
else
fprintf(fp, "(unused)\n");
fprintf(fp, " __SWP_OFFSET_SHIFT: %ld\n", ms->__SWP_OFFSET_SHIFT);
fprintf(fp, " __SWP_OFFSET_MASK: ");
if (ms->__SWP_OFFSET_MASK)
fprintf(fp, "%lx\n", ms->__SWP_OFFSET_MASK);
else
fprintf(fp, "(unused)\n");
fprintf(fp, " machine_kexec_start: %lx\n", ms->machine_kexec_start);
fprintf(fp, " machine_kexec_end: %lx\n", ms->machine_kexec_end);
fprintf(fp, " crash_kexec_start: %lx\n", ms->crash_kexec_start);
fprintf(fp, " crash_kexec_end: %lx\n", ms->crash_kexec_end);
fprintf(fp, " crash_save_cpu_start: %lx\n", ms->crash_save_cpu_start);
fprintf(fp, " crash_save_cpu_end: %lx\n", ms->crash_save_cpu_end);
fprintf(fp, " kernel_flags: %lx\n", ms->kernel_flags);
fprintf(fp, " irq_stackbuf: %lx\n", (ulong)ms->irq_stackbuf);
if (machdep->flags & IRQ_STACKS) {
fprintf(fp, " irq_stack_size: %ld\n", ms->irq_stack_size);
for (i = 0; i < kt->cpus; i++)
fprintf(fp, " irq_stacks[%d]: %lx\n",
i, ms->irq_stacks[i]);
} else {
fprintf(fp, " irq_stack_size: (unused)\n");
fprintf(fp, " irq_stacks: (unused)\n");
}
}
static int
arm64_parse_machdep_arg_l(char *argstring, char *param, ulong *value)
{
int len;
int megabytes = FALSE;
char *p;
len = strlen(param);
if (!STRNEQ(argstring, param) || (argstring[len] != '='))
return FALSE;
if ((LASTCHAR(argstring) == 'm') ||
(LASTCHAR(argstring) == 'M')) {
LASTCHAR(argstring) = NULLCHAR;
megabytes = TRUE;
}
p = argstring + len + 1;
if (strlen(p)) {
int flags = RETURN_ON_ERROR | QUIET;
int err = 0;
if (STRNEQ(argstring, "max_physmem_bits")) {
*value = dtol(p, flags, &err);
} else if (STRNEQ(argstring, "vabits_actual")) {
*value = dtol(p, flags, &err);
} else if (megabytes) {
*value = dtol(p, flags, &err);
if (!err)
*value = MEGABYTES(*value);
} else {
*value = htol(p, flags, &err);
}
if (!err)
return TRUE;
}
return FALSE;
}
/*
* Parse machine dependent command line arguments.
*
* Force the phys_offset address via:
*
* --machdep phys_offset=<address>
*/
static void
arm64_parse_cmdline_args(void)
{
int index, i, c;
char *arglist[MAXARGS];
char buf[BUFSIZE];
char *p;
for (index = 0; index < MAX_MACHDEP_ARGS; index++) {
if (!machdep->cmdline_args[index])
break;
if (!strstr(machdep->cmdline_args[index], "=")) {
error(WARNING, "ignoring --machdep option: %x\n",
machdep->cmdline_args[index]);
continue;
}
strcpy(buf, machdep->cmdline_args[index]);
for (p = buf; *p; p++) {
if (*p == ',')
*p = ' ';
}
c = parse_line(buf, arglist);
for (i = 0; i < c; i++) {
if (arm64_parse_machdep_arg_l(arglist[i], "phys_offset",
&machdep->machspec->phys_offset)) {
error(NOTE,
"setting phys_offset to: 0x%lx\n\n",
machdep->machspec->phys_offset);
machdep->flags |= PHYS_OFFSET;
continue;
} else if (arm64_parse_machdep_arg_l(arglist[i], "kimage_voffset",
&machdep->machspec->kimage_voffset)) {
error(NOTE,
"setting kimage_voffset to: 0x%lx\n\n",
machdep->machspec->kimage_voffset);
continue;
} else if (arm64_parse_machdep_arg_l(arglist[i], "max_physmem_bits",
&machdep->max_physmem_bits)) {
error(NOTE,
"setting max_physmem_bits to: %ld\n\n",
machdep->max_physmem_bits);
continue;
} else if (arm64_parse_machdep_arg_l(arglist[i], "vabits_actual",
&machdep->machspec->VA_BITS_ACTUAL)) {
error(NOTE,
"setting vabits_actual to: %ld\n\n",
machdep->machspec->VA_BITS_ACTUAL);
continue;
}
error(WARNING, "ignoring --machdep option: %s\n",
arglist[i]);
}
}
}
#define MIN_KIMG_ALIGN (0x00200000) /* kimage load address must be aligned 2M */
/*
* Traverse the entire dumpfile to find/verify kimage_voffset.
*/
static int
arm64_search_for_kimage_voffset(ulong phys_base)
{
ulong kimage_load_addr;
ulong phys_end;
struct machine_specific *ms = machdep->machspec;
if (!arm_kdump_phys_end(&phys_end))
return FALSE;
for (kimage_load_addr = phys_base;
kimage_load_addr <= phys_end; kimage_load_addr += MIN_KIMG_ALIGN) {
ms->kimage_voffset = ms->vmalloc_start_addr - kimage_load_addr;
if ((kt->flags2 & KASLR) && (kt->flags & RELOC_SET))
ms->kimage_voffset += (kt->relocate * - 1);
if (verify_kimage_voffset()) {
if (CRASHDEBUG(1))
error(INFO,
"dumpfile searched for kimage_voffset: %lx\n\n",
ms->kimage_voffset);
break;
}
}
if (kimage_load_addr > phys_end)
return FALSE;
return TRUE;
}
static int
verify_kimage_voffset(void)
{
ulong kimage_voffset;
if (!readmem(symbol_value("kimage_voffset"), KVADDR, &kimage_voffset,
sizeof(kimage_voffset), "verify kimage_voffset", QUIET|RETURN_ON_ERROR))
return FALSE;
return (machdep->machspec->kimage_voffset == kimage_voffset);
}
static void
arm64_calc_kimage_voffset(void)
{
struct machine_specific *ms = machdep->machspec;
ulong phys_addr = 0;
int errflag;
if (ms->kimage_voffset) /* vmcoreinfo, ioctl, or --machdep override */
return;
if (ACTIVE()) {
char buf[BUFSIZE];
char *p1;
FILE *iomem;
ulong kimage_voffset, vaddr;
if (pc->flags & PROC_KCORE) {
kimage_voffset = symbol_value_from_proc_kallsyms("kimage_voffset");
if ((kimage_voffset != BADVAL) &&
(READMEM(pc->mfd, &vaddr, sizeof(ulong),
kimage_voffset, KCORE_USE_VADDR) > 0)) {
ms->kimage_voffset = vaddr;
return;
}
}
if ((iomem = fopen("/proc/iomem", "r")) == NULL)
return;
errflag = 1;
while (fgets(buf, BUFSIZE, iomem)) {
if(strstr(buf, ": Kernel code")) {
errflag = 0;
break;
}
if (strstr(buf, ": System RAM")) {
clean_line(buf);
if (!(p1 = strstr(buf, "-")))
continue;
*p1 = NULLCHAR;
phys_addr = htol(buf, RETURN_ON_ERROR | QUIET, NULL);
if (phys_addr == BADADDR)
continue;
}
}
fclose(iomem);
if (errflag)
return;
} else if (KDUMP_DUMPFILE()) {
errflag = 1;
if (arm_kdump_phys_base(&phys_addr)) { /* Get start address of first memory block */
ms->kimage_voffset = ms->vmalloc_start_addr - phys_addr;
if ((kt->flags2 & KASLR) && (kt->flags & RELOC_SET))
ms->kimage_voffset += (kt->relocate * -1);
if (verify_kimage_voffset() || arm64_search_for_kimage_voffset(phys_addr))
errflag = 0;
}
if (errflag) {
error(WARNING,
"kimage_voffset cannot be determined from the dumpfile.\n");
error(CONT,
"Try using the command line option: --machdep kimage_voffset=<addr>\n");
}
return;
} else {
error(WARNING,
"kimage_voffset cannot be determined from the dumpfile.\n");
error(CONT,
"Using default value of 0. If this is not correct, then try\n");
error(CONT,
"using the command line option: --machdep kimage_voffset=<addr>\n");
return;
}
ms->kimage_voffset = ms->vmalloc_start_addr - phys_addr;
if ((kt->flags2 & KASLR) && (kt->flags & RELOC_SET))
ms->kimage_voffset += (kt->relocate * -1);
}
/*
* The physvirt_offset only exits in kernel [5.4, 5.10)
*
* 1) In kernel v5.4, the patch:
* "5383cc6efed137 arm64: mm: Introduce vabits_actual"
*
* introduced the physvirt_offset.
*
* 2) In kernel v5.10, the patch:
* "7bc1a0f9e17658 arm64: mm: use single quantity
* to represent the PA to VA translation"
* removed the physvirt_offset.
*/
static void
arm64_calc_physvirt_offset(void)
{
struct machine_specific *ms = machdep->machspec;
ulong physvirt_offset;
struct syment *sp;
ulong value;
if ((sp = kernel_symbol_search("physvirt_offset")) &&
machdep->machspec->kimage_voffset) {
if (pc->flags & PROC_KCORE) {
value = symbol_value_from_proc_kallsyms("physvirt_offset");
if ((value != BADVAL) &&
(READMEM(pc->mfd, &physvirt_offset, sizeof(ulong),
value, KCORE_USE_VADDR) > 0)) {
machdep->flags |= HAS_PHYSVIRT_OFFSET;
ms->physvirt_offset = physvirt_offset;
/* Update the ms->phys_offset which is wrong */
ms->phys_offset = ms->physvirt_offset + ms->page_offset;
return;
}
}
if (READMEM(pc->mfd, &physvirt_offset, sizeof(physvirt_offset),
sp->value, sp->value -
machdep->machspec->kimage_voffset) > 0) {
machdep->flags |= HAS_PHYSVIRT_OFFSET;
ms->physvirt_offset = physvirt_offset;
return;
}
}
/* Useless if no symbol 'physvirt_offset', just keep semantics */
ms->physvirt_offset = ms->phys_offset - ms->page_offset;
}
static void
arm64_calc_phys_offset(void)
{
struct machine_specific *ms = machdep->machspec;
ulong phys_offset;
if (machdep->flags & PHYS_OFFSET) /* --machdep override */
return;
/*
* Next determine suitable value for phys_offset. User can override this
* by passing valid '--machdep phys_offset=<addr>' option.
*/
ms->phys_offset = 0;
if (ACTIVE()) {
char buf[BUFSIZE];
char *p1;
int errflag;
FILE *iomem;
physaddr_t paddr;
ulong vaddr;
struct syment *sp;
if ((machdep->flags & NEW_VMEMMAP) &&
ms->kimage_voffset && (sp = kernel_symbol_search("memstart_addr"))) {
if (pc->flags & PROC_KCORE) {
if (arm64_get_vmcoreinfo(&ms->phys_offset, "NUMBER(PHYS_OFFSET)", NUM_HEX))
return;
vaddr = symbol_value_from_proc_kallsyms("memstart_addr");
if (vaddr == BADVAL)
vaddr = sp->value;
paddr = KCORE_USE_VADDR;
} else {
vaddr = sp->value;
paddr = sp->value - machdep->machspec->kimage_voffset;
}
if (READMEM(pc->mfd, &phys_offset, sizeof(phys_offset),
vaddr, paddr) > 0) {
ms->phys_offset = phys_offset;
return;
}
}
if ((iomem = fopen("/proc/iomem", "r")) == NULL)
return;
/*
* Memory regions are sorted in ascending order. We take the
* first region which should be correct for most uses.
*/
errflag = 1;
while (fgets(buf, BUFSIZE, iomem)) {
if (strstr(buf, ": System RAM")) {
clean_line(buf);
errflag = 0;
break;
}
}
fclose(iomem);
if (errflag)
return;
if (!(p1 = strstr(buf, "-")))
return;
*p1 = NULLCHAR;
phys_offset = htol(buf, RETURN_ON_ERROR | QUIET, &errflag);
if (errflag)
return;
ms->phys_offset = phys_offset;
} else if (DISKDUMP_DUMPFILE() && diskdump_phys_base(&phys_offset)) {
ms->phys_offset = phys_offset;
} else if (KDUMP_DUMPFILE() && arm64_kdump_phys_base(&phys_offset)) {
ms->phys_offset = phys_offset;
} else {
error(WARNING,
"phys_offset cannot be determined from the dumpfile.\n");
error(CONT,
"Using default value of 0. If this is not correct, then try\n");
error(CONT,
"using the command line option: --machdep phys_offset=<addr>\n");
}
if (CRASHDEBUG(1))
fprintf(fp, "using %lx as phys_offset\n", ms->phys_offset);
}
/*
* Determine SECTION_SIZE_BITS either by reading VMCOREINFO or the kernel
* config, otherwise use the 64-bit ARM default definiton.
*/
static void
arm64_get_section_size_bits(void)
{
int ret;
char *string;
if (THIS_KERNEL_VERSION >= LINUX(5,12,0)) {
if (machdep->pagesize == 65536)
machdep->section_size_bits = _SECTION_SIZE_BITS_5_12_64K;
else
machdep->section_size_bits = _SECTION_SIZE_BITS_5_12;
} else
machdep->section_size_bits = _SECTION_SIZE_BITS;
if (arm64_get_vmcoreinfo(&machdep->section_size_bits, "NUMBER(SECTION_SIZE_BITS)", NUM_DEC)) {
/* nothing */
} else if (kt->ikconfig_flags & IKCONFIG_AVAIL) {
if ((ret = get_kernel_config("CONFIG_MEMORY_HOTPLUG", NULL)) == IKCONFIG_Y) {
if ((ret = get_kernel_config("CONFIG_HOTPLUG_SIZE_BITS", &string)) == IKCONFIG_STR)
machdep->section_size_bits = atol(string);
}
}
if (CRASHDEBUG(1))
fprintf(fp, "SECTION_SIZE_BITS: %ld\n", machdep->section_size_bits);
}
/*
* Determine PHYS_OFFSET either by reading VMCOREINFO or the kernel
* symbol, otherwise borrow the 32-bit ARM functionality.
*/
static int
arm64_kdump_phys_base(ulong *phys_offset)
{
struct syment *sp;
physaddr_t paddr;
if (arm64_get_vmcoreinfo(phys_offset, "NUMBER(PHYS_OFFSET)", NUM_HEX))
return TRUE;
if ((machdep->flags & NEW_VMEMMAP) &&
machdep->machspec->kimage_voffset &&
(sp = kernel_symbol_search("memstart_addr"))) {
paddr = sp->value - machdep->machspec->kimage_voffset;
if (READMEM(-1, phys_offset, sizeof(*phys_offset),
sp->value, paddr) > 0)
return TRUE;
}
return arm_kdump_phys_base(phys_offset);
}
static void
arm64_init_kernel_pgd(void)
{
int i;
ulong value;
if (!kernel_symbol_exists("init_mm") ||
!readmem(symbol_value("init_mm") + OFFSET(mm_struct_pgd), KVADDR,
&value, sizeof(void *), "init_mm.pgd", RETURN_ON_ERROR)) {
if (kernel_symbol_exists("swapper_pg_dir"))
value = symbol_value("swapper_pg_dir");
else {
error(WARNING, "cannot determine kernel pgd location\n");
return;
}
}
for (i = 0; i < NR_CPUS; i++)
vt->kernel_pgd[i] = value;
}
ulong arm64_PTOV(ulong paddr)
{
struct machine_specific *ms = machdep->machspec;
/*
* Either older kernel before kernel has 'physvirt_offset' or newer
* kernel which removes 'physvirt_offset' has the same formula:
* #define __phys_to_virt(x) ((unsigned long)((x) - PHYS_OFFSET) | PAGE_OFFSET)
*/
if (!(machdep->flags & HAS_PHYSVIRT_OFFSET))
return (paddr - ms->phys_offset) | PAGE_OFFSET;
else
return paddr - ms->physvirt_offset;
}
ulong
arm64_VTOP(ulong addr)
{
if (is_mte_kvaddr(addr))
addr = mte_tag_reset(addr);
if (machdep->flags & NEW_VMEMMAP) {
if (machdep->machspec->VA_START &&
(addr >= machdep->machspec->kimage_text) &&
(addr <= machdep->machspec->kimage_end)) {
return addr - machdep->machspec->kimage_voffset;
}
if (addr >= machdep->machspec->page_offset) {
if (machdep->flags & HAS_PHYSVIRT_OFFSET) {
return addr + machdep->machspec->physvirt_offset;
} else {
/*
* Either older kernel before kernel has 'physvirt_offset' or newer
* kernel which removes 'physvirt_offset' has the same formula:
* #define __lm_to_phys(addr) (((addr) & ~PAGE_OFFSET) + PHYS_OFFSET)
*/
return (addr & ~PAGE_OFFSET) + machdep->machspec->phys_offset;
}
}
else if (machdep->machspec->kimage_voffset)
return addr - machdep->machspec->kimage_voffset;
else /* no randomness */
return machdep->machspec->phys_offset
+ (addr - machdep->machspec->vmalloc_start_addr);
} else {
return machdep->machspec->phys_offset
+ (addr - machdep->machspec->page_offset);
}
}
static int
arm64_kvtop(struct task_context *tc, ulong kvaddr, physaddr_t *paddr, int verbose)
{
ulong kernel_pgd;
if (!IS_KVADDR(kvaddr))
return FALSE;
if (!vt->vmalloc_start) {
*paddr = VTOP(kvaddr);
return TRUE;
}
if (!IS_VMALLOC_ADDR(kvaddr)) {
*paddr = VTOP(kvaddr);
if (!verbose)
return TRUE;
}
kernel_pgd = vt->kernel_pgd[0];
*paddr = 0;
switch (machdep->flags & (VM_L2_64K|VM_L3_64K|VM_L3_4K|VM_L4_4K|VM_L2_16K|VM_L3_16K|VM_L4_16K))
{
case VM_L2_64K:
return arm64_vtop_2level_64k(kernel_pgd, kvaddr, paddr, verbose);
case VM_L3_64K:
return arm64_vtop_3level_64k(kernel_pgd, kvaddr, paddr, verbose);
case VM_L3_4K:
return arm64_vtop_3level_4k(kernel_pgd, kvaddr, paddr, verbose);
case VM_L4_4K:
return arm64_vtop_4level_4k(kernel_pgd, kvaddr, paddr, verbose);
case VM_L2_16K:
return arm64_vtop_2level_16k(kernel_pgd, kvaddr, paddr, verbose);
case VM_L3_16K:
return arm64_vtop_3level_16k(kernel_pgd, kvaddr, paddr, verbose);
case VM_L4_16K:
return arm64_vtop_4level_16k(kernel_pgd, kvaddr, paddr, verbose);
default:
return FALSE;
}
}
static int
arm64_uvtop(struct task_context *tc, ulong uvaddr, physaddr_t *paddr, int verbose)
{
ulong user_pgd;
readmem(tc->mm_struct + OFFSET(mm_struct_pgd), KVADDR,
&user_pgd, sizeof(long), "user pgd", FAULT_ON_ERROR);
*paddr = 0;
switch (machdep->flags & (VM_L2_64K|VM_L3_64K|VM_L3_4K|VM_L4_4K|VM_L2_16K|VM_L3_16K|VM_L4_16K))
{
case VM_L2_64K:
return arm64_vtop_2level_64k(user_pgd, uvaddr, paddr, verbose);
case VM_L3_64K:
return arm64_vtop_3level_64k(user_pgd, uvaddr, paddr, verbose);
case VM_L3_4K:
return arm64_vtop_3level_4k(user_pgd, uvaddr, paddr, verbose);
case VM_L4_4K:
return arm64_vtop_4level_4k(user_pgd, uvaddr, paddr, verbose);
case VM_L2_16K:
return arm64_vtop_2level_16k(user_pgd, uvaddr, paddr, verbose);
case VM_L3_16K:
return arm64_vtop_3level_16k(user_pgd, uvaddr, paddr, verbose);
case VM_L4_16K:
return arm64_vtop_4level_16k(user_pgd, uvaddr, paddr, verbose);
default:
return FALSE;
}
}
#define PTE_ADDR_LOW ((((1UL) << (48 - machdep->pageshift)) - 1) << machdep->pageshift)
#define PTE_ADDR_HIGH ((0xfUL) << 12)
#define PTE_ADDR_HIGH_SHIFT 36
#define PTE_TO_PHYS(pteval) (machdep->max_physmem_bits == 52 ? \
(((pteval & PTE_ADDR_LOW) | ((pteval & PTE_ADDR_HIGH) << PTE_ADDR_HIGH_SHIFT))) : (pteval & PTE_ADDR_LOW))
#define PUD_TYPE_MASK 3
#define PUD_TYPE_SECT 1
#define PMD_TYPE_MASK 3
#define PMD_TYPE_SECT 1
#define PMD_TYPE_TABLE 2
#define SECTION_PAGE_MASK_2MB ((long)(~((MEGABYTES(2))-1)))
#define SECTION_PAGE_MASK_32MB ((long)(~((MEGABYTES(32))-1)))
#define SECTION_PAGE_MASK_512MB ((long)(~((MEGABYTES(512))-1)))
#define SECTION_PAGE_MASK_1GB ((long)(~((GIGABYTES(1))-1)))
#define SECTION_PAGE_MASK_64GB ((long)(~((GIGABYTES(64))-1)))
static int
arm64_vtop_2level_64k(ulong pgd, ulong vaddr, physaddr_t *paddr, int verbose)
{
ulong *pgd_base, *pgd_ptr, pgd_val;
ulong *pte_base, *pte_ptr, pte_val;
if (verbose)
fprintf(fp, "PAGE DIRECTORY: %lx\n", pgd);
pgd_base = (ulong *)pgd;
FILL_PGD(pgd_base, KVADDR, machdep->ptrs_per_pgd * sizeof(ulong));
pgd_ptr = pgd_base + (((vaddr) >> PGDIR_SHIFT_L2_64K) & (machdep->ptrs_per_pgd - 1));
pgd_val = ULONG(machdep->pgd + PAGEOFFSET(pgd_ptr));
if (verbose)
fprintf(fp, " PGD: %lx => %lx\n", (ulong)pgd_ptr, pgd_val);
if (!pgd_val)
goto no_page;
/*
* #define __PAGETABLE_PUD_FOLDED
* #define __PAGETABLE_PMD_FOLDED
*/
if ((pgd_val & PMD_TYPE_MASK) == PMD_TYPE_SECT) {
ulong sectionbase = PTE_TO_PHYS(pgd_val & SECTION_PAGE_MASK_512MB);
if (verbose) {
fprintf(fp, " PAGE: %lx (512MB%s)\n\n", sectionbase,
IS_ZEROPAGE(sectionbase) ? ", ZERO PAGE" : "");
arm64_translate_pte(pgd_val, 0, 0);
}
*paddr = sectionbase + (vaddr & ~SECTION_PAGE_MASK_512MB);
return TRUE;
}
pte_base = (ulong *)PTOV(PTE_TO_PHYS(pgd_val));
FILL_PTBL(pte_base, KVADDR, PTRS_PER_PTE_L2_64K * sizeof(ulong));
pte_ptr = pte_base + (((vaddr) >> machdep->pageshift) & (PTRS_PER_PTE_L2_64K - 1));
pte_val = ULONG(machdep->ptbl + PAGEOFFSET(pte_ptr));
if (verbose)
fprintf(fp, " PTE: %lx => %lx\n", (ulong)pte_ptr, pte_val);
if (!pte_val)
goto no_page;
if (pte_val & PTE_VALID) {
*paddr = PTE_TO_PHYS(pte_val) + PAGEOFFSET(vaddr);
if (verbose) {
fprintf(fp, " PAGE: %lx %s\n\n", PAGEBASE(*paddr),
IS_ZEROPAGE(PAGEBASE(*paddr)) ? "(ZERO PAGE)" : "");
arm64_translate_pte(pte_val, 0, 0);
}
} else {
if (IS_UVADDR(vaddr, NULL))
*paddr = pte_val;
if (verbose) {
fprintf(fp, "\n");
arm64_translate_pte(pte_val, 0, 0);
}
goto no_page;
}
return TRUE;
no_page:
return FALSE;
}
static int
arm64_vtop_3level_64k(ulong pgd, ulong vaddr, physaddr_t *paddr, int verbose)
{
ulong *pgd_base, *pgd_ptr, pgd_val;
ulong *pmd_base, *pmd_ptr, pmd_val;
ulong *pte_base, *pte_ptr, pte_val;
if (verbose)
fprintf(fp, "PAGE DIRECTORY: %lx\n", pgd);
pgd_base = (ulong *)pgd;
FILL_PGD(pgd_base, KVADDR, machdep->ptrs_per_pgd * sizeof(ulong));
/*
* Use machdep->ptrs_per_pgd to mask vaddr instead of using macro, because
* 48-bits and 52-bits have different size of ptrs_per_pgd.
*/
pgd_ptr = pgd_base + (((vaddr) >> PGDIR_SHIFT_L3_64K) & (machdep->ptrs_per_pgd - 1));
pgd_val = ULONG(machdep->pgd + PGDIR_OFFSET_L3_64K(pgd_ptr));
if (verbose)
fprintf(fp, " PGD: %lx => %lx\n", (ulong)pgd_ptr, pgd_val);
if (!pgd_val)
goto no_page;
/*
* #define __PAGETABLE_PUD_FOLDED
*/
pmd_base = (ulong *)PTOV(PTE_TO_PHYS(pgd_val));
FILL_PMD(pmd_base, KVADDR, PTRS_PER_PMD_L3_64K * sizeof(ulong));
pmd_ptr = pmd_base + (((vaddr) >> PMD_SHIFT_L3_64K) & (PTRS_PER_PMD_L3_64K - 1));
pmd_val = ULONG(machdep->pmd + PAGEOFFSET(pmd_ptr));
if (verbose)
fprintf(fp, " PMD: %lx => %lx\n", (ulong)pmd_ptr, pmd_val);
if (!pmd_val)
goto no_page;
if ((pmd_val & PMD_TYPE_MASK) == PMD_TYPE_SECT) {
ulong sectionbase = PTE_TO_PHYS(pmd_val & SECTION_PAGE_MASK_512MB);
if (verbose) {
fprintf(fp, " PAGE: %lx (512MB%s)\n\n", sectionbase,
IS_ZEROPAGE(sectionbase) ? ", ZERO PAGE" : "");
arm64_translate_pte(pmd_val, 0, 0);
}
*paddr = sectionbase + (vaddr & ~SECTION_PAGE_MASK_512MB);
return TRUE;
}
pte_base = (ulong *)PTOV(PTE_TO_PHYS(pmd_val));
FILL_PTBL(pte_base, KVADDR, PTRS_PER_PTE_L3_64K * sizeof(ulong));
pte_ptr = pte_base + (((vaddr) >> machdep->pageshift) & (PTRS_PER_PTE_L3_64K - 1));
pte_val = ULONG(machdep->ptbl + PAGEOFFSET(pte_ptr));
if (verbose)
fprintf(fp, " PTE: %lx => %lx\n", (ulong)pte_ptr, pte_val);
if (!pte_val)
goto no_page;
if (pte_val & PTE_VALID) {
*paddr = PTE_TO_PHYS(pte_val) + PAGEOFFSET(vaddr);
if (verbose) {
fprintf(fp, " PAGE: %lx %s\n\n", PAGEBASE(*paddr),
IS_ZEROPAGE(PAGEBASE(*paddr)) ? "(ZERO PAGE)" : "");
arm64_translate_pte(pte_val, 0, 0);
}
} else {
if (IS_UVADDR(vaddr, NULL))
*paddr = pte_val;
if (verbose) {
fprintf(fp, "\n");
arm64_translate_pte(pte_val, 0, 0);
}
goto no_page;
}
return TRUE;
no_page:
return FALSE;
}
static int
arm64_vtop_2level_16k(ulong pgd, ulong vaddr, physaddr_t *paddr, int verbose)
{
ulong *pgd_base, *pgd_ptr, pgd_val;
ulong *pte_base, *pte_ptr, pte_val;
if (verbose)
fprintf(fp, "PAGE DIRECTORY: %lx\n", pgd);
pgd_base = (ulong *)pgd;
FILL_PGD(pgd_base, KVADDR, machdep->ptrs_per_pgd * sizeof(ulong));
pgd_ptr = pgd_base + (((vaddr) >> PGDIR_SHIFT_L2_16K) & (machdep->ptrs_per_pgd - 1));
pgd_val = ULONG(machdep->pgd + PGDIR_OFFSET_L2_16K(pgd_ptr));
if (verbose)
fprintf(fp, " PGD: %lx => %lx\n", (ulong)pgd_ptr, pgd_val);
if (!pgd_val)
goto no_page;
/*
* #define __PAGETABLE_PUD_FOLDED
* #define __PAGETABLE_PMD_FOLDED
*/
if ((pgd_val & PMD_TYPE_MASK) == PMD_TYPE_SECT) {
ulong sectionbase = PTE_TO_PHYS(pgd_val & SECTION_PAGE_MASK_32MB);
if (verbose) {
fprintf(fp, " PAGE: %lx (32MB%s)\n\n", sectionbase,
IS_ZEROPAGE(sectionbase) ? ", ZERO PAGE" : "");
arm64_translate_pte(pgd_val, 0, 0);
}
*paddr = sectionbase + (vaddr & ~SECTION_PAGE_MASK_32MB);
return TRUE;
}
pte_base = (ulong *)PTOV(PTE_TO_PHYS(pgd_val));
FILL_PTBL(pte_base, KVADDR, PTRS_PER_PTE_L2_16K * sizeof(ulong));
pte_ptr = pte_base + (((vaddr) >> machdep->pageshift) & (PTRS_PER_PTE_L2_16K - 1));
pte_val = ULONG(machdep->ptbl + PAGEOFFSET(pte_ptr));
if (verbose)
fprintf(fp, " PTE: %lx => %lx\n", (ulong)pte_ptr, pte_val);
if (!pte_val)
goto no_page;
if (pte_val & PTE_VALID) {
*paddr = PTE_TO_PHYS(pte_val) + PAGEOFFSET(vaddr);
if (verbose) {
fprintf(fp, " PAGE: %lx %s\n\n", PAGEBASE(*paddr),
IS_ZEROPAGE(PAGEBASE(*paddr)) ? "(ZERO PAGE)" : "");
arm64_translate_pte(pte_val, 0, 0);
}
} else {
if (IS_UVADDR(vaddr, NULL))
*paddr = pte_val;
if (verbose) {
fprintf(fp, "\n");
arm64_translate_pte(pte_val, 0, 0);
}
goto no_page;
}
return TRUE;
no_page:
return FALSE;
}
static int
arm64_vtop_3level_16k(ulong pgd, ulong vaddr, physaddr_t *paddr, int verbose)
{
ulong *pgd_base, *pgd_ptr, pgd_val;
ulong *pmd_base, *pmd_ptr, pmd_val;
ulong *pte_base, *pte_ptr, pte_val;
if (verbose)
fprintf(fp, "PAGE DIRECTORY: %lx\n", pgd);
pgd_base = (ulong *)pgd;
FILL_PGD(pgd_base, KVADDR, machdep->ptrs_per_pgd * sizeof(ulong));
pgd_ptr = pgd_base + (((vaddr) >> PGDIR_SHIFT_L3_16K) & (machdep->ptrs_per_pgd - 1));
pgd_val = ULONG(machdep->pgd + PGDIR_OFFSET_L3_16K(pgd_ptr));
if (verbose)
fprintf(fp, " PGD: %lx => %lx\n", (ulong)pgd_ptr, pgd_val);
if (!pgd_val)
goto no_page;
/*
* #define __PAGETABLE_PUD_FOLDED
*/
pmd_base = (ulong *)PTOV(PTE_TO_PHYS(pgd_val));
FILL_PMD(pmd_base, KVADDR, PTRS_PER_PMD_L3_16K * sizeof(ulong));
pmd_ptr = pmd_base + (((vaddr) >> PMD_SHIFT_L3_16K) & (PTRS_PER_PMD_L3_16K - 1));
pmd_val = ULONG(machdep->pmd + PAGEOFFSET(pmd_ptr));
if (verbose)
fprintf(fp, " PMD: %lx => %lx\n", (ulong)pmd_ptr, pmd_val);
if (!pmd_val)
goto no_page;
if ((pmd_val & PMD_TYPE_MASK) == PMD_TYPE_SECT) {
ulong sectionbase = PTE_TO_PHYS(pmd_val & SECTION_PAGE_MASK_32MB);
if (verbose) {
fprintf(fp, " PAGE: %lx (32MB%s)\n\n", sectionbase,
IS_ZEROPAGE(sectionbase) ? ", ZERO PAGE" : "");
arm64_translate_pte(pmd_val, 0, 0);
}
*paddr = sectionbase + (vaddr & ~SECTION_PAGE_MASK_32MB);
return TRUE;
}
pte_base = (ulong *)PTOV(PTE_TO_PHYS(pmd_val));
FILL_PTBL(pte_base, KVADDR, PTRS_PER_PTE_L3_16K * sizeof(ulong));
pte_ptr = pte_base + (((vaddr) >> machdep->pageshift) & (PTRS_PER_PTE_L3_16K - 1));
pte_val = ULONG(machdep->ptbl + PAGEOFFSET(pte_ptr));
if (verbose)
fprintf(fp, " PTE: %lx => %lx\n", (ulong)pte_ptr, pte_val);
if (!pte_val)
goto no_page;
if (pte_val & PTE_VALID) {
*paddr = PTE_TO_PHYS(pte_val) + PAGEOFFSET(vaddr);
if (verbose) {
fprintf(fp, " PAGE: %lx %s\n\n", PAGEBASE(*paddr),
IS_ZEROPAGE(PAGEBASE(*paddr)) ? "(ZERO PAGE)" : "");
arm64_translate_pte(pte_val, 0, 0);
}
} else {
if (IS_UVADDR(vaddr, NULL))
*paddr = pte_val;
if (verbose) {
fprintf(fp, "\n");
arm64_translate_pte(pte_val, 0, 0);
}
goto no_page;
}
return TRUE;
no_page:
return FALSE;
}
static int
arm64_vtop_4level_16k(ulong pgd, ulong vaddr, physaddr_t *paddr, int verbose)
{
ulong *pgd_base, *pgd_ptr, pgd_val;
ulong *pud_base, *pud_ptr, pud_val;
ulong *pmd_base, *pmd_ptr, pmd_val;
ulong *pte_base, *pte_ptr, pte_val;
if (verbose)
fprintf(fp, "PAGE DIRECTORY: %lx\n", pgd);
pgd_base = (ulong *)pgd;
FILL_PGD(pgd_base, KVADDR, PTRS_PER_PGD_L4_16K * sizeof(ulong));
pgd_ptr = pgd_base + (((vaddr) >> PGDIR_SHIFT_L4_16K) & (PTRS_PER_PGD_L4_16K - 1));
pgd_val = ULONG(machdep->pgd + PGDIR_OFFSET_L4_16K(pgd_ptr));
if (verbose)
fprintf(fp, " PGD: %lx => %lx\n", (ulong)pgd_ptr, pgd_val);
if (!pgd_val)
goto no_page;
pud_base = (ulong *)PTOV(PTE_TO_PHYS(pgd_val));
FILL_PUD(pud_base, KVADDR, PTRS_PER_PUD_L4_16K * sizeof(ulong));
pud_ptr = pud_base + (((vaddr) >> PUD_SHIFT_L4_16K) & (PTRS_PER_PUD_L4_16K - 1));
pud_val = ULONG(machdep->pud + PAGEOFFSET(pud_ptr));
if (verbose)
fprintf(fp, " PUD: %lx => %lx\n", (ulong)pud_ptr, pud_val);
if (!pud_val)
goto no_page;
/*
* TODO:
* PUD Section mapping is only supported for LPA2 is enabled.
* LPA2 depends on CONFIG_ARM64_16K_PAGES and CONFIG_ARM64_PA_BITS_52.
*/
pmd_base = (ulong *)PTOV(PTE_TO_PHYS(pud_val));
FILL_PMD(pmd_base, KVADDR, PTRS_PER_PMD_L4_16K * sizeof(ulong));
pmd_ptr = pmd_base + (((vaddr) >> PMD_SHIFT_L4_16K) & (PTRS_PER_PMD_L4_16K - 1));
pmd_val = ULONG(machdep->pmd + PAGEOFFSET(pmd_ptr));
if (verbose)
fprintf(fp, " PMD: %lx => %lx\n", (ulong)pmd_ptr, pmd_val);
if (!pmd_val)
goto no_page;
if ((pmd_val & PMD_TYPE_MASK) == PMD_TYPE_SECT) {
ulong sectionbase = PTE_TO_PHYS(pmd_val) & SECTION_PAGE_MASK_32MB;
if (verbose) {
fprintf(fp, " PAGE: %lx (32MB%s)\n\n", sectionbase,
IS_ZEROPAGE(sectionbase) ? ", ZERO PAGE" : "");
arm64_translate_pte(pmd_val, 0, 0);
}
*paddr = sectionbase + (vaddr & ~SECTION_PAGE_MASK_32MB);
return TRUE;
}
pte_base = (ulong *)PTOV(PTE_TO_PHYS(pmd_val));
FILL_PTBL(pte_base, KVADDR, PTRS_PER_PTE_L4_16K * sizeof(ulong));
pte_ptr = pte_base + (((vaddr) >> machdep->pageshift) & (PTRS_PER_PTE_L4_16K - 1));
pte_val = ULONG(machdep->ptbl + PAGEOFFSET(pte_ptr));
if (verbose)
fprintf(fp, " PTE: %lx => %lx\n", (ulong)pte_ptr, pte_val);
if (!pte_val)
goto no_page;
if (pte_val & PTE_VALID) {
*paddr = PTE_TO_PHYS(pte_val) + PAGEOFFSET(vaddr);
if (verbose) {
fprintf(fp, " PAGE: %lx %s\n\n", PAGEBASE(*paddr),
IS_ZEROPAGE(PAGEBASE(*paddr)) ? "(ZERO PAGE)" : "");
arm64_translate_pte(pte_val, 0, 0);
}
} else {
if (IS_UVADDR(vaddr, NULL))
*paddr = pte_val;
if (verbose) {
fprintf(fp, "\n");
arm64_translate_pte(pte_val, 0, 0);
}
goto no_page;
}
return TRUE;
no_page:
return FALSE;
}
static int
arm64_vtop_3level_4k(ulong pgd, ulong vaddr, physaddr_t *paddr, int verbose)
{
ulong *pgd_base, *pgd_ptr, pgd_val;
ulong *pmd_base, *pmd_ptr, pmd_val;
ulong *pte_base, *pte_ptr, pte_val;
if (verbose)
fprintf(fp, "PAGE DIRECTORY: %lx\n", pgd);
pgd_base = (ulong *)pgd;
FILL_PGD(pgd_base, KVADDR, PTRS_PER_PGD_L3_4K * sizeof(ulong));
pgd_ptr = pgd_base + (((vaddr) >> PGDIR_SHIFT_L3_4K) & (PTRS_PER_PGD_L3_4K - 1));
pgd_val = ULONG(machdep->pgd + PAGEOFFSET(pgd_ptr));
if (verbose)
fprintf(fp, " PGD: %lx => %lx\n", (ulong)pgd_ptr, pgd_val);
if (!pgd_val)
goto no_page;
if ((pgd_val & PUD_TYPE_MASK) == PUD_TYPE_SECT) {
ulong sectionbase = PTE_TO_PHYS(pgd_val & SECTION_PAGE_MASK_1GB);
if (verbose) {
fprintf(fp, " PAGE: %lx (1GB)\n\n", sectionbase);
arm64_translate_pte(pgd_val, 0, 0);
}
*paddr = sectionbase + (vaddr & ~SECTION_PAGE_MASK_1GB);
return TRUE;
}
/*
* #define __PAGETABLE_PUD_FOLDED
*/
pmd_base = (ulong *)PTOV(PTE_TO_PHYS(pgd_val));
FILL_PMD(pmd_base, KVADDR, PTRS_PER_PMD_L3_4K * sizeof(ulong));
pmd_ptr = pmd_base + (((vaddr) >> PMD_SHIFT_L3_4K) & (PTRS_PER_PMD_L3_4K - 1));
pmd_val = ULONG(machdep->pmd + PAGEOFFSET(pmd_ptr));
if (verbose)
fprintf(fp, " PMD: %lx => %lx\n", (ulong)pmd_ptr, pmd_val);
if (!pmd_val)
goto no_page;
if ((pmd_val & PMD_TYPE_MASK) == PMD_TYPE_SECT) {
ulong sectionbase = PTE_TO_PHYS(pmd_val & SECTION_PAGE_MASK_2MB);
if (verbose) {
fprintf(fp, " PAGE: %lx (2MB%s)\n\n", sectionbase,
IS_ZEROPAGE(sectionbase) ? ", ZERO PAGE" : "");
arm64_translate_pte(pmd_val, 0, 0);
}
*paddr = sectionbase + (vaddr & ~SECTION_PAGE_MASK_2MB);
return TRUE;
}
pte_base = (ulong *)PTOV(PTE_TO_PHYS(pmd_val));
FILL_PTBL(pte_base, KVADDR, PTRS_PER_PTE_L3_4K * sizeof(ulong));
pte_ptr = pte_base + (((vaddr) >> machdep->pageshift) & (PTRS_PER_PTE_L3_4K - 1));
pte_val = ULONG(machdep->ptbl + PAGEOFFSET(pte_ptr));
if (verbose)
fprintf(fp, " PTE: %lx => %lx\n", (ulong)pte_ptr, pte_val);
if (!pte_val)
goto no_page;
if (pte_val & PTE_VALID) {
*paddr = PTE_TO_PHYS(pte_val) + PAGEOFFSET(vaddr);
if (verbose) {
fprintf(fp, " PAGE: %lx %s\n\n", PAGEBASE(*paddr),
IS_ZEROPAGE(PAGEBASE(*paddr)) ? "(ZERO PAGE)" : "");
arm64_translate_pte(pte_val, 0, 0);
}
} else {
if (IS_UVADDR(vaddr, NULL))
*paddr = pte_val;
if (verbose) {
fprintf(fp, "\n");
arm64_translate_pte(pte_val, 0, 0);
}
goto no_page;
}
return TRUE;
no_page:
return FALSE;
}
static int
arm64_vtop_4level_4k(ulong pgd, ulong vaddr, physaddr_t *paddr, int verbose)
{
ulong *pgd_base, *pgd_ptr, pgd_val;
ulong *pud_base, *pud_ptr, pud_val;
ulong *pmd_base, *pmd_ptr, pmd_val;
ulong *pte_base, *pte_ptr, pte_val;
if (verbose)
fprintf(fp, "PAGE DIRECTORY: %lx\n", pgd);
pgd_base = (ulong *)pgd;
FILL_PGD(pgd_base, KVADDR, PTRS_PER_PGD_L4_4K * sizeof(ulong));
pgd_ptr = pgd_base + (((vaddr) >> PGDIR_SHIFT_L4_4K) & (PTRS_PER_PGD_L4_4K - 1));
pgd_val = ULONG(machdep->pgd + PGDIR_OFFSET_48VA(pgd_ptr));
if (verbose)
fprintf(fp, " PGD: %lx => %lx\n", (ulong)pgd_ptr, pgd_val);
if (!pgd_val)
goto no_page;
pud_base = (ulong *)PTOV(PTE_TO_PHYS(pgd_val));
FILL_PUD(pud_base, KVADDR, PTRS_PER_PUD_L4_4K * sizeof(ulong));
pud_ptr = pud_base + (((vaddr) >> PUD_SHIFT_L4_4K) & (PTRS_PER_PUD_L4_4K - 1));
pud_val = ULONG(machdep->pud + PAGEOFFSET(pud_ptr));
if (verbose)
fprintf(fp, " PUD: %lx => %lx\n", (ulong)pud_ptr, pud_val);
if (!pud_val)
goto no_page;
if ((pud_val & PUD_TYPE_MASK) == PUD_TYPE_SECT) {
ulong sectionbase = PTE_TO_PHYS(pud_val & SECTION_PAGE_MASK_1GB);
if (verbose) {
fprintf(fp, " PAGE: %lx (1GB)\n\n", sectionbase);
arm64_translate_pte(pud_val, 0, 0);
}
*paddr = sectionbase + (vaddr & ~SECTION_PAGE_MASK_1GB);
return TRUE;
}
pmd_base = (ulong *)PTOV(PTE_TO_PHYS(pud_val));
FILL_PMD(pmd_base, KVADDR, PTRS_PER_PMD_L4_4K * sizeof(ulong));
pmd_ptr = pmd_base + (((vaddr) >> PMD_SHIFT_L4_4K) & (PTRS_PER_PMD_L4_4K - 1));
pmd_val = ULONG(machdep->pmd + PAGEOFFSET(pmd_ptr));
if (verbose)
fprintf(fp, " PMD: %lx => %lx\n", (ulong)pmd_ptr, pmd_val);
if (!pmd_val)
goto no_page;
if ((pmd_val & PMD_TYPE_MASK) == PMD_TYPE_SECT) {
ulong sectionbase = PTE_TO_PHYS(pmd_val & SECTION_PAGE_MASK_2MB);
if (verbose) {
fprintf(fp, " PAGE: %lx (2MB%s)\n\n", sectionbase,
IS_ZEROPAGE(sectionbase) ? ", ZERO PAGE" : "");
arm64_translate_pte(pmd_val, 0, 0);
}
*paddr = sectionbase + (vaddr & ~SECTION_PAGE_MASK_2MB);
return TRUE;
}
pte_base = (ulong *)PTOV(PTE_TO_PHYS(pmd_val));
FILL_PTBL(pte_base, KVADDR, PTRS_PER_PTE_L4_4K * sizeof(ulong));
pte_ptr = pte_base + (((vaddr) >> machdep->pageshift) & (PTRS_PER_PTE_L4_4K - 1));
pte_val = ULONG(machdep->ptbl + PAGEOFFSET(pte_ptr));
if (verbose)
fprintf(fp, " PTE: %lx => %lx\n", (ulong)pte_ptr, pte_val);
if (!pte_val)
goto no_page;
if (pte_val & PTE_VALID) {
*paddr = PTE_TO_PHYS(pte_val) + PAGEOFFSET(vaddr);
if (verbose) {
fprintf(fp, " PAGE: %lx %s\n\n", PAGEBASE(*paddr),
IS_ZEROPAGE(PAGEBASE(*paddr)) ? "(ZERO PAGE)" : "");
arm64_translate_pte(pte_val, 0, 0);
}
} else {
if (IS_UVADDR(vaddr, NULL))
*paddr = pte_val;
if (verbose) {
fprintf(fp, "\n");
arm64_translate_pte(pte_val, 0, 0);
}
goto no_page;
}
return TRUE;
no_page:
return FALSE;
}
static ulong
arm64_get_task_pgd(ulong task)
{
struct task_context *tc;
ulong pgd;
if ((tc = task_to_context(task)) &&
readmem(tc->mm_struct + OFFSET(mm_struct_pgd), KVADDR,
&pgd, sizeof(long), "user pgd", RETURN_ON_ERROR))
return pgd;
else
return NO_TASK;
}
static ulong
arm64_processor_speed(void)
{
return 0;
};
/*
* Gather IRQ stack values.
*/
static void
arm64_irq_stack_init(void)
{
int i;
struct syment *sp;
struct gnu_request request, *req;
struct machine_specific *ms = machdep->machspec;
ulong p, sz;
req = &request;
if (symbol_exists("irq_stack") &&
(sp = per_cpu_symbol_search("irq_stack")) &&
get_symbol_type("irq_stack", NULL, req)) {
/* before v4.14 or CONFIG_VMAP_STACK disabled */
if (CRASHDEBUG(1)) {
fprintf(fp, "irq_stack: \n");
fprintf(fp, " type: %s\n",
(req->typecode == TYPE_CODE_ARRAY) ?
"TYPE_CODE_ARRAY" : "other");
fprintf(fp, " target_typecode: %s\n",
req->target_typecode == TYPE_CODE_INT ?
"TYPE_CODE_INT" : "other");
fprintf(fp, " target_length: %ld\n",
req->target_length);
fprintf(fp, " length: %ld\n", req->length);
}
if (!(ms->irq_stacks = (ulong *)malloc((size_t)(kt->cpus * sizeof(ulong)))))
error(FATAL, "cannot malloc irq_stack addresses\n");
ms->irq_stack_size = req->length;
machdep->flags |= IRQ_STACKS;
for (i = 0; i < kt->cpus; i++)
ms->irq_stacks[i] = kt->__per_cpu_offset[i] + sp->value;
} else if (symbol_exists("irq_stack_ptr") &&
(sp = per_cpu_symbol_search("irq_stack_ptr")) &&
get_symbol_type("irq_stack_ptr", NULL, req)) {
/* v4.14 and later with CONFIG_VMAP_STACK enabled */
if (CRASHDEBUG(1)) {
fprintf(fp, "irq_stack_ptr: \n");
fprintf(fp, " type: %x, %s\n",
(int)req->typecode,
(req->typecode == TYPE_CODE_PTR) ?
"TYPE_CODE_PTR" : "other");
fprintf(fp, " target_typecode: %x, %s\n",
(int)req->target_typecode,
req->target_typecode == TYPE_CODE_INT ?
"TYPE_CODE_INT" : "other");
fprintf(fp, " target_length: %ld\n",
req->target_length);
fprintf(fp, " length: %ld\n", req->length);
}
if (!(ms->irq_stacks = (ulong *)malloc((size_t)(kt->cpus * sizeof(ulong)))))
error(FATAL, "cannot malloc irq_stack addresses\n");
/*
* Determining the IRQ_STACK_SIZE is tricky, but for now
* 4.14 kernel has:
*
* #define IRQ_STACK_SIZE THREAD_SIZE
*
* and finding a solid usage of THREAD_SIZE is hard, but:
*
* union thread_union {
* ...
* unsigned long stack[THREAD_SIZE/sizeof(long)];
* };
*/
if (MEMBER_EXISTS("thread_union", "stack")) {
if ((sz = MEMBER_SIZE("thread_union", "stack")) > 0)
ms->irq_stack_size = sz;
} else {
ulong res = arm64_set_irq_stack_size();
ms->irq_stack_size = (res > 0) ? res : ARM64_IRQ_STACK_SIZE;
}
machdep->flags |= IRQ_STACKS;
for (i = 0; i < kt->cpus; i++) {
p = kt->__per_cpu_offset[i] + sp->value;
readmem(p, KVADDR, &(ms->irq_stacks[i]), sizeof(ulong),
"IRQ stack pointer", RETURN_ON_ERROR);
}
}
}
/*
* Gather Overflow stack values.
*
* Overflow stack supported since 4.14, in commit 872d8327c
*/
static void
arm64_overflow_stack_init(void)
{
int i;
struct syment *sp;
struct gnu_request request, *req;
struct machine_specific *ms = machdep->machspec;
req = &request;
if (symbol_exists("overflow_stack") &&
(sp = per_cpu_symbol_search("overflow_stack")) &&
get_symbol_type("overflow_stack", NULL, req)) {
if (CRASHDEBUG(1)) {
fprintf(fp, "overflow_stack: \n");
fprintf(fp, " type: %x, %s\n",
(int)req->typecode,
(req->typecode == TYPE_CODE_ARRAY) ?
"TYPE_CODE_ARRAY" : "other");
fprintf(fp, " target_typecode: %x, %s\n",
(int)req->target_typecode,
req->target_typecode == TYPE_CODE_INT ?
"TYPE_CODE_INT" : "other");
fprintf(fp, " target_length: %ld\n",
req->target_length);
fprintf(fp, " length: %ld\n", req->length);
}
if (!(ms->overflow_stacks = (ulong *)malloc((size_t)(kt->cpus * sizeof(ulong)))))
error(FATAL, "cannot malloc overflow_stack addresses\n");
ms->overflow_stack_size = ARM64_OVERFLOW_STACK_SIZE;
machdep->flags |= OVERFLOW_STACKS;
for (i = 0; i < kt->cpus; i++)
ms->overflow_stacks[i] = kt->__per_cpu_offset[i] + sp->value;
}
}
/*
* Gather and verify all of the backtrace requirements.
*/
static void
arm64_stackframe_init(void)
{
long task_struct_thread;
long thread_struct_cpu_context;
long context_sp, context_pc, context_fp;
struct syment *sp1, *sp1n, *sp2, *sp2n, *sp3, *sp3n;
STRUCT_SIZE_INIT(note_buf, "note_buf_t");
STRUCT_SIZE_INIT(elf_prstatus, "elf_prstatus");
MEMBER_OFFSET_INIT(elf_prstatus_pr_pid, "elf_prstatus", "pr_pid");
MEMBER_OFFSET_INIT(elf_prstatus_pr_reg, "elf_prstatus", "pr_reg");
if (MEMBER_EXISTS("pt_regs", "stackframe")) {
machdep->machspec->user_eframe_offset = SIZE(pt_regs);
machdep->machspec->kern_eframe_offset = SIZE(pt_regs) - 16;
} else {
machdep->machspec->user_eframe_offset = SIZE(pt_regs) + 16;
machdep->machspec->kern_eframe_offset = SIZE(pt_regs);
}
if ((sp1 = kernel_symbol_search("__exception_text_start")) &&
(sp2 = kernel_symbol_search("__exception_text_end"))) {
machdep->machspec->__exception_text_start = sp1->value;
machdep->machspec->__exception_text_end = sp2->value;
}
if ((sp1 = kernel_symbol_search("__irqentry_text_start")) &&
(sp2 = kernel_symbol_search("__irqentry_text_end"))) {
machdep->machspec->__irqentry_text_start = sp1->value;
machdep->machspec->__irqentry_text_end = sp2->value;
}
if ((sp1 = kernel_symbol_search("vectors")) &&
(sp1n = kernel_symbol_search("cpu_switch_to")) &&
(sp2 = kernel_symbol_search("ret_fast_syscall")) &&
(sp2n = kernel_symbol_search("sys_rt_sigreturn_wrapper"))) {
machdep->machspec->exp_entry1_start = sp1->value;
machdep->machspec->exp_entry1_end = sp1n->value;
machdep->machspec->exp_entry2_start = sp2->value;
machdep->machspec->exp_entry2_end = sp2n->value;
}
if ((sp1 = kernel_symbol_search("crash_kexec")) &&
(sp1n = next_symbol(NULL, sp1)) &&
(sp2 = kernel_symbol_search("crash_save_cpu")) &&
(sp2n = next_symbol(NULL, sp2)) &&
(sp3 = kernel_symbol_search("machine_kexec")) &&
(sp3n = next_symbol(NULL, sp3))) {
machdep->machspec->crash_kexec_start = sp1->value;
machdep->machspec->crash_kexec_end = sp1n->value;
machdep->machspec->crash_save_cpu_start = sp2->value;
machdep->machspec->crash_save_cpu_end = sp2n->value;
machdep->machspec->machine_kexec_start = sp3->value;
machdep->machspec->machine_kexec_end = sp3n->value;
machdep->flags |= KDUMP_ENABLED;
}
task_struct_thread = MEMBER_OFFSET("task_struct", "thread");
thread_struct_cpu_context = MEMBER_OFFSET("thread_struct", "cpu_context");
if ((task_struct_thread == INVALID_OFFSET) ||
(thread_struct_cpu_context == INVALID_OFFSET)) {
error(INFO,
"cannot determine task_struct.thread.context offset\n");
return;
}
/*
* Pay for the convenience of using a hardcopy of a kernel structure.
*/
if (offsetof(struct arm64_stackframe, sp) !=
MEMBER_OFFSET("stackframe", "sp")) {
if (CRASHDEBUG(1))
error(INFO, "builtin stackframe.sp offset differs from kernel version\n");
}
if (offsetof(struct arm64_stackframe, fp) !=
MEMBER_OFFSET("stackframe", "fp")) {
if (CRASHDEBUG(1))
error(INFO, "builtin stackframe.fp offset differs from kernel version\n");
}
if (offsetof(struct arm64_stackframe, pc) !=
MEMBER_OFFSET("stackframe", "pc")) {
if (CRASHDEBUG(1))
error(INFO, "builtin stackframe.pc offset differs from kernel version\n");
}
if (!MEMBER_EXISTS("stackframe", "sp"))
machdep->flags |= UNW_4_14;
context_sp = MEMBER_OFFSET("cpu_context", "sp");
context_fp = MEMBER_OFFSET("cpu_context", "fp");
context_pc = MEMBER_OFFSET("cpu_context", "pc");
if (context_sp == INVALID_OFFSET) {
error(INFO, "cannot determine cpu_context.sp offset\n");
return;
}
if (context_fp == INVALID_OFFSET) {
error(INFO, "cannot determine cpu_context.fp offset\n");
return;
}
if (context_pc == INVALID_OFFSET) {
error(INFO, "cannot determine cpu_context.pc offset\n");
return;
}
ASSIGN_OFFSET(task_struct_thread_context_sp) =
task_struct_thread + thread_struct_cpu_context + context_sp;
ASSIGN_OFFSET(task_struct_thread_context_fp) =
task_struct_thread + thread_struct_cpu_context + context_fp;
ASSIGN_OFFSET(task_struct_thread_context_pc) =
task_struct_thread + thread_struct_cpu_context + context_pc;
}
#define KERNEL_MODE (1)
#define USER_MODE (2)
#define USER_EFRAME_OFFSET (machdep->machspec->user_eframe_offset)
#define KERN_EFRAME_OFFSET (machdep->machspec->kern_eframe_offset)
/*
* PSR bits
*/
#define PSR_MODE_EL0t 0x00000000
#define PSR_MODE_EL1t 0x00000004
#define PSR_MODE_EL1h 0x00000005
#define PSR_MODE_EL2t 0x00000008
#define PSR_MODE_EL2h 0x00000009
#define PSR_MODE_EL3t 0x0000000c
#define PSR_MODE_EL3h 0x0000000d
#define PSR_MODE_MASK 0x0000000f
/* Architecturally defined mapping between AArch32 and AArch64 registers */
#define compat_usr(x) regs[(x)]
#define compat_fp regs[11]
#define compat_sp regs[13]
#define compat_lr regs[14]
#define user_mode(ptregs) \
(((ptregs)->pstate & PSR_MODE_MASK) == PSR_MODE_EL0t)
#define compat_user_mode(ptregs) \
(((ptregs)->pstate & (PSR_MODE32_BIT | PSR_MODE_MASK)) == \
(PSR_MODE32_BIT | PSR_MODE_EL0t))
#define user_stack_pointer(ptregs) \
(!compat_user_mode(ptregs) ? (ptregs)->sp : (ptregs)->compat_sp)
#define user_frame_pointer(ptregs) \
(!compat_user_mode(ptregs) ? (ptregs)->regs[29] : (ptregs)->compat_fp)
static int
arm64_is_kernel_exception_frame(struct bt_info *bt, ulong stkptr)
{
struct arm64_pt_regs *regs;
struct machine_specific *ms = machdep->machspec;
if (stkptr > STACKSIZE() && !INSTACK(stkptr, bt)) {
if (CRASHDEBUG(1))
error(WARNING, "stkptr: %lx is outside the kernel stack range\n", stkptr);
return FALSE;
}
regs = (struct arm64_pt_regs *)&bt->stackbuf[(ulong)(STACK_OFFSET_TYPE(stkptr))];
if (INSTACK(regs->sp, bt) && INSTACK(regs->regs[29], bt) &&
!(regs->pstate & (0xffffffff00000000ULL | PSR_MODE32_BIT)) &&
is_kernel_text(regs->pc) &&
is_kernel_text(regs->regs[30] | ms->CONFIG_ARM64_KERNELPACMASK)) {
switch (regs->pstate & PSR_MODE_MASK)
{
case PSR_MODE_EL1t:
case PSR_MODE_EL1h:
case PSR_MODE_EL2t:
case PSR_MODE_EL2h:
return TRUE;
}
}
return FALSE;
}
static int
arm64_eframe_search(struct bt_info *bt)
{
int c;
ulong ptr, count;
struct machine_specific *ms;
if (bt->flags & BT_EFRAME_SEARCH2) {
if (!(machdep->flags & IRQ_STACKS))
error(FATAL, "IRQ stacks do not exist in this kernel\n");
ms = machdep->machspec;
for (c = 0; c < kt->cpus; c++) {
if ((bt->flags & BT_CPUMASK) &&
!(NUM_IN_BITMAP(bt->cpumask, c)))
continue;
fprintf(fp, "CPU %d IRQ STACK:", c);
bt->stackbase = ms->irq_stacks[c];
bt->stacktop = bt->stackbase + ms->irq_stack_size;
alter_stackbuf(bt);
count = 0;
for (ptr = bt->stackbase; ptr < bt->stacktop - SIZE(pt_regs); ptr++) {
if (arm64_is_kernel_exception_frame(bt, ptr)) {
fprintf(fp, "%s\nKERNEL-MODE EXCEPTION FRAME AT: %lx\n",
count ? "" : "\n", ptr);
arm64_print_exception_frame(bt, ptr, KERNEL_MODE, fp);
count++;
}
}
if (count)
fprintf(fp, "\n");
else
fprintf(fp, "(none found)\n\n");
}
return 0;
}
count = 0;
for (ptr = bt->stackbase; ptr < bt->stacktop - SIZE(pt_regs); ptr++) {
if (arm64_is_kernel_exception_frame(bt, ptr)) {
fprintf(fp, "\nKERNEL-MODE EXCEPTION FRAME AT: %lx\n", ptr);
arm64_print_exception_frame(bt, ptr, KERNEL_MODE, fp);
count++;
}
}
if (is_kernel_thread(bt->tc->task))
return count;
ptr = bt->stacktop - USER_EFRAME_OFFSET;
fprintf(fp, "%sUSER-MODE EXCEPTION FRAME AT: %lx\n",
count++ ? "\n" : "", ptr);
arm64_print_exception_frame(bt, ptr, USER_MODE, fp);
return count;
}
static char *arm64_exception_functions[] = {
"do_undefinstr",
"do_sysinstr",
"do_debug_exception",
"do_mem_abort",
"do_el0_irq_bp_hardening",
"do_sp_pc_abort",
"handle_bad_stack",
NULL
};
static int
arm64_in_exception_text(ulong ptr)
{
struct machine_specific *ms = machdep->machspec;
char *name, **func;
if (ms->__irqentry_text_start && ms->__irqentry_text_end &&
((ptr >= ms->__irqentry_text_start) &&
(ptr < ms->__irqentry_text_end)))
return TRUE;
if (ms->__exception_text_start && ms->__exception_text_end) {
if ((ptr >= ms->__exception_text_start) &&
(ptr < ms->__exception_text_end))
return TRUE;
}
name = closest_symbol(ptr);
if (name != NULL) { /* Linux 5.5 and later */
for (func = &arm64_exception_functions[0]; *func; func++) {
if (STREQ(name, *func))
return TRUE;
}
}
return FALSE;
}
static int
arm64_in_exp_entry(ulong addr)
{
struct machine_specific *ms;
ms = machdep->machspec;
if ((ms->exp_entry1_start <= addr) && (addr < ms->exp_entry1_end))
return TRUE;
if ((ms->exp_entry2_start <= addr) && (addr < ms->exp_entry2_end))
return TRUE;
return FALSE;
}
#define BACKTRACE_CONTINUE (1)
#define BACKTRACE_COMPLETE_KERNEL (2)
#define BACKTRACE_COMPLETE_USER (3)
static int
arm64_print_stackframe_entry(struct bt_info *bt, int level, struct arm64_stackframe *frame, FILE *ofp)
{
char *name, *name_plus_offset;
ulong branch_pc, symbol_offset;
struct syment *sp;
struct load_module *lm;
char buf[BUFSIZE];
/*
* if pc comes from a saved lr, it actually points to an instruction
* after branch. To avoid any confusion, decrement pc by 4.
* See, for example, "bl schedule" before ret_to_user().
*/
branch_pc = frame->pc - 4;
name = closest_symbol(branch_pc);
name_plus_offset = NULL;
if (bt->flags & BT_SYMBOL_OFFSET) {
sp = value_search(branch_pc, &symbol_offset);
if (sp && symbol_offset)
name_plus_offset =
value_to_symstr(branch_pc, buf, bt->radix);
}
if (!INSTACK(frame->fp, bt) && IN_TASK_VMA(bt->task, frame->fp))
frame->fp = 0;
if (bt->flags & BT_FULL) {
if (level)
arm64_display_full_frame(bt, frame->fp);
bt->frameptr = frame->fp;
}
fprintf(ofp, "%s#%d [%8lx] %s at %lx", level < 10 ? " " : "", level,
frame->fp ? frame->fp : bt->stacktop - USER_EFRAME_OFFSET,
name_plus_offset ? name_plus_offset : name, branch_pc);
if (BT_REFERENCE_CHECK(bt)) {
arm64_do_bt_reference_check(bt, frame->pc, closest_symbol(frame->pc));
arm64_do_bt_reference_check(bt, branch_pc, name);
}
if (module_symbol(branch_pc, NULL, &lm, NULL, 0))
fprintf(ofp, " [%s]", lm->mod_name);
fprintf(ofp, "\n");
if (bt->flags & BT_LINE_NUMBERS) {
get_line_number(branch_pc, buf, FALSE);
if (strlen(buf))
fprintf(ofp, " %s\n", buf);
}
if (STREQ(name, "start_kernel") || STREQ(name, "secondary_start_kernel") ||
STREQ(name, "kthread") || STREQ(name, "kthreadd"))
return BACKTRACE_COMPLETE_KERNEL;
return BACKTRACE_CONTINUE;
}
static int
arm64_print_stackframe_entry_v2(struct bt_info *bt, int level, struct arm64_stackframe *frame, FILE *ofp)
{
char *name, *name_plus_offset;
ulong pc, symbol_offset;
struct syment *sp;
struct load_module *lm;
char buf[BUFSIZE];
/*
* if pc comes from a saved lr, it actually points to an instruction
* after branch. To avoid any confusion, decrement pc by 4.
* See, for example, "bl schedule" before ret_to_user().
*/
pc = frame->pc - 0x4;
name = closest_symbol(pc);
name_plus_offset = NULL;
if (bt->flags & BT_SYMBOL_OFFSET) {
sp = value_search(pc, &symbol_offset);
if (sp && symbol_offset)
name_plus_offset = value_to_symstr(pc, buf, bt->radix);
}
if (bt->flags & BT_USER_EFRAME)
frame->fp = 0;
fprintf(ofp, "%s#%d [%8lx] %s at %lx", level < 10 ? " " : "", level,
frame->fp ? frame->fp : bt->stacktop - USER_EFRAME_OFFSET,
name_plus_offset ? name_plus_offset : name, pc);
if (BT_REFERENCE_CHECK(bt))
arm64_do_bt_reference_check(bt, pc, name);
if (module_symbol(pc, NULL, &lm, NULL, 0))
fprintf(ofp, " [%s]", lm->mod_name);
fprintf(ofp, "\n");
if (bt->flags & BT_LINE_NUMBERS) {
get_line_number(pc, buf, FALSE);
if (strlen(buf))
fprintf(ofp, " %s\n", buf);
}
if (STREQ(name, "start_kernel") ||
STREQ(name, "secondary_start_kernel") ||
STREQ(name, "kthread") || STREQ(name, "kthreadd"))
return BACKTRACE_COMPLETE_KERNEL;
return BACKTRACE_CONTINUE;
}
static void
arm64_display_full_frame(struct bt_info *bt, ulong sp)
{
int i, u_idx;
ulong *up;
ulong words, addr;
char buf[BUFSIZE];
if (bt->frameptr == sp)
return;
if (INSTACK(bt->frameptr, bt)) {
if (INSTACK(sp, bt)) {
; /* normal case */
} else {
if (sp == 0)
/* interrupt in user mode */
sp = bt->stacktop - USER_EFRAME_OFFSET;
else
/* interrupt in kernel mode */
sp = bt->stacktop;
}
} else {
/* This is a transition case from irq to process stack. */
return;
}
words = (sp - bt->frameptr) / sizeof(ulong);
addr = bt->frameptr;
u_idx = (bt->frameptr - bt->stackbase)/sizeof(ulong);
for (i = 0; i < words; i++, u_idx++) {
if (!(i & 1))
fprintf(fp, "%s %lx: ", i ? "\n" : "", addr);
up = (ulong *)(&bt->stackbuf[u_idx*sizeof(ulong)]);
fprintf(fp, "%s ", format_stack_entry(bt, buf, *up, 0));
addr += sizeof(ulong);
}
fprintf(fp, "\n");
}
static void
arm64_display_full_frame_v2(struct bt_info *bt, struct arm64_stackframe *cur,
struct arm64_stackframe *next)
{
struct machine_specific *ms;
ulong next_fp, stackbase;
char *stackbuf;
int i, u_idx;
ulong *up;
ulong words, addr;
char buf[BUFSIZE];
stackbase = bt->stackbase;
stackbuf = bt->stackbuf;
ms = machdep->machspec;
/* Calc next fp for dump */
if (next->fp == 0)
/* last stackframe on kernel tack */
next_fp = bt->stacktop - 0x10;
else if (!INSTACK(cur->sp, bt)) {
/* We have just switched over stacks */
next_fp = ms->irq_stacks[bt->tc->processor]
+ ms->irq_stack_size - 0x10;
/*
* We are already buffering a process stack.
* So use an old buffer for IRQ stack.
*/
stackbase = ms->irq_stacks[bt->tc->processor];
stackbuf = ms->irq_stackbuf;
} else
next_fp = next->fp;
if (CRASHDEBUG(1))
fprintf(fp, " frame <%016lx:%016lx>\n", cur->fp, next_fp);
/* Check here because we want to see a debug message above. */
if (!(bt->flags & BT_FULL))
return;
if (next_fp <= cur->fp)
return;
/* Dump */
words = (next_fp - cur->fp) / sizeof(ulong);
addr = cur->fp;
u_idx = (cur->fp - stackbase)/sizeof(ulong);
for (i = 0; i < words; i++, u_idx++) {
if (!(i & 1))
fprintf(fp, "%s %lx: ", i ? "\n" : "", addr);
up = (ulong *)(&stackbuf[u_idx*sizeof(ulong)]);
fprintf(fp, "%s ", format_stack_entry(bt, buf, *up, 0));
addr += sizeof(ulong);
}
fprintf(fp, "\n");
if (stackbuf == ms->irq_stackbuf)
FREEBUF(stackbuf);
}
static int
arm64_unwind_frame(struct bt_info *bt, struct arm64_stackframe *frame)
{
unsigned long high, low, fp;
unsigned long stack_mask;
unsigned long irq_stack_ptr, orig_sp;
struct arm64_pt_regs *ptregs;
struct machine_specific *ms = machdep->machspec;
stack_mask = (unsigned long)(ARM64_STACK_SIZE) - 1;
fp = frame->fp;
low = frame->sp;
high = (low + stack_mask) & ~(stack_mask);
if (fp < low || fp > high || fp & 0xf || !INSTACK(fp, bt))
return FALSE;
frame->sp = fp + 0x10;
frame->fp = GET_STACK_ULONG(fp);
frame->pc = GET_STACK_ULONG(fp + 8);
if (is_kernel_text(frame->pc | ms->CONFIG_ARM64_KERNELPACMASK))
frame->pc |= ms->CONFIG_ARM64_KERNELPACMASK;
if ((frame->fp == 0) && (frame->pc == 0))
return FALSE;
if (!(machdep->flags & (IRQ_STACKS | OVERFLOW_STACKS)))
return TRUE;
if (machdep->flags & UNW_4_14) {
if (((bt->flags & BT_IRQSTACK) &&
!arm64_on_irq_stack(bt->tc->processor, frame->fp)) ||
((bt->flags & BT_OVERFLOW_STACK) &&
!arm64_on_overflow_stack(bt->tc->processor, frame->fp))) {
if (arm64_on_process_stack(bt, frame->fp)) {
arm64_set_process_stack(bt);
frame->sp = frame->fp - KERN_EFRAME_OFFSET;
/*
* for switch_stack
* fp still points to irq stack
*/
bt->bptr = fp;
/*
* for display_full_frame
* sp points to process stack
*
* If we want to see pt_regs,
* comment out the below.
* bt->frameptr = frame->sp;
*/
} else {
/* irq -> user */
return FALSE;
}
}
return TRUE;
}
/*
* The kernel's manner of determining the end of the IRQ stack:
*
* #define THREAD_SIZE 16384
* #define THREAD_START_SP (THREAD_SIZE - 16)
* #define IRQ_STACK_START_SP THREAD_START_SP
* #define IRQ_STACK_PTR(cpu) ((unsigned long)per_cpu(irq_stack, cpu) + IRQ_STACK_START_SP)
* #define IRQ_STACK_TO_TASK_STACK(ptr) (*((unsigned long *)((ptr) - 0x08)))
*
* irq_stack_ptr = IRQ_STACK_PTR(raw_smp_processor_id());
* orig_sp = IRQ_STACK_TO_TASK_STACK(irq_stack_ptr); (pt_regs pointer on process stack)
*/
irq_stack_ptr = ms->irq_stacks[bt->tc->processor] + ms->irq_stack_size - 16;
if (frame->sp == irq_stack_ptr) {
orig_sp = GET_STACK_ULONG(irq_stack_ptr - 8);
arm64_set_process_stack(bt);
if (INSTACK(orig_sp, bt) && (INSTACK(frame->fp, bt) || (frame->fp == 0))) {
ptregs = (struct arm64_pt_regs *)&bt->stackbuf[(ulong)(STACK_OFFSET_TYPE(orig_sp))];
frame->sp = orig_sp;
frame->pc = ptregs->pc;
bt->bptr = fp;
if (CRASHDEBUG(1))
error(INFO,
"arm64_unwind_frame: switch stacks: fp: %lx sp: %lx pc: %lx\n",
frame->fp, frame->sp, frame->pc);
} else {
error(WARNING,
"arm64_unwind_frame: on IRQ stack: oriq_sp: %lx%s fp: %lx%s\n",
orig_sp, INSTACK(orig_sp, bt) ? "" : " (?)",
frame->fp, INSTACK(frame->fp, bt) ? "" : " (?)");
return FALSE;
}
}
return TRUE;
}
/*
* The following figure shows how unwinding can be done.
* Here we assume that the callstack order is:
* #(X-1) ppc (previous PC)
* #X cpc (current PC)
* < #(X+ 1) epc (Exception entry) >
* #(X+1/2) npc (Next PC)
* #(X+2/3) Npc (One before Next)
* #(X+3/4) NNpc (One before 'Npc')
* and unwind frames from #X to #(X+1).
* When we add a faked frame for exception entry (exception frame)
* as #(X+1), the next frame for npc will be recognized as #(x+2).
*
* (1)Normal stackframe:
* +------+
* | pfp |
* | cpc |
* psp + +
* | |
* | |
* pfp +------+ <--- :prev stackframe = <pfp, psp, ppc>
* | cfp |
* | npc |
* csp + +
* | |
* | |
* cfp +------+ <--- :curr stackframe = <cfp, csp, cpc>
* | nfp | cfp = *pfp
* | Npc | csp = pfp + 0x10
* nsp + +
* | |
* | |
* nfp +------+ <--- :next stackframe = <nfp, nsp, npc>
* | |
*
* (2)Exception on the same (IRQ or process) stack:
* +------+
* | pfp |
* | cpc |
* psp + +
* | |
* | |
* pfp +------+ <--- :prev stackframe = <pfp, psp, ppc>
* | cfp |
* | npc |
* csp + +
* | |
* | |
* cfp +------+ <--- :curr stackframe = <cfp, csp, cpc>
* | nfp |
* | epc |
* + +
* | |
* | | faked(*)
* esp +------+ <--- :excp stackframe = <---, esp, epc
* | | esp = nsp - sizeof(pt_regs)
* | |
* | Npc | (*) If we didn't add this frame, the next frame
* | nfp | would be
* | nsp | <nfp, nfp + 0x10, epc>
* | npc | and the frame below for npc would be lost.
* nsp + +
* | |
* nfp +------+ <--- :task stackframe = <nfp, nsp, npc>
* | Nfp |
* | NNpc |
* Nsp + +
* | |
* Nfp +------+ <--- :task stackframe = <Nfp, Nsp, Npc>
* | NNfp |
*
* (3)Interrupt:
* +------+
* | cfp |
* | ipc |
* csp + +
* | |
* | |
* cfp +------+ <--- :curr stackframe = <cfp, csp, cpc>
* | ifp |
* | epc |
* isp + +
* | |
* | | (*)
* ifp +------+ <--- :irq stackframe = <ifp, isp, epc>
* | nfp | ifp == IRQ_STACK_PTR
* | esp | (*) Before the kernel enters an irq handler, frame
* top +------+ pointer moves to the top of IRQ stack.
* IRQ stack So we have to skip this frame in unwinding.
*
* faked
* esp +------+ <--- :excp stackframe = <---, esp, epc>
* | | esp = nsp - sizeof(pt_regs)
* | |
* | Npc |
* | nfp |
* | nsp |
* | npc |
* nsp + +
* | |
* nfp +------+ <--- :task stackframe = <nfp, nsp, npc>
* | Nfp |
* | NNpc |
* Nsp + +
* | |
* Nfp +------+ <--- :task stackframe = <Nfp, Nsp, Npc>
* | NNfp |
*/
static struct arm64_stackframe ext_frame;
static int
arm64_unwind_frame_v2(struct bt_info *bt, struct arm64_stackframe *frame,
FILE *ofp)
{
unsigned long high, low, fp;
unsigned long stack_mask;
unsigned long irq_stack_ptr;
struct machine_specific *ms;
stack_mask = (unsigned long)(ARM64_STACK_SIZE) - 1;
fp = frame->fp;
low = frame->sp;
high = (low + stack_mask) & ~(stack_mask);
if (fp < low || fp > high || fp & 0xf || !INSTACK(fp, bt))
return FALSE;
if (CRASHDEBUG(1))
fprintf(ofp, " cur fp:%016lx sp:%016lx pc:%016lx\n",
frame->fp, frame->sp, frame->pc);
if (ext_frame.pc) {
/*
* The previous frame was a dummy for exception entry.
* So complement a missing (task) stackframe now.
*/
frame->fp = ext_frame.fp;
frame->sp = ext_frame.sp;
frame->pc = ext_frame.pc;
ext_frame.pc = 0; /* back to normal unwinding */
goto unwind_done;
}
frame->pc = GET_STACK_ULONG(fp + 8);
if (!arm64_in_exp_entry(frame->pc)) {
/* (1) Normal stack frame */
frame->sp = fp + 0x10;
frame->fp = GET_STACK_ULONG(fp);
} else {
/*
* We are in exception entry code, and so
* - add a faked frame for exception entry, and
* - prepare for a stackframe hidden by exception
*/
ext_frame.fp = GET_STACK_ULONG(fp);
/*
* Note:
* In the following code, we determine a stack pointer for
* exception entry based on ext_frame.fp because we have
* no way to know a ext_frame.sp.
* Fortunately, this will work fine for most functions
* in the kernel.
*/
if (ext_frame.fp == 0) {
/*
* (2)
* Either on process stack or on IRQ stack,
* the next frame is the last one on process stack.
*/
frame->sp = bt->stacktop
- sizeof(struct arm64_pt_regs) - 0x10;
frame->fp = frame->sp;
} else if (!arm64_on_irq_stack(bt->tc->processor, frame->sp)) {
/*
* (2)
* We are on process stack. Just add a faked frame
*/
if (!arm64_on_irq_stack(bt->tc->processor, ext_frame.fp))
frame->sp = ext_frame.fp
- sizeof(struct arm64_pt_regs);
else {
/*
* FIXME: very exceptional case
* We are already back on process stack, but
* a saved frame pointer indicates that we are
* on IRQ stack. Unfortunately this can happen
* when some functions are called after
* an irq handler is done because irq_exit()
* doesn't restore a frame pointer (x29).
* Those functions include
* - do_notify_resume()
* - trace_hardirqs_off()
* - schedule()
*
* We have no perfect way to determine a true
* stack pointer value here.
* 0x20 is a stackframe size of schedule().
* Really ugly
*/
frame->sp = frame->fp + 0x20;
fprintf(ofp, " (Next exception frame might be wrong)\n");
}
frame->fp = frame->sp;
} else {
/* We are on IRQ stack */
ms = machdep->machspec;
irq_stack_ptr = ms->irq_stacks[bt->tc->processor]
+ ms->irq_stack_size - 0x20;
if (ext_frame.fp != irq_stack_ptr) {
/* (2) Just add a faked frame */
frame->sp = ext_frame.fp
- sizeof(struct arm64_pt_regs);
frame->fp = frame->sp;
} else {
/*
* (3)
* Switch from IRQ stack to process stack
*/
frame->sp = GET_STACK_ULONG(irq_stack_ptr + 8);
frame->fp = frame->sp;
/*
* Keep a buffer for a while until
* displaying the last frame on IRQ stack
* at next arm64_print_stackframe_entry_v2()
*/
if (bt->flags & BT_FULL)
ms->irq_stackbuf = bt->stackbuf;
arm64_set_process_stack(bt);
}
}
/* prepare for a stackframe hidden by exception */
arm64_gen_hidden_frame(bt, frame->sp, &ext_frame);
}
unwind_done:
if (CRASHDEBUG(1))
fprintf(ofp, " nxt fp:%016lx sp:%016lx pc:%016lx\n",
frame->fp, frame->sp, frame->pc);
return TRUE;
}
/*
* A layout of a stack frame in a function looks like:
*
* stack grows to lower addresses.
* /|\
* |
* | |
* new sp +------+ <---
* |dyn | |
* | vars | |
* new fp +- - - + |
* |old fp| | a function's stack frame
* |old lr| |
* |static| |
* | vars| |
* old sp +------+ <---
* |dyn |
* | vars |
* old fp +------+
* | |
*
* - On function entry, sp is decremented down to new fp.
*
* - and old fp and sp are saved into this stack frame.
* "Static" local variables are allocated at the same time.
*
* - Later on, "dynamic" local variables may be allocated on a stack.
* But those dynamic variables are rarely used in the kernel image,
* and, as a matter of fact, sp is equal to fp in almost all functions.
* (not 100% though.)
*
* - Currently, sp is determined in arm64_unwind_frame() by
* sp = a callee's fp + 0x10
* where 0x10 stands for a saved area for fp and sp
*
* - As you can see, however, this calculated sp still points to the top of
* callee's static local variables and doesn't match with a *real* sp.
*
* - So, generally, dumping a stack from this calculated sp to the next frame's
* sp shows "callee's static local variables", old fp and sp.
*
* Diagram and explanation courtesy of Takahiro Akashi
*/
static void
arm64_back_trace_cmd(struct bt_info *bt)
{
struct arm64_stackframe stackframe;
int level;
ulong exception_frame;
FILE *ofp;
if (bt->flags & BT_OPT_BACK_TRACE) {
if (machdep->flags & UNW_4_14) {
option_not_supported('o');
return;
}
arm64_back_trace_cmd_v2(bt);
return;
}
ofp = BT_REFERENCE_CHECK(bt) ? pc->nullfp : fp;
/*
* stackframes are created from 3 contiguous stack addresses:
*
* x: contains stackframe.fp -- points to next triplet
* x+8: contains stackframe.pc -- text return address
* x+16: is the stackframe.sp address
*/
if (bt->flags & BT_KDUMP_ADJUST) {
if (arm64_on_irq_stack(bt->tc->processor, bt->bptr)) {
arm64_set_irq_stack(bt);
bt->flags |= BT_IRQSTACK;
}
stackframe.fp = GET_STACK_ULONG(bt->bptr - 8);
stackframe.pc = GET_STACK_ULONG(bt->bptr);
stackframe.sp = bt->bptr + 8;
bt->frameptr = stackframe.sp;
} else if (bt->hp && bt->hp->esp) {
if (arm64_on_irq_stack(bt->tc->processor, bt->hp->esp)) {
arm64_set_irq_stack(bt);
bt->flags |= BT_IRQSTACK;
}
stackframe.fp = GET_STACK_ULONG(bt->hp->esp - 8);
stackframe.pc = bt->hp->eip ?
bt->hp->eip : GET_STACK_ULONG(bt->hp->esp);
stackframe.sp = bt->hp->esp + 8;
bt->flags &= ~BT_REGS_NOT_FOUND;
} else {
if (arm64_on_irq_stack(bt->tc->processor, bt->frameptr)) {
arm64_set_irq_stack(bt);
bt->flags |= BT_IRQSTACK;
} else if (arm64_on_overflow_stack(bt->tc->processor, bt->frameptr)) {
arm64_set_overflow_stack(bt);
bt->flags |= BT_OVERFLOW_STACK;
}
stackframe.sp = bt->stkptr;
stackframe.pc = bt->instptr;
stackframe.fp = bt->frameptr;
}
if (bt->flags & BT_TEXT_SYMBOLS) {
arm64_print_text_symbols(bt, &stackframe, ofp);
if (BT_REFERENCE_FOUND(bt)) {
print_task_header(fp, task_to_context(bt->task), 0);
arm64_print_text_symbols(bt, &stackframe, fp);
fprintf(fp, "\n");
}
return;
}
if (bt->flags & BT_REGS_NOT_FOUND)
return;
if (!(bt->flags & BT_KDUMP_ADJUST)) {
if (bt->flags & BT_USER_SPACE)
goto complete_user;
if (DUMPFILE() && is_task_active(bt->task)) {
exception_frame = stackframe.fp - KERN_EFRAME_OFFSET;
if (arm64_is_kernel_exception_frame(bt, exception_frame))
arm64_print_exception_frame(bt, exception_frame,
KERNEL_MODE, ofp);
}
}
level = exception_frame = 0;
while (1) {
bt->instptr = stackframe.pc;
switch (arm64_print_stackframe_entry(bt, level, &stackframe, ofp))
{
case BACKTRACE_COMPLETE_KERNEL:
return;
case BACKTRACE_COMPLETE_USER:
goto complete_user;
case BACKTRACE_CONTINUE:
break;
}
if (exception_frame) {
arm64_print_exception_frame(bt, exception_frame, KERNEL_MODE, ofp);
exception_frame = 0;
}
if (!arm64_unwind_frame(bt, &stackframe))
break;
if (arm64_in_exception_text(bt->instptr) && INSTACK(stackframe.fp, bt)) {
if (bt->flags & BT_OVERFLOW_STACK) {
exception_frame = stackframe.fp - KERN_EFRAME_OFFSET;
} else if (!(bt->flags & BT_IRQSTACK) ||
((stackframe.sp + SIZE(pt_regs)) < bt->stacktop)) {
if (arm64_is_kernel_exception_frame(bt, stackframe.fp - KERN_EFRAME_OFFSET))
exception_frame = stackframe.fp - KERN_EFRAME_OFFSET;
}
}
if ((bt->flags & BT_IRQSTACK) &&
!arm64_on_irq_stack(bt->tc->processor, stackframe.fp)) {
bt->flags &= ~BT_IRQSTACK;
if (arm64_switch_stack(bt, &stackframe, ofp) == USER_MODE)
break;
}
if ((bt->flags & BT_OVERFLOW_STACK) &&
!arm64_on_overflow_stack(bt->tc->processor, stackframe.fp)) {
bt->flags &= ~BT_OVERFLOW_STACK;
if (arm64_switch_stack_from_overflow(bt, &stackframe, ofp) == USER_MODE)
break;
}
level++;
}
if (is_kernel_thread(bt->tc->task))
return;
complete_user:
exception_frame = bt->stacktop - USER_EFRAME_OFFSET;
arm64_print_exception_frame(bt, exception_frame, USER_MODE, ofp);
if ((bt->flags & (BT_USER_SPACE|BT_KDUMP_ADJUST)) == BT_USER_SPACE)
fprintf(ofp, " #0 [user space]\n");
}
static void
arm64_back_trace_cmd_v2(struct bt_info *bt)
{
struct arm64_stackframe stackframe, cur_frame;
int level, mode;
ulong exception_frame;
FILE *ofp;
ofp = BT_REFERENCE_CHECK(bt) ? pc->nullfp : fp;
/*
* stackframes are created from 3 contiguous stack addresses:
*
* x: contains stackframe.fp -- points to next triplet
* x+8: contains stackframe.pc -- text return address
* x+16: is the stackframe.sp address
*/
if (bt->flags & BT_KDUMP_ADJUST) {
if (arm64_on_irq_stack(bt->tc->processor, bt->bptr)) {
arm64_set_irq_stack(bt);
bt->flags |= BT_IRQSTACK;
}
stackframe.fp = GET_STACK_ULONG(bt->bptr);
stackframe.pc = GET_STACK_ULONG(bt->bptr + 8);
stackframe.sp = bt->bptr + 16;
bt->frameptr = stackframe.fp;
} else {
if (arm64_on_irq_stack(bt->tc->processor, bt->frameptr)) {
arm64_set_irq_stack(bt);
bt->flags |= BT_IRQSTACK;
}
stackframe.sp = bt->stkptr;
stackframe.pc = bt->instptr;
stackframe.fp = bt->frameptr;
}
if (bt->flags & BT_TEXT_SYMBOLS) {
arm64_print_text_symbols(bt, &stackframe, ofp);
if (BT_REFERENCE_FOUND(bt)) {
print_task_header(fp, task_to_context(bt->task), 0);
arm64_print_text_symbols(bt, &stackframe, fp);
fprintf(fp, "\n");
}
return;
}
if (bt->flags & BT_REGS_NOT_FOUND)
return;
if (!(bt->flags & BT_KDUMP_ADJUST)) {
if (bt->flags & BT_USER_SPACE) {
user_space:
exception_frame = bt->stacktop - USER_EFRAME_OFFSET;
arm64_print_exception_frame(bt, exception_frame,
USER_MODE, ofp);
// fprintf(ofp, " #0 [user space]\n");
return;
}
if (DUMPFILE() && is_task_active(bt->task)) {
exception_frame = stackframe.fp - SIZE(pt_regs);
if (arm64_is_kernel_exception_frame(bt, exception_frame))
arm64_print_exception_frame(bt, exception_frame,
KERNEL_MODE, ofp);
}
}
for (level = 0;; level++) {
bt->instptr = stackframe.pc;
/*
* Show one-line stackframe info
*/
if (arm64_print_stackframe_entry_v2(bt, level, &stackframe, ofp)
== BACKTRACE_COMPLETE_KERNEL)
break;
cur_frame = stackframe;
if (!arm64_unwind_frame_v2(bt, &stackframe, ofp))
break;
/*
* Dump the contents of the current stackframe.
* We need to know the next stackframe to determine
* the dump range:
* <cur_frame.fp:stackframe.fp>
*/
arm64_display_full_frame_v2(bt, &cur_frame, &stackframe);
/*
* If we are in a normal stackframe, just continue,
* otherwise show an exception frame.
* Since exception entry code doesn't have a real
* stackframe, we fake a dummy frame here.
*/
if (!arm64_in_exp_entry(stackframe.pc))
continue;
if (!INSTACK(cur_frame.sp, bt))
fprintf(ofp, "--- <IRQ stack> ---\n");
arm64_print_stackframe_entry_v2(bt, ++level, &stackframe, ofp);
if (bt->flags & BT_USER_EFRAME)
goto user_space;
cur_frame = stackframe;
arm64_unwind_frame_v2(bt, &stackframe, ofp);
/*
* and don't show the contenxts. Instead,
* show an exception frame below
*/
if (!INSTACK(cur_frame.sp, bt)) {
/* This check is a safeguard. See unwind_frame(). */
error(WARNING,
"stack pointer for exception frame is wrong\n");
return;
}
mode = (stackframe.pc < machdep->machspec->userspace_top) ?
USER_MODE : KERNEL_MODE;
// fprintf(ofp, "--- <Exception in %s> ---\n",
// mode == KERNEL_MODE ? "kernel" : "user");
arm64_print_exception_frame(bt, cur_frame.sp, mode, ofp);
if (mode == USER_MODE)
break;
}
}
static void
arm64_print_text_symbols(struct bt_info *bt, struct arm64_stackframe *frame, FILE *ofp)
{
int i;
ulong *up;
struct load_module *lm;
char buf1[BUFSIZE];
char buf2[BUFSIZE];
char *name;
ulong start;
ulong val;
struct machine_specific *ms = machdep->machspec;
if (bt->flags & BT_TEXT_SYMBOLS_ALL)
start = bt->stackbase;
else {
start = frame->sp - 8;
fprintf(ofp, "%sSTART: %s at %lx\n",
space(VADDR_PRLEN > 8 ? 14 : 6),
bt->flags & BT_SYMBOL_OFFSET ?
value_to_symstr(frame->pc, buf2, bt->radix) :
closest_symbol(frame->pc), frame->pc);
}
for (i = (start - bt->stackbase)/sizeof(ulong); i < LONGS_PER_STACK; i++) {
up = (ulong *)(&bt->stackbuf[i*sizeof(ulong)]);
val = *up;
if (is_kernel_text(val | ms->CONFIG_ARM64_KERNELPACMASK)) {
val |= ms->CONFIG_ARM64_KERNELPACMASK;
name = closest_symbol(val);
fprintf(ofp, " %s[%s] %s at %lx",
bt->flags & BT_ERROR_MASK ?
" " : "",
mkstring(buf1, VADDR_PRLEN,
RJUST|LONG_HEX,
MKSTR(bt->stackbase +
(i * sizeof(long)))),
bt->flags & BT_SYMBOL_OFFSET ?
value_to_symstr(val, buf2, bt->radix) :
name, val);
if (module_symbol(val, NULL, &lm, NULL, 0))
fprintf(ofp, " [%s]", lm->mod_name);
fprintf(ofp, "\n");
if (BT_REFERENCE_CHECK(bt))
arm64_do_bt_reference_check(bt, val, name);
}
}
}
static int
arm64_in_kdump_text(struct bt_info *bt, struct arm64_stackframe *frame)
{
ulong *ptr, *start, *base;
struct machine_specific *ms;
ulong crash_kexec_frame;
if (!(machdep->flags & KDUMP_ENABLED))
return FALSE;
base = (ulong *)&bt->stackbuf[(ulong)(STACK_OFFSET_TYPE(bt->stackbase))];
if (bt->flags & BT_USER_SPACE)
start = (ulong *)&bt->stackbuf[(ulong)(STACK_OFFSET_TYPE(bt->stacktop))];
else {
if (INSTACK(frame->fp, bt))
start = (ulong *)&bt->stackbuf[(ulong)(STACK_OFFSET_TYPE(frame->fp))];
else
start = (ulong *)&bt->stackbuf[(ulong)(STACK_OFFSET_TYPE(bt->stacktop))];
}
crash_kexec_frame = 0;
ms = machdep->machspec;
for (ptr = start - 8; ptr >= base; ptr--) {
if (bt->flags & BT_OPT_BACK_TRACE) {
if ((*ptr > ms->crash_kexec_start) &&
(*ptr < ms->crash_kexec_end) &&
INSTACK(*(ptr - 1), bt)) {
bt->bptr = ((ulong)(ptr - 1) - (ulong)base)
+ task_to_stackbase(bt->tc->task);
if (CRASHDEBUG(1))
fprintf(fp, "%lx: %lx (crash_kexec)\n", bt->bptr, *ptr);
return TRUE;
}
if ((*ptr > ms->crash_save_cpu_start) &&
(*ptr < ms->crash_save_cpu_end) &&
INSTACK(*(ptr - 1), bt)) {
bt->bptr = ((ulong)(ptr - 1) - (ulong)base)
+ task_to_stackbase(bt->tc->task);
if (CRASHDEBUG(1))
fprintf(fp, "%lx: %lx (crash_save_cpu)\n", bt->bptr, *ptr);
return TRUE;
}
} else {
if ((*ptr > ms->machine_kexec_start) && (*ptr < ms->machine_kexec_end)) {
bt->bptr = ((ulong)ptr - (ulong)base)
+ task_to_stackbase(bt->tc->task);
if (CRASHDEBUG(1))
fprintf(fp, "%lx: %lx (machine_kexec)\n", bt->bptr, *ptr);
return TRUE;
}
if ((*ptr > ms->crash_kexec_start) && (*ptr < ms->crash_kexec_end)) {
/*
* Stash the first crash_kexec frame in case the machine_kexec
* frame is not found.
*/
if (!crash_kexec_frame) {
crash_kexec_frame = ((ulong)ptr - (ulong)base)
+ task_to_stackbase(bt->tc->task);
if (CRASHDEBUG(1))
fprintf(fp, "%lx: %lx (crash_kexec)\n",
bt->bptr, *ptr);
}
continue;
}
if ((*ptr > ms->crash_save_cpu_start) && (*ptr < ms->crash_save_cpu_end)) {
bt->bptr = ((ulong)ptr - (ulong)base)
+ task_to_stackbase(bt->tc->task);
if (CRASHDEBUG(1))
fprintf(fp, "%lx: %lx (crash_save_cpu)\n", bt->bptr, *ptr);
return TRUE;
}
}
}
if (crash_kexec_frame) {
bt->bptr = crash_kexec_frame;
return TRUE;
}
return FALSE;
}
static int
arm64_in_kdump_text_on_irq_stack(struct bt_info *bt)
{
int cpu;
ulong stackbase;
char *stackbuf;
ulong *ptr, *start, *base;
struct machine_specific *ms;
if ((machdep->flags & (IRQ_STACKS|KDUMP_ENABLED)) != (IRQ_STACKS|KDUMP_ENABLED))
return FALSE;
ms = machdep->machspec;
cpu = bt->tc->processor;
stackbase = ms->irq_stacks[cpu];
stackbuf = GETBUF(ms->irq_stack_size);
if (!readmem(stackbase, KVADDR, stackbuf,
ms->irq_stack_size, "IRQ stack contents", RETURN_ON_ERROR)) {
error(INFO, "read of IRQ stack at %lx failed\n", stackbase);
FREEBUF(stackbuf);
return FALSE;
}
base = (ulong *)stackbuf;
start = (ulong *)(stackbuf + ms->irq_stack_size);
for (ptr = start - 8; ptr >= base; ptr--) {
if (bt->flags & BT_OPT_BACK_TRACE) {
if ((*ptr > ms->crash_kexec_start) &&
(*ptr < ms->crash_kexec_end) &&
INSTACK(*(ptr - 1), bt)) {
bt->bptr = ((ulong)(ptr - 1) - (ulong)base) + stackbase;
if (CRASHDEBUG(1))
fprintf(fp, "%lx: %lx (crash_kexec on IRQ stack)\n",
bt->bptr, *ptr);
FREEBUF(stackbuf);
return TRUE;
}
if ((*ptr > ms->crash_save_cpu_start) &&
(*ptr < ms->crash_save_cpu_end) &&
INSTACK(*(ptr - 1), bt)) {
bt->bptr = ((ulong)(ptr - 1) - (ulong)base) + stackbase;
if (CRASHDEBUG(1))
fprintf(fp, "%lx: %lx (crash_save_cpu on IRQ stack)\n",
bt->bptr, *ptr);
FREEBUF(stackbuf);
return TRUE;
}
} else {
if ((*ptr > ms->crash_kexec_start) && (*ptr < ms->crash_kexec_end)) {
bt->bptr = ((ulong)ptr - (ulong)base) + stackbase;
if (CRASHDEBUG(1))
fprintf(fp, "%lx: %lx (crash_kexec on IRQ stack)\n",
bt->bptr, *ptr);
FREEBUF(stackbuf);
return TRUE;
}
if ((*ptr > ms->crash_save_cpu_start) && (*ptr < ms->crash_save_cpu_end)) {
bt->bptr = ((ulong)ptr - (ulong)base) + stackbase;
if (CRASHDEBUG(1))
fprintf(fp, "%lx: %lx (crash_save_cpu on IRQ stack)\n",
bt->bptr, *ptr);
FREEBUF(stackbuf);
return TRUE;
}
}
}
FREEBUF(stackbuf);
return FALSE;
}
static int
arm64_switch_stack(struct bt_info *bt, struct arm64_stackframe *frame, FILE *ofp)
{
int i;
ulong stacktop, words, addr;
ulong *stackbuf;
char buf[BUFSIZE];
struct machine_specific *ms = machdep->machspec;
if (bt->flags & BT_FULL) {
stacktop = ms->irq_stacks[bt->tc->processor] + ms->irq_stack_size;
words = (stacktop - bt->bptr) / sizeof(ulong);
stackbuf = (ulong *)GETBUF(words * sizeof(ulong));
readmem(bt->bptr, KVADDR, stackbuf, words * sizeof(long),
"top of IRQ stack", FAULT_ON_ERROR);
addr = bt->bptr;
for (i = 0; i < words; i++) {
if (!(i & 1))
fprintf(ofp, "%s %lx: ", i ? "\n" : "", addr);
fprintf(ofp, "%s ", format_stack_entry(bt, buf, stackbuf[i], 0));
addr += sizeof(ulong);
}
fprintf(ofp, "\n");
FREEBUF(stackbuf);
}
fprintf(ofp, "--- <IRQ stack> ---\n");
if (frame->fp == 0)
return USER_MODE;
if (!(machdep->flags & UNW_4_14))
arm64_print_exception_frame(bt, frame->sp, KERNEL_MODE, ofp);
return KERNEL_MODE;
}
static int
arm64_switch_stack_from_overflow(struct bt_info *bt, struct arm64_stackframe *frame, FILE *ofp)
{
int i;
ulong stacktop, words, addr;
ulong *stackbuf;
char buf[BUFSIZE];
struct machine_specific *ms = machdep->machspec;
if (bt->flags & BT_FULL) {
stacktop = ms->overflow_stacks[bt->tc->processor] + ms->overflow_stack_size;
words = (stacktop - bt->bptr) / sizeof(ulong);
stackbuf = (ulong *)GETBUF(words * sizeof(ulong));
readmem(bt->bptr, KVADDR, stackbuf, words * sizeof(long),
"top of overflow stack", FAULT_ON_ERROR);
addr = bt->bptr;
for (i = 0; i < words; i++) {
if (!(i & 1))
fprintf(ofp, "%s %lx: ", i ? "\n" : "", addr);
fprintf(ofp, "%s ", format_stack_entry(bt, buf, stackbuf[i], 0));
addr += sizeof(ulong);
}
fprintf(ofp, "\n");
FREEBUF(stackbuf);
}
fprintf(ofp, "--- <Overflow stack> ---\n");
if (frame->fp == 0)
return USER_MODE;
if (!(machdep->flags & UNW_4_14))
arm64_print_exception_frame(bt, frame->sp, KERNEL_MODE, ofp);
return KERNEL_MODE;
}
static int
arm64_get_dumpfile_stackframe(struct bt_info *bt, struct arm64_stackframe *frame)
{
struct machine_specific *ms = machdep->machspec;
struct arm64_pt_regs *ptregs;
bool skip = false;
if (bt->flags & BT_SKIP_IDLE) {
skip = true;
bt->flags &= ~BT_SKIP_IDLE;
}
if (!ms->panic_task_regs ||
(!ms->panic_task_regs[bt->tc->processor].sp &&
!ms->panic_task_regs[bt->tc->processor].pc)) {
bt->flags |= BT_REGS_NOT_FOUND;
return FALSE;
}
ptregs = &ms->panic_task_regs[bt->tc->processor];
frame->pc = ptregs->pc;
if (user_mode(ptregs)) {
frame->sp = user_stack_pointer(ptregs);
frame->fp = user_frame_pointer(ptregs);
if (is_kernel_text(frame->pc) ||
!in_user_stack(bt->tc->task, frame->sp)) {
error(WARNING,
"corrupt NT_PRSTATUS? pstate: 0x%lx, but no user frame found\n",
ptregs->pstate);
if (is_kernel_text(frame->pc) &&
INSTACK(frame->sp, bt) && INSTACK(frame->fp, bt))
goto try_kernel;
bt->flags |= BT_REGS_NOT_FOUND;
return FALSE;
}
bt->flags |= BT_USER_SPACE;
} else {
try_kernel:
frame->sp = ptregs->sp;
frame->fp = ptregs->regs[29];
bt->machdep = ptregs;
}
if (arm64_in_kdump_text(bt, frame) ||
arm64_in_kdump_text_on_irq_stack(bt)) {
bt->flags |= BT_KDUMP_ADJUST;
if (skip && is_idle_thread(bt->task))
bt->flags |= BT_SKIP_IDLE;
}
return TRUE;
}
static int
arm64_get_stackframe(struct bt_info *bt, struct arm64_stackframe *frame)
{
if (!fill_task_struct(bt->task))
return FALSE;
frame->sp = ULONG(tt->task_struct + OFFSET(task_struct_thread_context_sp));
frame->pc = ULONG(tt->task_struct + OFFSET(task_struct_thread_context_pc));
frame->fp = ULONG(tt->task_struct + OFFSET(task_struct_thread_context_fp));
return TRUE;
}
static void
arm64_get_stack_frame(struct bt_info *bt, ulong *pcp, ulong *spp)
{
struct user_regs_bitmap_struct *ur_bitmap;
struct arm64_stackframe stackframe = { 0 };
if (DUMPFILE() && is_task_active(bt->task)) {
arm64_get_dumpfile_stackframe(bt, &stackframe);
bt->need_free = FALSE;
} else {
if (bt->flags & BT_SKIP_IDLE)
bt->flags &= ~BT_SKIP_IDLE;
arm64_get_stackframe(bt, &stackframe);
ur_bitmap = (struct user_regs_bitmap_struct *)GETBUF(sizeof(*ur_bitmap));
memset(ur_bitmap, 0, sizeof(*ur_bitmap));
ur_bitmap->ur.pc = stackframe.pc;
ur_bitmap->ur.sp = stackframe.sp;
ur_bitmap->ur.regs[29] = stackframe.fp;
SET_BIT(ur_bitmap->bitmap, REG_SEQ(arm64_pt_regs, pc));
SET_BIT(ur_bitmap->bitmap, REG_SEQ(arm64_pt_regs, sp));
SET_BIT(ur_bitmap->bitmap, REG_SEQ(arm64_pt_regs, regs[0])
+ X29_REGNUM - X0_REGNUM);
bt->machdep = ur_bitmap;
bt->need_free = TRUE;
}
bt->frameptr = stackframe.fp;
if (pcp)
*pcp = stackframe.pc;
if (spp)
*spp = stackframe.sp;
}
static void
arm64_gen_hidden_frame(struct bt_info *bt, ulong sp,
struct arm64_stackframe *frame)
{
struct arm64_pt_regs *ptregs;
if (IN_TASK_VMA(bt->task, sp)) {
bt->flags |= BT_USER_EFRAME;
return;
}
ptregs = (struct arm64_pt_regs *)
&bt->stackbuf[(ulong)(STACK_OFFSET_TYPE(sp))];
frame->pc = ptregs->pc;
frame->fp = ptregs->regs[29];
frame->sp = ptregs->sp;
}
static void
arm64_print_exception_frame(struct bt_info *bt, ulong pt_regs, int mode, FILE *ofp)
{
int i, r, rows, top_reg, is_64_bit;
struct arm64_pt_regs *regs;
struct syment *sp;
ulong LR, SP, offset;
char buf[BUFSIZE];
struct machine_specific *ms = machdep->machspec;
if (CRASHDEBUG(1))
fprintf(ofp, "pt_regs: %lx\n", pt_regs);
regs = (struct arm64_pt_regs *)
&bt->stackbuf[(ulong)(STACK_OFFSET_TYPE(pt_regs))];
if ((mode == USER_MODE) && (regs->pstate & PSR_MODE32_BIT)) {
LR = regs->regs[14];
SP = regs->regs[13];
top_reg = 12;
is_64_bit = FALSE;
rows = 4;
} else {
LR = regs->regs[30];
if (is_kernel_text (LR | ms->CONFIG_ARM64_KERNELPACMASK))
LR |= ms->CONFIG_ARM64_KERNELPACMASK;
SP = regs->sp;
top_reg = 29;
is_64_bit = TRUE;
rows = 3;
}
switch (mode) {
case USER_MODE:
if (is_64_bit)
fprintf(ofp,
" PC: %016lx LR: %016lx SP: %016lx\n ",
(ulong)regs->pc, LR, SP);
else
fprintf(ofp,
" PC: %08lx LR: %08lx SP: %08lx PSTATE: %08lx\n ",
(ulong)regs->pc, LR, SP, (ulong)regs->pstate);
break;
case KERNEL_MODE:
fprintf(ofp, " PC: %016lx ", (ulong)regs->pc);
if (is_kernel_text(regs->pc) &&
(sp = value_search(regs->pc, &offset))) {
fprintf(ofp, "[%s", sp->name);
if (offset)
fprintf(ofp, (*gdb_output_radix == 16) ?
"+0x%lx" : "+%ld",
offset);
fprintf(ofp, "]\n");
} else
fprintf(ofp, "[unknown or invalid address]\n");
fprintf(ofp, " LR: %016lx ", LR);
if (is_kernel_text(LR) &&
(sp = value_search(LR, &offset))) {
fprintf(ofp, "[%s", sp->name);
if (offset)
fprintf(ofp, (*gdb_output_radix == 16) ?
"+0x%lx" : "+%ld",
offset);
fprintf(ofp, "]\n");
} else
fprintf(ofp, "[unknown or invalid address]\n");
fprintf(ofp, " SP: %016lx PSTATE: %08lx\n ",
SP, (ulong)regs->pstate);
break;
}
for (i = top_reg, r = 1; i >= 0; r++, i--) {
fprintf(ofp, "%sX%d: ",
i < 10 ? " " : "", i);
fprintf(ofp, is_64_bit ? "%016lx" : "%08lx",
(ulong)regs->regs[i]);
if ((i == 0) && !is_64_bit)
fprintf(ofp, "\n");
else if ((i == 0) || ((r % rows) == 0))
fprintf(ofp, "\n%s",
(i == 0) && (mode == KERNEL_MODE) ? "" : " ");
else
fprintf(ofp, "%s", is_64_bit ? " " : " ");
}
if (is_64_bit) {
if (mode == USER_MODE) {
fprintf(ofp, "ORIG_X0: %016lx SYSCALLNO: %lx",
(ulong)regs->orig_x0, (ulong)regs->syscallno);
fprintf(ofp, " PSTATE: %08lx\n", (ulong)regs->pstate);
}
}
if (is_kernel_text(regs->pc) && (bt->flags & BT_LINE_NUMBERS)) {
get_line_number(regs->pc, buf, FALSE);
if (strlen(buf))
fprintf(ofp, " %s\n", buf);
}
if (BT_REFERENCE_CHECK(bt)) {
arm64_do_bt_reference_check(bt, regs->pc, NULL);
if ((sp = value_search(regs->pc, &offset)))
arm64_do_bt_reference_check(bt, 0, sp->name);
arm64_do_bt_reference_check(bt, LR, NULL);
arm64_do_bt_reference_check(bt, SP, NULL);
arm64_do_bt_reference_check(bt, regs->pstate, NULL);
for (i = 0; i <= top_reg; i++)
arm64_do_bt_reference_check(bt, regs->regs[i], NULL);
if (is_64_bit) {
arm64_do_bt_reference_check(bt, regs->orig_x0, NULL);
arm64_do_bt_reference_check(bt, regs->syscallno, NULL);
}
}
}
/*
* Check a frame for a requested reference.
*/
static void
arm64_do_bt_reference_check(struct bt_info *bt, ulong text, char *name)
{
ulong offset;
struct syment *sp = NULL;
if (!name)
sp = value_search(text, &offset);
else if (!text)
sp = symbol_search(name);
switch (bt->ref->cmdflags & (BT_REF_SYMBOL|BT_REF_HEXVAL))
{
case BT_REF_SYMBOL:
if (name) {
if (STREQ(name, bt->ref->str))
bt->ref->cmdflags |= BT_REF_FOUND;
} else {
if (sp && !offset && STREQ(sp->name, bt->ref->str))
bt->ref->cmdflags |= BT_REF_FOUND;
}
break;
case BT_REF_HEXVAL:
if (text) {
if (bt->ref->hexval == text)
bt->ref->cmdflags |= BT_REF_FOUND;
} else if (sp && (bt->ref->hexval == sp->value))
bt->ref->cmdflags |= BT_REF_FOUND;
else if (!name && !text && (bt->ref->hexval == 0))
bt->ref->cmdflags |= BT_REF_FOUND;
break;
}
}
/*
* Translate a PTE, returning TRUE if the page is present.
* If a physaddr pointer is passed in, don't print anything.
*/
static int
arm64_translate_pte(ulong pte, void *physaddr, ulonglong unused)
{
int c, others, len1, len2, len3;
ulong paddr;
char buf1[BUFSIZE];
char buf2[BUFSIZE];
char buf3[BUFSIZE];
char ptebuf[BUFSIZE];
char physbuf[BUFSIZE];
char *arglist[MAXARGS];
int page_present;
paddr = PTE_TO_PHYS(pte);
page_present = pte & (PTE_VALID | machdep->machspec->PTE_PROT_NONE);
if (physaddr) {
*((ulong *)physaddr) = paddr;
return page_present;
}
sprintf(ptebuf, "%lx", pte);
len1 = MAX(strlen(ptebuf), strlen("PTE"));
fprintf(fp, "%s ", mkstring(buf1, len1, CENTER|LJUST, "PTE"));
if (!page_present) {
swap_location(pte, buf1);
if ((c = parse_line(buf1, arglist)) != 3)
error(FATAL, "cannot determine swap location\n");
len2 = MAX(strlen(arglist[0]), strlen("SWAP"));
len3 = MAX(strlen(arglist[2]), strlen("OFFSET"));
fprintf(fp, "%s %s\n",
mkstring(buf2, len2, CENTER|LJUST, "SWAP"),
mkstring(buf3, len3, CENTER|LJUST, "OFFSET"));
strcpy(buf2, arglist[0]);
strcpy(buf3, arglist[2]);
fprintf(fp, "%s %s %s\n",
mkstring(ptebuf, len1, CENTER|RJUST, NULL),
mkstring(buf2, len2, CENTER|RJUST, NULL),
mkstring(buf3, len3, CENTER|RJUST, NULL));
return page_present;
}
sprintf(physbuf, "%lx", paddr);
len2 = MAX(strlen(physbuf), strlen("PHYSICAL"));
fprintf(fp, "%s ", mkstring(buf1, len2, CENTER|LJUST, "PHYSICAL"));
fprintf(fp, "FLAGS\n");
fprintf(fp, "%s %s ",
mkstring(ptebuf, len1, CENTER|RJUST, NULL),
mkstring(physbuf, len2, CENTER|RJUST, NULL));
fprintf(fp, "(");
others = 0;
if (pte) {
if (pte & PTE_VALID)
fprintf(fp, "%sVALID", others++ ? "|" : "");
if (pte & machdep->machspec->PTE_FILE)
fprintf(fp, "%sFILE", others++ ? "|" : "");
if (pte & machdep->machspec->PTE_PROT_NONE)
fprintf(fp, "%sPROT_NONE", others++ ? "|" : "");
if (pte & PTE_USER)
fprintf(fp, "%sUSER", others++ ? "|" : "");
if (pte & PTE_RDONLY)
fprintf(fp, "%sRDONLY", others++ ? "|" : "");
if (pte & PTE_SHARED)
fprintf(fp, "%sSHARED", others++ ? "|" : "");
if (pte & PTE_AF)
fprintf(fp, "%sAF", others++ ? "|" : "");
if (pte & PTE_NG)
fprintf(fp, "%sNG", others++ ? "|" : "");
if (pte & PTE_PXN)
fprintf(fp, "%sPXN", others++ ? "|" : "");
if (pte & PTE_UXN)
fprintf(fp, "%sUXN", others++ ? "|" : "");
if (pte & PTE_DIRTY)
fprintf(fp, "%sDIRTY", others++ ? "|" : "");
if (pte & PTE_SPECIAL)
fprintf(fp, "%sSPECIAL", others++ ? "|" : "");
} else {
fprintf(fp, "no mapping");
}
fprintf(fp, ")\n");
return (page_present);
}
static ulong
arm64_vmalloc_start(void)
{
return machdep->machspec->vmalloc_start_addr;
}
/*
* Not so accurate since thread_info introduction.
*/
static int
arm64_is_task_addr(ulong task)
{
if (tt->flags & THREAD_INFO)
return IS_KVADDR(task);
else
return (IS_KVADDR(task) && (ALIGNED_STACK_OFFSET(task) == 0));
}
static ulong
PLT_veneer_to_kvaddr(ulong value)
{
uint32_t insn;
ulong addr = 0;
int i;
/*
* PLT veneer always looks:
* movn x16, #0x....
* movk x16, #0x...., lsl #16
* movk x16, #0x...., lsl #32
* br x16
*/
for (i = 0; i < 4; i++) {
if (!readmem(value + i * sizeof(insn), KVADDR, &insn,
sizeof(insn), "PLT veneer", RETURN_ON_ERROR)) {
error(WARNING, "cannot read PLT veneer instruction at %lx\n",
value + i * sizeof(insn));
return value;
}
switch (i) {
case 0:
if ((insn & 0xffe0001f) != 0x92800010)
goto not_plt;
addr = ~((ulong)(insn & 0x1fffe0) >> 5);
break;
case 1:
if ((insn & 0xffe0001f) != 0xf2a00010)
goto not_plt;
addr &= 0xffffffff0000ffff;
addr |= (ulong)(insn & 0x1fffe0) << (16 - 5);
break;
case 2:
if ((insn & 0xffe0001f) != 0xf2c00010)
goto not_plt;
addr &= 0xffff0000ffffffff;
addr |= (ulong)(insn & 0x1fffe0) << (32 - 5);
break;
case 3:
if (insn != 0xd61f0200)
goto not_plt;
break;
default:
return value; /* to avoid any warnings */
}
}
return addr;
not_plt:
return value;
}
/*
* Filter dissassembly output if the output radix is not gdb's default 10
*/
static int
arm64_dis_filter(ulong vaddr, char *inbuf, unsigned int output_radix)
{
char buf1[BUFSIZE];
char buf2[BUFSIZE];
char *colon, *p1;
int argc;
char *argv[MAXARGS];
ulong value;
if (!inbuf)
return TRUE;
console("IN: %s", inbuf);
colon = strstr(inbuf, ":");
if (colon) {
sprintf(buf1, "0x%lx <%s>", vaddr,
value_to_symstr(vaddr, buf2, output_radix));
sprintf(buf2, "%s%s", buf1, colon);
strcpy(inbuf, buf2);
}
strcpy(buf1, inbuf);
argc = parse_line(buf1, argv);
if ((FIRSTCHAR(argv[argc-1]) == '<') &&
(LASTCHAR(argv[argc-1]) == '>')) {
p1 = rindex(inbuf, '<');
while ((p1 > inbuf) && !(STRNEQ(p1, " 0x") || STRNEQ(p1, "\t0x")))
p1--;
if (!(STRNEQ(p1, " 0x") || STRNEQ(p1, "\t0x")))
return FALSE;
p1++;
if (!extract_hex(p1, &value, NULLCHAR, TRUE))
return FALSE;
sprintf(buf1, "0x%lx <%s>\n", value,
value_to_symstr(value, buf2, output_radix));
sprintf(p1, "%s", buf1);
}
if (IS_MODULE_VADDR(vaddr)) {
ulong orig_value;
p1 = &inbuf[strlen(inbuf)-1];
strcpy(buf1, inbuf);
argc = parse_line(buf1, argv);
if ((STREQ(argv[argc-2], "b") || STREQ(argv[argc-2], "bl")) &&
extract_hex(argv[argc-1], &orig_value, NULLCHAR, TRUE)) {
value = PLT_veneer_to_kvaddr(orig_value);
sprintf(p1, " <%s%s>\n",
value == orig_value ? "" : "plt:",
value_to_symstr(value, buf2, output_radix));
}
}
console(" %s", inbuf);
return TRUE;
}
/*
* Machine dependent command.
*/
static void
arm64_cmd_mach(void)
{
int c;
while ((c = getopt(argcnt, args, "cm")) != -1) {
switch (c) {
case 'c':
case 'm':
option_not_supported(c);
break;
default:
argerrs++;
break;
}
}
if (argerrs)
cmd_usage(pc->curcmd, SYNOPSIS);
arm64_display_machine_stats();
}
static void
arm64_display_machine_stats(void)
{
int i, pad;
struct new_utsname *uts;
char buf[BUFSIZE];
ulong mhz;
uts = &kt->utsname;
fprintf(fp, " MACHINE TYPE: %s\n", uts->machine);
fprintf(fp, " MEMORY SIZE: %s\n", get_memory_size(buf));
fprintf(fp, " CPUS: %d\n", get_cpus_to_display());
if ((mhz = machdep->processor_speed()))
fprintf(fp, " PROCESSOR SPEED: %ld Mhz\n", mhz);
fprintf(fp, " HZ: %d\n", machdep->hz);
fprintf(fp, " PAGE SIZE: %d\n", PAGESIZE());
fprintf(fp, "KERNEL VIRTUAL BASE: %lx\n", machdep->machspec->page_offset);
fprintf(fp, "KERNEL MODULES BASE: %lx\n", machdep->machspec->modules_vaddr);
fprintf(fp, "KERNEL VMALLOC BASE: %lx\n", machdep->machspec->vmalloc_start_addr);
fprintf(fp, "KERNEL VMEMMAP BASE: %lx\n", machdep->machspec->vmemmap_vaddr);
fprintf(fp, " KERNEL STACK SIZE: %ld\n", STACKSIZE());
if (machdep->machspec->irq_stack_size) {
fprintf(fp, " IRQ STACK SIZE: %ld\n",
machdep->machspec->irq_stack_size);
fprintf(fp, " IRQ STACKS:\n");
for (i = 0; i < kt->cpus; i++) {
pad = (i < 10) ? 3 : (i < 100) ? 2 : (i < 1000) ? 1 : 0;
fprintf(fp, "%s CPU %d: %lx\n", space(pad), i,
machdep->machspec->irq_stacks[i]);
}
}
if (machdep->machspec->overflow_stack_size) {
fprintf(fp, "OVERFLOW STACK SIZE: %ld\n",
machdep->machspec->overflow_stack_size);
fprintf(fp, " OVERFLOW STACKS:\n");
for (i = 0; i < kt->cpus; i++) {
pad = (i < 10) ? 3 : (i < 100) ? 2 : (i < 1000) ? 1 : 0;
fprintf(fp, "%s CPU %d: %lx\n", space(pad), i,
machdep->machspec->overflow_stacks[i]);
}
}
}
static int
arm64_get_smp_cpus(void)
{
int cpus;
if ((cpus = get_cpus_present()))
return cpus;
else
return MAX(get_cpus_online(), get_highest_cpu_online()+1);
}
/*
* Retrieve task registers for the time of the crash.
*/
static void
arm64_get_crash_notes(void)
{
struct machine_specific *ms = machdep->machspec;
ulong crash_notes;
Elf64_Nhdr *note = NULL;
ulong offset;
char *buf, *p;
ulong *notes_ptrs;
ulong i, found;
if (!symbol_exists("crash_notes")) {
if (DISKDUMP_DUMPFILE() || KDUMP_DUMPFILE()) {
if (!(ms->panic_task_regs = calloc((size_t)kt->cpus, sizeof(struct arm64_pt_regs))))
error(FATAL, "cannot calloc panic_task_regs space\n");
for (i = found = 0; i < kt->cpus; i++) {
if (DISKDUMP_DUMPFILE())
note = diskdump_get_prstatus_percpu(i);
else if (KDUMP_DUMPFILE())
note = netdump_get_prstatus_percpu(i);
if (!note) {
error(WARNING, "cpu %d: cannot find NT_PRSTATUS note\n", i);
continue;
}
/*
* Find correct location of note data. This contains elf_prstatus
* structure which has registers etc. for the crashed task.
*/
offset = sizeof(Elf64_Nhdr);
offset = roundup(offset + note->n_namesz, 4);
p = (char *)note + offset; /* start of elf_prstatus */
BCOPY(p + OFFSET(elf_prstatus_pr_reg), &ms->panic_task_regs[i],
sizeof(struct arm64_pt_regs));
found++;
}
if (!found) {
free(ms->panic_task_regs);
ms->panic_task_regs = NULL;
}
}
return;
}
crash_notes = symbol_value("crash_notes");
notes_ptrs = (ulong *)GETBUF(kt->cpus*sizeof(notes_ptrs[0]));
/*
* Read crash_notes for the first CPU. crash_notes are in standard ELF
* note format.
*/
if (!readmem(crash_notes, KVADDR, &notes_ptrs[kt->cpus-1],
sizeof(notes_ptrs[kt->cpus-1]), "crash_notes", RETURN_ON_ERROR)) {
error(WARNING, "cannot read \"crash_notes\"\n");
FREEBUF(notes_ptrs);
return;
}
if (symbol_exists("__per_cpu_offset")) {
/*
* Add __per_cpu_offset for each cpu to form the notes pointer.
*/
for (i = 0; i<kt->cpus; i++)
notes_ptrs[i] = notes_ptrs[kt->cpus-1] + kt->__per_cpu_offset[i];
}
buf = GETBUF(SIZE(note_buf));
if (!(ms->panic_task_regs = calloc((size_t)kt->cpus, sizeof(struct arm64_pt_regs))))
error(FATAL, "cannot calloc panic_task_regs space\n");
for (i = found = 0; i < kt->cpus; i++) {
if (!readmem(notes_ptrs[i], KVADDR, buf, SIZE(note_buf),
"note_buf_t", RETURN_ON_ERROR)) {
error(WARNING, "cpu %d: cannot read NT_PRSTATUS note\n", i);
continue;
}
/*
* Do some sanity checks for this note before reading registers from it.
*/
note = (Elf64_Nhdr *)buf;
p = buf + sizeof(Elf64_Nhdr);
/*
* dumpfiles created with qemu won't have crash_notes, but there will
* be elf notes; dumpfiles created by kdump do not create notes for
* offline cpus.
*/
if (note->n_namesz == 0 && (DISKDUMP_DUMPFILE() || KDUMP_DUMPFILE())) {
if (DISKDUMP_DUMPFILE())
note = diskdump_get_prstatus_percpu(i);
else if (KDUMP_DUMPFILE())
note = netdump_get_prstatus_percpu(i);
if (note) {
/*
* SIZE(note_buf) accounts for a "final note", which is a
* trailing empty elf note header.
*/
long notesz = SIZE(note_buf) - sizeof(Elf64_Nhdr);
if (sizeof(Elf64_Nhdr) + roundup(note->n_namesz, 4) +
note->n_descsz == notesz)
BCOPY((char *)note, buf, notesz);
} else {
error(WARNING, "cpu %d: cannot find NT_PRSTATUS note\n", i);
continue;
}
}
/*
* Check the sanity of NT_PRSTATUS note only for each online cpu.
* If this cpu has invalid note, continue to find the crash notes
* for other online cpus.
*/
if (note->n_type != NT_PRSTATUS) {
error(WARNING, "cpu %d: invalid NT_PRSTATUS note (n_type != NT_PRSTATUS)\n", i);
continue;
}
if (!STRNEQ(p, "CORE")) {
error(WARNING, "cpu %d: invalid NT_PRSTATUS note (name != \"CORE\")\n", i);
continue;
}
/*
* Find correct location of note data. This contains elf_prstatus
* structure which has registers etc. for the crashed task.
*/
offset = sizeof(Elf64_Nhdr);
offset = roundup(offset + note->n_namesz, 4);
p = buf + offset; /* start of elf_prstatus */
BCOPY(p + OFFSET(elf_prstatus_pr_reg), &ms->panic_task_regs[i],
sizeof(struct arm64_pt_regs));
found++;
}
FREEBUF(buf);
FREEBUF(notes_ptrs);
if (!found) {
free(ms->panic_task_regs);
ms->panic_task_regs = NULL;
}
}
static void
arm64_clear_machdep_cache(void) {
/*
* TBD: probably not necessary...
*/
return;
}
static int
arm64_on_process_stack(struct bt_info *bt, ulong stkptr)
{
ulong stackbase, stacktop;
stackbase = GET_STACKBASE(bt->task);
stacktop = GET_STACKTOP(bt->task);
if ((stkptr >= stackbase) && (stkptr < stacktop))
return TRUE;
return FALSE;
}
static int
arm64_in_alternate_stackv(int cpu, ulong stkptr, ulong *stacks, ulong stack_size)
{
if ((cpu >= kt->cpus) || (stacks == NULL) || !stack_size)
return FALSE;
if ((stkptr >= stacks[cpu]) &&
(stkptr < (stacks[cpu] + stack_size)))
return TRUE;
return FALSE;
}
static int
arm64_in_alternate_stack(int cpu, ulong stkptr)
{
return (arm64_on_irq_stack(cpu, stkptr) ||
arm64_on_overflow_stack(cpu, stkptr));
}
static int
arm64_on_irq_stack(int cpu, ulong stkptr)
{
struct machine_specific *ms = machdep->machspec;
return arm64_in_alternate_stackv(cpu, stkptr,
ms->irq_stacks, ms->irq_stack_size);
}
static int
arm64_on_overflow_stack(int cpu, ulong stkptr)
{
struct machine_specific *ms = machdep->machspec;
return arm64_in_alternate_stackv(cpu, stkptr,
ms->overflow_stacks, ms->overflow_stack_size);
}
static void
arm64_set_irq_stack(struct bt_info *bt)
{
struct machine_specific *ms = machdep->machspec;
bt->stackbase = ms->irq_stacks[bt->tc->processor];
bt->stacktop = bt->stackbase + ms->irq_stack_size;
alter_stackbuf(bt);
}
static void
arm64_set_overflow_stack(struct bt_info *bt)
{
struct machine_specific *ms = machdep->machspec;
bt->stackbase = ms->overflow_stacks[bt->tc->processor];
bt->stacktop = bt->stackbase + ms->overflow_stack_size;
alter_stackbuf(bt);
}
static void
arm64_set_process_stack(struct bt_info *bt)
{
bt->stackbase = GET_STACKBASE(bt->task);
bt->stacktop = GET_STACKTOP(bt->task);
alter_stackbuf(bt);
}
static int
compare_kvaddr(const void *v1, const void *v2)
{
struct vaddr_range *r1, *r2;
r1 = (struct vaddr_range *)v1;
r2 = (struct vaddr_range *)v2;
return (r1->start < r2->start ? -1 :
r1->start == r2->start ? 0 : 1);
}
static int
arm64_get_kvaddr_ranges(struct vaddr_range *vrp)
{
int cnt;
cnt = 0;
vrp[cnt].type = KVADDR_UNITY_MAP;
vrp[cnt].start = machdep->machspec->page_offset;
vrp[cnt++].end = vt->high_memory;
vrp[cnt].type = KVADDR_VMALLOC;
vrp[cnt].start = machdep->machspec->vmalloc_start_addr;
vrp[cnt++].end = last_vmalloc_address();
if (st->mods_installed) {
vrp[cnt].type = KVADDR_MODULES;
vrp[cnt].start = lowest_module_address();
vrp[cnt++].end = roundup(highest_module_address(),
PAGESIZE());
}
if (machdep->flags & VMEMMAP) {
vrp[cnt].type = KVADDR_VMEMMAP;
vrp[cnt].start = machdep->machspec->vmemmap_vaddr;
vrp[cnt++].end = vt->node_table[vt->numnodes-1].mem_map +
(vt->node_table[vt->numnodes-1].size * SIZE(page));
}
qsort(vrp, cnt, sizeof(struct vaddr_range), compare_kvaddr);
return cnt;
}
/*
* Include both vmalloc'd, module and vmemmap address space as VMALLOC space.
*/
int
arm64_IS_VMALLOC_ADDR(ulong vaddr)
{
struct machine_specific *ms = machdep->machspec;
if (is_mte_kvaddr(vaddr))
vaddr = mte_tag_reset(vaddr);
if ((machdep->flags & NEW_VMEMMAP) &&
(vaddr >= machdep->machspec->kimage_text) &&
(vaddr <= machdep->machspec->kimage_end))
return FALSE;
if (ms->VA_START && (vaddr >= ms->VA_START))
return TRUE;
return ((vaddr >= ms->vmalloc_start_addr && vaddr <= ms->vmalloc_end) ||
((machdep->flags & VMEMMAP) &&
((vaddr >= ms->vmemmap_vaddr && vaddr <= ms->vmemmap_end) ||
(vaddr >= ms->vmalloc_end && vaddr <= ms->vmemmap_vaddr))) ||
(vaddr >= ms->modules_vaddr && vaddr <= ms->modules_end));
}
/* Return TRUE if we succeed, return FALSE on failure. */
static int
arm64_set_va_bits_by_tcr(void)
{
ulong value;
if (arm64_get_vmcoreinfo(&value, "NUMBER(TCR_EL1_T1SZ)", NUM_HEX) ||
arm64_get_vmcoreinfo(&value, "NUMBER(tcr_el1_t1sz)", NUM_HEX)) {
/* See ARMv8 ARM for the description of
* TCR_EL1.T1SZ and how it can be used
* to calculate the vabits_actual
* supported by underlying kernel.
*
* Basically:
* vabits_actual = 64 - T1SZ;
*/
value = 64 - value;
if (CRASHDEBUG(1))
fprintf(fp, "vmcoreinfo : vabits_actual: %ld\n", value);
machdep->machspec->VA_BITS_ACTUAL = value;
machdep->machspec->VA_BITS = value;
machdep->machspec->VA_START = _VA_START(machdep->machspec->VA_BITS_ACTUAL);
return TRUE;
}
return FALSE;
}
static void
arm64_calc_VA_BITS(void)
{
int bitval;
struct syment *sp;
ulong vabits_actual, value;
arm64_get_vmcoreinfo(&machdep->machspec->CONFIG_ARM64_VA_BITS, "NUMBER(VA_BITS)", NUM_DEC);
if (kernel_symbol_exists("vabits_actual")) {
if (pc->flags & PROC_KCORE) {
vabits_actual = symbol_value_from_proc_kallsyms("vabits_actual");
if ((vabits_actual != BADVAL) && (READMEM(pc->mfd, &value, sizeof(ulong),
vabits_actual, KCORE_USE_VADDR) > 0)) {
if (CRASHDEBUG(1))
fprintf(fp,
"/proc/kcore: vabits_actual: %ld\n", value);
machdep->machspec->VA_BITS_ACTUAL = value;
machdep->machspec->VA_BITS = value;
machdep->machspec->VA_START = _VA_START(machdep->machspec->VA_BITS_ACTUAL);
} else
error(FATAL, "/proc/kcore: cannot read vabits_actual\n");
} else if (ACTIVE())
error(FATAL, "cannot determine VA_BITS_ACTUAL: please use /proc/kcore\n");
else {
if (arm64_set_va_bits_by_tcr()) {
/* nothing */
} else if (machdep->machspec->VA_BITS_ACTUAL) {
machdep->machspec->VA_BITS = machdep->machspec->VA_BITS_ACTUAL;
machdep->machspec->VA_START = _VA_START(machdep->machspec->VA_BITS_ACTUAL);
} else if (machdep->machspec->CONFIG_ARM64_VA_BITS) {
/* guess */
machdep->machspec->VA_BITS_ACTUAL = machdep->machspec->CONFIG_ARM64_VA_BITS;
machdep->machspec->VA_BITS = machdep->machspec->CONFIG_ARM64_VA_BITS;
machdep->machspec->VA_START = _VA_START(machdep->machspec->VA_BITS_ACTUAL);
} else
error(FATAL, "cannot determine VA_BITS_ACTUAL\n");
}
if (machdep->machspec->CONFIG_ARM64_VA_BITS)
machdep->machspec->VA_BITS = machdep->machspec->CONFIG_ARM64_VA_BITS;
/*
* The mm flip commit is introduced before 52-bits VA, which is before the
* commit to export NUMBER(TCR_EL1_T1SZ)
*/
machdep->flags |= FLIPPED_VM;
return;
} else if (arm64_set_va_bits_by_tcr()) {
return;
} else if (machdep->machspec->VA_BITS_ACTUAL) {
machdep->machspec->VA_BITS = machdep->machspec->VA_BITS_ACTUAL;
machdep->machspec->VA_START = _VA_START(machdep->machspec->VA_BITS_ACTUAL);
return;
}
if (!(sp = symbol_search("swapper_pg_dir")) &&
!(sp = symbol_search("idmap_pg_dir")) &&
!(sp = symbol_search("_text")) &&
!(sp = symbol_search("stext"))) {
for (sp = st->symtable; sp < st->symend; sp++) {
if (highest_bit_long(sp->value) == 63)
break;
}
}
if (sp)
value = sp->value;
else
value = kt->vmcoreinfo.log_buf_SYMBOL; /* crash --log */
for (bitval = highest_bit_long(value); bitval; bitval--) {
if ((value & (1UL << bitval)) == 0) {
if (machdep->flags & NEW_VMEMMAP)
machdep->machspec->VA_BITS = bitval + 1;
else
machdep->machspec->VA_BITS = bitval + 2;
break;
}
}
/*
* Verify against dumpfiles that export VA_BITS in vmcoreinfo
*/
if (machdep->machspec->CONFIG_ARM64_VA_BITS &&
(machdep->machspec->VA_BITS != machdep->machspec->CONFIG_ARM64_VA_BITS)) {
error(WARNING, "VA_BITS: calculated: %ld vmcoreinfo: %ld\n",
machdep->machspec->VA_BITS, machdep->machspec->CONFIG_ARM64_VA_BITS);
machdep->machspec->VA_BITS = machdep->machspec->CONFIG_ARM64_VA_BITS;
}
if (CRASHDEBUG(1))
fprintf(fp, "VA_BITS: %ld\n", machdep->machspec->VA_BITS);
}
/*
* The size and end of the vmalloc range is dependent upon the kernel's
* VMEMMAP_SIZE value, and the vmemmap range is dependent upon the end
* of the vmalloc range as well as the VMEMMAP_SIZE:
*
* #define VMEMMAP_SIZE ALIGN((1UL << (VA_BITS - PAGE_SHIFT)) * sizeof(struct page), PUD_SIZE)
* #define VMALLOC_START (UL(0xffffffffffffffff) << VA_BITS)
* #define VMALLOC_END (PAGE_OFFSET - PUD_SIZE - VMEMMAP_SIZE - SZ_64K)
*
* Since VMEMMAP_SIZE is dependent upon the size of a struct page,
* the two ranges cannot be determined until POST_GDB.
*
* Since 52-bit VA was introduced:
*
* #define STRUCT_PAGE_MAX_SHIFT 6
* #define VMEMMAP_SIZE (UL(1) << (VA_BITS - PAGE_SHIFT - 1 + STRUCT_PAGE_MAX_SHIFT))
* #define VMEMMAP_START (-VMEMMAP_SIZE)
* #define VMALLOC_START (MODULES_END)
* #define VMALLOC_END (- PUD_SIZE - VMEMMAP_SIZE - SZ_64K)
* #define vmemmap ((struct page *)VMEMMAP_START - (memstart_addr >> PAGE_SHIFT))
*/
#define ALIGN(x, a) __ALIGN_KERNEL((x), (a))
#define __ALIGN_KERNEL(x, a) __ALIGN_KERNEL_MASK(x, (typeof(x))(a) - 1)
#define __ALIGN_KERNEL_MASK(x, mask) (((x) + (mask)) & ~(mask))
static void
arm64_calc_virtual_memory_ranges(void)
{
struct machine_specific *ms = machdep->machspec;
ulong value, vmemmap_start, vmemmap_end, vmemmap_size, vmalloc_end;
char *string;
int ret;
ulong PUD_SIZE = UNINITIALIZED;
if (!machdep->machspec->CONFIG_ARM64_VA_BITS) {
if (arm64_get_vmcoreinfo(&value, "NUMBER(VA_BITS)", NUM_DEC)) {
machdep->machspec->CONFIG_ARM64_VA_BITS = value;
} else if (kt->ikconfig_flags & IKCONFIG_AVAIL) {
if ((ret = get_kernel_config("CONFIG_ARM64_VA_BITS",
&string)) == IKCONFIG_STR)
machdep->machspec->CONFIG_ARM64_VA_BITS = atol(string);
}
}
if (THIS_KERNEL_VERSION < LINUX(3,17,0)) /* use original hardwired values */
return;
STRUCT_SIZE_INIT(page, "page");
ms->struct_page_size = SIZE(page);
switch (machdep->flags & (VM_L2_64K|VM_L3_64K|VM_L3_4K|VM_L4_4K))
{
case VM_L2_64K:
case VM_L3_64K:
PUD_SIZE = PGDIR_SIZE_L2_64K;
break;
case VM_L3_4K:
PUD_SIZE = PGDIR_SIZE_L3_4K;
case VM_L4_4K:
PUD_SIZE = PUD_SIZE_L4_4K;
break;
}
#define STRUCT_PAGE_MAX_SHIFT 6
if (ms->VA_BITS_ACTUAL) {
ulong va_bits_min = 48;
if (machdep->machspec->CONFIG_ARM64_VA_BITS < 48)
va_bits_min = ms->CONFIG_ARM64_VA_BITS;
vmemmap_size = (1UL) << (va_bits_min - machdep->pageshift - 1 + STRUCT_PAGE_MAX_SHIFT);
vmalloc_end = (- PUD_SIZE - vmemmap_size - KILOBYTES(64));
vmemmap_start = (-vmemmap_size - MEGABYTES(2));
ms->vmalloc_end = vmalloc_end - 1;
ms->vmemmap_vaddr = vmemmap_start;
ms->vmemmap_end = vmemmap_start + vmemmap_size;
return;
}
if (machdep->flags & NEW_VMEMMAP)
vmemmap_size = 1UL << (ms->VA_BITS - machdep->pageshift - 1
+ STRUCT_PAGE_MAX_SHIFT);
else
vmemmap_size = ALIGN((1UL << (ms->VA_BITS - machdep->pageshift)) * SIZE(page), PUD_SIZE);
vmalloc_end = (ms->page_offset - PUD_SIZE - vmemmap_size - SZ_64K);
if (machdep->flags & NEW_VMEMMAP) {
vmemmap_start = ms->page_offset - vmemmap_size;
vmemmap_end = ms->page_offset;
} else {
vmemmap_start = vmalloc_end + SZ_64K;
vmemmap_end = vmemmap_start + vmemmap_size;
}
ms->vmalloc_end = vmalloc_end - 1;
ms->vmemmap_vaddr = vmemmap_start;
ms->vmemmap_end = vmemmap_end - 1;
}
static int
arm64_is_uvaddr(ulong addr, struct task_context *tc)
{
return (addr < machdep->machspec->userspace_top);
}
ulong
arm64_swp_type(ulong pte)
{
struct machine_specific *ms = machdep->machspec;
pte >>= ms->__SWP_TYPE_SHIFT;
pte &= ms->__SWP_TYPE_MASK;
return pte;
}
ulong
arm64_swp_offset(ulong pte)
{
struct machine_specific *ms = machdep->machspec;
pte >>= ms->__SWP_OFFSET_SHIFT;
if (ms->__SWP_OFFSET_MASK)
pte &= ms->__SWP_OFFSET_MASK;
return pte;
}
static void arm64_calc_KERNELPACMASK(void)
{
ulong value;
if (arm64_get_vmcoreinfo(&value, "NUMBER(KERNELPACMASK)", NUM_HEX)) {
machdep->machspec->CONFIG_ARM64_KERNELPACMASK = value;
if (CRASHDEBUG(1))
fprintf(fp, "CONFIG_ARM64_KERNELPACMASK: %lx\n", value);
}
}
#define GENMASK_UL(h, l) \
(((~0UL) << (l)) & (~0UL >> (BITS_PER_LONG - 1 - (h))))
static void arm64_recalc_KERNELPACMASK(void){
/*
* Check if PAC is enabled according to the existence of
* kernel symbol 'ptrauth_keys_kernel'.
*/
if (STRUCT_EXISTS("ptrauth_keys_kernel") &&
machdep->machspec->VA_BITS_ACTUAL){
machdep->machspec->CONFIG_ARM64_KERNELPACMASK =
GENMASK_UL(63, machdep->machspec->VA_BITS_ACTUAL);
if (CRASHDEBUG(1))
fprintf(fp, "CONFIG_ARM64_KERNELPACMASK: %lx\n",
machdep->machspec->CONFIG_ARM64_KERNELPACMASK);
}
}
static ulong arm64_set_irq_stack_size(void)
{
int min_thread_shift = 14;
ulong thread_shift = 0;
char buf1[BUFSIZE];
char *pos1, *pos2;
int errflag = 0;
if (kernel_symbol_exists("vmcoreinfo_data") &&
kernel_symbol_exists("vmcoreinfo_size")) {
/*
* Referring to arch/arm64/include/asm/memory.h
*/
if (kernel_symbol_exists("kasan_enable_current"))
min_thread_shift += 1;
if (MEMBER_EXISTS("task_struct", "stack_vm_area") &&
(min_thread_shift < machdep->pageshift))
thread_shift = machdep->pageshift;
else
thread_shift = min_thread_shift;
} else {
sprintf(buf1, "x/32i vectors");
open_tmpfile();
if (!gdb_pass_through(buf1, pc->tmpfile, GNU_RETURN_ON_ERROR))
goto out;
rewind(pc->tmpfile);
while (fgets(buf1, BUFSIZE, pc->tmpfile)) {
if ((pos1 = strstr(buf1, "tbnz"))) {
if ((pos2 = strchr(pos1, '#'))) {
pos2 += 1;
for (pos1 = pos2; *pos2 != '\0' && *pos2 != ','; pos2++);
*pos2 = '\0';
thread_shift = stol(pos1, RETURN_ON_ERROR|QUIET, &errflag);
if (errflag)
thread_shift = 0;
break;
}
}
}
out:
close_tmpfile();
}
if (thread_shift)
return ((1UL) << thread_shift);
return 0;
}
#endif /* ARM64 */