forked from Mirrors/opensbi

This code has nothing to do with the ISA's registers, it's about the format of ELF relocations. As such, __SIZEOF_LONG__, being a language / ABI-level property, is a more appropriate constant to use. This also makes it easier to support CHERI, where general-purpose registers are extended to be capabilities, not just integers, and so the register size is not the same as the machine word size. This also happens to make it more correct for RV64ILP32, where the registers are 64-bit integers but the ABI is 32-bit (both for long and for the ELF format), though properly supporting that ABI is not part of the motivation here, just a consequence of improving the code for CHERI. Signed-off-by: Jessica Clarke <jrtc27@jrtc27.com> Reviewed-by: Anup Patel <anup@brainfault.org> Link: https://lore.kernel.org/r/20250709232932.37622-2-jrtc27@jrtc27.com Signed-off-by: Anup Patel <anup@brainfault.org>
770 lines
18 KiB
ArmAsm
770 lines
18 KiB
ArmAsm
/*
|
|
* SPDX-License-Identifier: BSD-2-Clause
|
|
*
|
|
* Copyright (c) 2019 Western Digital Corporation or its affiliates.
|
|
*
|
|
* Authors:
|
|
* Anup Patel <anup.patel@wdc.com>
|
|
*/
|
|
|
|
#include <sbi/riscv_asm.h>
|
|
#include <sbi/riscv_encoding.h>
|
|
#include <sbi/riscv_elf.h>
|
|
#include <sbi/sbi_platform.h>
|
|
#include <sbi/sbi_scratch.h>
|
|
#include <sbi/sbi_trap.h>
|
|
|
|
#define BOOT_LOTTERY_ACQUIRED 1
|
|
#define BOOT_STATUS_BOOT_HART_DONE 1
|
|
|
|
.macro MOV_3R __d0, __s0, __d1, __s1, __d2, __s2
|
|
add \__d0, \__s0, zero
|
|
add \__d1, \__s1, zero
|
|
add \__d2, \__s2, zero
|
|
.endm
|
|
|
|
.macro MOV_5R __d0, __s0, __d1, __s1, __d2, __s2, __d3, __s3, __d4, __s4
|
|
add \__d0, \__s0, zero
|
|
add \__d1, \__s1, zero
|
|
add \__d2, \__s2, zero
|
|
add \__d3, \__s3, zero
|
|
add \__d4, \__s4, zero
|
|
.endm
|
|
|
|
.macro CLEAR_MDT tmp
|
|
#if __riscv_xlen == 32
|
|
li \tmp, MSTATUSH_MDT
|
|
csrc CSR_MSTATUSH, \tmp
|
|
#else
|
|
li \tmp, MSTATUS_MDT
|
|
csrc CSR_MSTATUS, \tmp
|
|
#endif
|
|
.endm
|
|
|
|
.section .entry, "ax", %progbits
|
|
.align 3
|
|
.globl _start
|
|
.globl _start_warm
|
|
_start:
|
|
/* Find preferred boot HART id */
|
|
MOV_3R s0, a0, s1, a1, s2, a2
|
|
call fw_boot_hart
|
|
add a6, a0, zero
|
|
MOV_3R a0, s0, a1, s1, a2, s2
|
|
li a7, -1
|
|
beq a6, a7, _try_lottery
|
|
/* Jump to relocation wait loop if we are not boot hart */
|
|
bne a0, a6, _wait_for_boot_hart
|
|
_try_lottery:
|
|
/* Jump to relocation wait loop if we don't get relocation lottery */
|
|
lla a6, _boot_lottery
|
|
li a7, BOOT_LOTTERY_ACQUIRED
|
|
#ifdef __riscv_atomic
|
|
amoswap.w a6, a7, (a6)
|
|
bnez a6, _wait_for_boot_hart
|
|
#elif __riscv_zalrsc
|
|
_sc_fail:
|
|
lr.w t0, (a6)
|
|
sc.w t1, a7, (a6)
|
|
bnez t1, _sc_fail
|
|
bnez t0, _wait_for_boot_hart
|
|
#else
|
|
#error "need a or zalrsc"
|
|
#endif
|
|
|
|
/* relocate the global table content */
|
|
li t0, FW_TEXT_START /* link start */
|
|
lla t1, _fw_start /* load start */
|
|
sub t2, t1, t0 /* load offset */
|
|
lla t0, __rela_dyn_start
|
|
lla t1, __rela_dyn_end
|
|
beq t0, t1, _relocate_done
|
|
2:
|
|
REG_L t5, __SIZEOF_LONG__(t0) /* t5 <-- relocation info:type */
|
|
li t3, R_RISCV_RELATIVE /* reloc type R_RISCV_RELATIVE */
|
|
bne t5, t3, 3f
|
|
REG_L t3, 0(t0)
|
|
REG_L t5, (__SIZEOF_LONG__ * 2)(t0) /* t5 <-- addend */
|
|
add t5, t5, t2
|
|
add t3, t3, t2
|
|
REG_S t5, 0(t3) /* store runtime address to the GOT entry */
|
|
|
|
3:
|
|
addi t0, t0, (__SIZEOF_LONG__ * 3)
|
|
blt t0, t1, 2b
|
|
_relocate_done:
|
|
/* At this point we are running from link address */
|
|
|
|
/* Reset all registers except ra, a0, a1, a2, a3 and a4 for boot HART */
|
|
li ra, 0
|
|
call _reset_regs
|
|
|
|
/* Zero-out BSS */
|
|
lla s4, _bss_start
|
|
lla s5, _bss_end
|
|
_bss_zero:
|
|
REG_S zero, (s4)
|
|
add s4, s4, __SIZEOF_POINTER__
|
|
blt s4, s5, _bss_zero
|
|
|
|
/* Setup temporary trap handler */
|
|
lla s4, _start_hang
|
|
csrw CSR_MTVEC, s4
|
|
|
|
/*
|
|
* While at this point, trap handling is rudimentary, if a trap happens,
|
|
* it will end up in _start_hang which is enough to hook up a GDB. Clear
|
|
* MDT to avoid generating a double trap and thus entering a
|
|
* critical-error state.
|
|
*/
|
|
CLEAR_MDT t0
|
|
|
|
/* Setup temporary stack */
|
|
lla s4, _fw_end
|
|
li s5, (SBI_SCRATCH_SIZE * 2)
|
|
add sp, s4, s5
|
|
|
|
/* Allow main firmware to save info */
|
|
MOV_5R s0, a0, s1, a1, s2, a2, s3, a3, s4, a4
|
|
call fw_save_info
|
|
MOV_5R a0, s0, a1, s1, a2, s2, a3, s3, a4, s4
|
|
|
|
#ifdef FW_FDT_PATH
|
|
/* Override previous arg1 */
|
|
lla a1, fw_fdt_bin
|
|
#endif
|
|
|
|
/*
|
|
* Initialize platform
|
|
* Note: The a0 to a4 registers passed to the
|
|
* firmware are parameters to this function.
|
|
*/
|
|
MOV_5R s0, a0, s1, a1, s2, a2, s3, a3, s4, a4
|
|
call fw_platform_init
|
|
add t0, a0, zero
|
|
MOV_5R a0, s0, a1, s1, a2, s2, a3, s3, a4, s4
|
|
add a1, t0, zero
|
|
|
|
/* Preload HART details
|
|
* s7 -> HART Count
|
|
* s8 -> HART Stack Size
|
|
* s9 -> Heap Size
|
|
* s10 -> Heap Offset
|
|
*/
|
|
lla a4, platform
|
|
#if __riscv_xlen > 32
|
|
lwu s7, SBI_PLATFORM_HART_COUNT_OFFSET(a4)
|
|
lwu s8, SBI_PLATFORM_HART_STACK_SIZE_OFFSET(a4)
|
|
lwu s9, SBI_PLATFORM_HEAP_SIZE_OFFSET(a4)
|
|
#else
|
|
lw s7, SBI_PLATFORM_HART_COUNT_OFFSET(a4)
|
|
lw s8, SBI_PLATFORM_HART_STACK_SIZE_OFFSET(a4)
|
|
lw s9, SBI_PLATFORM_HEAP_SIZE_OFFSET(a4)
|
|
#endif
|
|
|
|
/* Setup scratch space for all the HARTs*/
|
|
lla tp, _fw_end
|
|
mul a5, s7, s8
|
|
add tp, tp, a5
|
|
/* Setup heap base address */
|
|
lla s10, _fw_start
|
|
sub s10, tp, s10
|
|
add tp, tp, s9
|
|
/* Keep a copy of tp */
|
|
add t3, tp, zero
|
|
/* Counter */
|
|
li t2, 1
|
|
/* hartid 0 is mandated by ISA */
|
|
li t1, 0
|
|
_scratch_init:
|
|
/*
|
|
* The following registers hold values that are computed before
|
|
* entering this block, and should remain unchanged.
|
|
*
|
|
* t3 -> the firmware end address
|
|
* s7 -> HART count
|
|
* s8 -> HART stack size
|
|
* s9 -> Heap Size
|
|
* s10 -> Heap Offset
|
|
*/
|
|
add tp, t3, zero
|
|
sub tp, tp, s9
|
|
mul a5, s8, t1
|
|
sub tp, tp, a5
|
|
li a5, SBI_SCRATCH_SIZE
|
|
sub tp, tp, a5
|
|
|
|
/* Initialize scratch space */
|
|
/* Store fw_start and fw_size in scratch space */
|
|
lla a4, _fw_start
|
|
sub a5, t3, a4
|
|
REG_S a4, SBI_SCRATCH_FW_START_OFFSET(tp)
|
|
REG_S a5, SBI_SCRATCH_FW_SIZE_OFFSET(tp)
|
|
|
|
/* Store R/W section's offset in scratch space */
|
|
lla a5, _fw_rw_start
|
|
sub a5, a5, a4
|
|
REG_S a5, SBI_SCRATCH_FW_RW_OFFSET(tp)
|
|
|
|
/* Store fw_heap_offset and fw_heap_size in scratch space */
|
|
REG_S s10, SBI_SCRATCH_FW_HEAP_OFFSET(tp)
|
|
REG_S s9, SBI_SCRATCH_FW_HEAP_SIZE_OFFSET(tp)
|
|
|
|
/* Store next arg1 in scratch space */
|
|
MOV_3R s0, a0, s1, a1, s2, a2
|
|
call fw_next_arg1
|
|
REG_S a0, SBI_SCRATCH_NEXT_ARG1_OFFSET(tp)
|
|
MOV_3R a0, s0, a1, s1, a2, s2
|
|
/* Store next address in scratch space */
|
|
MOV_3R s0, a0, s1, a1, s2, a2
|
|
call fw_next_addr
|
|
REG_S a0, SBI_SCRATCH_NEXT_ADDR_OFFSET(tp)
|
|
MOV_3R a0, s0, a1, s1, a2, s2
|
|
/* Store next mode in scratch space */
|
|
MOV_3R s0, a0, s1, a1, s2, a2
|
|
call fw_next_mode
|
|
REG_S a0, SBI_SCRATCH_NEXT_MODE_OFFSET(tp)
|
|
MOV_3R a0, s0, a1, s1, a2, s2
|
|
/* Store warm_boot address in scratch space */
|
|
lla a4, _start_warm
|
|
REG_S a4, SBI_SCRATCH_WARMBOOT_ADDR_OFFSET(tp)
|
|
/* Store platform address in scratch space */
|
|
lla a4, platform
|
|
REG_S a4, SBI_SCRATCH_PLATFORM_ADDR_OFFSET(tp)
|
|
/* Store hartid-to-scratch function address in scratch space */
|
|
lla a4, _hartid_to_scratch
|
|
REG_S a4, SBI_SCRATCH_HARTID_TO_SCRATCH_OFFSET(tp)
|
|
/* Clear trap_context and tmp0 in scratch space */
|
|
REG_S zero, SBI_SCRATCH_TRAP_CONTEXT_OFFSET(tp)
|
|
REG_S zero, SBI_SCRATCH_TMP0_OFFSET(tp)
|
|
/* Store firmware options in scratch space */
|
|
MOV_3R s0, a0, s1, a1, s2, a2
|
|
#ifdef FW_OPTIONS
|
|
li a0, FW_OPTIONS
|
|
#else
|
|
call fw_options
|
|
#endif
|
|
REG_S a0, SBI_SCRATCH_OPTIONS_OFFSET(tp)
|
|
MOV_3R a0, s0, a1, s1, a2, s2
|
|
/* Store hart index in scratch space */
|
|
REG_S t1, SBI_SCRATCH_HARTINDEX_OFFSET(tp)
|
|
/* Move to next scratch space */
|
|
add t1, t1, t2
|
|
blt t1, s7, _scratch_init
|
|
|
|
/*
|
|
* Relocate Flatened Device Tree (FDT)
|
|
* source FDT address = previous arg1
|
|
* destination FDT address = next arg1
|
|
*
|
|
* Note: We will preserve a0 and a1 passed by
|
|
* previous booting stage.
|
|
*/
|
|
beqz a1, _fdt_reloc_done
|
|
/* Mask values in a4 */
|
|
li a4, 0xff
|
|
/* t1 = destination FDT start address */
|
|
MOV_3R s0, a0, s1, a1, s2, a2
|
|
call fw_next_arg1
|
|
add t1, a0, zero
|
|
MOV_3R a0, s0, a1, s1, a2, s2
|
|
beqz t1, _fdt_reloc_done
|
|
beq t1, a1, _fdt_reloc_done
|
|
/* t0 = source FDT start address */
|
|
add t0, a1, zero
|
|
/* t2 = source FDT size (convert from big-endian) */
|
|
lbu t2, 7(t0)
|
|
lbu t3, 6(t0)
|
|
lbu t4, 5(t0)
|
|
lbu t5, 4(t0)
|
|
slli t3, t3, 8
|
|
slli t4, t4, 16
|
|
slli t5, t5, 24
|
|
or t2, t2, t3
|
|
or t2, t2, t4
|
|
or t2, t2, t5
|
|
/* t2 = destination FDT end address */
|
|
add t2, t1, t2
|
|
/* FDT copy loop */
|
|
ble t2, t1, _fdt_reloc_done
|
|
_fdt_reloc_again:
|
|
REG_L t3, 0(t0)
|
|
REG_S t3, 0(t1)
|
|
add t0, t0, __SIZEOF_POINTER__
|
|
add t1, t1, __SIZEOF_POINTER__
|
|
blt t1, t2, _fdt_reloc_again
|
|
_fdt_reloc_done:
|
|
|
|
/* mark boot hart done */
|
|
li t0, BOOT_STATUS_BOOT_HART_DONE
|
|
lla t1, _boot_status
|
|
fence rw, rw
|
|
REG_S t0, 0(t1)
|
|
j _start_warm
|
|
|
|
/* waiting for boot hart to be done (_boot_status == BOOT_STATUS_BOOT_HART_DONE) */
|
|
_wait_for_boot_hart:
|
|
li t0, BOOT_STATUS_BOOT_HART_DONE
|
|
lla t1, _boot_status
|
|
REG_L t1, 0(t1)
|
|
/* Reduce the bus traffic so that boot hart may proceed faster */
|
|
div t2, t2, zero
|
|
div t2, t2, zero
|
|
div t2, t2, zero
|
|
bne t0, t1, _wait_for_boot_hart
|
|
|
|
_start_warm:
|
|
/* Reset all registers except ra, a0, a1, a2, a3 and a4 for non-boot HART */
|
|
li ra, 0
|
|
call _reset_regs
|
|
|
|
/* Disable all interrupts */
|
|
csrw CSR_MIE, zero
|
|
|
|
/* Find HART count and HART stack size */
|
|
lla a4, platform
|
|
#if __riscv_xlen > 32
|
|
lwu s7, SBI_PLATFORM_HART_COUNT_OFFSET(a4)
|
|
lwu s8, SBI_PLATFORM_HART_STACK_SIZE_OFFSET(a4)
|
|
#else
|
|
lw s7, SBI_PLATFORM_HART_COUNT_OFFSET(a4)
|
|
lw s8, SBI_PLATFORM_HART_STACK_SIZE_OFFSET(a4)
|
|
#endif
|
|
REG_L s9, SBI_PLATFORM_HART_INDEX2ID_OFFSET(a4)
|
|
|
|
/* Find HART id */
|
|
csrr s6, CSR_MHARTID
|
|
|
|
/* Find HART index */
|
|
beqz s9, 3f
|
|
li a4, 0
|
|
1:
|
|
#if __riscv_xlen > 32
|
|
lwu a5, (s9)
|
|
#else
|
|
lw a5, (s9)
|
|
#endif
|
|
beq a5, s6, 2f
|
|
add s9, s9, 4
|
|
add a4, a4, 1
|
|
blt a4, s7, 1b
|
|
2: add s6, a4, zero
|
|
3: bge s6, s7, _start_hang
|
|
|
|
/* Find the scratch space based on HART index */
|
|
lla tp, _fw_end
|
|
mul a5, s7, s8
|
|
add tp, tp, a5
|
|
mul a5, s8, s6
|
|
sub tp, tp, a5
|
|
li a5, SBI_SCRATCH_SIZE
|
|
sub tp, tp, a5
|
|
|
|
/* update the mscratch */
|
|
csrw CSR_MSCRATCH, tp
|
|
|
|
/* Setup stack */
|
|
add sp, tp, zero
|
|
|
|
/* Setup trap handler */
|
|
lla a4, _trap_handler
|
|
csrr a5, CSR_MISA
|
|
srli a5, a5, ('H' - 'A')
|
|
andi a5, a5, 0x1
|
|
beq a5, zero, _skip_trap_handler_hyp
|
|
lla a4, _trap_handler_hyp
|
|
_skip_trap_handler_hyp:
|
|
csrw CSR_MTVEC, a4
|
|
|
|
/* Clear MDT here again for all harts */
|
|
CLEAR_MDT t0
|
|
|
|
/* Initialize SBI runtime */
|
|
csrr a0, CSR_MSCRATCH
|
|
call sbi_init
|
|
|
|
/* We don't expect to reach here hence just hang */
|
|
j _start_hang
|
|
|
|
.data
|
|
.align 3
|
|
_boot_lottery:
|
|
RISCV_PTR 0
|
|
_boot_status:
|
|
RISCV_PTR 0
|
|
|
|
.section .entry, "ax", %progbits
|
|
.align 3
|
|
.globl _hartid_to_scratch
|
|
_hartid_to_scratch:
|
|
/*
|
|
* a0 -> HART ID (passed by caller)
|
|
* a1 -> HART Index (passed by caller)
|
|
* t0 -> HART Stack Size
|
|
* t1 -> HART Stack End
|
|
* t2 -> Temporary
|
|
*/
|
|
lla t2, platform
|
|
#if __riscv_xlen > 32
|
|
lwu t0, SBI_PLATFORM_HART_STACK_SIZE_OFFSET(t2)
|
|
lwu t2, SBI_PLATFORM_HART_COUNT_OFFSET(t2)
|
|
#else
|
|
lw t0, SBI_PLATFORM_HART_STACK_SIZE_OFFSET(t2)
|
|
lw t2, SBI_PLATFORM_HART_COUNT_OFFSET(t2)
|
|
#endif
|
|
sub t2, t2, a1
|
|
mul t2, t2, t0
|
|
lla t1, _fw_end
|
|
add t1, t1, t2
|
|
li t2, SBI_SCRATCH_SIZE
|
|
sub a0, t1, t2
|
|
ret
|
|
|
|
.section .entry, "ax", %progbits
|
|
.align 3
|
|
.globl _start_hang
|
|
_start_hang:
|
|
wfi
|
|
j _start_hang
|
|
|
|
.section .entry, "ax", %progbits
|
|
.align 3
|
|
.weak fw_platform_init
|
|
fw_platform_init:
|
|
add a0, a1, zero
|
|
ret
|
|
|
|
/* Map implicit memcpy() added by compiler to sbi_memcpy() */
|
|
.section .text
|
|
.align 3
|
|
.globl memcpy
|
|
memcpy:
|
|
tail sbi_memcpy
|
|
|
|
/* Map implicit memset() added by compiler to sbi_memset() */
|
|
.section .text
|
|
.align 3
|
|
.globl memset
|
|
memset:
|
|
tail sbi_memset
|
|
|
|
/* Map implicit memmove() added by compiler to sbi_memmove() */
|
|
.section .text
|
|
.align 3
|
|
.globl memmove
|
|
memmove:
|
|
tail sbi_memmove
|
|
|
|
/* Map implicit memcmp() added by compiler to sbi_memcmp() */
|
|
.section .text
|
|
.align 3
|
|
.globl memcmp
|
|
memcmp:
|
|
tail sbi_memcmp
|
|
|
|
.macro TRAP_SAVE_AND_SETUP_SP_T0
|
|
/* Swap TP and MSCRATCH */
|
|
csrrw tp, CSR_MSCRATCH, tp
|
|
|
|
/* Save T0 in scratch space */
|
|
REG_S t0, SBI_SCRATCH_TMP0_OFFSET(tp)
|
|
|
|
/*
|
|
* Set T0 to appropriate exception stack
|
|
*
|
|
* Came_From_M_Mode = ((MSTATUS.MPP < PRV_M) ? 1 : 0) - 1;
|
|
* Exception_Stack = TP ^ (Came_From_M_Mode & (SP ^ TP))
|
|
*
|
|
* Came_From_M_Mode = 0 ==> Exception_Stack = TP
|
|
* Came_From_M_Mode = -1 ==> Exception_Stack = SP
|
|
*/
|
|
csrr t0, CSR_MSTATUS
|
|
srl t0, t0, MSTATUS_MPP_SHIFT
|
|
and t0, t0, PRV_M
|
|
slti t0, t0, PRV_M
|
|
add t0, t0, -1
|
|
xor sp, sp, tp
|
|
and t0, t0, sp
|
|
xor sp, sp, tp
|
|
xor t0, tp, t0
|
|
|
|
/* Save original SP on exception stack */
|
|
REG_S sp, (SBI_TRAP_REGS_OFFSET(sp) - SBI_TRAP_CONTEXT_SIZE)(t0)
|
|
|
|
/* Set SP to exception stack and make room for trap context */
|
|
add sp, t0, -(SBI_TRAP_CONTEXT_SIZE)
|
|
|
|
/* Restore T0 from scratch space */
|
|
REG_L t0, SBI_SCRATCH_TMP0_OFFSET(tp)
|
|
|
|
/* Save T0 on stack */
|
|
REG_S t0, SBI_TRAP_REGS_OFFSET(t0)(sp)
|
|
|
|
/* Swap TP and MSCRATCH */
|
|
csrrw tp, CSR_MSCRATCH, tp
|
|
.endm
|
|
|
|
.macro TRAP_SAVE_MEPC_MSTATUS have_mstatush
|
|
/* Save MEPC and MSTATUS CSRs */
|
|
csrr t0, CSR_MEPC
|
|
REG_S t0, SBI_TRAP_REGS_OFFSET(mepc)(sp)
|
|
csrr t0, CSR_MSTATUS
|
|
REG_S t0, SBI_TRAP_REGS_OFFSET(mstatus)(sp)
|
|
.if \have_mstatush
|
|
csrr t0, CSR_MSTATUSH
|
|
REG_S t0, SBI_TRAP_REGS_OFFSET(mstatusH)(sp)
|
|
.else
|
|
REG_S zero, SBI_TRAP_REGS_OFFSET(mstatusH)(sp)
|
|
.endif
|
|
.endm
|
|
|
|
.macro TRAP_SAVE_GENERAL_REGS_EXCEPT_SP_T0
|
|
/* Save all general regisers except SP and T0 */
|
|
REG_S zero, SBI_TRAP_REGS_OFFSET(zero)(sp)
|
|
REG_S ra, SBI_TRAP_REGS_OFFSET(ra)(sp)
|
|
REG_S gp, SBI_TRAP_REGS_OFFSET(gp)(sp)
|
|
REG_S tp, SBI_TRAP_REGS_OFFSET(tp)(sp)
|
|
REG_S t1, SBI_TRAP_REGS_OFFSET(t1)(sp)
|
|
REG_S t2, SBI_TRAP_REGS_OFFSET(t2)(sp)
|
|
REG_S s0, SBI_TRAP_REGS_OFFSET(s0)(sp)
|
|
REG_S s1, SBI_TRAP_REGS_OFFSET(s1)(sp)
|
|
REG_S a0, SBI_TRAP_REGS_OFFSET(a0)(sp)
|
|
REG_S a1, SBI_TRAP_REGS_OFFSET(a1)(sp)
|
|
REG_S a2, SBI_TRAP_REGS_OFFSET(a2)(sp)
|
|
REG_S a3, SBI_TRAP_REGS_OFFSET(a3)(sp)
|
|
REG_S a4, SBI_TRAP_REGS_OFFSET(a4)(sp)
|
|
REG_S a5, SBI_TRAP_REGS_OFFSET(a5)(sp)
|
|
REG_S a6, SBI_TRAP_REGS_OFFSET(a6)(sp)
|
|
REG_S a7, SBI_TRAP_REGS_OFFSET(a7)(sp)
|
|
REG_S s2, SBI_TRAP_REGS_OFFSET(s2)(sp)
|
|
REG_S s3, SBI_TRAP_REGS_OFFSET(s3)(sp)
|
|
REG_S s4, SBI_TRAP_REGS_OFFSET(s4)(sp)
|
|
REG_S s5, SBI_TRAP_REGS_OFFSET(s5)(sp)
|
|
REG_S s6, SBI_TRAP_REGS_OFFSET(s6)(sp)
|
|
REG_S s7, SBI_TRAP_REGS_OFFSET(s7)(sp)
|
|
REG_S s8, SBI_TRAP_REGS_OFFSET(s8)(sp)
|
|
REG_S s9, SBI_TRAP_REGS_OFFSET(s9)(sp)
|
|
REG_S s10, SBI_TRAP_REGS_OFFSET(s10)(sp)
|
|
REG_S s11, SBI_TRAP_REGS_OFFSET(s11)(sp)
|
|
REG_S t3, SBI_TRAP_REGS_OFFSET(t3)(sp)
|
|
REG_S t4, SBI_TRAP_REGS_OFFSET(t4)(sp)
|
|
REG_S t5, SBI_TRAP_REGS_OFFSET(t5)(sp)
|
|
REG_S t6, SBI_TRAP_REGS_OFFSET(t6)(sp)
|
|
.endm
|
|
|
|
.macro TRAP_SAVE_INFO have_mstatush have_h_extension
|
|
csrr t0, CSR_MCAUSE
|
|
REG_S t0, (SBI_TRAP_REGS_SIZE + SBI_TRAP_INFO_OFFSET(cause))(sp)
|
|
csrr t0, CSR_MTVAL
|
|
REG_S t0, (SBI_TRAP_REGS_SIZE + SBI_TRAP_INFO_OFFSET(tval))(sp)
|
|
.if \have_h_extension
|
|
csrr t0, CSR_MTVAL2
|
|
REG_S t0, (SBI_TRAP_REGS_SIZE + SBI_TRAP_INFO_OFFSET(tval2))(sp)
|
|
csrr t0, CSR_MTINST
|
|
REG_S t0, (SBI_TRAP_REGS_SIZE + SBI_TRAP_INFO_OFFSET(tinst))(sp)
|
|
.if \have_mstatush
|
|
csrr t0, CSR_MSTATUSH
|
|
srli t0, t0, MSTATUSH_GVA_SHIFT
|
|
.else
|
|
csrr t0, CSR_MSTATUS
|
|
srli t0, t0, MSTATUS_GVA_SHIFT
|
|
.endif
|
|
and t0, t0, 0x1
|
|
.else
|
|
REG_S zero, (SBI_TRAP_REGS_SIZE + SBI_TRAP_INFO_OFFSET(tval2))(sp)
|
|
REG_S zero, (SBI_TRAP_REGS_SIZE + SBI_TRAP_INFO_OFFSET(tinst))(sp)
|
|
li t0, 0
|
|
.endif
|
|
REG_S t0, (SBI_TRAP_REGS_SIZE + SBI_TRAP_INFO_OFFSET(gva))(sp)
|
|
|
|
/* We are ready to take another trap, clear MDT */
|
|
CLEAR_MDT t0
|
|
.endm
|
|
|
|
.macro TRAP_CALL_C_ROUTINE
|
|
/* Call C routine */
|
|
add a0, sp, zero
|
|
call sbi_trap_handler
|
|
.endm
|
|
|
|
.macro TRAP_RESTORE_GENERAL_REGS_EXCEPT_A0_T0
|
|
/* Restore all general regisers except A0 and T0 */
|
|
REG_L ra, SBI_TRAP_REGS_OFFSET(ra)(a0)
|
|
REG_L sp, SBI_TRAP_REGS_OFFSET(sp)(a0)
|
|
REG_L gp, SBI_TRAP_REGS_OFFSET(gp)(a0)
|
|
REG_L tp, SBI_TRAP_REGS_OFFSET(tp)(a0)
|
|
REG_L t1, SBI_TRAP_REGS_OFFSET(t1)(a0)
|
|
REG_L t2, SBI_TRAP_REGS_OFFSET(t2)(a0)
|
|
REG_L s0, SBI_TRAP_REGS_OFFSET(s0)(a0)
|
|
REG_L s1, SBI_TRAP_REGS_OFFSET(s1)(a0)
|
|
REG_L a1, SBI_TRAP_REGS_OFFSET(a1)(a0)
|
|
REG_L a2, SBI_TRAP_REGS_OFFSET(a2)(a0)
|
|
REG_L a3, SBI_TRAP_REGS_OFFSET(a3)(a0)
|
|
REG_L a4, SBI_TRAP_REGS_OFFSET(a4)(a0)
|
|
REG_L a5, SBI_TRAP_REGS_OFFSET(a5)(a0)
|
|
REG_L a6, SBI_TRAP_REGS_OFFSET(a6)(a0)
|
|
REG_L a7, SBI_TRAP_REGS_OFFSET(a7)(a0)
|
|
REG_L s2, SBI_TRAP_REGS_OFFSET(s2)(a0)
|
|
REG_L s3, SBI_TRAP_REGS_OFFSET(s3)(a0)
|
|
REG_L s4, SBI_TRAP_REGS_OFFSET(s4)(a0)
|
|
REG_L s5, SBI_TRAP_REGS_OFFSET(s5)(a0)
|
|
REG_L s6, SBI_TRAP_REGS_OFFSET(s6)(a0)
|
|
REG_L s7, SBI_TRAP_REGS_OFFSET(s7)(a0)
|
|
REG_L s8, SBI_TRAP_REGS_OFFSET(s8)(a0)
|
|
REG_L s9, SBI_TRAP_REGS_OFFSET(s9)(a0)
|
|
REG_L s10, SBI_TRAP_REGS_OFFSET(s10)(a0)
|
|
REG_L s11, SBI_TRAP_REGS_OFFSET(s11)(a0)
|
|
REG_L t3, SBI_TRAP_REGS_OFFSET(t3)(a0)
|
|
REG_L t4, SBI_TRAP_REGS_OFFSET(t4)(a0)
|
|
REG_L t5, SBI_TRAP_REGS_OFFSET(t5)(a0)
|
|
REG_L t6, SBI_TRAP_REGS_OFFSET(t6)(a0)
|
|
.endm
|
|
|
|
.macro TRAP_RESTORE_MEPC_MSTATUS have_mstatush
|
|
/*
|
|
* Restore MSTATUS and MEPC CSRs starting with MSTATUS/H to set MDT
|
|
* flags since we can not take a trap now or MEPC would be cloberred
|
|
*/
|
|
.if \have_mstatush
|
|
REG_L t0, SBI_TRAP_REGS_OFFSET(mstatusH)(a0)
|
|
csrw CSR_MSTATUSH, t0
|
|
.endif
|
|
REG_L t0, SBI_TRAP_REGS_OFFSET(mstatus)(a0)
|
|
csrw CSR_MSTATUS, t0
|
|
REG_L t0, SBI_TRAP_REGS_OFFSET(mepc)(a0)
|
|
csrw CSR_MEPC, t0
|
|
.endm
|
|
|
|
.macro TRAP_RESTORE_A0_T0
|
|
/* Restore T0 */
|
|
REG_L t0, SBI_TRAP_REGS_OFFSET(t0)(a0)
|
|
|
|
/* Restore A0 */
|
|
REG_L a0, SBI_TRAP_REGS_OFFSET(a0)(a0)
|
|
.endm
|
|
|
|
.section .entry, "ax", %progbits
|
|
.align 3
|
|
.globl _trap_handler
|
|
_trap_handler:
|
|
TRAP_SAVE_AND_SETUP_SP_T0
|
|
|
|
TRAP_SAVE_MEPC_MSTATUS 0
|
|
|
|
TRAP_SAVE_GENERAL_REGS_EXCEPT_SP_T0
|
|
|
|
TRAP_SAVE_INFO 0 0
|
|
|
|
TRAP_CALL_C_ROUTINE
|
|
|
|
TRAP_RESTORE_GENERAL_REGS_EXCEPT_A0_T0
|
|
|
|
TRAP_RESTORE_MEPC_MSTATUS 0
|
|
|
|
TRAP_RESTORE_A0_T0
|
|
|
|
mret
|
|
|
|
.section .entry, "ax", %progbits
|
|
.align 3
|
|
.globl _trap_handler_hyp
|
|
_trap_handler_hyp:
|
|
TRAP_SAVE_AND_SETUP_SP_T0
|
|
|
|
#if __riscv_xlen == 32
|
|
TRAP_SAVE_MEPC_MSTATUS 1
|
|
#else
|
|
TRAP_SAVE_MEPC_MSTATUS 0
|
|
#endif
|
|
|
|
TRAP_SAVE_GENERAL_REGS_EXCEPT_SP_T0
|
|
|
|
#if __riscv_xlen == 32
|
|
TRAP_SAVE_INFO 1 1
|
|
#else
|
|
TRAP_SAVE_INFO 0 1
|
|
#endif
|
|
|
|
TRAP_CALL_C_ROUTINE
|
|
|
|
TRAP_RESTORE_GENERAL_REGS_EXCEPT_A0_T0
|
|
|
|
#if __riscv_xlen == 32
|
|
TRAP_RESTORE_MEPC_MSTATUS 1
|
|
#else
|
|
TRAP_RESTORE_MEPC_MSTATUS 0
|
|
#endif
|
|
|
|
TRAP_RESTORE_A0_T0
|
|
|
|
mret
|
|
|
|
.section .entry, "ax", %progbits
|
|
.align 3
|
|
.globl _reset_regs
|
|
_reset_regs:
|
|
|
|
/* flush the instruction cache */
|
|
fence.i
|
|
/* Reset all registers except ra, a0, a1, a2, a3 and a4 */
|
|
li sp, 0
|
|
li gp, 0
|
|
li tp, 0
|
|
li t0, 0
|
|
li t1, 0
|
|
li t2, 0
|
|
li s0, 0
|
|
li s1, 0
|
|
li a5, 0
|
|
li a6, 0
|
|
li a7, 0
|
|
li s2, 0
|
|
li s3, 0
|
|
li s4, 0
|
|
li s5, 0
|
|
li s6, 0
|
|
li s7, 0
|
|
li s8, 0
|
|
li s9, 0
|
|
li s10, 0
|
|
li s11, 0
|
|
li t3, 0
|
|
li t4, 0
|
|
li t5, 0
|
|
li t6, 0
|
|
csrw CSR_MSCRATCH, 0
|
|
|
|
ret
|
|
|
|
.section .rodata
|
|
.Lstack_corrupt_msg:
|
|
.string "stack smashing detected\n"
|
|
|
|
/* This will be called when the stack corruption is detected */
|
|
.section .text
|
|
.align 3
|
|
.globl __stack_chk_fail
|
|
.type __stack_chk_fail, %function
|
|
__stack_chk_fail:
|
|
la a0, .Lstack_corrupt_msg
|
|
call sbi_panic
|
|
|
|
/* Initial value of the stack guard variable */
|
|
.section .data
|
|
.align 3
|
|
.globl __stack_chk_guard
|
|
.type __stack_chk_guard, %object
|
|
__stack_chk_guard:
|
|
RISCV_PTR 0x95B5FF5A
|
|
|
|
#ifdef FW_FDT_PATH
|
|
.section .rodata
|
|
.align 4
|
|
.globl fw_fdt_bin
|
|
fw_fdt_bin:
|
|
.incbin FW_FDT_PATH
|
|
#ifdef FW_FDT_PADDING
|
|
.fill FW_FDT_PADDING, 1, 0
|
|
#endif
|
|
#endif
|