mirror of
https://github.com/espressif/esp-idf.git
synced 2024-10-05 20:47:46 -04:00
a8b1475fe7
This commit mainly targets the ESP32-P4. It adds supports for coprocessors on RISC-V based targets. The coprocessor save area, describing the used coprocessors is stored at the end of the stack of each task (highest address) whereas each coprocessor save area is allocated at the beginning of the task (lowest address). The context of each coprocessor is saved lazily, by the task that want to use it.
437 lines
17 KiB
ArmAsm
437 lines
17 KiB
ArmAsm
/*
|
|
* SPDX-FileCopyrightText: 2015-2023 Espressif Systems (Shanghai) CO LTD
|
|
*
|
|
* SPDX-License-Identifier: Apache-2.0
|
|
*/
|
|
#include "sdkconfig.h"
|
|
#include "portmacro.h"
|
|
#include "freertos/FreeRTOSConfig.h"
|
|
#include "soc/soc_caps.h"
|
|
#include "riscv/rvruntime-frames.h"
|
|
|
|
.extern pxCurrentTCBs
|
|
|
|
#if CONFIG_ESP_SYSTEM_HW_STACK_GUARD
|
|
#include "esp_private/hw_stack_guard.h"
|
|
#endif
|
|
|
|
.global port_uxInterruptNesting
|
|
.global port_xSchedulerRunning
|
|
.global xIsrStackTop
|
|
.global pxCurrentTCBs
|
|
.global vTaskSwitchContext
|
|
.global xPortSwitchFlag
|
|
#if CONFIG_ESP_SYSTEM_HW_STACK_GUARD
|
|
.global xIsrStackBottom
|
|
.global esp_hw_stack_guard_monitor_stop
|
|
.global esp_hw_stack_guard_monitor_start
|
|
.global esp_hw_stack_guard_set_bounds
|
|
#endif /* CONFIG_ESP_SYSTEM_HW_STACK_GUARD */
|
|
|
|
.section .text
|
|
|
|
|
|
#if SOC_CPU_COPROC_NUM > 0
|
|
|
|
#if SOC_CPU_HAS_FPU
|
|
|
|
/* Bit to set in mstatus to enable the FPU */
|
|
#define CSR_MSTATUS_FPU_ENABLE (1 << 13)
|
|
/* Bit to clear in mstatus to disable the FPU */
|
|
#define CSR_MSTATUS_FPU_DISABLE (3 << 13)
|
|
|
|
.macro save_fpu_regs frame=sp
|
|
fsw ft0, RV_FPU_FT0(\frame)
|
|
fsw ft1, RV_FPU_FT1(\frame)
|
|
fsw ft2, RV_FPU_FT2(\frame)
|
|
fsw ft3, RV_FPU_FT3(\frame)
|
|
fsw ft4, RV_FPU_FT4(\frame)
|
|
fsw ft5, RV_FPU_FT5(\frame)
|
|
fsw ft6, RV_FPU_FT6(\frame)
|
|
fsw ft7, RV_FPU_FT7(\frame)
|
|
fsw fs0, RV_FPU_FS0(\frame)
|
|
fsw fs1, RV_FPU_FS1(\frame)
|
|
fsw fa0, RV_FPU_FA0(\frame)
|
|
fsw fa1, RV_FPU_FA1(\frame)
|
|
fsw fa2, RV_FPU_FA2(\frame)
|
|
fsw fa3, RV_FPU_FA3(\frame)
|
|
fsw fa4, RV_FPU_FA4(\frame)
|
|
fsw fa5, RV_FPU_FA5(\frame)
|
|
fsw fa6, RV_FPU_FA6(\frame)
|
|
fsw fa7, RV_FPU_FA7(\frame)
|
|
fsw fs2, RV_FPU_FS2(\frame)
|
|
fsw fs3, RV_FPU_FS3(\frame)
|
|
fsw fs4, RV_FPU_FS4(\frame)
|
|
fsw fs5, RV_FPU_FS5(\frame)
|
|
fsw fs6, RV_FPU_FS6(\frame)
|
|
fsw fs7, RV_FPU_FS7(\frame)
|
|
fsw fs8, RV_FPU_FS8(\frame)
|
|
fsw fs9, RV_FPU_FS9(\frame)
|
|
fsw fs10, RV_FPU_FS10(\frame)
|
|
fsw fs11, RV_FPU_FS11(\frame)
|
|
fsw ft8, RV_FPU_FT8 (\frame)
|
|
fsw ft9, RV_FPU_FT9 (\frame)
|
|
fsw ft10, RV_FPU_FT10(\frame)
|
|
fsw ft11, RV_FPU_FT11(\frame)
|
|
.endm
|
|
|
|
.macro restore_fpu_regs frame=sp
|
|
flw ft0, RV_FPU_FT0(\frame)
|
|
flw ft1, RV_FPU_FT1(\frame)
|
|
flw ft2, RV_FPU_FT2(\frame)
|
|
flw ft3, RV_FPU_FT3(\frame)
|
|
flw ft4, RV_FPU_FT4(\frame)
|
|
flw ft5, RV_FPU_FT5(\frame)
|
|
flw ft6, RV_FPU_FT6(\frame)
|
|
flw ft7, RV_FPU_FT7(\frame)
|
|
flw fs0, RV_FPU_FS0(\frame)
|
|
flw fs1, RV_FPU_FS1(\frame)
|
|
flw fa0, RV_FPU_FA0(\frame)
|
|
flw fa1, RV_FPU_FA1(\frame)
|
|
flw fa2, RV_FPU_FA2(\frame)
|
|
flw fa3, RV_FPU_FA3(\frame)
|
|
flw fa4, RV_FPU_FA4(\frame)
|
|
flw fa5, RV_FPU_FA5(\frame)
|
|
flw fa6, RV_FPU_FA6(\frame)
|
|
flw fa7, RV_FPU_FA7(\frame)
|
|
flw fs2, RV_FPU_FS2(\frame)
|
|
flw fs3, RV_FPU_FS3(\frame)
|
|
flw fs4, RV_FPU_FS4(\frame)
|
|
flw fs5, RV_FPU_FS5(\frame)
|
|
flw fs6, RV_FPU_FS6(\frame)
|
|
flw fs7, RV_FPU_FS7(\frame)
|
|
flw fs8, RV_FPU_FS8(\frame)
|
|
flw fs9, RV_FPU_FS9(\frame)
|
|
flw fs10, RV_FPU_FS10(\frame)
|
|
flw fs11, RV_FPU_FS11(\frame)
|
|
flw ft8, RV_FPU_FT8(\frame)
|
|
flw ft9, RV_FPU_FT9(\frame)
|
|
flw ft10, RV_FPU_FT10(\frame)
|
|
flw ft11, RV_FPU_FT11(\frame)
|
|
.endm
|
|
|
|
|
|
.macro fpu_read_dirty_bit reg
|
|
csrr \reg, mstatus
|
|
srli \reg, \reg, 13
|
|
andi \reg, \reg, 1
|
|
.endm
|
|
|
|
|
|
.macro fpu_clear_dirty_bit reg
|
|
li \reg, 1 << 13
|
|
csrc mstatus, \reg
|
|
.endm
|
|
|
|
|
|
.macro fpu_enable reg
|
|
li \reg, CSR_MSTATUS_FPU_ENABLE
|
|
csrs mstatus, \reg
|
|
.endm
|
|
|
|
|
|
.macro fpu_disable reg
|
|
li \reg, CSR_MSTATUS_FPU_DISABLE
|
|
csrc mstatus, \reg
|
|
.endm
|
|
|
|
.global vPortTaskPinToCore
|
|
.global vPortCoprocUsedInISR
|
|
.global pxPortUpdateCoprocOwner
|
|
|
|
/**
|
|
* @brief Save the current FPU context in the FPU owner's save area
|
|
*
|
|
* @param sp Interuptee's RvExcFrame address
|
|
*
|
|
* Note: Since this routine is ONLY meant to be called from _panic_handler routine,
|
|
* it is possible to alter `s0-s11` registers
|
|
*/
|
|
.global rtos_save_fpu_coproc
|
|
.type rtos_save_fpu_coproc, @function
|
|
rtos_save_fpu_coproc:
|
|
/* If we are in an interrupt context, we have to abort. We don't allow using the FPU from ISR */
|
|
#if ( configNUM_CORES > 1 )
|
|
csrr a2, mhartid /* a2 = coreID */
|
|
slli a2, a2, 2 /* a2 = coreID * 4 */
|
|
la a1, port_uxInterruptNesting /* a1 = &port_uxInterruptNesting */
|
|
add a1, a1, a2 /* a1 = &port_uxInterruptNesting[coreID] */
|
|
lw a1, 0(a1) /* a1 = port_uxInterruptNesting[coreID] */
|
|
#else /* ( configNUM_CORES <= 1 ) */
|
|
lw a1, (port_uxInterruptNesting) /* a1 = port_uxInterruptNesting */
|
|
#endif /* ( configNUM_CORES > 1 ) */
|
|
/* SP still contains the RvExcFrame address */
|
|
mv a0, sp
|
|
bnez a1, vPortCoprocUsedInISR
|
|
/* Enable the FPU needed by the current task */
|
|
fpu_enable a1
|
|
mv s0, ra
|
|
call rtos_current_tcb
|
|
/* If the current TCB is NULL, the FPU is used during initialization, even before
|
|
* the scheduler started. Consider this a valid usage, the FPU will be disabled
|
|
* as soon as the scheduler is started anyway*/
|
|
beqz a0, rtos_save_fpu_coproc_norestore
|
|
mv s1, a0 /* s1 = pxCurrentTCBs */
|
|
/* Prepare parameters of pxPortUpdateCoprocOwner */
|
|
mv a2, a0
|
|
li a1, FPU_COPROC_IDX
|
|
csrr a0, mhartid
|
|
call pxPortUpdateCoprocOwner
|
|
/* If the save area is NULL, no need to save context */
|
|
beqz a0, rtos_save_fpu_coproc_nosave
|
|
/* Save the FPU context in the structure */
|
|
lw a0, RV_COPROC_SA+FPU_COPROC_IDX*4(a0) /* a0 = RvCoprocSaveArea->sa_coprocs[FPU_COPROC_IDX] */
|
|
save_fpu_regs a0
|
|
csrr a1, fcsr
|
|
sw a1, RV_FPU_FCSR(a0)
|
|
rtos_save_fpu_coproc_nosave:
|
|
/* Pin current task to current core */
|
|
mv a0, s1
|
|
csrr a1, mhartid
|
|
call vPortTaskPinToCore
|
|
/* Check if we have to restore a previous FPU context from the current TCB */
|
|
mv a0, s1
|
|
call pxPortGetCoprocArea
|
|
/* Get the enable flags from the coprocessor save area */
|
|
lw a1, RV_COPROC_ENABLE(a0)
|
|
/* To avoid having branches below, set the FPU enable flag now */
|
|
ori a2, a1, 1 << FPU_COPROC_IDX
|
|
sw a2, RV_COPROC_ENABLE(a0)
|
|
/* Check if the former FPU enable bit was set */
|
|
andi a2, a1, 1 << FPU_COPROC_IDX
|
|
beqz a2, rtos_save_fpu_coproc_norestore
|
|
/* FPU enable bit was set, restore the FPU context */
|
|
lw a0, RV_COPROC_SA+FPU_COPROC_IDX*4(a0) /* a0 = RvCoprocSaveArea->sa_coprocs[FPU_COPROC_IDX] */
|
|
restore_fpu_regs a0
|
|
lw a1, RV_FPU_FCSR(a0)
|
|
csrw fcsr, a1
|
|
rtos_save_fpu_coproc_norestore:
|
|
/* Return from routine via s0, instead of ra */
|
|
jr s0
|
|
.size rtos_save_fpu_coproc, .-rtos_save_fpu_coproc
|
|
|
|
#endif /* SOC_CPU_HAS_FPU */
|
|
|
|
#endif /* SOC_CPU_COPROC_NUM > 0 */
|
|
|
|
|
|
/**
|
|
* @brief Get current TCB on current core
|
|
*/
|
|
.type rtos_current_tcb, @function
|
|
rtos_current_tcb:
|
|
#if ( configNUM_CORES > 1 )
|
|
csrr a1, mhartid
|
|
slli a1, a1, 2
|
|
la a0, pxCurrentTCBs /* a0 = &pxCurrentTCBs */
|
|
add a0, a0, a1 /* a0 = &pxCurrentTCBs[coreID] */
|
|
lw a0, 0(a0) /* a0 = pxCurrentTCBs[coreID] */
|
|
#else
|
|
/* Recover the stack of next task */
|
|
lw a0, pxCurrentTCBs
|
|
#endif /* ( configNUM_CORES > 1 ) */
|
|
ret
|
|
.size, .-rtos_current_tcb
|
|
|
|
|
|
/**
|
|
* This function makes the RTOS aware about an ISR entering. It takes the
|
|
* current task stack pointer and places it into the pxCurrentTCBs.
|
|
* It then loads the ISR stack into sp.
|
|
* TODO: ISR nesting code improvements ?
|
|
* In the routines below, let's use a0-a5 registers to let the compiler generate
|
|
* 16-bit instructions.
|
|
*/
|
|
.global rtos_int_enter
|
|
.type rtos_int_enter, @function
|
|
rtos_int_enter:
|
|
#if ( configNUM_CORES > 1 )
|
|
csrr a5, mhartid /* a5 = coreID */
|
|
slli a5, a5, 2 /* a5 = coreID * 4 */
|
|
la a0, port_xSchedulerRunning /* a0 = &port_xSchedulerRunning */
|
|
add a0, a0, a5 /* a0 = &port_xSchedulerRunning[coreID] */
|
|
lw a0, (a0) /* a0 = port_xSchedulerRunning[coreID] */
|
|
#else
|
|
lw a0, port_xSchedulerRunning /* a0 = port_xSchedulerRunning */
|
|
#endif /* ( configNUM_CORES > 1 ) */
|
|
beqz a0, rtos_int_enter_end /* if (port_xSchedulerRunning[coreID] == 0) jump to rtos_int_enter_end */
|
|
|
|
/* Increment the ISR nesting count */
|
|
la a0, port_uxInterruptNesting /* a0 = &port_uxInterruptNesting */
|
|
#if ( configNUM_CORES > 1 )
|
|
add a0, a0, a5 /* a0 = &port_uxInterruptNesting[coreID] // a5 already contains coreID * 4 */
|
|
#endif /* ( configNUM_CORES > 1 ) */
|
|
lw a1, 0(a0) /* a1 = port_uxInterruptNesting[coreID] */
|
|
addi a2, a1, 1 /* a2 = a1 + 1 */
|
|
sw a2, 0(a0) /* port_uxInterruptNesting[coreID] = a2 */
|
|
|
|
/* If we reached here from another low-priority ISR, i.e, port_uxInterruptNesting[coreID] > 0, then skip stack pushing to TCB */
|
|
bnez a1, rtos_int_enter_end /* if (port_uxInterruptNesting[coreID] > 0) jump to rtos_int_enter_end */
|
|
|
|
#if SOC_CPU_COPROC_NUM > 0
|
|
/* Disable the FPU to forbid the ISR from using it. We don't need to re-enable it manually since the caller
|
|
* will restore `mstatus` before returning from interrupt. */
|
|
fpu_disable a0
|
|
#endif /* SOC_CPU_COPROC_NUM > 0 */
|
|
|
|
|
|
#if CONFIG_ESP_SYSTEM_HW_STACK_GUARD
|
|
/* esp_hw_stack_guard_monitor_stop(); pass the scratch registers */
|
|
ESP_HW_STACK_GUARD_MONITOR_STOP_CUR_CORE a0 a1
|
|
#endif /* CONFIG_ESP_SYSTEM_HW_STACK_GUARD */
|
|
|
|
/* Save the current sp in pxCurrentTCBs[coreID] and load the ISR stack on to sp */
|
|
#if ( configNUM_CORES > 1 )
|
|
la a0, pxCurrentTCBs /* a0 = &pxCurrentTCBs */
|
|
add a0, a0, a5 /* a0 = &pxCurrentTCBs[coreID] // a5 already contains coreID * 4 */
|
|
lw a0, (a0) /* a0 = pxCurrentTCBs[coreID] */
|
|
sw sp, 0(a0) /* pxCurrentTCBs[coreID] = sp */
|
|
la a0, xIsrStackTop /* a0 = &xIsrStackTop */
|
|
add a0, a0, a5 /* a0 = &xIsrStackTop[coreID] // a5 already contains coreID * 4 */
|
|
lw sp, (a0) /* sp = xIsrStackTop[coreID] */
|
|
#else
|
|
lw a0, pxCurrentTCBs /* a0 = pxCurrentTCBs */
|
|
sw sp, 0(a0) /* pxCurrentTCBs[0] = sp */
|
|
lw sp, xIsrStackTop /* sp = xIsrStackTop */
|
|
#endif /* ( configNUM_CORES > 1 ) */
|
|
|
|
#if CONFIG_ESP_SYSTEM_HW_STACK_GUARD
|
|
/* Prepare the parameters for esp_hw_stack_guard_set_bounds(xIsrStackBottom, xIsrStackTop); */
|
|
#if ( configNUM_CORES > 1 )
|
|
/* Load the xIsrStack for the current core and set the new bounds */
|
|
la a0, xIsrStackBottom
|
|
add a0, a0, a5 /* a0 = &xIsrStackBottom[coreID] */
|
|
lw a0, (a0) /* a0 = xIsrStackBottom[coreID] */
|
|
#else
|
|
lw a0, xIsrStackBottom
|
|
#endif /* ( configNUM_CORES > 1 ) */
|
|
mv a1, sp
|
|
/* esp_hw_stack_guard_set_bounds(xIsrStackBottom[coreID], xIsrStackTop[coreID]);
|
|
*/
|
|
ESP_HW_STACK_GUARD_SET_BOUNDS_CUR_CORE a2
|
|
ESP_HW_STACK_GUARD_MONITOR_START_CUR_CORE a0 a1
|
|
#endif /* CONFIG_ESP_SYSTEM_HW_STACK_GUARD */
|
|
|
|
rtos_int_enter_end:
|
|
ret
|
|
|
|
/**
|
|
* @brief Restore the stack pointer of the next task to run.
|
|
*
|
|
* @param a0 Former mstatus
|
|
*
|
|
* @returns New mstatus (potentially with coprocessors disabled)
|
|
*/
|
|
.global rtos_int_exit
|
|
.type rtos_int_exit, @function
|
|
rtos_int_exit:
|
|
/* To speed up this routine and because this current routine is only meant to be called from the interrupt
|
|
* handler, let's use callee-saved registers instead of stack space. Registers `s3-s11` are not used by
|
|
* the caller */
|
|
mv s11, a0
|
|
#if ( configNUM_CORES > 1 )
|
|
csrr a1, mhartid /* a1 = coreID */
|
|
slli a1, a1, 2 /* a1 = a1 * 4 */
|
|
la a0, port_xSchedulerRunning /* a0 = &port_xSchedulerRunning */
|
|
add a0, a0, a1 /* a0 = &port_xSchedulerRunning[coreID] */
|
|
lw a0, (a0) /* a0 = port_xSchedulerRunning[coreID] */
|
|
#else
|
|
lw a0, port_xSchedulerRunning /* a0 = port_xSchedulerRunning */
|
|
#endif /* ( configNUM_CORES > 1 ) */
|
|
beqz a0, rtos_int_exit_end /* if (port_uxSchedulerRunning == 0) jump to rtos_int_exit_end */
|
|
|
|
/* Update nesting interrupts counter */
|
|
la a2, port_uxInterruptNesting /* a2 = &port_uxInterruptNesting */
|
|
#if ( configNUM_CORES > 1 )
|
|
add a2, a2, a1 /* a2 = &port_uxInterruptNesting[coreID] // a1 already contains coreID * 4 */
|
|
#endif /* ( configNUM_CORES > 1 ) */
|
|
lw a0, 0(a2) /* a0 = port_uxInterruptNesting[coreID] */
|
|
|
|
/* Already zero, protect against underflow */
|
|
beqz a0, isr_skip_decrement /* if (port_uxInterruptNesting[coreID] == 0) jump to isr_skip_decrement */
|
|
addi a0, a0, -1 /* a0 = a0 - 1 */
|
|
sw a0, 0(a2) /* port_uxInterruptNesting[coreID] = a0 */
|
|
/* May still have interrupts pending, skip section below and exit */
|
|
bnez a0, rtos_int_exit_end
|
|
|
|
isr_skip_decrement:
|
|
/* If the CPU reached this label, a2 (uxInterruptNesting) is 0 for sure */
|
|
|
|
/* Schedule the next task if a yield is pending */
|
|
la a0, xPortSwitchFlag /* a0 = &xPortSwitchFlag */
|
|
#if ( configNUM_CORES > 1 )
|
|
add a0, a0, a1 /* a0 = &xPortSwitchFlag[coreID] // a1 already contains coreID * 4 */
|
|
#endif /* ( configNUM_CORES > 1 ) */
|
|
lw a2, 0(a0) /* a2 = xPortSwitchFlag[coreID] */
|
|
beqz a2, no_switch /* if (xPortSwitchFlag[coreID] == 0) jump to no_switch */
|
|
|
|
/* Preserve return address and schedule next task. To speed up the process, and because this current routine
|
|
* is only meant to be called from the interrupt handle, let's save some speed and space by using callee-saved
|
|
* registers instead of stack space. Registers `s3-s11` are not used by the caller */
|
|
mv s10, ra
|
|
#if ( SOC_CPU_COPROC_NUM > 0 )
|
|
/* In the cases where the newly scheduled task is different from the previously running one,
|
|
* we have to disable the coprocessor(s) to let them trigger an exception on first use.
|
|
* Else, if the same task is scheduled, do not change the coprocessor(s) state. */
|
|
call rtos_current_tcb
|
|
mv s9, a0
|
|
call vTaskSwitchContext
|
|
call rtos_current_tcb
|
|
beq a0, s9, rtos_int_exit_no_change
|
|
/* Disable the coprocessors in s11 register (former mstatus) */
|
|
li a0, ~CSR_MSTATUS_FPU_DISABLE
|
|
and s11, s11, a0
|
|
rtos_int_exit_no_change:
|
|
#else /* ( SOC_CPU_COPROC_NUM == 0 ) */
|
|
call vTaskSwitchContext
|
|
#endif /* ( SOC_CPU_COPROC_NUM > 0 ) */
|
|
mv ra, s10
|
|
|
|
/* Clears the switch pending flag */
|
|
la a0, xPortSwitchFlag /* a0 = &xPortSwitchFlag */
|
|
#if ( configNUM_CORES > 1 )
|
|
/* C routine vTaskSwitchContext may change the temp registers, so we read again */
|
|
csrr a1, mhartid /* a1 = coreID */
|
|
slli a1, a1, 2 /* a1 = a1 * 4 */
|
|
add a0, a0, a1 /* a0 = &xPortSwitchFlag[coreID]; */
|
|
#endif /* ( configNUM_CORES > 1 ) */
|
|
sw zero, 0(a0) /* xPortSwitchFlag[coreID] = 0; */
|
|
|
|
no_switch:
|
|
|
|
#if CONFIG_ESP_SYSTEM_HW_STACK_GUARD
|
|
/* esp_hw_stack_guard_monitor_stop(); pass the scratch registers */
|
|
ESP_HW_STACK_GUARD_MONITOR_STOP_CUR_CORE a0 a1
|
|
#endif /* CONFIG_ESP_SYSTEM_HW_STACK_GUARD */
|
|
|
|
|
|
#if ( configNUM_CORES > 1 )
|
|
/* Recover the stack of next task and prepare to exit */
|
|
csrr a1, mhartid
|
|
slli a1, a1, 2
|
|
la a0, pxCurrentTCBs /* a0 = &pxCurrentTCBs */
|
|
add a0, a0, a1 /* a0 = &pxCurrentTCBs[coreID] */
|
|
lw a0, 0(a0) /* a0 = pxCurrentTCBs[coreID] */
|
|
lw sp, 0(a0) /* sp = previous sp */
|
|
#else
|
|
/* Recover the stack of next task */
|
|
lw a0, pxCurrentTCBs
|
|
lw sp, 0(a0)
|
|
#endif /* ( configNUM_CORES > 1 ) */
|
|
|
|
|
|
#if CONFIG_ESP_SYSTEM_HW_STACK_GUARD
|
|
/* esp_hw_stack_guard_set_bounds(pxCurrentTCBs[0]->pxStack,
|
|
* pxCurrentTCBs[0]->pxEndOfStack);
|
|
*/
|
|
lw a1, PORT_OFFSET_PX_END_OF_STACK(a0)
|
|
lw a0, PORT_OFFSET_PX_STACK(a0)
|
|
ESP_HW_STACK_GUARD_SET_BOUNDS_CUR_CORE a2
|
|
/* esp_hw_stack_guard_monitor_start(); */
|
|
ESP_HW_STACK_GUARD_MONITOR_START_CUR_CORE a0 a1
|
|
#endif /* CONFIG_ESP_SYSTEM_HW_STACK_GUARD */
|
|
|
|
rtos_int_exit_end:
|
|
mv a0, s11 /* a0 = new mstatus */
|
|
ret
|