|
|
|
|
@ -6,6 +6,7 @@
|
|
|
|
|
*/
|
|
|
|
|
|
|
|
|
|
#include <linux/context_tracking.h>
|
|
|
|
|
#include <linux/irq-entry-common.h>
|
|
|
|
|
#include <linux/kasan.h>
|
|
|
|
|
#include <linux/linkage.h>
|
|
|
|
|
#include <linux/livepatch.h>
|
|
|
|
|
@ -29,13 +30,6 @@
|
|
|
|
|
#include <asm/sysreg.h>
|
|
|
|
|
#include <asm/system_misc.h>
|
|
|
|
|
|
|
|
|
|
typedef struct irqentry_state {
|
|
|
|
|
union {
|
|
|
|
|
bool exit_rcu;
|
|
|
|
|
bool lockdep;
|
|
|
|
|
};
|
|
|
|
|
} arm64_irqentry_state_t;
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
* Handle IRQ/context state management when entering from kernel mode.
|
|
|
|
|
* Before this function is called it is not safe to call regular kernel code,
|
|
|
|
|
@ -44,31 +38,14 @@ typedef struct irqentry_state {
|
|
|
|
|
* This is intended to match the logic in irqentry_enter(), handling the kernel
|
|
|
|
|
* mode transitions only.
|
|
|
|
|
*/
|
|
|
|
|
static __always_inline arm64_irqentry_state_t __enter_from_kernel_mode(struct pt_regs *regs)
|
|
|
|
|
static __always_inline irqentry_state_t __enter_from_kernel_mode(struct pt_regs *regs)
|
|
|
|
|
{
|
|
|
|
|
arm64_irqentry_state_t state = {
|
|
|
|
|
.exit_rcu = false,
|
|
|
|
|
};
|
|
|
|
|
|
|
|
|
|
if (!IS_ENABLED(CONFIG_TINY_RCU) && is_idle_task(current)) {
|
|
|
|
|
lockdep_hardirqs_off(CALLER_ADDR0);
|
|
|
|
|
ct_irq_enter();
|
|
|
|
|
trace_hardirqs_off_finish();
|
|
|
|
|
|
|
|
|
|
state.exit_rcu = true;
|
|
|
|
|
return state;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
lockdep_hardirqs_off(CALLER_ADDR0);
|
|
|
|
|
rcu_irq_enter_check_tick();
|
|
|
|
|
trace_hardirqs_off_finish();
|
|
|
|
|
|
|
|
|
|
return state;
|
|
|
|
|
return irqentry_enter(regs);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static noinstr arm64_irqentry_state_t enter_from_kernel_mode(struct pt_regs *regs)
|
|
|
|
|
static noinstr irqentry_state_t enter_from_kernel_mode(struct pt_regs *regs)
|
|
|
|
|
{
|
|
|
|
|
arm64_irqentry_state_t state;
|
|
|
|
|
irqentry_state_t state;
|
|
|
|
|
|
|
|
|
|
state = __enter_from_kernel_mode(regs);
|
|
|
|
|
mte_check_tfsr_entry();
|
|
|
|
|
@ -77,51 +54,6 @@ static noinstr arm64_irqentry_state_t enter_from_kernel_mode(struct pt_regs *reg
|
|
|
|
|
return state;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static inline bool arm64_preempt_schedule_irq(void)
|
|
|
|
|
{
|
|
|
|
|
/*
|
|
|
|
|
* DAIF.DA are cleared at the start of IRQ/FIQ handling, and when GIC
|
|
|
|
|
* priority masking is used the GIC irqchip driver will clear DAIF.IF
|
|
|
|
|
* using gic_arch_enable_irqs() for normal IRQs. If anything is set in
|
|
|
|
|
* DAIF we must have handled an NMI, so skip preemption.
|
|
|
|
|
*/
|
|
|
|
|
if (system_uses_irq_prio_masking() && read_sysreg(daif))
|
|
|
|
|
return false;
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
* Preempting a task from an IRQ means we leave copies of PSTATE
|
|
|
|
|
* on the stack. cpufeature's enable calls may modify PSTATE, but
|
|
|
|
|
* resuming one of these preempted tasks would undo those changes.
|
|
|
|
|
*
|
|
|
|
|
* Only allow a task to be preempted once cpufeatures have been
|
|
|
|
|
* enabled.
|
|
|
|
|
*/
|
|
|
|
|
if (!system_capabilities_finalized())
|
|
|
|
|
return false;
|
|
|
|
|
|
|
|
|
|
return true;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
#ifdef CONFIG_PREEMPTION
|
|
|
|
|
void raw_irqentry_exit_cond_resched(void)
|
|
|
|
|
{
|
|
|
|
|
if (!preempt_count()) {
|
|
|
|
|
if (need_resched() && arm64_preempt_schedule_irq())
|
|
|
|
|
preempt_schedule_irq();
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
#endif
|
|
|
|
|
|
|
|
|
|
#ifdef CONFIG_PREEMPT_DYNAMIC
|
|
|
|
|
DEFINE_STATIC_KEY_TRUE(sk_dynamic_irqentry_exit_cond_resched);
|
|
|
|
|
void dynamic_irqentry_exit_cond_resched(void)
|
|
|
|
|
{
|
|
|
|
|
if (!static_branch_unlikely(&sk_dynamic_irqentry_exit_cond_resched))
|
|
|
|
|
return;
|
|
|
|
|
raw_irqentry_exit_cond_resched();
|
|
|
|
|
}
|
|
|
|
|
#endif
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
* Handle IRQ/context state management when exiting to kernel mode.
|
|
|
|
|
* After this function returns it is not safe to call regular kernel code,
|
|
|
|
|
@ -131,31 +63,13 @@ void dynamic_irqentry_exit_cond_resched(void)
|
|
|
|
|
* mode transitions only, and with preemption handled elsewhere.
|
|
|
|
|
*/
|
|
|
|
|
static __always_inline void __exit_to_kernel_mode(struct pt_regs *regs,
|
|
|
|
|
arm64_irqentry_state_t state)
|
|
|
|
|
irqentry_state_t state)
|
|
|
|
|
{
|
|
|
|
|
lockdep_assert_irqs_disabled();
|
|
|
|
|
|
|
|
|
|
if (!regs_irqs_disabled(regs)) {
|
|
|
|
|
if (state.exit_rcu) {
|
|
|
|
|
trace_hardirqs_on_prepare();
|
|
|
|
|
lockdep_hardirqs_on_prepare();
|
|
|
|
|
ct_irq_exit();
|
|
|
|
|
lockdep_hardirqs_on(CALLER_ADDR0);
|
|
|
|
|
return;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
if (IS_ENABLED(CONFIG_PREEMPTION))
|
|
|
|
|
irqentry_exit_cond_resched();
|
|
|
|
|
|
|
|
|
|
trace_hardirqs_on();
|
|
|
|
|
} else {
|
|
|
|
|
if (state.exit_rcu)
|
|
|
|
|
ct_irq_exit();
|
|
|
|
|
}
|
|
|
|
|
irqentry_exit(regs, state);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static void noinstr exit_to_kernel_mode(struct pt_regs *regs,
|
|
|
|
|
arm64_irqentry_state_t state)
|
|
|
|
|
irqentry_state_t state)
|
|
|
|
|
{
|
|
|
|
|
mte_check_tfsr_exit();
|
|
|
|
|
__exit_to_kernel_mode(regs, state);
|
|
|
|
|
@ -166,18 +80,15 @@ static void noinstr exit_to_kernel_mode(struct pt_regs *regs,
|
|
|
|
|
* Before this function is called it is not safe to call regular kernel code,
|
|
|
|
|
* instrumentable code, or any code which may trigger an exception.
|
|
|
|
|
*/
|
|
|
|
|
static __always_inline void __enter_from_user_mode(void)
|
|
|
|
|
static __always_inline void __enter_from_user_mode(struct pt_regs *regs)
|
|
|
|
|
{
|
|
|
|
|
lockdep_hardirqs_off(CALLER_ADDR0);
|
|
|
|
|
CT_WARN_ON(ct_state() != CT_STATE_USER);
|
|
|
|
|
user_exit_irqoff();
|
|
|
|
|
trace_hardirqs_off_finish();
|
|
|
|
|
enter_from_user_mode(regs);
|
|
|
|
|
mte_disable_tco_entry(current);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static __always_inline void enter_from_user_mode(struct pt_regs *regs)
|
|
|
|
|
static __always_inline void arm64_enter_from_user_mode(struct pt_regs *regs)
|
|
|
|
|
{
|
|
|
|
|
__enter_from_user_mode();
|
|
|
|
|
__enter_from_user_mode(regs);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
@ -185,116 +96,19 @@ static __always_inline void enter_from_user_mode(struct pt_regs *regs)
|
|
|
|
|
* After this function returns it is not safe to call regular kernel code,
|
|
|
|
|
* instrumentable code, or any code which may trigger an exception.
|
|
|
|
|
*/
|
|
|
|
|
static __always_inline void __exit_to_user_mode(void)
|
|
|
|
|
|
|
|
|
|
static __always_inline void arm64_exit_to_user_mode(struct pt_regs *regs)
|
|
|
|
|
{
|
|
|
|
|
trace_hardirqs_on_prepare();
|
|
|
|
|
lockdep_hardirqs_on_prepare();
|
|
|
|
|
user_enter_irqoff();
|
|
|
|
|
lockdep_hardirqs_on(CALLER_ADDR0);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static void do_notify_resume(struct pt_regs *regs, unsigned long thread_flags)
|
|
|
|
|
{
|
|
|
|
|
do {
|
|
|
|
|
local_irq_enable();
|
|
|
|
|
|
|
|
|
|
if (thread_flags & (_TIF_NEED_RESCHED | _TIF_NEED_RESCHED_LAZY))
|
|
|
|
|
schedule();
|
|
|
|
|
|
|
|
|
|
if (thread_flags & _TIF_UPROBE)
|
|
|
|
|
uprobe_notify_resume(regs);
|
|
|
|
|
|
|
|
|
|
if (thread_flags & _TIF_MTE_ASYNC_FAULT) {
|
|
|
|
|
clear_thread_flag(TIF_MTE_ASYNC_FAULT);
|
|
|
|
|
send_sig_fault(SIGSEGV, SEGV_MTEAERR,
|
|
|
|
|
(void __user *)NULL, current);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
if (thread_flags & _TIF_PATCH_PENDING)
|
|
|
|
|
klp_update_patch_state(current);
|
|
|
|
|
|
|
|
|
|
if (thread_flags & (_TIF_SIGPENDING | _TIF_NOTIFY_SIGNAL))
|
|
|
|
|
do_signal(regs);
|
|
|
|
|
|
|
|
|
|
if (thread_flags & _TIF_NOTIFY_RESUME)
|
|
|
|
|
resume_user_mode_work(regs);
|
|
|
|
|
|
|
|
|
|
if (thread_flags & _TIF_FOREIGN_FPSTATE)
|
|
|
|
|
fpsimd_restore_current_state();
|
|
|
|
|
|
|
|
|
|
local_irq_disable();
|
|
|
|
|
thread_flags = read_thread_flags();
|
|
|
|
|
} while (thread_flags & _TIF_WORK_MASK);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static __always_inline void exit_to_user_mode_prepare(struct pt_regs *regs)
|
|
|
|
|
{
|
|
|
|
|
unsigned long flags;
|
|
|
|
|
|
|
|
|
|
local_irq_disable();
|
|
|
|
|
|
|
|
|
|
flags = read_thread_flags();
|
|
|
|
|
if (unlikely(flags & _TIF_WORK_MASK))
|
|
|
|
|
do_notify_resume(regs, flags);
|
|
|
|
|
|
|
|
|
|
local_daif_mask();
|
|
|
|
|
|
|
|
|
|
lockdep_sys_exit();
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static __always_inline void exit_to_user_mode(struct pt_regs *regs)
|
|
|
|
|
{
|
|
|
|
|
exit_to_user_mode_prepare(regs);
|
|
|
|
|
local_daif_mask();
|
|
|
|
|
mte_check_tfsr_exit();
|
|
|
|
|
__exit_to_user_mode();
|
|
|
|
|
exit_to_user_mode();
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
asmlinkage void noinstr asm_exit_to_user_mode(struct pt_regs *regs)
|
|
|
|
|
{
|
|
|
|
|
exit_to_user_mode(regs);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
* Handle IRQ/context state management when entering an NMI from user/kernel
|
|
|
|
|
* mode. Before this function is called it is not safe to call regular kernel
|
|
|
|
|
* code, instrumentable code, or any code which may trigger an exception.
|
|
|
|
|
*/
|
|
|
|
|
static noinstr arm64_irqentry_state_t arm64_enter_nmi(struct pt_regs *regs)
|
|
|
|
|
{
|
|
|
|
|
arm64_irqentry_state_t state;
|
|
|
|
|
|
|
|
|
|
state.lockdep = lockdep_hardirqs_enabled();
|
|
|
|
|
|
|
|
|
|
__nmi_enter();
|
|
|
|
|
lockdep_hardirqs_off(CALLER_ADDR0);
|
|
|
|
|
lockdep_hardirq_enter();
|
|
|
|
|
ct_nmi_enter();
|
|
|
|
|
|
|
|
|
|
trace_hardirqs_off_finish();
|
|
|
|
|
ftrace_nmi_enter();
|
|
|
|
|
|
|
|
|
|
return state;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
* Handle IRQ/context state management when exiting an NMI from user/kernel
|
|
|
|
|
* mode. After this function returns it is not safe to call regular kernel
|
|
|
|
|
* code, instrumentable code, or any code which may trigger an exception.
|
|
|
|
|
*/
|
|
|
|
|
static void noinstr arm64_exit_nmi(struct pt_regs *regs,
|
|
|
|
|
arm64_irqentry_state_t state)
|
|
|
|
|
{
|
|
|
|
|
ftrace_nmi_exit();
|
|
|
|
|
if (state.lockdep) {
|
|
|
|
|
trace_hardirqs_on_prepare();
|
|
|
|
|
lockdep_hardirqs_on_prepare();
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
ct_nmi_exit();
|
|
|
|
|
lockdep_hardirq_exit();
|
|
|
|
|
if (state.lockdep)
|
|
|
|
|
lockdep_hardirqs_on(CALLER_ADDR0);
|
|
|
|
|
__nmi_exit();
|
|
|
|
|
arm64_exit_to_user_mode(regs);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
@ -302,9 +116,9 @@ static void noinstr arm64_exit_nmi(struct pt_regs *regs,
|
|
|
|
|
* kernel mode. Before this function is called it is not safe to call regular
|
|
|
|
|
* kernel code, instrumentable code, or any code which may trigger an exception.
|
|
|
|
|
*/
|
|
|
|
|
static noinstr arm64_irqentry_state_t arm64_enter_el1_dbg(struct pt_regs *regs)
|
|
|
|
|
static noinstr irqentry_state_t arm64_enter_el1_dbg(struct pt_regs *regs)
|
|
|
|
|
{
|
|
|
|
|
arm64_irqentry_state_t state;
|
|
|
|
|
irqentry_state_t state;
|
|
|
|
|
|
|
|
|
|
state.lockdep = lockdep_hardirqs_enabled();
|
|
|
|
|
|
|
|
|
|
@ -322,7 +136,7 @@ static noinstr arm64_irqentry_state_t arm64_enter_el1_dbg(struct pt_regs *regs)
|
|
|
|
|
* kernel code, instrumentable code, or any code which may trigger an exception.
|
|
|
|
|
*/
|
|
|
|
|
static void noinstr arm64_exit_el1_dbg(struct pt_regs *regs,
|
|
|
|
|
arm64_irqentry_state_t state)
|
|
|
|
|
irqentry_state_t state)
|
|
|
|
|
{
|
|
|
|
|
if (state.lockdep) {
|
|
|
|
|
trace_hardirqs_on_prepare();
|
|
|
|
|
@ -353,7 +167,7 @@ extern void (*handle_arch_fiq)(struct pt_regs *);
|
|
|
|
|
static void noinstr __panic_unhandled(struct pt_regs *regs, const char *vector,
|
|
|
|
|
unsigned long esr)
|
|
|
|
|
{
|
|
|
|
|
arm64_enter_nmi(regs);
|
|
|
|
|
irqentry_nmi_enter(regs);
|
|
|
|
|
|
|
|
|
|
console_verbose();
|
|
|
|
|
|
|
|
|
|
@ -504,7 +318,7 @@ UNHANDLED(el1t, 64, error)
|
|
|
|
|
static void noinstr el1_abort(struct pt_regs *regs, unsigned long esr)
|
|
|
|
|
{
|
|
|
|
|
unsigned long far = read_sysreg(far_el1);
|
|
|
|
|
arm64_irqentry_state_t state;
|
|
|
|
|
irqentry_state_t state;
|
|
|
|
|
|
|
|
|
|
state = enter_from_kernel_mode(regs);
|
|
|
|
|
local_daif_inherit(regs);
|
|
|
|
|
@ -516,7 +330,7 @@ static void noinstr el1_abort(struct pt_regs *regs, unsigned long esr)
|
|
|
|
|
static void noinstr el1_pc(struct pt_regs *regs, unsigned long esr)
|
|
|
|
|
{
|
|
|
|
|
unsigned long far = read_sysreg(far_el1);
|
|
|
|
|
arm64_irqentry_state_t state;
|
|
|
|
|
irqentry_state_t state;
|
|
|
|
|
|
|
|
|
|
state = enter_from_kernel_mode(regs);
|
|
|
|
|
local_daif_inherit(regs);
|
|
|
|
|
@ -527,7 +341,7 @@ static void noinstr el1_pc(struct pt_regs *regs, unsigned long esr)
|
|
|
|
|
|
|
|
|
|
static void noinstr el1_undef(struct pt_regs *regs, unsigned long esr)
|
|
|
|
|
{
|
|
|
|
|
arm64_irqentry_state_t state;
|
|
|
|
|
irqentry_state_t state;
|
|
|
|
|
|
|
|
|
|
state = enter_from_kernel_mode(regs);
|
|
|
|
|
local_daif_inherit(regs);
|
|
|
|
|
@ -538,7 +352,7 @@ static void noinstr el1_undef(struct pt_regs *regs, unsigned long esr)
|
|
|
|
|
|
|
|
|
|
static void noinstr el1_bti(struct pt_regs *regs, unsigned long esr)
|
|
|
|
|
{
|
|
|
|
|
arm64_irqentry_state_t state;
|
|
|
|
|
irqentry_state_t state;
|
|
|
|
|
|
|
|
|
|
state = enter_from_kernel_mode(regs);
|
|
|
|
|
local_daif_inherit(regs);
|
|
|
|
|
@ -549,7 +363,7 @@ static void noinstr el1_bti(struct pt_regs *regs, unsigned long esr)
|
|
|
|
|
|
|
|
|
|
static void noinstr el1_gcs(struct pt_regs *regs, unsigned long esr)
|
|
|
|
|
{
|
|
|
|
|
arm64_irqentry_state_t state;
|
|
|
|
|
irqentry_state_t state;
|
|
|
|
|
|
|
|
|
|
state = enter_from_kernel_mode(regs);
|
|
|
|
|
local_daif_inherit(regs);
|
|
|
|
|
@ -560,7 +374,7 @@ static void noinstr el1_gcs(struct pt_regs *regs, unsigned long esr)
|
|
|
|
|
|
|
|
|
|
static void noinstr el1_mops(struct pt_regs *regs, unsigned long esr)
|
|
|
|
|
{
|
|
|
|
|
arm64_irqentry_state_t state;
|
|
|
|
|
irqentry_state_t state;
|
|
|
|
|
|
|
|
|
|
state = enter_from_kernel_mode(regs);
|
|
|
|
|
local_daif_inherit(regs);
|
|
|
|
|
@ -571,7 +385,7 @@ static void noinstr el1_mops(struct pt_regs *regs, unsigned long esr)
|
|
|
|
|
|
|
|
|
|
static void noinstr el1_breakpt(struct pt_regs *regs, unsigned long esr)
|
|
|
|
|
{
|
|
|
|
|
arm64_irqentry_state_t state;
|
|
|
|
|
irqentry_state_t state;
|
|
|
|
|
|
|
|
|
|
state = arm64_enter_el1_dbg(regs);
|
|
|
|
|
debug_exception_enter(regs);
|
|
|
|
|
@ -582,7 +396,7 @@ static void noinstr el1_breakpt(struct pt_regs *regs, unsigned long esr)
|
|
|
|
|
|
|
|
|
|
static void noinstr el1_softstp(struct pt_regs *regs, unsigned long esr)
|
|
|
|
|
{
|
|
|
|
|
arm64_irqentry_state_t state;
|
|
|
|
|
irqentry_state_t state;
|
|
|
|
|
|
|
|
|
|
state = arm64_enter_el1_dbg(regs);
|
|
|
|
|
if (!cortex_a76_erratum_1463225_debug_handler(regs)) {
|
|
|
|
|
@ -604,7 +418,7 @@ static void noinstr el1_watchpt(struct pt_regs *regs, unsigned long esr)
|
|
|
|
|
{
|
|
|
|
|
/* Watchpoints are the only debug exception to write FAR_EL1 */
|
|
|
|
|
unsigned long far = read_sysreg(far_el1);
|
|
|
|
|
arm64_irqentry_state_t state;
|
|
|
|
|
irqentry_state_t state;
|
|
|
|
|
|
|
|
|
|
state = arm64_enter_el1_dbg(regs);
|
|
|
|
|
debug_exception_enter(regs);
|
|
|
|
|
@ -615,7 +429,7 @@ static void noinstr el1_watchpt(struct pt_regs *regs, unsigned long esr)
|
|
|
|
|
|
|
|
|
|
static void noinstr el1_brk64(struct pt_regs *regs, unsigned long esr)
|
|
|
|
|
{
|
|
|
|
|
arm64_irqentry_state_t state;
|
|
|
|
|
irqentry_state_t state;
|
|
|
|
|
|
|
|
|
|
state = arm64_enter_el1_dbg(regs);
|
|
|
|
|
debug_exception_enter(regs);
|
|
|
|
|
@ -626,7 +440,7 @@ static void noinstr el1_brk64(struct pt_regs *regs, unsigned long esr)
|
|
|
|
|
|
|
|
|
|
static void noinstr el1_fpac(struct pt_regs *regs, unsigned long esr)
|
|
|
|
|
{
|
|
|
|
|
arm64_irqentry_state_t state;
|
|
|
|
|
irqentry_state_t state;
|
|
|
|
|
|
|
|
|
|
state = enter_from_kernel_mode(regs);
|
|
|
|
|
local_daif_inherit(regs);
|
|
|
|
|
@ -687,17 +501,17 @@ asmlinkage void noinstr el1h_64_sync_handler(struct pt_regs *regs)
|
|
|
|
|
static __always_inline void __el1_pnmi(struct pt_regs *regs,
|
|
|
|
|
void (*handler)(struct pt_regs *))
|
|
|
|
|
{
|
|
|
|
|
arm64_irqentry_state_t state;
|
|
|
|
|
irqentry_state_t state;
|
|
|
|
|
|
|
|
|
|
state = arm64_enter_nmi(regs);
|
|
|
|
|
state = irqentry_nmi_enter(regs);
|
|
|
|
|
do_interrupt_handler(regs, handler);
|
|
|
|
|
arm64_exit_nmi(regs, state);
|
|
|
|
|
irqentry_nmi_exit(regs, state);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static __always_inline void __el1_irq(struct pt_regs *regs,
|
|
|
|
|
void (*handler)(struct pt_regs *))
|
|
|
|
|
{
|
|
|
|
|
arm64_irqentry_state_t state;
|
|
|
|
|
irqentry_state_t state;
|
|
|
|
|
|
|
|
|
|
state = enter_from_kernel_mode(regs);
|
|
|
|
|
|
|
|
|
|
@ -731,22 +545,22 @@ asmlinkage void noinstr el1h_64_fiq_handler(struct pt_regs *regs)
|
|
|
|
|
asmlinkage void noinstr el1h_64_error_handler(struct pt_regs *regs)
|
|
|
|
|
{
|
|
|
|
|
unsigned long esr = read_sysreg(esr_el1);
|
|
|
|
|
arm64_irqentry_state_t state;
|
|
|
|
|
irqentry_state_t state;
|
|
|
|
|
|
|
|
|
|
local_daif_restore(DAIF_ERRCTX);
|
|
|
|
|
state = arm64_enter_nmi(regs);
|
|
|
|
|
state = irqentry_nmi_enter(regs);
|
|
|
|
|
do_serror(regs, esr);
|
|
|
|
|
arm64_exit_nmi(regs, state);
|
|
|
|
|
irqentry_nmi_exit(regs, state);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static void noinstr el0_da(struct pt_regs *regs, unsigned long esr)
|
|
|
|
|
{
|
|
|
|
|
unsigned long far = read_sysreg(far_el1);
|
|
|
|
|
|
|
|
|
|
enter_from_user_mode(regs);
|
|
|
|
|
arm64_enter_from_user_mode(regs);
|
|
|
|
|
local_daif_restore(DAIF_PROCCTX);
|
|
|
|
|
do_mem_abort(far, esr, regs);
|
|
|
|
|
exit_to_user_mode(regs);
|
|
|
|
|
arm64_exit_to_user_mode(regs);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static void noinstr el0_ia(struct pt_regs *regs, unsigned long esr)
|
|
|
|
|
@ -761,50 +575,50 @@ static void noinstr el0_ia(struct pt_regs *regs, unsigned long esr)
|
|
|
|
|
if (!is_ttbr0_addr(far))
|
|
|
|
|
arm64_apply_bp_hardening();
|
|
|
|
|
|
|
|
|
|
enter_from_user_mode(regs);
|
|
|
|
|
arm64_enter_from_user_mode(regs);
|
|
|
|
|
local_daif_restore(DAIF_PROCCTX);
|
|
|
|
|
do_mem_abort(far, esr, regs);
|
|
|
|
|
exit_to_user_mode(regs);
|
|
|
|
|
arm64_exit_to_user_mode(regs);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static void noinstr el0_fpsimd_acc(struct pt_regs *regs, unsigned long esr)
|
|
|
|
|
{
|
|
|
|
|
enter_from_user_mode(regs);
|
|
|
|
|
arm64_enter_from_user_mode(regs);
|
|
|
|
|
local_daif_restore(DAIF_PROCCTX);
|
|
|
|
|
do_fpsimd_acc(esr, regs);
|
|
|
|
|
exit_to_user_mode(regs);
|
|
|
|
|
arm64_exit_to_user_mode(regs);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static void noinstr el0_sve_acc(struct pt_regs *regs, unsigned long esr)
|
|
|
|
|
{
|
|
|
|
|
enter_from_user_mode(regs);
|
|
|
|
|
arm64_enter_from_user_mode(regs);
|
|
|
|
|
local_daif_restore(DAIF_PROCCTX);
|
|
|
|
|
do_sve_acc(esr, regs);
|
|
|
|
|
exit_to_user_mode(regs);
|
|
|
|
|
arm64_exit_to_user_mode(regs);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static void noinstr el0_sme_acc(struct pt_regs *regs, unsigned long esr)
|
|
|
|
|
{
|
|
|
|
|
enter_from_user_mode(regs);
|
|
|
|
|
arm64_enter_from_user_mode(regs);
|
|
|
|
|
local_daif_restore(DAIF_PROCCTX);
|
|
|
|
|
do_sme_acc(esr, regs);
|
|
|
|
|
exit_to_user_mode(regs);
|
|
|
|
|
arm64_exit_to_user_mode(regs);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static void noinstr el0_fpsimd_exc(struct pt_regs *regs, unsigned long esr)
|
|
|
|
|
{
|
|
|
|
|
enter_from_user_mode(regs);
|
|
|
|
|
arm64_enter_from_user_mode(regs);
|
|
|
|
|
local_daif_restore(DAIF_PROCCTX);
|
|
|
|
|
do_fpsimd_exc(esr, regs);
|
|
|
|
|
exit_to_user_mode(regs);
|
|
|
|
|
arm64_exit_to_user_mode(regs);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static void noinstr el0_sys(struct pt_regs *regs, unsigned long esr)
|
|
|
|
|
{
|
|
|
|
|
enter_from_user_mode(regs);
|
|
|
|
|
arm64_enter_from_user_mode(regs);
|
|
|
|
|
local_daif_restore(DAIF_PROCCTX);
|
|
|
|
|
do_el0_sys(esr, regs);
|
|
|
|
|
exit_to_user_mode(regs);
|
|
|
|
|
arm64_exit_to_user_mode(regs);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static void noinstr el0_pc(struct pt_regs *regs, unsigned long esr)
|
|
|
|
|
@ -814,58 +628,58 @@ static void noinstr el0_pc(struct pt_regs *regs, unsigned long esr)
|
|
|
|
|
if (!is_ttbr0_addr(instruction_pointer(regs)))
|
|
|
|
|
arm64_apply_bp_hardening();
|
|
|
|
|
|
|
|
|
|
enter_from_user_mode(regs);
|
|
|
|
|
arm64_enter_from_user_mode(regs);
|
|
|
|
|
local_daif_restore(DAIF_PROCCTX);
|
|
|
|
|
do_sp_pc_abort(far, esr, regs);
|
|
|
|
|
exit_to_user_mode(regs);
|
|
|
|
|
arm64_exit_to_user_mode(regs);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static void noinstr el0_sp(struct pt_regs *regs, unsigned long esr)
|
|
|
|
|
{
|
|
|
|
|
enter_from_user_mode(regs);
|
|
|
|
|
arm64_enter_from_user_mode(regs);
|
|
|
|
|
local_daif_restore(DAIF_PROCCTX);
|
|
|
|
|
do_sp_pc_abort(regs->sp, esr, regs);
|
|
|
|
|
exit_to_user_mode(regs);
|
|
|
|
|
arm64_exit_to_user_mode(regs);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static void noinstr el0_undef(struct pt_regs *regs, unsigned long esr)
|
|
|
|
|
{
|
|
|
|
|
enter_from_user_mode(regs);
|
|
|
|
|
arm64_enter_from_user_mode(regs);
|
|
|
|
|
local_daif_restore(DAIF_PROCCTX);
|
|
|
|
|
do_el0_undef(regs, esr);
|
|
|
|
|
exit_to_user_mode(regs);
|
|
|
|
|
arm64_exit_to_user_mode(regs);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static void noinstr el0_bti(struct pt_regs *regs)
|
|
|
|
|
{
|
|
|
|
|
enter_from_user_mode(regs);
|
|
|
|
|
arm64_enter_from_user_mode(regs);
|
|
|
|
|
local_daif_restore(DAIF_PROCCTX);
|
|
|
|
|
do_el0_bti(regs);
|
|
|
|
|
exit_to_user_mode(regs);
|
|
|
|
|
arm64_exit_to_user_mode(regs);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static void noinstr el0_mops(struct pt_regs *regs, unsigned long esr)
|
|
|
|
|
{
|
|
|
|
|
enter_from_user_mode(regs);
|
|
|
|
|
arm64_enter_from_user_mode(regs);
|
|
|
|
|
local_daif_restore(DAIF_PROCCTX);
|
|
|
|
|
do_el0_mops(regs, esr);
|
|
|
|
|
exit_to_user_mode(regs);
|
|
|
|
|
arm64_exit_to_user_mode(regs);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static void noinstr el0_gcs(struct pt_regs *regs, unsigned long esr)
|
|
|
|
|
{
|
|
|
|
|
enter_from_user_mode(regs);
|
|
|
|
|
arm64_enter_from_user_mode(regs);
|
|
|
|
|
local_daif_restore(DAIF_PROCCTX);
|
|
|
|
|
do_el0_gcs(regs, esr);
|
|
|
|
|
exit_to_user_mode(regs);
|
|
|
|
|
arm64_exit_to_user_mode(regs);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static void noinstr el0_inv(struct pt_regs *regs, unsigned long esr)
|
|
|
|
|
{
|
|
|
|
|
enter_from_user_mode(regs);
|
|
|
|
|
arm64_enter_from_user_mode(regs);
|
|
|
|
|
local_daif_restore(DAIF_PROCCTX);
|
|
|
|
|
bad_el0_sync(regs, 0, esr);
|
|
|
|
|
exit_to_user_mode(regs);
|
|
|
|
|
arm64_exit_to_user_mode(regs);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static void noinstr el0_breakpt(struct pt_regs *regs, unsigned long esr)
|
|
|
|
|
@ -873,12 +687,12 @@ static void noinstr el0_breakpt(struct pt_regs *regs, unsigned long esr)
|
|
|
|
|
if (!is_ttbr0_addr(regs->pc))
|
|
|
|
|
arm64_apply_bp_hardening();
|
|
|
|
|
|
|
|
|
|
enter_from_user_mode(regs);
|
|
|
|
|
arm64_enter_from_user_mode(regs);
|
|
|
|
|
debug_exception_enter(regs);
|
|
|
|
|
do_breakpoint(esr, regs);
|
|
|
|
|
debug_exception_exit(regs);
|
|
|
|
|
local_daif_restore(DAIF_PROCCTX);
|
|
|
|
|
exit_to_user_mode(regs);
|
|
|
|
|
arm64_exit_to_user_mode(regs);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static void noinstr el0_softstp(struct pt_regs *regs, unsigned long esr)
|
|
|
|
|
@ -886,7 +700,7 @@ static void noinstr el0_softstp(struct pt_regs *regs, unsigned long esr)
|
|
|
|
|
if (!is_ttbr0_addr(regs->pc))
|
|
|
|
|
arm64_apply_bp_hardening();
|
|
|
|
|
|
|
|
|
|
enter_from_user_mode(regs);
|
|
|
|
|
arm64_enter_from_user_mode(regs);
|
|
|
|
|
/*
|
|
|
|
|
* After handling a breakpoint, we suspend the breakpoint
|
|
|
|
|
* and use single-step to move to the next instruction.
|
|
|
|
|
@ -897,7 +711,7 @@ static void noinstr el0_softstp(struct pt_regs *regs, unsigned long esr)
|
|
|
|
|
local_daif_restore(DAIF_PROCCTX);
|
|
|
|
|
do_el0_softstep(esr, regs);
|
|
|
|
|
}
|
|
|
|
|
exit_to_user_mode(regs);
|
|
|
|
|
arm64_exit_to_user_mode(regs);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static void noinstr el0_watchpt(struct pt_regs *regs, unsigned long esr)
|
|
|
|
|
@ -905,39 +719,39 @@ static void noinstr el0_watchpt(struct pt_regs *regs, unsigned long esr)
|
|
|
|
|
/* Watchpoints are the only debug exception to write FAR_EL1 */
|
|
|
|
|
unsigned long far = read_sysreg(far_el1);
|
|
|
|
|
|
|
|
|
|
enter_from_user_mode(regs);
|
|
|
|
|
arm64_enter_from_user_mode(regs);
|
|
|
|
|
debug_exception_enter(regs);
|
|
|
|
|
do_watchpoint(far, esr, regs);
|
|
|
|
|
debug_exception_exit(regs);
|
|
|
|
|
local_daif_restore(DAIF_PROCCTX);
|
|
|
|
|
exit_to_user_mode(regs);
|
|
|
|
|
arm64_exit_to_user_mode(regs);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static void noinstr el0_brk64(struct pt_regs *regs, unsigned long esr)
|
|
|
|
|
{
|
|
|
|
|
enter_from_user_mode(regs);
|
|
|
|
|
arm64_enter_from_user_mode(regs);
|
|
|
|
|
local_daif_restore(DAIF_PROCCTX);
|
|
|
|
|
do_el0_brk64(esr, regs);
|
|
|
|
|
exit_to_user_mode(regs);
|
|
|
|
|
arm64_exit_to_user_mode(regs);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static void noinstr el0_svc(struct pt_regs *regs)
|
|
|
|
|
{
|
|
|
|
|
enter_from_user_mode(regs);
|
|
|
|
|
arm64_enter_from_user_mode(regs);
|
|
|
|
|
cortex_a76_erratum_1463225_svc_handler();
|
|
|
|
|
fpsimd_syscall_enter();
|
|
|
|
|
local_daif_restore(DAIF_PROCCTX);
|
|
|
|
|
do_el0_svc(regs);
|
|
|
|
|
exit_to_user_mode(regs);
|
|
|
|
|
arm64_exit_to_user_mode(regs);
|
|
|
|
|
fpsimd_syscall_exit();
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static void noinstr el0_fpac(struct pt_regs *regs, unsigned long esr)
|
|
|
|
|
{
|
|
|
|
|
enter_from_user_mode(regs);
|
|
|
|
|
arm64_enter_from_user_mode(regs);
|
|
|
|
|
local_daif_restore(DAIF_PROCCTX);
|
|
|
|
|
do_el0_fpac(regs, esr);
|
|
|
|
|
exit_to_user_mode(regs);
|
|
|
|
|
arm64_exit_to_user_mode(regs);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
asmlinkage void noinstr el0t_64_sync_handler(struct pt_regs *regs)
|
|
|
|
|
@ -1011,7 +825,7 @@ asmlinkage void noinstr el0t_64_sync_handler(struct pt_regs *regs)
|
|
|
|
|
static void noinstr el0_interrupt(struct pt_regs *regs,
|
|
|
|
|
void (*handler)(struct pt_regs *))
|
|
|
|
|
{
|
|
|
|
|
enter_from_user_mode(regs);
|
|
|
|
|
arm64_enter_from_user_mode(regs);
|
|
|
|
|
|
|
|
|
|
write_sysreg(DAIF_PROCCTX_NOIRQ, daif);
|
|
|
|
|
|
|
|
|
|
@ -1022,7 +836,7 @@ static void noinstr el0_interrupt(struct pt_regs *regs,
|
|
|
|
|
do_interrupt_handler(regs, handler);
|
|
|
|
|
irq_exit_rcu();
|
|
|
|
|
|
|
|
|
|
exit_to_user_mode(regs);
|
|
|
|
|
arm64_exit_to_user_mode(regs);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static void noinstr __el0_irq_handler_common(struct pt_regs *regs)
|
|
|
|
|
@ -1048,15 +862,15 @@ asmlinkage void noinstr el0t_64_fiq_handler(struct pt_regs *regs)
|
|
|
|
|
static void noinstr __el0_error_handler_common(struct pt_regs *regs)
|
|
|
|
|
{
|
|
|
|
|
unsigned long esr = read_sysreg(esr_el1);
|
|
|
|
|
arm64_irqentry_state_t state;
|
|
|
|
|
irqentry_state_t state;
|
|
|
|
|
|
|
|
|
|
enter_from_user_mode(regs);
|
|
|
|
|
arm64_enter_from_user_mode(regs);
|
|
|
|
|
local_daif_restore(DAIF_ERRCTX);
|
|
|
|
|
state = arm64_enter_nmi(regs);
|
|
|
|
|
state = irqentry_nmi_enter(regs);
|
|
|
|
|
do_serror(regs, esr);
|
|
|
|
|
arm64_exit_nmi(regs, state);
|
|
|
|
|
irqentry_nmi_exit(regs, state);
|
|
|
|
|
local_daif_restore(DAIF_PROCCTX);
|
|
|
|
|
exit_to_user_mode(regs);
|
|
|
|
|
arm64_exit_to_user_mode(regs);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
asmlinkage void noinstr el0t_64_error_handler(struct pt_regs *regs)
|
|
|
|
|
@ -1067,27 +881,27 @@ asmlinkage void noinstr el0t_64_error_handler(struct pt_regs *regs)
|
|
|
|
|
#ifdef CONFIG_COMPAT
|
|
|
|
|
static void noinstr el0_cp15(struct pt_regs *regs, unsigned long esr)
|
|
|
|
|
{
|
|
|
|
|
enter_from_user_mode(regs);
|
|
|
|
|
arm64_enter_from_user_mode(regs);
|
|
|
|
|
local_daif_restore(DAIF_PROCCTX);
|
|
|
|
|
do_el0_cp15(esr, regs);
|
|
|
|
|
exit_to_user_mode(regs);
|
|
|
|
|
arm64_exit_to_user_mode(regs);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static void noinstr el0_svc_compat(struct pt_regs *regs)
|
|
|
|
|
{
|
|
|
|
|
enter_from_user_mode(regs);
|
|
|
|
|
arm64_enter_from_user_mode(regs);
|
|
|
|
|
cortex_a76_erratum_1463225_svc_handler();
|
|
|
|
|
local_daif_restore(DAIF_PROCCTX);
|
|
|
|
|
do_el0_svc_compat(regs);
|
|
|
|
|
exit_to_user_mode(regs);
|
|
|
|
|
arm64_exit_to_user_mode(regs);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static void noinstr el0_bkpt32(struct pt_regs *regs, unsigned long esr)
|
|
|
|
|
{
|
|
|
|
|
enter_from_user_mode(regs);
|
|
|
|
|
arm64_enter_from_user_mode(regs);
|
|
|
|
|
local_daif_restore(DAIF_PROCCTX);
|
|
|
|
|
do_bkpt32(esr, regs);
|
|
|
|
|
exit_to_user_mode(regs);
|
|
|
|
|
arm64_exit_to_user_mode(regs);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
asmlinkage void noinstr el0t_32_sync_handler(struct pt_regs *regs)
|
|
|
|
|
@ -1166,7 +980,7 @@ asmlinkage void noinstr __noreturn handle_bad_stack(struct pt_regs *regs)
|
|
|
|
|
unsigned long esr = read_sysreg(esr_el1);
|
|
|
|
|
unsigned long far = read_sysreg(far_el1);
|
|
|
|
|
|
|
|
|
|
arm64_enter_nmi(regs);
|
|
|
|
|
irqentry_nmi_enter(regs);
|
|
|
|
|
panic_bad_stack(regs, esr, far);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
@ -1174,7 +988,7 @@ asmlinkage void noinstr __noreturn handle_bad_stack(struct pt_regs *regs)
|
|
|
|
|
asmlinkage noinstr unsigned long
|
|
|
|
|
__sdei_handler(struct pt_regs *regs, struct sdei_registered_event *arg)
|
|
|
|
|
{
|
|
|
|
|
arm64_irqentry_state_t state;
|
|
|
|
|
irqentry_state_t state;
|
|
|
|
|
unsigned long ret;
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
@ -1199,9 +1013,9 @@ __sdei_handler(struct pt_regs *regs, struct sdei_registered_event *arg)
|
|
|
|
|
else if (cpu_has_pan())
|
|
|
|
|
set_pstate_pan(0);
|
|
|
|
|
|
|
|
|
|
state = arm64_enter_nmi(regs);
|
|
|
|
|
state = irqentry_nmi_enter(regs);
|
|
|
|
|
ret = do_sdei_event(regs, arg);
|
|
|
|
|
arm64_exit_nmi(regs, state);
|
|
|
|
|
irqentry_nmi_exit(regs, state);
|
|
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
|
}
|
|
|
|
|
|