2020-05-21 16:05:23 -04:00
|
|
|
/* SPDX-License-Identifier: GPL-2.0 */
|
|
|
|
#ifndef _ASM_X86_IRQ_STACK_H
|
|
|
|
#define _ASM_X86_IRQ_STACK_H
|
|
|
|
|
|
|
|
#include <linux/ptrace.h>
|
2022-03-14 13:05:52 -04:00
|
|
|
#include <linux/objtool.h>
|
2020-05-21 16:05:23 -04:00
|
|
|
|
|
|
|
#include <asm/processor.h>
|
|
|
|
|
|
|
|
#ifdef CONFIG_X86_64
|
2021-02-09 18:40:46 -05:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Macro to inline switching to an interrupt stack and invoking function
|
|
|
|
* calls from there. The following rules apply:
|
|
|
|
*
|
|
|
|
* - Ordering:
|
|
|
|
*
|
|
|
|
* 1. Write the stack pointer into the top most place of the irq
|
|
|
|
* stack. This ensures that the various unwinders can link back to the
|
|
|
|
* original stack.
|
|
|
|
*
|
|
|
|
* 2. Switch the stack pointer to the top of the irq stack.
|
|
|
|
*
|
|
|
|
* 3. Invoke whatever needs to be done (@asm_call argument)
|
|
|
|
*
|
|
|
|
* 4. Pop the original stack pointer from the top of the irq stack
|
|
|
|
* which brings it back to the original stack where it left off.
|
|
|
|
*
|
|
|
|
* - Function invocation:
|
|
|
|
*
|
|
|
|
* To allow flexible usage of the macro, the actual function code including
|
|
|
|
* the store of the arguments in the call ABI registers is handed in via
|
|
|
|
* the @asm_call argument.
|
|
|
|
*
|
|
|
|
* - Local variables:
|
|
|
|
*
|
|
|
|
* @tos:
|
|
|
|
* The @tos variable holds a pointer to the top of the irq stack and
|
|
|
|
* _must_ be allocated in a non-callee saved register as this is a
|
|
|
|
* restriction coming from objtool.
|
|
|
|
*
|
|
|
|
* Note, that (tos) is both in input and output constraints to ensure
|
|
|
|
* that the compiler does not assume that R11 is left untouched in
|
|
|
|
* case this macro is used in some place where the per cpu interrupt
|
|
|
|
* stack pointer is used again afterwards
|
|
|
|
*
|
|
|
|
* - Function arguments:
|
|
|
|
* The function argument(s), if any, have to be defined in register
|
|
|
|
* variables at the place where this is invoked. Storing the
|
|
|
|
* argument(s) in the proper register(s) is part of the @asm_call
|
|
|
|
*
|
|
|
|
* - Constraints:
|
|
|
|
*
|
|
|
|
* The constraints have to be done very carefully because the compiler
|
|
|
|
* does not know about the assembly call.
|
|
|
|
*
|
|
|
|
* output:
|
|
|
|
* As documented already above the @tos variable is required to be in
|
|
|
|
* the output constraints to make the compiler aware that R11 cannot be
|
|
|
|
* reused after the asm() statement.
|
|
|
|
*
|
2021-10-05 15:48:30 -04:00
|
|
|
* For builds with CONFIG_UNWINDER_FRAME_POINTER, ASM_CALL_CONSTRAINT is
|
2021-02-09 18:40:46 -05:00
|
|
|
* required as well as this prevents certain creative GCC variants from
|
|
|
|
* misplacing the ASM code.
|
|
|
|
*
|
|
|
|
* input:
|
|
|
|
* - func:
|
|
|
|
* Immediate, which tells the compiler that the function is referenced.
|
|
|
|
*
|
|
|
|
* - tos:
|
|
|
|
* Register. The actual register is defined by the variable declaration.
|
|
|
|
*
|
|
|
|
* - function arguments:
|
|
|
|
* The constraints are handed in via the 'argconstr' argument list. They
|
|
|
|
* describe the register arguments which are used in @asm_call.
|
|
|
|
*
|
|
|
|
* clobbers:
|
|
|
|
* Function calls can clobber anything except the callee-saved
|
|
|
|
* registers. Tell the compiler.
|
|
|
|
*/
|
2021-09-15 11:12:59 -04:00
|
|
|
#define call_on_stack(stack, func, asm_call, argconstr...) \
|
2021-02-09 18:40:46 -05:00
|
|
|
{ \
|
|
|
|
register void *tos asm("r11"); \
|
|
|
|
\
|
2021-09-15 11:12:59 -04:00
|
|
|
tos = ((void *)(stack)); \
|
2021-02-09 18:40:46 -05:00
|
|
|
\
|
|
|
|
asm_inline volatile( \
|
|
|
|
"movq %%rsp, (%[tos]) \n" \
|
|
|
|
"movq %[tos], %%rsp \n" \
|
|
|
|
\
|
|
|
|
asm_call \
|
|
|
|
\
|
|
|
|
"popq %%rsp \n" \
|
|
|
|
\
|
|
|
|
: "+r" (tos), ASM_CALL_CONSTRAINT \
|
|
|
|
: [__func] "i" (func), [tos] "r" (tos) argconstr \
|
|
|
|
: "cc", "rax", "rcx", "rdx", "rsi", "rdi", "r8", "r9", "r10", \
|
|
|
|
"memory" \
|
|
|
|
); \
|
|
|
|
}
|
|
|
|
|
2021-09-15 11:12:59 -04:00
|
|
|
#define ASM_CALL_ARG0 \
|
2022-03-08 10:30:50 -05:00
|
|
|
"call %P[__func] \n" \
|
|
|
|
ASM_REACHABLE
|
2021-09-15 11:12:59 -04:00
|
|
|
|
|
|
|
#define ASM_CALL_ARG1 \
|
|
|
|
"movq %[arg1], %%rdi \n" \
|
|
|
|
ASM_CALL_ARG0
|
|
|
|
|
|
|
|
#define ASM_CALL_ARG2 \
|
|
|
|
"movq %[arg2], %%rsi \n" \
|
|
|
|
ASM_CALL_ARG1
|
|
|
|
|
|
|
|
#define ASM_CALL_ARG3 \
|
|
|
|
"movq %[arg3], %%rdx \n" \
|
|
|
|
ASM_CALL_ARG2
|
|
|
|
|
|
|
|
#define call_on_irqstack(func, asm_call, argconstr...) \
|
2022-09-15 07:11:05 -04:00
|
|
|
call_on_stack(__this_cpu_read(pcpu_hot.hardirq_stack_ptr), \
|
2021-09-15 11:12:59 -04:00
|
|
|
func, asm_call, argconstr)
|
|
|
|
|
2021-02-09 18:40:46 -05:00
|
|
|
/* Macros to assert type correctness for run_*_on_irqstack macros */
|
|
|
|
#define assert_function_type(func, proto) \
|
|
|
|
static_assert(__builtin_types_compatible_p(typeof(&func), proto))
|
|
|
|
|
|
|
|
#define assert_arg_type(arg, proto) \
|
|
|
|
static_assert(__builtin_types_compatible_p(typeof(arg), proto))
|
|
|
|
|
2021-02-09 18:40:47 -05:00
|
|
|
/*
|
|
|
|
* Macro to invoke system vector and device interrupt C handlers.
|
|
|
|
*/
|
|
|
|
#define call_on_irqstack_cond(func, regs, asm_call, constr, c_args...) \
|
|
|
|
{ \
|
|
|
|
/* \
|
|
|
|
* User mode entry and interrupt on the irq stack do not \
|
|
|
|
* switch stacks. If from user mode the task stack is empty. \
|
|
|
|
*/ \
|
2022-09-15 07:11:05 -04:00
|
|
|
if (user_mode(regs) || __this_cpu_read(pcpu_hot.hardirq_stack_inuse)) { \
|
2021-02-09 18:40:47 -05:00
|
|
|
irq_enter_rcu(); \
|
|
|
|
func(c_args); \
|
|
|
|
irq_exit_rcu(); \
|
|
|
|
} else { \
|
|
|
|
/* \
|
|
|
|
* Mark the irq stack inuse _before_ and unmark _after_ \
|
|
|
|
* switching stacks. Interrupts are disabled in both \
|
|
|
|
* places. Invoke the stack switch macro with the call \
|
|
|
|
* sequence which matches the above direct invocation. \
|
|
|
|
*/ \
|
2022-09-15 07:11:05 -04:00
|
|
|
__this_cpu_write(pcpu_hot.hardirq_stack_inuse, true); \
|
2021-02-09 18:40:47 -05:00
|
|
|
call_on_irqstack(func, asm_call, constr); \
|
2022-09-15 07:11:05 -04:00
|
|
|
__this_cpu_write(pcpu_hot.hardirq_stack_inuse, false); \
|
2021-02-09 18:40:47 -05:00
|
|
|
} \
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Function call sequence for __call_on_irqstack() for system vectors.
|
|
|
|
*
|
|
|
|
* Note that irq_enter_rcu() and irq_exit_rcu() do not use the input
|
|
|
|
* mechanism because these functions are global and cannot be optimized out
|
|
|
|
* when compiling a particular source file which uses one of these macros.
|
|
|
|
*
|
|
|
|
* The argument (regs) does not need to be pushed or stashed in a callee
|
|
|
|
* saved register to be safe vs. the irq_enter_rcu() call because the
|
|
|
|
* clobbers already prevent the compiler from storing it in a callee
|
|
|
|
* clobbered register. As the compiler has to preserve @regs for the final
|
|
|
|
* call to idtentry_exit() anyway, it's likely that it does not cause extra
|
|
|
|
* effort for this asm magic.
|
|
|
|
*/
|
|
|
|
#define ASM_CALL_SYSVEC \
|
|
|
|
"call irq_enter_rcu \n" \
|
2021-09-15 11:12:59 -04:00
|
|
|
ASM_CALL_ARG1 \
|
2021-02-09 18:40:47 -05:00
|
|
|
"call irq_exit_rcu \n"
|
|
|
|
|
|
|
|
#define SYSVEC_CONSTRAINTS , [arg1] "r" (regs)
|
|
|
|
|
|
|
|
#define run_sysvec_on_irqstack_cond(func, regs) \
|
|
|
|
{ \
|
|
|
|
assert_function_type(func, void (*)(struct pt_regs *)); \
|
|
|
|
assert_arg_type(regs, struct pt_regs *); \
|
|
|
|
\
|
|
|
|
call_on_irqstack_cond(func, regs, ASM_CALL_SYSVEC, \
|
|
|
|
SYSVEC_CONSTRAINTS, regs); \
|
|
|
|
}
|
|
|
|
|
2021-02-09 18:40:48 -05:00
|
|
|
/*
|
|
|
|
* As in ASM_CALL_SYSVEC above the clobbers force the compiler to store
|
|
|
|
* @regs and @vector in callee saved registers.
|
|
|
|
*/
|
|
|
|
#define ASM_CALL_IRQ \
|
|
|
|
"call irq_enter_rcu \n" \
|
2021-09-15 11:12:59 -04:00
|
|
|
ASM_CALL_ARG2 \
|
2021-02-09 18:40:48 -05:00
|
|
|
"call irq_exit_rcu \n"
|
|
|
|
|
2021-09-15 11:12:59 -04:00
|
|
|
#define IRQ_CONSTRAINTS , [arg1] "r" (regs), [arg2] "r" ((unsigned long)vector)
|
2021-02-09 18:40:48 -05:00
|
|
|
|
|
|
|
#define run_irq_on_irqstack_cond(func, regs, vector) \
|
|
|
|
{ \
|
|
|
|
assert_function_type(func, void (*)(struct pt_regs *, u32)); \
|
|
|
|
assert_arg_type(regs, struct pt_regs *); \
|
|
|
|
assert_arg_type(vector, u32); \
|
|
|
|
\
|
|
|
|
call_on_irqstack_cond(func, regs, ASM_CALL_IRQ, \
|
|
|
|
IRQ_CONSTRAINTS, regs, vector); \
|
|
|
|
}
|
|
|
|
|
2022-08-25 04:25:05 -04:00
|
|
|
#ifdef CONFIG_SOFTIRQ_ON_OWN_STACK
|
2021-02-09 18:40:50 -05:00
|
|
|
/*
|
2021-02-09 18:40:51 -05:00
|
|
|
* Macro to invoke __do_softirq on the irq stack. This is only called from
|
2021-03-18 10:28:01 -04:00
|
|
|
* task context when bottom halves are about to be reenabled and soft
|
2021-02-09 18:40:51 -05:00
|
|
|
* interrupts are pending to be processed. The interrupt stack cannot be in
|
|
|
|
* use here.
|
2021-02-09 18:40:50 -05:00
|
|
|
*/
|
2021-02-09 18:40:54 -05:00
|
|
|
#define do_softirq_own_stack() \
|
2021-02-09 18:40:50 -05:00
|
|
|
{ \
|
2022-09-15 07:11:05 -04:00
|
|
|
__this_cpu_write(pcpu_hot.hardirq_stack_inuse, true); \
|
2021-09-15 11:12:59 -04:00
|
|
|
call_on_irqstack(__do_softirq, ASM_CALL_ARG0); \
|
2022-09-15 07:11:05 -04:00
|
|
|
__this_cpu_write(pcpu_hot.hardirq_stack_inuse, false); \
|
2020-09-22 03:58:52 -04:00
|
|
|
}
|
|
|
|
|
2021-09-24 12:12:45 -04:00
|
|
|
#endif
|
|
|
|
|
2020-05-21 16:05:23 -04:00
|
|
|
#else /* CONFIG_X86_64 */
|
2021-02-09 18:40:47 -05:00
|
|
|
/* System vector handlers always run on the stack they interrupted. */
|
|
|
|
#define run_sysvec_on_irqstack_cond(func, regs) \
|
|
|
|
{ \
|
|
|
|
irq_enter_rcu(); \
|
|
|
|
func(regs); \
|
|
|
|
irq_exit_rcu(); \
|
|
|
|
}
|
|
|
|
|
2021-02-09 18:40:48 -05:00
|
|
|
/* Switches to the irq stack within func() */
|
|
|
|
#define run_irq_on_irqstack_cond(func, regs, vector) \
|
|
|
|
{ \
|
|
|
|
irq_enter_rcu(); \
|
|
|
|
func(regs, vector); \
|
|
|
|
irq_exit_rcu(); \
|
|
|
|
}
|
|
|
|
|
2020-05-21 16:05:23 -04:00
|
|
|
#endif /* !CONFIG_X86_64 */
|
|
|
|
|
|
|
|
#endif
|