#ifndef _ASM_X86_SWITCH_TO_H #define _ASM_X86_SWITCH_TO_H struct task_struct; /* one of the stranger aspects of C forward declarations */ struct task_struct *__switch_to_asm(struct task_struct *prev, struct task_struct *next); __visible struct task_struct *__switch_to(struct task_struct *prev, struct task_struct *next); /* This runs runs on the previous thread's stack. */ static inline void prepare_switch_to(struct task_struct *prev, struct task_struct *next) { #ifdef CONFIG_VMAP_STACK /* * If we switch to a stack that has a top-level paging entry * that is not present in the current mm, the resulting #PF will * will be promoted to a double-fault and we'll panic. Probe * the new stack now so that vmalloc_fault can fix up the page * tables if needed. This can only happen if we use a stack * in vmap space. * * We assume that the stack is aligned so that it never spans * more than one top-level paging entry. * * To minimize cache pollution, just follow the stack pointer. */ READ_ONCE(*(unsigned char *)next->thread.sp); #endif } asmlinkage void ret_from_fork(void); /* data that is pointed to by thread.sp */ struct inactive_task_frame { unsigned long flags; #ifdef CONFIG_X86_64 unsigned long r15; unsigned long r14; unsigned long r13; unsigned long r12; #else unsigned long si; unsigned long di; #endif unsigned long bx; unsigned long bp; unsigned long ret_addr; }; struct fork_frame { struct inactive_task_frame frame; struct pt_regs regs; }; #define switch_to(prev, next, last) \ do { \ prepare_switch_to(prev, next); \ \ ((last) = __switch_to_asm((prev), (next))); \ } while (0) #endif /* _ASM_X86_SWITCH_TO_H */