/* SPDX-License-Identifier: GPL-2.0 */ /* * Copyright (C) 1999 Cort Dougan <cort@cs.nmt.edu> */ #ifndef _ASM_POWERPC_SWITCH_TO_H #define _ASM_POWERPC_SWITCH_TO_H #include <linux/sched.h> #include <asm/reg.h> struct thread_struct; struct task_struct; struct pt_regs; extern struct task_struct *__switch_to(struct task_struct *, struct task_struct *); #define switch_to(prev, next, last) ((last) = __switch_to((prev), (next))) extern struct task_struct *_switch(struct thread_struct *prev, struct thread_struct *next); extern void switch_booke_debug_regs(struct debug_reg *new_debug); extern int emulate_altivec(struct pt_regs *); #ifdef CONFIG_PPC_BOOK3S_64 void restore_math(struct pt_regs *regs); #else static inline void restore_math(struct pt_regs *regs) { } #endif void restore_tm_state(struct pt_regs *regs); extern void flush_all_to_thread(struct task_struct *); extern void giveup_all(struct task_struct *); #ifdef CONFIG_PPC_FPU extern void enable_kernel_fp(void); extern void flush_fp_to_thread(struct task_struct *); extern void giveup_fpu(struct task_struct *); extern void save_fpu(struct task_struct *); static inline void disable_kernel_fp(void) { msr_check_and_clear(MSR_FP); } #else static inline void save_fpu(struct task_struct *t) { } static inline void flush_fp_to_thread(struct task_struct *t) { } #endif #ifdef CONFIG_ALTIVEC extern void enable_kernel_altivec(void); extern void flush_altivec_to_thread(struct task_struct *); extern void giveup_altivec(struct task_struct *); extern void save_altivec(struct task_struct *); static inline void disable_kernel_altivec(void) { msr_check_and_clear(MSR_VEC); } #else static inline void save_altivec(struct task_struct *t) { } static inline void __giveup_altivec(struct task_struct *t) { } static inline void enable_kernel_altivec(void) { BUILD_BUG(); } static inline void disable_kernel_altivec(void) { BUILD_BUG(); } #endif #ifdef CONFIG_VSX extern void enable_kernel_vsx(void); extern void flush_vsx_to_thread(struct task_struct *); static inline void disable_kernel_vsx(void) { msr_check_and_clear(MSR_FP|MSR_VEC|MSR_VSX); } #else static inline void enable_kernel_vsx(void) { BUILD_BUG(); } static inline void disable_kernel_vsx(void) { BUILD_BUG(); } #endif #ifdef CONFIG_SPE extern void enable_kernel_spe(void); extern void flush_spe_to_thread(struct task_struct *); extern void giveup_spe(struct task_struct *); extern void __giveup_spe(struct task_struct *); static inline void disable_kernel_spe(void) { msr_check_and_clear(MSR_SPE); } #else static inline void __giveup_spe(struct task_struct *t) { } #endif static inline void clear_task_ebb(struct task_struct *t) { #ifdef CONFIG_PPC_BOOK3S_64 /* EBB perf events are not inherited, so clear all EBB state. */ t->thread.ebbrr = 0; t->thread.ebbhr = 0; t->thread.bescr = 0; t->thread.mmcr2 = 0; t->thread.mmcr0 = 0; t->thread.siar = 0; t->thread.sdar = 0; t->thread.sier = 0; t->thread.used_ebb = 0; #endif } void kvmppc_save_user_regs(void); void kvmppc_save_current_sprs(void); extern int set_thread_tidr(struct task_struct *t); #endif /* _ASM_POWERPC_SWITCH_TO_H */