2 * Copyright (C) 1994 Linus Torvalds
4 * Pentium III FXSR, SSE support
5 * General FPU state handling cleanups
6 * Gareth Hughes <gareth@valinux.com>, May 2000
7 * x86-64 work by Andi Kleen 2002
10 #ifndef _ASM_X86_I387_H
11 #define _ASM_X86_I387_H
15 #include <linux/sched.h>
16 #include <linux/hardirq.h>
19 struct user_i387_struct
;
21 extern int init_fpu(struct task_struct
*child
);
22 extern void fpu_finit(struct fpu
*fpu
);
23 extern int dump_fpu(struct pt_regs
*, struct user_i387_struct
*);
24 extern void math_state_restore(void);
26 extern bool irq_fpu_usable(void);
29 * Careful: __kernel_fpu_begin/end() must be called with preempt disabled
30 * and they don't touch the preempt state on their own.
31 * If you enable preemption after __kernel_fpu_begin(), preempt notifier
32 * should call the __kernel_fpu_end() to prevent the kernel/user FPU
33 * state from getting corrupted. KVM for example uses this model.
35 * All other cases use kernel_fpu_begin/end() which disable preemption
36 * during kernel FPU usage.
38 extern void __kernel_fpu_begin(void);
39 extern void __kernel_fpu_end(void);
41 static inline void kernel_fpu_begin(void)
43 WARN_ON_ONCE(!irq_fpu_usable());
48 static inline void kernel_fpu_end(void)
55 * Some instructions like VIA's padlock instructions generate a spurious
56 * DNA fault but don't modify SSE registers. And these instructions
57 * get used from interrupt context as well. To prevent these kernel instructions
58 * in interrupt context interacting wrongly with other user/kernel fpu usage, we
59 * should use them only in the context of irq_ts_save/restore()
61 static inline int irq_ts_save(void)
64 * If in process context and not atomic, we can take a spurious DNA fault.
65 * Otherwise, doing clts() in process context requires disabling preemption
66 * or some heavy lifting like kernel_fpu_begin()
71 if (read_cr0() & X86_CR0_TS
) {
79 static inline void irq_ts_restore(int TS_state
)
86 * The question "does this thread have fpu access?"
87 * is slightly racy, since preemption could come in
88 * and revoke it immediately after the test.
90 * However, even in that very unlikely scenario,
91 * we can just assume we have FPU access - typically
92 * to save the FP state - we'll just take a #NM
93 * fault and get the FPU access back.
95 static inline int user_has_fpu(void)
97 return current
->thread
.fpu
.has_fpu
;
100 extern void unlazy_fpu(struct task_struct
*tsk
);
102 #endif /* __ASSEMBLY__ */
104 #endif /* _ASM_X86_I387_H */