blob: 89ae3e051741354e9474cff0dc1a14501dc2a9a0 [file] [log] [blame]
Roland McGrath1eeaed72008-01-30 13:31:51 +01001/*
2 * Copyright (C) 1994 Linus Torvalds
3 *
4 * Pentium III FXSR, SSE support
5 * General FPU state handling cleanups
6 * Gareth Hughes <gareth@valinux.com>, May 2000
7 * x86-64 work by Andi Kleen 2002
8 */
9
H. Peter Anvin1965aae2008-10-22 22:26:29 -070010#ifndef _ASM_X86_I387_H
11#define _ASM_X86_I387_H
Roland McGrath1eeaed72008-01-30 13:31:51 +010012
Herbert Xu3b0d6592009-11-03 09:11:15 -050013#ifndef __ASSEMBLY__
14
Roland McGrath1eeaed72008-01-30 13:31:51 +010015#include <linux/sched.h>
Suresh Siddhae4914012008-08-13 22:02:26 +100016#include <linux/hardirq.h>
Roland McGrath1eeaed72008-01-30 13:31:51 +010017
Linus Torvalds1361b832012-02-21 13:19:22 -080018struct pt_regs;
19struct user_i387_struct;
20
Ingo Molnar97185c92015-04-03 12:02:02 +020021extern int fpstate_alloc_init(struct task_struct *curr);
Ingo Molnarc0ee2cf2015-04-03 13:01:52 +020022extern void fpstate_init(struct fpu *fpu);
Ingo Molnar81683cc2015-04-22 11:52:13 +020023extern void fpu__flush_thread(struct task_struct *tsk);
Ingo Molnar97185c92015-04-03 12:02:02 +020024
Jaswinder Singh36454932008-07-21 22:31:57 +053025extern int dump_fpu(struct pt_regs *, struct user_i387_struct *);
Ingo Molnar3a0aee42015-04-22 13:16:47 +020026extern void fpu__restore(void);
Ingo Molnar4d164092015-04-22 13:44:25 +020027extern void fpu__init_check_bugs(void);
Roland McGrath1eeaed72008-01-30 13:31:51 +010028
Linus Torvalds8546c002012-02-21 10:25:45 -080029extern bool irq_fpu_usable(void);
Suresh Siddhab1a74bf2012-09-20 11:01:49 -070030
31/*
32 * Careful: __kernel_fpu_begin/end() must be called with preempt disabled
33 * and they don't touch the preempt state on their own.
34 * If you enable preemption after __kernel_fpu_begin(), preempt notifier
35 * should call the __kernel_fpu_end() to prevent the kernel/user FPU
36 * state from getting corrupted. KVM for example uses this model.
37 *
38 * All other cases use kernel_fpu_begin/end() which disable preemption
39 * during kernel FPU usage.
40 */
41extern void __kernel_fpu_begin(void);
42extern void __kernel_fpu_end(void);
43
44static inline void kernel_fpu_begin(void)
45{
Suresh Siddhab1a74bf2012-09-20 11:01:49 -070046 preempt_disable();
Oleg Nesterov14e153e2015-01-15 20:19:43 +010047 WARN_ON_ONCE(!irq_fpu_usable());
Suresh Siddhab1a74bf2012-09-20 11:01:49 -070048 __kernel_fpu_begin();
49}
50
51static inline void kernel_fpu_end(void)
52{
53 __kernel_fpu_end();
54 preempt_enable();
55}
Roland McGrath1eeaed72008-01-30 13:31:51 +010056
Oleg Nesterov75756372015-01-15 20:20:28 +010057/* Must be called with preempt disabled */
58extern void kernel_fpu_disable(void);
59extern void kernel_fpu_enable(void);
60
Suresh Siddhae4914012008-08-13 22:02:26 +100061/*
62 * Some instructions like VIA's padlock instructions generate a spurious
63 * DNA fault but don't modify SSE registers. And these instructions
Chuck Ebbert0b8c3d52009-06-09 10:40:50 -040064 * get used from interrupt context as well. To prevent these kernel instructions
65 * in interrupt context interacting wrongly with other user/kernel fpu usage, we
Suresh Siddhae4914012008-08-13 22:02:26 +100066 * should use them only in the context of irq_ts_save/restore()
67 */
68static inline int irq_ts_save(void)
69{
70 /*
Chuck Ebbert0b8c3d52009-06-09 10:40:50 -040071 * If in process context and not atomic, we can take a spurious DNA fault.
72 * Otherwise, doing clts() in process context requires disabling preemption
73 * or some heavy lifting like kernel_fpu_begin()
Suresh Siddhae4914012008-08-13 22:02:26 +100074 */
Chuck Ebbert0b8c3d52009-06-09 10:40:50 -040075 if (!in_atomic())
Suresh Siddhae4914012008-08-13 22:02:26 +100076 return 0;
77
78 if (read_cr0() & X86_CR0_TS) {
79 clts();
80 return 1;
81 }
82
83 return 0;
84}
85
86static inline void irq_ts_restore(int TS_state)
87{
88 if (TS_state)
89 stts();
90}
91
Roland McGrath1eeaed72008-01-30 13:31:51 +010092/*
Linus Torvalds15d87912012-02-16 09:15:04 -080093 * The question "does this thread have fpu access?"
94 * is slightly racy, since preemption could come in
95 * and revoke it immediately after the test.
96 *
97 * However, even in that very unlikely scenario,
98 * we can just assume we have FPU access - typically
99 * to save the FP state - we'll just take a #NM
100 * fault and get the FPU access back.
Linus Torvalds15d87912012-02-16 09:15:04 -0800101 */
102static inline int user_has_fpu(void)
103{
Linus Torvalds1361b832012-02-21 13:19:22 -0800104 return current->thread.fpu.has_fpu;
Roland McGrath1eeaed72008-01-30 13:31:51 +0100105}
106
Ingo Molnar0a781552015-04-03 10:58:52 +0200107extern void fpu__save(struct task_struct *tsk);
Roland McGrath1eeaed72008-01-30 13:31:51 +0100108
Herbert Xu3b0d6592009-11-03 09:11:15 -0500109#endif /* __ASSEMBLY__ */
110
H. Peter Anvin1965aae2008-10-22 22:26:29 -0700111#endif /* _ASM_X86_I387_H */