x86/alternatives: Discard dynamic check after init
[deliverable/linux.git] / arch / x86 / include / asm / cpufeature.h
1 #ifndef _ASM_X86_CPUFEATURE_H
2 #define _ASM_X86_CPUFEATURE_H
3
4 #include <asm/processor.h>
5
6 #if defined(__KERNEL__) && !defined(__ASSEMBLY__)
7
8 #include <asm/asm.h>
9 #include <linux/bitops.h>
10
11 enum cpuid_leafs
12 {
13 CPUID_1_EDX = 0,
14 CPUID_8000_0001_EDX,
15 CPUID_8086_0001_EDX,
16 CPUID_LNX_1,
17 CPUID_1_ECX,
18 CPUID_C000_0001_EDX,
19 CPUID_8000_0001_ECX,
20 CPUID_LNX_2,
21 CPUID_LNX_3,
22 CPUID_7_0_EBX,
23 CPUID_D_1_EAX,
24 CPUID_F_0_EDX,
25 CPUID_F_1_EDX,
26 CPUID_8000_0008_EBX,
27 CPUID_6_EAX,
28 CPUID_8000_000A_EDX,
29 };
30
31 #ifdef CONFIG_X86_FEATURE_NAMES
32 extern const char * const x86_cap_flags[NCAPINTS*32];
33 extern const char * const x86_power_flags[32];
34 #define X86_CAP_FMT "%s"
35 #define x86_cap_flag(flag) x86_cap_flags[flag]
36 #else
37 #define X86_CAP_FMT "%d:%d"
38 #define x86_cap_flag(flag) ((flag) >> 5), ((flag) & 31)
39 #endif
40
41 /*
42 * In order to save room, we index into this array by doing
43 * X86_BUG_<name> - NCAPINTS*32.
44 */
45 extern const char * const x86_bug_flags[NBUGINTS*32];
46
47 #define test_cpu_cap(c, bit) \
48 test_bit(bit, (unsigned long *)((c)->x86_capability))
49
50 #define REQUIRED_MASK_BIT_SET(bit) \
51 ( (((bit)>>5)==0 && (1UL<<((bit)&31) & REQUIRED_MASK0)) || \
52 (((bit)>>5)==1 && (1UL<<((bit)&31) & REQUIRED_MASK1)) || \
53 (((bit)>>5)==2 && (1UL<<((bit)&31) & REQUIRED_MASK2)) || \
54 (((bit)>>5)==3 && (1UL<<((bit)&31) & REQUIRED_MASK3)) || \
55 (((bit)>>5)==4 && (1UL<<((bit)&31) & REQUIRED_MASK4)) || \
56 (((bit)>>5)==5 && (1UL<<((bit)&31) & REQUIRED_MASK5)) || \
57 (((bit)>>5)==6 && (1UL<<((bit)&31) & REQUIRED_MASK6)) || \
58 (((bit)>>5)==7 && (1UL<<((bit)&31) & REQUIRED_MASK7)) || \
59 (((bit)>>5)==8 && (1UL<<((bit)&31) & REQUIRED_MASK8)) || \
60 (((bit)>>5)==9 && (1UL<<((bit)&31) & REQUIRED_MASK9)) )
61
62 #define DISABLED_MASK_BIT_SET(bit) \
63 ( (((bit)>>5)==0 && (1UL<<((bit)&31) & DISABLED_MASK0)) || \
64 (((bit)>>5)==1 && (1UL<<((bit)&31) & DISABLED_MASK1)) || \
65 (((bit)>>5)==2 && (1UL<<((bit)&31) & DISABLED_MASK2)) || \
66 (((bit)>>5)==3 && (1UL<<((bit)&31) & DISABLED_MASK3)) || \
67 (((bit)>>5)==4 && (1UL<<((bit)&31) & DISABLED_MASK4)) || \
68 (((bit)>>5)==5 && (1UL<<((bit)&31) & DISABLED_MASK5)) || \
69 (((bit)>>5)==6 && (1UL<<((bit)&31) & DISABLED_MASK6)) || \
70 (((bit)>>5)==7 && (1UL<<((bit)&31) & DISABLED_MASK7)) || \
71 (((bit)>>5)==8 && (1UL<<((bit)&31) & DISABLED_MASK8)) || \
72 (((bit)>>5)==9 && (1UL<<((bit)&31) & DISABLED_MASK9)) )
73
74 #define cpu_has(c, bit) \
75 (__builtin_constant_p(bit) && REQUIRED_MASK_BIT_SET(bit) ? 1 : \
76 test_cpu_cap(c, bit))
77
78 #define this_cpu_has(bit) \
79 (__builtin_constant_p(bit) && REQUIRED_MASK_BIT_SET(bit) ? 1 : \
80 x86_this_cpu_test_bit(bit, (unsigned long *)&cpu_info.x86_capability))
81
82 /*
83 * This macro is for detection of features which need kernel
84 * infrastructure to be used. It may *not* directly test the CPU
85 * itself. Use the cpu_has() family if you want true runtime
86 * testing of CPU features, like in hypervisor code where you are
87 * supporting a possible guest feature where host support for it
88 * is not relevant.
89 */
90 #define cpu_feature_enabled(bit) \
91 (__builtin_constant_p(bit) && DISABLED_MASK_BIT_SET(bit) ? 0 : \
92 cpu_has(&boot_cpu_data, bit))
93
94 #define boot_cpu_has(bit) cpu_has(&boot_cpu_data, bit)
95
96 #define set_cpu_cap(c, bit) set_bit(bit, (unsigned long *)((c)->x86_capability))
97 #define clear_cpu_cap(c, bit) clear_bit(bit, (unsigned long *)((c)->x86_capability))
98 #define setup_clear_cpu_cap(bit) do { \
99 clear_cpu_cap(&boot_cpu_data, bit); \
100 set_bit(bit, (unsigned long *)cpu_caps_cleared); \
101 } while (0)
102 #define setup_force_cpu_cap(bit) do { \
103 set_cpu_cap(&boot_cpu_data, bit); \
104 set_bit(bit, (unsigned long *)cpu_caps_set); \
105 } while (0)
106
107 #define cpu_has_fpu boot_cpu_has(X86_FEATURE_FPU)
108 #define cpu_has_pse boot_cpu_has(X86_FEATURE_PSE)
109 #define cpu_has_tsc boot_cpu_has(X86_FEATURE_TSC)
110 #define cpu_has_pge boot_cpu_has(X86_FEATURE_PGE)
111 #define cpu_has_apic boot_cpu_has(X86_FEATURE_APIC)
112 #define cpu_has_fxsr boot_cpu_has(X86_FEATURE_FXSR)
113 #define cpu_has_xmm boot_cpu_has(X86_FEATURE_XMM)
114 #define cpu_has_xmm2 boot_cpu_has(X86_FEATURE_XMM2)
115 #define cpu_has_aes boot_cpu_has(X86_FEATURE_AES)
116 #define cpu_has_avx boot_cpu_has(X86_FEATURE_AVX)
117 #define cpu_has_avx2 boot_cpu_has(X86_FEATURE_AVX2)
118 #define cpu_has_clflush boot_cpu_has(X86_FEATURE_CLFLUSH)
119 #define cpu_has_gbpages boot_cpu_has(X86_FEATURE_GBPAGES)
120 #define cpu_has_arch_perfmon boot_cpu_has(X86_FEATURE_ARCH_PERFMON)
121 #define cpu_has_pat boot_cpu_has(X86_FEATURE_PAT)
122 #define cpu_has_x2apic boot_cpu_has(X86_FEATURE_X2APIC)
123 #define cpu_has_xsave boot_cpu_has(X86_FEATURE_XSAVE)
124 #define cpu_has_xsaves boot_cpu_has(X86_FEATURE_XSAVES)
125 #define cpu_has_osxsave boot_cpu_has(X86_FEATURE_OSXSAVE)
126 #define cpu_has_hypervisor boot_cpu_has(X86_FEATURE_HYPERVISOR)
127 /*
128 * Do not add any more of those clumsy macros - use static_cpu_has() for
129 * fast paths and boot_cpu_has() otherwise!
130 */
131
132 #if defined(CC_HAVE_ASM_GOTO) && defined(CONFIG_X86_FAST_FEATURE_TESTS)
133 /*
134 * Static testing of CPU features. Used the same as boot_cpu_has().
135 * These will statically patch the target code for additional
136 * performance.
137 */
138 static __always_inline __pure bool _static_cpu_has(u16 bit)
139 {
140 asm_volatile_goto("1: jmp 6f\n"
141 "2:\n"
142 ".skip -(((5f-4f) - (2b-1b)) > 0) * "
143 "((5f-4f) - (2b-1b)),0x90\n"
144 "3:\n"
145 ".section .altinstructions,\"a\"\n"
146 " .long 1b - .\n" /* src offset */
147 " .long 4f - .\n" /* repl offset */
148 " .word %P1\n" /* always replace */
149 " .byte 3b - 1b\n" /* src len */
150 " .byte 5f - 4f\n" /* repl len */
151 " .byte 3b - 2b\n" /* pad len */
152 ".previous\n"
153 ".section .altinstr_replacement,\"ax\"\n"
154 "4: jmp %l[t_no]\n"
155 "5:\n"
156 ".previous\n"
157 ".section .altinstructions,\"a\"\n"
158 " .long 1b - .\n" /* src offset */
159 " .long 0\n" /* no replacement */
160 " .word %P0\n" /* feature bit */
161 " .byte 3b - 1b\n" /* src len */
162 " .byte 0\n" /* repl len */
163 " .byte 0\n" /* pad len */
164 ".previous\n"
165 ".section .altinstr_aux,\"ax\"\n"
166 "6:\n"
167 " testb %[bitnum],%[cap_byte]\n"
168 " jnz %l[t_yes]\n"
169 " jmp %l[t_no]\n"
170 ".previous\n"
171 : : "i" (bit), "i" (X86_FEATURE_ALWAYS),
172 [bitnum] "i" (1 << (bit & 7)),
173 [cap_byte] "m" (((const char *)boot_cpu_data.x86_capability)[bit >> 3])
174 : : t_yes, t_no);
175 t_yes:
176 return true;
177 t_no:
178 return false;
179 }
180
181 #define static_cpu_has(bit) \
182 ( \
183 __builtin_constant_p(boot_cpu_has(bit)) ? \
184 boot_cpu_has(bit) : \
185 _static_cpu_has(bit) \
186 )
187 #else
188 /*
189 * Fall back to dynamic for gcc versions which don't support asm goto. Should be
190 * a minority now anyway.
191 */
192 #define static_cpu_has(bit) boot_cpu_has(bit)
193 #endif
194
195 #define cpu_has_bug(c, bit) cpu_has(c, (bit))
196 #define set_cpu_bug(c, bit) set_cpu_cap(c, (bit))
197 #define clear_cpu_bug(c, bit) clear_cpu_cap(c, (bit))
198
199 #define static_cpu_has_bug(bit) static_cpu_has((bit))
200 #define boot_cpu_has_bug(bit) cpu_has_bug(&boot_cpu_data, (bit))
201
202 #define MAX_CPU_FEATURES (NCAPINTS * 32)
203 #define cpu_have_feature boot_cpu_has
204
205 #define CPU_FEATURE_TYPEFMT "x86,ven%04Xfam%04Xmod%04X"
206 #define CPU_FEATURE_TYPEVAL boot_cpu_data.x86_vendor, boot_cpu_data.x86, \
207 boot_cpu_data.x86_model
208
209 #endif /* defined(__KERNEL__) && !defined(__ASSEMBLY__) */
210 #endif /* _ASM_X86_CPUFEATURE_H */
This page took 0.061692 seconds and 5 git commands to generate.