[PATCH] x86-64: fixed size remaining fake nodes
[deliverable/linux.git] / arch / x86_64 / kernel / setup.c
CommitLineData
1da177e4
LT
1/*
2 * linux/arch/x86-64/kernel/setup.c
3 *
4 * Copyright (C) 1995 Linus Torvalds
5 *
6 * Nov 2001 Dave Jones <davej@suse.de>
7 * Forked from i386 setup code.
1da177e4
LT
8 */
9
10/*
11 * This file handles the architecture-dependent parts of initialization
12 */
13
14#include <linux/errno.h>
15#include <linux/sched.h>
16#include <linux/kernel.h>
17#include <linux/mm.h>
18#include <linux/stddef.h>
19#include <linux/unistd.h>
20#include <linux/ptrace.h>
21#include <linux/slab.h>
22#include <linux/user.h>
23#include <linux/a.out.h>
894673ee 24#include <linux/screen_info.h>
1da177e4
LT
25#include <linux/ioport.h>
26#include <linux/delay.h>
1da177e4
LT
27#include <linux/init.h>
28#include <linux/initrd.h>
29#include <linux/highmem.h>
30#include <linux/bootmem.h>
31#include <linux/module.h>
32#include <asm/processor.h>
33#include <linux/console.h>
34#include <linux/seq_file.h>
aac04b32 35#include <linux/crash_dump.h>
1da177e4
LT
36#include <linux/root_dev.h>
37#include <linux/pci.h>
38#include <linux/acpi.h>
39#include <linux/kallsyms.h>
40#include <linux/edd.h>
bbfceef4 41#include <linux/mmzone.h>
5f5609df 42#include <linux/kexec.h>
95235ca2 43#include <linux/cpufreq.h>
e9928674 44#include <linux/dmi.h>
17a941d8 45#include <linux/dma-mapping.h>
681558fd 46#include <linux/ctype.h>
bbfceef4 47
1da177e4
LT
48#include <asm/mtrr.h>
49#include <asm/uaccess.h>
50#include <asm/system.h>
51#include <asm/io.h>
52#include <asm/smp.h>
53#include <asm/msr.h>
54#include <asm/desc.h>
55#include <video/edid.h>
56#include <asm/e820.h>
57#include <asm/dma.h>
58#include <asm/mpspec.h>
59#include <asm/mmu_context.h>
60#include <asm/bootsetup.h>
61#include <asm/proto.h>
62#include <asm/setup.h>
63#include <asm/mach_apic.h>
64#include <asm/numa.h>
2bc0414e 65#include <asm/sections.h>
f2d3efed 66#include <asm/dmi.h>
1da177e4
LT
67
68/*
69 * Machine setup..
70 */
71
6c231b7b 72struct cpuinfo_x86 boot_cpu_data __read_mostly;
2ee60e17 73EXPORT_SYMBOL(boot_cpu_data);
1da177e4
LT
74
75unsigned long mmu_cr4_features;
76
1da177e4
LT
77/* Boot loader ID as an integer, for the benefit of proc_dointvec */
78int bootloader_type;
79
80unsigned long saved_video_mode;
81
f2d3efed
AK
82/*
83 * Early DMI memory
84 */
85int dmi_alloc_index;
86char dmi_alloc_data[DMI_MAX_DATA];
87
1da177e4
LT
88/*
89 * Setup options
90 */
1da177e4 91struct screen_info screen_info;
2ee60e17 92EXPORT_SYMBOL(screen_info);
1da177e4
LT
93struct sys_desc_table_struct {
94 unsigned short length;
95 unsigned char table[0];
96};
97
98struct edid_info edid_info;
ba70710e 99EXPORT_SYMBOL_GPL(edid_info);
1da177e4
LT
100
101extern int root_mountflags;
1da177e4 102
adf48856 103char __initdata command_line[COMMAND_LINE_SIZE];
1da177e4
LT
104
105struct resource standard_io_resources[] = {
106 { .name = "dma1", .start = 0x00, .end = 0x1f,
107 .flags = IORESOURCE_BUSY | IORESOURCE_IO },
108 { .name = "pic1", .start = 0x20, .end = 0x21,
109 .flags = IORESOURCE_BUSY | IORESOURCE_IO },
110 { .name = "timer0", .start = 0x40, .end = 0x43,
111 .flags = IORESOURCE_BUSY | IORESOURCE_IO },
112 { .name = "timer1", .start = 0x50, .end = 0x53,
113 .flags = IORESOURCE_BUSY | IORESOURCE_IO },
114 { .name = "keyboard", .start = 0x60, .end = 0x6f,
115 .flags = IORESOURCE_BUSY | IORESOURCE_IO },
116 { .name = "dma page reg", .start = 0x80, .end = 0x8f,
117 .flags = IORESOURCE_BUSY | IORESOURCE_IO },
118 { .name = "pic2", .start = 0xa0, .end = 0xa1,
119 .flags = IORESOURCE_BUSY | IORESOURCE_IO },
120 { .name = "dma2", .start = 0xc0, .end = 0xdf,
121 .flags = IORESOURCE_BUSY | IORESOURCE_IO },
122 { .name = "fpu", .start = 0xf0, .end = 0xff,
123 .flags = IORESOURCE_BUSY | IORESOURCE_IO }
124};
125
1da177e4
LT
126#define IORESOURCE_RAM (IORESOURCE_BUSY | IORESOURCE_MEM)
127
128struct resource data_resource = {
129 .name = "Kernel data",
130 .start = 0,
131 .end = 0,
132 .flags = IORESOURCE_RAM,
133};
134struct resource code_resource = {
135 .name = "Kernel code",
136 .start = 0,
137 .end = 0,
138 .flags = IORESOURCE_RAM,
139};
140
2c8c0e6b
AK
141#ifdef CONFIG_PROC_VMCORE
142/* elfcorehdr= specifies the location of elf core header
143 * stored by the crashed kernel. This option will be passed
144 * by kexec loader to the capture kernel.
145 */
146static int __init setup_elfcorehdr(char *arg)
681558fd 147{
2c8c0e6b
AK
148 char *end;
149 if (!arg)
150 return -EINVAL;
151 elfcorehdr_addr = memparse(arg, &end);
152 return end > arg ? 0 : -EINVAL;
681558fd 153}
2c8c0e6b 154early_param("elfcorehdr", setup_elfcorehdr);
e2c03888
AK
155#endif
156
2b97690f 157#ifndef CONFIG_NUMA
bbfceef4
MT
158static void __init
159contig_initmem_init(unsigned long start_pfn, unsigned long end_pfn)
1da177e4 160{
bbfceef4
MT
161 unsigned long bootmap_size, bootmap;
162
bbfceef4
MT
163 bootmap_size = bootmem_bootmap_pages(end_pfn)<<PAGE_SHIFT;
164 bootmap = find_e820_area(0, end_pfn<<PAGE_SHIFT, bootmap_size);
165 if (bootmap == -1L)
166 panic("Cannot find bootmem map of size %ld\n",bootmap_size);
167 bootmap_size = init_bootmem(bootmap >> PAGE_SHIFT, end_pfn);
5cb248ab
MG
168 e820_register_active_regions(0, start_pfn, end_pfn);
169 free_bootmem_with_active_regions(0, end_pfn);
bbfceef4 170 reserve_bootmem(bootmap, bootmap_size);
1da177e4
LT
171}
172#endif
173
1da177e4
LT
174#if defined(CONFIG_EDD) || defined(CONFIG_EDD_MODULE)
175struct edd edd;
176#ifdef CONFIG_EDD_MODULE
177EXPORT_SYMBOL(edd);
178#endif
179/**
180 * copy_edd() - Copy the BIOS EDD information
181 * from boot_params into a safe place.
182 *
183 */
184static inline void copy_edd(void)
185{
186 memcpy(edd.mbr_signature, EDD_MBR_SIGNATURE, sizeof(edd.mbr_signature));
187 memcpy(edd.edd_info, EDD_BUF, sizeof(edd.edd_info));
188 edd.mbr_signature_nr = EDD_MBR_SIG_NR;
189 edd.edd_info_nr = EDD_NR;
190}
191#else
192static inline void copy_edd(void)
193{
194}
195#endif
196
197#define EBDA_ADDR_POINTER 0x40E
ac71d12c
AK
198
199unsigned __initdata ebda_addr;
200unsigned __initdata ebda_size;
201
202static void discover_ebda(void)
1da177e4 203{
ac71d12c 204 /*
1da177e4
LT
205 * there is a real-mode segmented pointer pointing to the
206 * 4K EBDA area at 0x40E
207 */
bdb96a66 208 ebda_addr = *(unsigned short *)__va(EBDA_ADDR_POINTER);
ac71d12c
AK
209 ebda_addr <<= 4;
210
bdb96a66 211 ebda_size = *(unsigned short *)__va(ebda_addr);
ac71d12c
AK
212
213 /* Round EBDA up to pages */
214 if (ebda_size == 0)
215 ebda_size = 1;
216 ebda_size <<= 10;
217 ebda_size = round_up(ebda_size + (ebda_addr & ~PAGE_MASK), PAGE_SIZE);
218 if (ebda_size > 64*1024)
219 ebda_size = 64*1024;
1da177e4
LT
220}
221
222void __init setup_arch(char **cmdline_p)
223{
adf48856 224 printk(KERN_INFO "Command line: %s\n", boot_command_line);
43c85c9c 225
1da177e4 226 ROOT_DEV = old_decode_dev(ORIG_ROOT_DEV);
1da177e4
LT
227 screen_info = SCREEN_INFO;
228 edid_info = EDID_INFO;
229 saved_video_mode = SAVED_VIDEO_MODE;
230 bootloader_type = LOADER_TYPE;
231
232#ifdef CONFIG_BLK_DEV_RAM
233 rd_image_start = RAMDISK_FLAGS & RAMDISK_IMAGE_START_MASK;
234 rd_prompt = ((RAMDISK_FLAGS & RAMDISK_PROMPT_FLAG) != 0);
235 rd_doload = ((RAMDISK_FLAGS & RAMDISK_LOAD_FLAG) != 0);
236#endif
237 setup_memory_region();
238 copy_edd();
239
240 if (!MOUNT_ROOT_RDONLY)
241 root_mountflags &= ~MS_RDONLY;
242 init_mm.start_code = (unsigned long) &_text;
243 init_mm.end_code = (unsigned long) &_etext;
244 init_mm.end_data = (unsigned long) &_edata;
245 init_mm.brk = (unsigned long) &_end;
0dbf7028 246 init_mm.pgd = __va(__pa_symbol(&init_level4_pgt));
1da177e4 247
0dbf7028
VG
248 code_resource.start = __pa_symbol(&_text);
249 code_resource.end = __pa_symbol(&_etext)-1;
250 data_resource.start = __pa_symbol(&_etext);
251 data_resource.end = __pa_symbol(&_edata)-1;
1da177e4 252
1da177e4
LT
253 early_identify_cpu(&boot_cpu_data);
254
adf48856 255 strlcpy(command_line, boot_command_line, COMMAND_LINE_SIZE);
2c8c0e6b
AK
256 *cmdline_p = command_line;
257
258 parse_early_param();
259
260 finish_e820_parsing();
9ca33eb6 261
5cb248ab 262 e820_register_active_regions(0, 0, -1UL);
1da177e4
LT
263 /*
264 * partially used pages are not usable - thus
265 * we are rounding upwards:
266 */
267 end_pfn = e820_end_of_ram();
caff0710 268 num_physpages = end_pfn;
1da177e4
LT
269
270 check_efer();
271
ac71d12c
AK
272 discover_ebda();
273
1da177e4
LT
274 init_memory_mapping(0, (end_pfn_map << PAGE_SHIFT));
275
f2d3efed
AK
276 dmi_scan_machine();
277
888ba6c6 278#ifdef CONFIG_ACPI
1da177e4
LT
279 /*
280 * Initialize the ACPI boot-time table parser (gets the RSDP and SDT).
281 * Call this early for SRAT node setup.
282 */
283 acpi_boot_table_init();
284#endif
285
caff0710
JB
286 /* How many end-of-memory variables you have, grandma! */
287 max_low_pfn = end_pfn;
288 max_pfn = end_pfn;
289 high_memory = (void *)__va(end_pfn * PAGE_SIZE - 1) + 1;
290
5cb248ab
MG
291 /* Remove active ranges so rediscovery with NUMA-awareness happens */
292 remove_all_active_ranges();
293
1da177e4
LT
294#ifdef CONFIG_ACPI_NUMA
295 /*
296 * Parse SRAT to discover nodes.
297 */
298 acpi_numa_init();
299#endif
300
2b97690f 301#ifdef CONFIG_NUMA
1da177e4
LT
302 numa_initmem_init(0, end_pfn);
303#else
bbfceef4 304 contig_initmem_init(0, end_pfn);
1da177e4
LT
305#endif
306
307 /* Reserve direct mapping */
308 reserve_bootmem_generic(table_start << PAGE_SHIFT,
309 (table_end - table_start) << PAGE_SHIFT);
310
311 /* reserve kernel */
ceee8822
AK
312 reserve_bootmem_generic(__pa_symbol(&_text),
313 __pa_symbol(&_end) - __pa_symbol(&_text));
1da177e4
LT
314
315 /*
316 * reserve physical page 0 - it's a special BIOS page on many boxes,
317 * enabling clean reboots, SMP operation, laptop functions.
318 */
319 reserve_bootmem_generic(0, PAGE_SIZE);
320
321 /* reserve ebda region */
ac71d12c
AK
322 if (ebda_addr)
323 reserve_bootmem_generic(ebda_addr, ebda_size);
076422d2
AS
324#ifdef CONFIG_NUMA
325 /* reserve nodemap region */
326 if (nodemap_addr)
327 reserve_bootmem_generic(nodemap_addr, nodemap_size);
328#endif
1da177e4
LT
329
330#ifdef CONFIG_SMP
1da177e4 331 /* Reserve SMP trampoline */
90b1c208 332 reserve_bootmem_generic(SMP_TRAMPOLINE_BASE, 2*PAGE_SIZE);
1da177e4
LT
333#endif
334
335#ifdef CONFIG_ACPI_SLEEP
336 /*
337 * Reserve low memory region for sleep support.
338 */
339 acpi_reserve_bootmem();
340#endif
1da177e4
LT
341 /*
342 * Find and reserve possible boot-time SMP configuration:
343 */
344 find_smp_config();
1da177e4
LT
345#ifdef CONFIG_BLK_DEV_INITRD
346 if (LOADER_TYPE && INITRD_START) {
347 if (INITRD_START + INITRD_SIZE <= (end_pfn << PAGE_SHIFT)) {
348 reserve_bootmem_generic(INITRD_START, INITRD_SIZE);
19e5d9c0 349 initrd_start = INITRD_START + PAGE_OFFSET;
1da177e4
LT
350 initrd_end = initrd_start+INITRD_SIZE;
351 }
352 else {
353 printk(KERN_ERR "initrd extends beyond end of memory "
354 "(0x%08lx > 0x%08lx)\ndisabling initrd\n",
355 (unsigned long)(INITRD_START + INITRD_SIZE),
356 (unsigned long)(end_pfn << PAGE_SHIFT));
357 initrd_start = 0;
358 }
359 }
360#endif
5f5609df
EB
361#ifdef CONFIG_KEXEC
362 if (crashk_res.start != crashk_res.end) {
00212fef 363 reserve_bootmem_generic(crashk_res.start,
5f5609df
EB
364 crashk_res.end - crashk_res.start + 1);
365 }
366#endif
0d317fb7 367
1da177e4
LT
368 paging_init();
369
f157cbb1 370#ifdef CONFIG_PCI
dfa4698c 371 early_quirks();
f157cbb1 372#endif
1da177e4 373
51f62e18
AR
374 /*
375 * set this early, so we dont allocate cpu0
376 * if MADT list doesnt list BSP first
377 * mpparse.c/MP_processor_info() allocates logical cpu numbers.
378 */
379 cpu_set(0, cpu_present_map);
888ba6c6 380#ifdef CONFIG_ACPI
1da177e4
LT
381 /*
382 * Read APIC and some other early information from ACPI tables.
383 */
384 acpi_boot_init();
385#endif
386
05b3cbd8
RT
387 init_cpu_to_node();
388
1da177e4
LT
389 /*
390 * get boot-time SMP configuration:
391 */
392 if (smp_found_config)
393 get_smp_config();
394 init_apic_mappings();
1da177e4
LT
395
396 /*
fc986db4
AK
397 * We trust e820 completely. No explicit ROM probing in memory.
398 */
1da177e4 399 e820_reserve_resources();
e8eff5ac 400 e820_mark_nosave_regions();
1da177e4 401
1da177e4
LT
402 {
403 unsigned i;
404 /* request I/O space for devices used on all i[345]86 PCs */
9d0ef4fd 405 for (i = 0; i < ARRAY_SIZE(standard_io_resources); i++)
1da177e4
LT
406 request_resource(&ioport_resource, &standard_io_resources[i]);
407 }
408
a1e97782 409 e820_setup_gap();
1da177e4 410
1da177e4
LT
411#ifdef CONFIG_VT
412#if defined(CONFIG_VGA_CONSOLE)
413 conswitchp = &vga_con;
414#elif defined(CONFIG_DUMMY_CONSOLE)
415 conswitchp = &dummy_con;
416#endif
417#endif
418}
419
e6982c67 420static int __cpuinit get_model_name(struct cpuinfo_x86 *c)
1da177e4
LT
421{
422 unsigned int *v;
423
ebfcaa96 424 if (c->extended_cpuid_level < 0x80000004)
1da177e4
LT
425 return 0;
426
427 v = (unsigned int *) c->x86_model_id;
428 cpuid(0x80000002, &v[0], &v[1], &v[2], &v[3]);
429 cpuid(0x80000003, &v[4], &v[5], &v[6], &v[7]);
430 cpuid(0x80000004, &v[8], &v[9], &v[10], &v[11]);
431 c->x86_model_id[48] = 0;
432 return 1;
433}
434
435
e6982c67 436static void __cpuinit display_cacheinfo(struct cpuinfo_x86 *c)
1da177e4
LT
437{
438 unsigned int n, dummy, eax, ebx, ecx, edx;
439
ebfcaa96 440 n = c->extended_cpuid_level;
1da177e4
LT
441
442 if (n >= 0x80000005) {
443 cpuid(0x80000005, &dummy, &ebx, &ecx, &edx);
444 printk(KERN_INFO "CPU: L1 I Cache: %dK (%d bytes/line), D cache %dK (%d bytes/line)\n",
445 edx>>24, edx&0xFF, ecx>>24, ecx&0xFF);
446 c->x86_cache_size=(ecx>>24)+(edx>>24);
447 /* On K8 L1 TLB is inclusive, so don't count it */
448 c->x86_tlbsize = 0;
449 }
450
451 if (n >= 0x80000006) {
452 cpuid(0x80000006, &dummy, &ebx, &ecx, &edx);
453 ecx = cpuid_ecx(0x80000006);
454 c->x86_cache_size = ecx >> 16;
455 c->x86_tlbsize += ((ebx >> 16) & 0xfff) + (ebx & 0xfff);
456
457 printk(KERN_INFO "CPU: L2 Cache: %dK (%d bytes/line)\n",
458 c->x86_cache_size, ecx & 0xFF);
459 }
460
461 if (n >= 0x80000007)
462 cpuid(0x80000007, &dummy, &dummy, &dummy, &c->x86_power);
463 if (n >= 0x80000008) {
464 cpuid(0x80000008, &eax, &dummy, &dummy, &dummy);
465 c->x86_virt_bits = (eax >> 8) & 0xff;
466 c->x86_phys_bits = eax & 0xff;
467 }
468}
469
3f098c26
AK
470#ifdef CONFIG_NUMA
471static int nearby_node(int apicid)
472{
473 int i;
474 for (i = apicid - 1; i >= 0; i--) {
475 int node = apicid_to_node[i];
476 if (node != NUMA_NO_NODE && node_online(node))
477 return node;
478 }
479 for (i = apicid + 1; i < MAX_LOCAL_APIC; i++) {
480 int node = apicid_to_node[i];
481 if (node != NUMA_NO_NODE && node_online(node))
482 return node;
483 }
484 return first_node(node_online_map); /* Shouldn't happen */
485}
486#endif
487
63518644
AK
488/*
489 * On a AMD dual core setup the lower bits of the APIC id distingush the cores.
490 * Assumes number of cores is a power of two.
491 */
492static void __init amd_detect_cmp(struct cpuinfo_x86 *c)
493{
494#ifdef CONFIG_SMP
b41e2939 495 unsigned bits;
3f098c26 496#ifdef CONFIG_NUMA
f3fa8ebc 497 int cpu = smp_processor_id();
3f098c26 498 int node = 0;
60c1bc82 499 unsigned apicid = hard_smp_processor_id();
3f098c26 500#endif
faee9a5d 501 unsigned ecx = cpuid_ecx(0x80000008);
b41e2939 502
faee9a5d 503 c->x86_max_cores = (ecx & 0xff) + 1;
b41e2939 504
faee9a5d
AK
505 /* CPU telling us the core id bits shift? */
506 bits = (ecx >> 12) & 0xF;
507
508 /* Otherwise recompute */
509 if (bits == 0) {
510 while ((1 << bits) < c->x86_max_cores)
511 bits++;
512 }
b41e2939
AK
513
514 /* Low order bits define the core id (index of core in socket) */
f3fa8ebc 515 c->cpu_core_id = c->phys_proc_id & ((1 << bits)-1);
b41e2939 516 /* Convert the APIC ID into the socket ID */
f3fa8ebc 517 c->phys_proc_id = phys_pkg_id(bits);
63518644
AK
518
519#ifdef CONFIG_NUMA
f3fa8ebc 520 node = c->phys_proc_id;
3f098c26
AK
521 if (apicid_to_node[apicid] != NUMA_NO_NODE)
522 node = apicid_to_node[apicid];
523 if (!node_online(node)) {
524 /* Two possibilities here:
525 - The CPU is missing memory and no node was created.
526 In that case try picking one from a nearby CPU
527 - The APIC IDs differ from the HyperTransport node IDs
528 which the K8 northbridge parsing fills in.
529 Assume they are all increased by a constant offset,
530 but in the same order as the HT nodeids.
531 If that doesn't result in a usable node fall back to the
532 path for the previous case. */
f3fa8ebc 533 int ht_nodeid = apicid - (cpu_data[0].phys_proc_id << bits);
3f098c26
AK
534 if (ht_nodeid >= 0 &&
535 apicid_to_node[ht_nodeid] != NUMA_NO_NODE)
536 node = apicid_to_node[ht_nodeid];
537 /* Pick a nearby node */
538 if (!node_online(node))
539 node = nearby_node(apicid);
540 }
69d81fcd 541 numa_set_node(cpu, node);
3f098c26 542
e42f9437 543 printk(KERN_INFO "CPU %d/%x -> Node %d\n", cpu, apicid, node);
63518644 544#endif
63518644
AK
545#endif
546}
1da177e4 547
ed77504b 548static void __cpuinit init_amd(struct cpuinfo_x86 *c)
1da177e4 549{
7bcd3f34 550 unsigned level;
1da177e4 551
bc5e8fdf
LT
552#ifdef CONFIG_SMP
553 unsigned long value;
554
7d318d77
AK
555 /*
556 * Disable TLB flush filter by setting HWCR.FFDIS on K8
557 * bit 6 of msr C001_0015
558 *
559 * Errata 63 for SH-B3 steppings
560 * Errata 122 for all steppings (F+ have it disabled by default)
561 */
562 if (c->x86 == 15) {
563 rdmsrl(MSR_K8_HWCR, value);
564 value |= 1 << 6;
565 wrmsrl(MSR_K8_HWCR, value);
566 }
bc5e8fdf
LT
567#endif
568
1da177e4
LT
569 /* Bit 31 in normal CPUID used for nonstandard 3DNow ID;
570 3DNow is IDd by bit 31 in extended CPUID (1*32+31) anyway */
571 clear_bit(0*32+31, &c->x86_capability);
572
7bcd3f34
AK
573 /* On C+ stepping K8 rep microcode works well for copy/memset */
574 level = cpuid_eax(1);
575 if (c->x86 == 15 && ((level >= 0x0f48 && level < 0x0f50) || level >= 0x0f58))
576 set_bit(X86_FEATURE_REP_GOOD, &c->x86_capability);
577
18bd057b
AK
578 /* Enable workaround for FXSAVE leak */
579 if (c->x86 >= 6)
580 set_bit(X86_FEATURE_FXSAVE_LEAK, &c->x86_capability);
581
e42f9437
RS
582 level = get_model_name(c);
583 if (!level) {
1da177e4
LT
584 switch (c->x86) {
585 case 15:
586 /* Should distinguish Models here, but this is only
587 a fallback anyways. */
588 strcpy(c->x86_model_id, "Hammer");
589 break;
590 }
591 }
592 display_cacheinfo(c);
593
130951cc
AK
594 /* c->x86_power is 8000_0007 edx. Bit 8 is constant TSC */
595 if (c->x86_power & (1<<8))
596 set_bit(X86_FEATURE_CONSTANT_TSC, &c->x86_capability);
597
faee9a5d
AK
598 /* Multi core CPU? */
599 if (c->extended_cpuid_level >= 0x80000008)
63518644 600 amd_detect_cmp(c);
1da177e4 601
240cd6a8
AK
602 /* Fix cpuid4 emulation for more */
603 num_cache_leaves = 3;
2049336f 604
61677965
AK
605 /* RDTSC can be speculated around */
606 clear_bit(X86_FEATURE_SYNC_RDTSC, &c->x86_capability);
1da177e4
LT
607}
608
e6982c67 609static void __cpuinit detect_ht(struct cpuinfo_x86 *c)
1da177e4
LT
610{
611#ifdef CONFIG_SMP
612 u32 eax, ebx, ecx, edx;
94605eff 613 int index_msb, core_bits;
94605eff
SS
614
615 cpuid(1, &eax, &ebx, &ecx, &edx);
616
94605eff 617
e42f9437 618 if (!cpu_has(c, X86_FEATURE_HT))
1da177e4 619 return;
e42f9437
RS
620 if (cpu_has(c, X86_FEATURE_CMP_LEGACY))
621 goto out;
1da177e4 622
1da177e4 623 smp_num_siblings = (ebx & 0xff0000) >> 16;
94605eff 624
1da177e4
LT
625 if (smp_num_siblings == 1) {
626 printk(KERN_INFO "CPU: Hyper-Threading is disabled\n");
94605eff
SS
627 } else if (smp_num_siblings > 1 ) {
628
1da177e4
LT
629 if (smp_num_siblings > NR_CPUS) {
630 printk(KERN_WARNING "CPU: Unsupported number of the siblings %d", smp_num_siblings);
631 smp_num_siblings = 1;
632 return;
633 }
94605eff
SS
634
635 index_msb = get_count_order(smp_num_siblings);
f3fa8ebc 636 c->phys_proc_id = phys_pkg_id(index_msb);
3dd9d514 637
94605eff 638 smp_num_siblings = smp_num_siblings / c->x86_max_cores;
3dd9d514 639
94605eff
SS
640 index_msb = get_count_order(smp_num_siblings) ;
641
642 core_bits = get_count_order(c->x86_max_cores);
3dd9d514 643
f3fa8ebc 644 c->cpu_core_id = phys_pkg_id(index_msb) &
94605eff 645 ((1 << core_bits) - 1);
1da177e4 646 }
e42f9437
RS
647out:
648 if ((c->x86_max_cores * smp_num_siblings) > 1) {
649 printk(KERN_INFO "CPU: Physical Processor ID: %d\n", c->phys_proc_id);
650 printk(KERN_INFO "CPU: Processor Core ID: %d\n", c->cpu_core_id);
651 }
652
1da177e4
LT
653#endif
654}
655
3dd9d514
AK
656/*
657 * find out the number of processor cores on the die
658 */
e6982c67 659static int __cpuinit intel_num_cpu_cores(struct cpuinfo_x86 *c)
3dd9d514 660{
2bbc419f 661 unsigned int eax, t;
3dd9d514
AK
662
663 if (c->cpuid_level < 4)
664 return 1;
665
2bbc419f 666 cpuid_count(4, 0, &eax, &t, &t, &t);
3dd9d514
AK
667
668 if (eax & 0x1f)
669 return ((eax >> 26) + 1);
670 else
671 return 1;
672}
673
df0cc26b
AK
674static void srat_detect_node(void)
675{
676#ifdef CONFIG_NUMA
ddea7be0 677 unsigned node;
df0cc26b 678 int cpu = smp_processor_id();
e42f9437 679 int apicid = hard_smp_processor_id();
df0cc26b
AK
680
681 /* Don't do the funky fallback heuristics the AMD version employs
682 for now. */
e42f9437 683 node = apicid_to_node[apicid];
df0cc26b 684 if (node == NUMA_NO_NODE)
0d015324 685 node = first_node(node_online_map);
69d81fcd 686 numa_set_node(cpu, node);
df0cc26b 687
c31fbb1a 688 printk(KERN_INFO "CPU %d/%x -> Node %d\n", cpu, apicid, node);
df0cc26b
AK
689#endif
690}
691
e6982c67 692static void __cpuinit init_intel(struct cpuinfo_x86 *c)
1da177e4
LT
693{
694 /* Cache sizes */
695 unsigned n;
696
697 init_intel_cacheinfo(c);
0080e667
VP
698 if (c->cpuid_level > 9 ) {
699 unsigned eax = cpuid_eax(10);
700 /* Check for version and the number of counters */
701 if ((eax & 0xff) && (((eax>>8) & 0xff) > 1))
702 set_bit(X86_FEATURE_ARCH_PERFMON, &c->x86_capability);
703 }
704
36b2a8d5
SE
705 if (cpu_has_ds) {
706 unsigned int l1, l2;
707 rdmsr(MSR_IA32_MISC_ENABLE, l1, l2);
ee58fad5
SE
708 if (!(l1 & (1<<11)))
709 set_bit(X86_FEATURE_BTS, c->x86_capability);
36b2a8d5
SE
710 if (!(l1 & (1<<12)))
711 set_bit(X86_FEATURE_PEBS, c->x86_capability);
712 }
713
ebfcaa96 714 n = c->extended_cpuid_level;
1da177e4
LT
715 if (n >= 0x80000008) {
716 unsigned eax = cpuid_eax(0x80000008);
717 c->x86_virt_bits = (eax >> 8) & 0xff;
718 c->x86_phys_bits = eax & 0xff;
af9c142d
SL
719 /* CPUID workaround for Intel 0F34 CPU */
720 if (c->x86_vendor == X86_VENDOR_INTEL &&
721 c->x86 == 0xF && c->x86_model == 0x3 &&
722 c->x86_mask == 0x4)
723 c->x86_phys_bits = 36;
1da177e4
LT
724 }
725
726 if (c->x86 == 15)
727 c->x86_cache_alignment = c->x86_clflush_size * 2;
39b3a791
AK
728 if ((c->x86 == 0xf && c->x86_model >= 0x03) ||
729 (c->x86 == 0x6 && c->x86_model >= 0x0e))
c29601e9 730 set_bit(X86_FEATURE_CONSTANT_TSC, &c->x86_capability);
27fbe5b2
AK
731 if (c->x86 == 6)
732 set_bit(X86_FEATURE_REP_GOOD, &c->x86_capability);
f3d73707
AV
733 if (c->x86 == 15)
734 set_bit(X86_FEATURE_SYNC_RDTSC, &c->x86_capability);
735 else
736 clear_bit(X86_FEATURE_SYNC_RDTSC, &c->x86_capability);
94605eff 737 c->x86_max_cores = intel_num_cpu_cores(c);
df0cc26b
AK
738
739 srat_detect_node();
1da177e4
LT
740}
741
672289e9 742static void __cpuinit get_cpu_vendor(struct cpuinfo_x86 *c)
1da177e4
LT
743{
744 char *v = c->x86_vendor_id;
745
746 if (!strcmp(v, "AuthenticAMD"))
747 c->x86_vendor = X86_VENDOR_AMD;
748 else if (!strcmp(v, "GenuineIntel"))
749 c->x86_vendor = X86_VENDOR_INTEL;
750 else
751 c->x86_vendor = X86_VENDOR_UNKNOWN;
752}
753
754struct cpu_model_info {
755 int vendor;
756 int family;
757 char *model_names[16];
758};
759
760/* Do some early cpuid on the boot CPU to get some parameter that are
761 needed before check_bugs. Everything advanced is in identify_cpu
762 below. */
e6982c67 763void __cpuinit early_identify_cpu(struct cpuinfo_x86 *c)
1da177e4
LT
764{
765 u32 tfms;
766
767 c->loops_per_jiffy = loops_per_jiffy;
768 c->x86_cache_size = -1;
769 c->x86_vendor = X86_VENDOR_UNKNOWN;
770 c->x86_model = c->x86_mask = 0; /* So far unknown... */
771 c->x86_vendor_id[0] = '\0'; /* Unset */
772 c->x86_model_id[0] = '\0'; /* Unset */
773 c->x86_clflush_size = 64;
774 c->x86_cache_alignment = c->x86_clflush_size;
94605eff 775 c->x86_max_cores = 1;
ebfcaa96 776 c->extended_cpuid_level = 0;
1da177e4
LT
777 memset(&c->x86_capability, 0, sizeof c->x86_capability);
778
779 /* Get vendor name */
780 cpuid(0x00000000, (unsigned int *)&c->cpuid_level,
781 (unsigned int *)&c->x86_vendor_id[0],
782 (unsigned int *)&c->x86_vendor_id[8],
783 (unsigned int *)&c->x86_vendor_id[4]);
784
785 get_cpu_vendor(c);
786
787 /* Initialize the standard set of capabilities */
788 /* Note that the vendor-specific code below might override */
789
790 /* Intel-defined flags: level 0x00000001 */
791 if (c->cpuid_level >= 0x00000001) {
792 __u32 misc;
793 cpuid(0x00000001, &tfms, &misc, &c->x86_capability[4],
794 &c->x86_capability[0]);
795 c->x86 = (tfms >> 8) & 0xf;
796 c->x86_model = (tfms >> 4) & 0xf;
797 c->x86_mask = tfms & 0xf;
f5f786d0 798 if (c->x86 == 0xf)
1da177e4 799 c->x86 += (tfms >> 20) & 0xff;
f5f786d0 800 if (c->x86 >= 0x6)
1da177e4 801 c->x86_model += ((tfms >> 16) & 0xF) << 4;
1da177e4
LT
802 if (c->x86_capability[0] & (1<<19))
803 c->x86_clflush_size = ((misc >> 8) & 0xff) * 8;
1da177e4
LT
804 } else {
805 /* Have CPUID level 0 only - unheard of */
806 c->x86 = 4;
807 }
a158608b
AK
808
809#ifdef CONFIG_SMP
f3fa8ebc 810 c->phys_proc_id = (cpuid_ebx(1) >> 24) & 0xff;
a158608b 811#endif
1da177e4
LT
812}
813
814/*
815 * This does the hard work of actually picking apart the CPU stuff...
816 */
e6982c67 817void __cpuinit identify_cpu(struct cpuinfo_x86 *c)
1da177e4
LT
818{
819 int i;
820 u32 xlvl;
821
822 early_identify_cpu(c);
823
824 /* AMD-defined flags: level 0x80000001 */
825 xlvl = cpuid_eax(0x80000000);
ebfcaa96 826 c->extended_cpuid_level = xlvl;
1da177e4
LT
827 if ((xlvl & 0xffff0000) == 0x80000000) {
828 if (xlvl >= 0x80000001) {
829 c->x86_capability[1] = cpuid_edx(0x80000001);
5b7abc6f 830 c->x86_capability[6] = cpuid_ecx(0x80000001);
1da177e4
LT
831 }
832 if (xlvl >= 0x80000004)
833 get_model_name(c); /* Default name */
834 }
835
836 /* Transmeta-defined flags: level 0x80860001 */
837 xlvl = cpuid_eax(0x80860000);
838 if ((xlvl & 0xffff0000) == 0x80860000) {
839 /* Don't set x86_cpuid_level here for now to not confuse. */
840 if (xlvl >= 0x80860001)
841 c->x86_capability[2] = cpuid_edx(0x80860001);
842 }
843
1e9f28fa
SS
844 c->apicid = phys_pkg_id(0);
845
1da177e4
LT
846 /*
847 * Vendor-specific initialization. In this section we
848 * canonicalize the feature flags, meaning if there are
849 * features a certain CPU supports which CPUID doesn't
850 * tell us, CPUID claiming incorrect flags, or other bugs,
851 * we handle them here.
852 *
853 * At the end of this section, c->x86_capability better
854 * indicate the features this CPU genuinely supports!
855 */
856 switch (c->x86_vendor) {
857 case X86_VENDOR_AMD:
858 init_amd(c);
859 break;
860
861 case X86_VENDOR_INTEL:
862 init_intel(c);
863 break;
864
865 case X86_VENDOR_UNKNOWN:
866 default:
867 display_cacheinfo(c);
868 break;
869 }
870
871 select_idle_routine(c);
872 detect_ht(c);
1da177e4
LT
873
874 /*
875 * On SMP, boot_cpu_data holds the common feature set between
876 * all CPUs; so make sure that we indicate which features are
877 * common between the CPUs. The first time this routine gets
878 * executed, c == &boot_cpu_data.
879 */
880 if (c != &boot_cpu_data) {
881 /* AND the already accumulated flags with these */
882 for (i = 0 ; i < NCAPINTS ; i++)
883 boot_cpu_data.x86_capability[i] &= c->x86_capability[i];
884 }
885
886#ifdef CONFIG_X86_MCE
887 mcheck_init(c);
888#endif
3b520b23
SL
889 if (c == &boot_cpu_data)
890 mtrr_bp_init();
891 else
892 mtrr_ap_init();
1da177e4 893#ifdef CONFIG_NUMA
3019e8eb 894 numa_add_cpu(smp_processor_id());
1da177e4
LT
895#endif
896}
897
898
e6982c67 899void __cpuinit print_cpu_info(struct cpuinfo_x86 *c)
1da177e4
LT
900{
901 if (c->x86_model_id[0])
902 printk("%s", c->x86_model_id);
903
904 if (c->x86_mask || c->cpuid_level >= 0)
905 printk(" stepping %02x\n", c->x86_mask);
906 else
907 printk("\n");
908}
909
910/*
911 * Get CPU information for use by the procfs.
912 */
913
914static int show_cpuinfo(struct seq_file *m, void *v)
915{
916 struct cpuinfo_x86 *c = v;
917
918 /*
919 * These flag bits must match the definitions in <asm/cpufeature.h>.
920 * NULL means this bit is undefined or reserved; either way it doesn't
921 * have meaning as far as Linux is concerned. Note that it's important
922 * to realize there is a difference between this table and CPUID -- if
923 * applications want to get the raw CPUID data, they should access
924 * /dev/cpu/<cpu_nr>/cpuid instead.
925 */
926 static char *x86_cap_flags[] = {
927 /* Intel-defined */
928 "fpu", "vme", "de", "pse", "tsc", "msr", "pae", "mce",
929 "cx8", "apic", NULL, "sep", "mtrr", "pge", "mca", "cmov",
930 "pat", "pse36", "pn", "clflush", NULL, "dts", "acpi", "mmx",
931 "fxsr", "sse", "sse2", "ss", "ht", "tm", "ia64", NULL,
932
933 /* AMD-defined */
3c3b73b6 934 NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL,
1da177e4
LT
935 NULL, NULL, NULL, "syscall", NULL, NULL, NULL, NULL,
936 NULL, NULL, NULL, NULL, "nx", NULL, "mmxext", NULL,
f790cd30
AK
937 NULL, "fxsr_opt", "pdpe1gb", "rdtscp", NULL, "lm",
938 "3dnowext", "3dnow",
1da177e4
LT
939
940 /* Transmeta-defined */
941 "recovery", "longrun", NULL, "lrti", NULL, NULL, NULL, NULL,
942 NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL,
943 NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL,
944 NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL,
945
946 /* Other (Linux-defined) */
622dcaf9 947 "cxmmx", NULL, "cyrix_arr", "centaur_mcr", NULL,
c29601e9 948 "constant_tsc", NULL, NULL,
d167a518 949 "up", NULL, NULL, NULL, NULL, NULL, NULL, NULL,
1da177e4
LT
950 NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL,
951 NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL,
952
953 /* Intel-defined (#2) */
9d95dd84 954 "pni", NULL, NULL, "monitor", "ds_cpl", "vmx", "smx", "est",
dcf10307 955 "tm2", "ssse3", "cid", NULL, NULL, "cx16", "xtpr", NULL,
f790cd30 956 NULL, NULL, "dca", NULL, NULL, NULL, NULL, "popcnt",
1da177e4
LT
957 NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL,
958
5b7abc6f
PA
959 /* VIA/Cyrix/Centaur-defined */
960 NULL, NULL, "rng", "rng_en", NULL, NULL, "ace", "ace_en",
961 NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL,
962 NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL,
963 NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL,
964
1da177e4 965 /* AMD-defined (#2) */
f790cd30
AK
966 "lahf_lm", "cmp_legacy", "svm", "extapic", "cr8_legacy",
967 "altmovcr8", "abm", "sse4a",
968 "misalignsse", "3dnowprefetch",
969 "osvw", "ibs", NULL, NULL, NULL, NULL,
1da177e4 970 NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL,
5b7abc6f 971 NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL,
1da177e4
LT
972 };
973 static char *x86_power_flags[] = {
974 "ts", /* temperature sensor */
975 "fid", /* frequency id control */
976 "vid", /* voltage id control */
977 "ttp", /* thermal trip */
978 "tm",
3f98bc49 979 "stc",
f790cd30
AK
980 "100mhzsteps",
981 "hwpstate",
982 NULL, /* tsc invariant mapped to constant_tsc */
3f98bc49 983 NULL,
39b3a791 984 /* nothing */ /* constant_tsc - moved to flags */
1da177e4
LT
985 };
986
987
988#ifdef CONFIG_SMP
989 if (!cpu_online(c-cpu_data))
990 return 0;
991#endif
992
993 seq_printf(m,"processor\t: %u\n"
994 "vendor_id\t: %s\n"
995 "cpu family\t: %d\n"
996 "model\t\t: %d\n"
997 "model name\t: %s\n",
998 (unsigned)(c-cpu_data),
999 c->x86_vendor_id[0] ? c->x86_vendor_id : "unknown",
1000 c->x86,
1001 (int)c->x86_model,
1002 c->x86_model_id[0] ? c->x86_model_id : "unknown");
1003
1004 if (c->x86_mask || c->cpuid_level >= 0)
1005 seq_printf(m, "stepping\t: %d\n", c->x86_mask);
1006 else
1007 seq_printf(m, "stepping\t: unknown\n");
1008
1009 if (cpu_has(c,X86_FEATURE_TSC)) {
95235ca2
VP
1010 unsigned int freq = cpufreq_quick_get((unsigned)(c-cpu_data));
1011 if (!freq)
1012 freq = cpu_khz;
1da177e4 1013 seq_printf(m, "cpu MHz\t\t: %u.%03u\n",
95235ca2 1014 freq / 1000, (freq % 1000));
1da177e4
LT
1015 }
1016
1017 /* Cache size */
1018 if (c->x86_cache_size >= 0)
1019 seq_printf(m, "cache size\t: %d KB\n", c->x86_cache_size);
1020
1021#ifdef CONFIG_SMP
94605eff 1022 if (smp_num_siblings * c->x86_max_cores > 1) {
db468681 1023 int cpu = c - cpu_data;
f3fa8ebc 1024 seq_printf(m, "physical id\t: %d\n", c->phys_proc_id);
94605eff 1025 seq_printf(m, "siblings\t: %d\n", cpus_weight(cpu_core_map[cpu]));
f3fa8ebc 1026 seq_printf(m, "core id\t\t: %d\n", c->cpu_core_id);
94605eff 1027 seq_printf(m, "cpu cores\t: %d\n", c->booted_cores);
db468681 1028 }
1da177e4
LT
1029#endif
1030
1031 seq_printf(m,
1032 "fpu\t\t: yes\n"
1033 "fpu_exception\t: yes\n"
1034 "cpuid level\t: %d\n"
1035 "wp\t\t: yes\n"
1036 "flags\t\t:",
1037 c->cpuid_level);
1038
1039 {
1040 int i;
1041 for ( i = 0 ; i < 32*NCAPINTS ; i++ )
3d1712c9 1042 if (cpu_has(c, i) && x86_cap_flags[i] != NULL)
1da177e4
LT
1043 seq_printf(m, " %s", x86_cap_flags[i]);
1044 }
1045
1046 seq_printf(m, "\nbogomips\t: %lu.%02lu\n",
1047 c->loops_per_jiffy/(500000/HZ),
1048 (c->loops_per_jiffy/(5000/HZ)) % 100);
1049
1050 if (c->x86_tlbsize > 0)
1051 seq_printf(m, "TLB size\t: %d 4K pages\n", c->x86_tlbsize);
1052 seq_printf(m, "clflush size\t: %d\n", c->x86_clflush_size);
1053 seq_printf(m, "cache_alignment\t: %d\n", c->x86_cache_alignment);
1054
1055 seq_printf(m, "address sizes\t: %u bits physical, %u bits virtual\n",
1056 c->x86_phys_bits, c->x86_virt_bits);
1057
1058 seq_printf(m, "power management:");
1059 {
1060 unsigned i;
1061 for (i = 0; i < 32; i++)
1062 if (c->x86_power & (1 << i)) {
3f98bc49
AK
1063 if (i < ARRAY_SIZE(x86_power_flags) &&
1064 x86_power_flags[i])
1065 seq_printf(m, "%s%s",
1066 x86_power_flags[i][0]?" ":"",
1067 x86_power_flags[i]);
1da177e4
LT
1068 else
1069 seq_printf(m, " [%d]", i);
1070 }
1071 }
1da177e4 1072
d31ddaa1 1073 seq_printf(m, "\n\n");
1da177e4
LT
1074
1075 return 0;
1076}
1077
1078static void *c_start(struct seq_file *m, loff_t *pos)
1079{
1080 return *pos < NR_CPUS ? cpu_data + *pos : NULL;
1081}
1082
1083static void *c_next(struct seq_file *m, void *v, loff_t *pos)
1084{
1085 ++*pos;
1086 return c_start(m, pos);
1087}
1088
1089static void c_stop(struct seq_file *m, void *v)
1090{
1091}
1092
1093struct seq_operations cpuinfo_op = {
1094 .start =c_start,
1095 .next = c_next,
1096 .stop = c_stop,
1097 .show = show_cpuinfo,
1098};
This page took 0.297907 seconds and 5 git commands to generate.