Commit | Line | Data |
---|---|---|
03501dab PM |
1 | /* |
2 | * Common boot and setup code for both 32-bit and 64-bit. | |
3 | * Extracted from arch/powerpc/kernel/setup_64.c. | |
4 | * | |
5 | * Copyright (C) 2001 PPC64 Team, IBM Corp | |
6 | * | |
7 | * This program is free software; you can redistribute it and/or | |
8 | * modify it under the terms of the GNU General Public License | |
9 | * as published by the Free Software Foundation; either version | |
10 | * 2 of the License, or (at your option) any later version. | |
11 | */ | |
e8222502 BH |
12 | |
13 | #undef DEBUG | |
14 | ||
4b16f8e2 | 15 | #include <linux/export.h> |
03501dab PM |
16 | #include <linux/string.h> |
17 | #include <linux/sched.h> | |
18 | #include <linux/init.h> | |
19 | #include <linux/kernel.h> | |
20 | #include <linux/reboot.h> | |
21 | #include <linux/delay.h> | |
22 | #include <linux/initrd.h> | |
e5c6c8e4 | 23 | #include <linux/platform_device.h> |
03501dab PM |
24 | #include <linux/seq_file.h> |
25 | #include <linux/ioport.h> | |
26 | #include <linux/console.h> | |
894673ee | 27 | #include <linux/screen_info.h> |
03501dab PM |
28 | #include <linux/root_dev.h> |
29 | #include <linux/notifier.h> | |
30 | #include <linux/cpu.h> | |
31 | #include <linux/unistd.h> | |
32 | #include <linux/serial.h> | |
33 | #include <linux/serial_8250.h> | |
94a3807c | 34 | #include <linux/debugfs.h> |
8d089085 | 35 | #include <linux/percpu.h> |
95f72d1e | 36 | #include <linux/memblock.h> |
d746286c | 37 | #include <linux/of_platform.h> |
b1923caa | 38 | #include <linux/hugetlb.h> |
03501dab | 39 | #include <asm/io.h> |
1426d5a3 | 40 | #include <asm/paca.h> |
03501dab PM |
41 | #include <asm/prom.h> |
42 | #include <asm/processor.h> | |
a7f290da | 43 | #include <asm/vdso_datapage.h> |
03501dab | 44 | #include <asm/pgtable.h> |
03501dab PM |
45 | #include <asm/smp.h> |
46 | #include <asm/elf.h> | |
47 | #include <asm/machdep.h> | |
48 | #include <asm/time.h> | |
49 | #include <asm/cputable.h> | |
50 | #include <asm/sections.h> | |
e8222502 | 51 | #include <asm/firmware.h> |
03501dab PM |
52 | #include <asm/btext.h> |
53 | #include <asm/nvram.h> | |
54 | #include <asm/setup.h> | |
03501dab PM |
55 | #include <asm/rtas.h> |
56 | #include <asm/iommu.h> | |
57 | #include <asm/serial.h> | |
58 | #include <asm/cache.h> | |
59 | #include <asm/page.h> | |
60 | #include <asm/mmu.h> | |
fca5dcd4 | 61 | #include <asm/xmon.h> |
8d089085 | 62 | #include <asm/cputhreads.h> |
f465df81 | 63 | #include <mm/mmu_decl.h> |
ebaeb5ae | 64 | #include <asm/fadump.h> |
b1923caa BH |
65 | #include <asm/udbg.h> |
66 | #include <asm/hugetlb.h> | |
67 | #include <asm/livepatch.h> | |
68 | #include <asm/mmu_context.h> | |
b92a226e | 69 | #include <asm/cpu_has_feature.h> |
b1923caa BH |
70 | |
71 | #include "setup.h" | |
03501dab | 72 | |
03501dab | 73 | #ifdef DEBUG |
f9e4ec57 | 74 | #include <asm/udbg.h> |
03501dab PM |
75 | #define DBG(fmt...) udbg_printf(fmt) |
76 | #else | |
77 | #define DBG(fmt...) | |
78 | #endif | |
79 | ||
e8222502 BH |
80 | /* The main machine-dep calls structure |
81 | */ | |
82 | struct machdep_calls ppc_md; | |
83 | EXPORT_SYMBOL(ppc_md); | |
84 | struct machdep_calls *machine_id; | |
85 | EXPORT_SYMBOL(machine_id); | |
799d6046 | 86 | |
36ae37e3 BH |
87 | int boot_cpuid = -1; |
88 | EXPORT_SYMBOL_GPL(boot_cpuid); | |
89 | ||
49b09853 PM |
90 | unsigned long klimit = (unsigned long) _end; |
91 | ||
03501dab PM |
92 | /* |
93 | * This still seems to be needed... -- paulus | |
94 | */ | |
95 | struct screen_info screen_info = { | |
96 | .orig_x = 0, | |
97 | .orig_y = 25, | |
98 | .orig_video_cols = 80, | |
99 | .orig_video_lines = 25, | |
100 | .orig_video_isVGA = 1, | |
101 | .orig_video_points = 16 | |
102 | }; | |
e1802b06 AB |
103 | #if defined(CONFIG_FB_VGA16_MODULE) |
104 | EXPORT_SYMBOL(screen_info); | |
105 | #endif | |
03501dab | 106 | |
540c6c39 MW |
107 | /* Variables required to store legacy IO irq routing */ |
108 | int of_i8042_kbd_irq; | |
ee110066 | 109 | EXPORT_SYMBOL_GPL(of_i8042_kbd_irq); |
540c6c39 | 110 | int of_i8042_aux_irq; |
ee110066 | 111 | EXPORT_SYMBOL_GPL(of_i8042_aux_irq); |
540c6c39 | 112 | |
03501dab PM |
113 | #ifdef __DO_IRQ_CANON |
114 | /* XXX should go elsewhere eventually */ | |
115 | int ppc_do_canonicalize_irqs; | |
116 | EXPORT_SYMBOL(ppc_do_canonicalize_irqs); | |
117 | #endif | |
118 | ||
119 | /* also used by kexec */ | |
120 | void machine_shutdown(void) | |
121 | { | |
67b43b9d MS |
122 | #ifdef CONFIG_FA_DUMP |
123 | /* | |
124 | * if fadump is active, cleanup the fadump registration before we | |
125 | * shutdown. | |
126 | */ | |
127 | fadump_cleanup(); | |
128 | #endif | |
129 | ||
3d1229d6 ME |
130 | if (ppc_md.machine_shutdown) |
131 | ppc_md.machine_shutdown(); | |
03501dab PM |
132 | } |
133 | ||
134 | void machine_restart(char *cmd) | |
135 | { | |
136 | machine_shutdown(); | |
b8e383d5 KG |
137 | if (ppc_md.restart) |
138 | ppc_md.restart(cmd); | |
03501dab | 139 | smp_send_stop(); |
03501dab PM |
140 | printk(KERN_EMERG "System Halted, OK to turn off power\n"); |
141 | local_irq_disable(); | |
142 | while (1) ; | |
143 | } | |
144 | ||
145 | void machine_power_off(void) | |
146 | { | |
147 | machine_shutdown(); | |
9178ba29 AG |
148 | if (pm_power_off) |
149 | pm_power_off(); | |
03501dab | 150 | smp_send_stop(); |
03501dab PM |
151 | printk(KERN_EMERG "System Halted, OK to turn off power\n"); |
152 | local_irq_disable(); | |
153 | while (1) ; | |
154 | } | |
155 | /* Used by the G5 thermal driver */ | |
156 | EXPORT_SYMBOL_GPL(machine_power_off); | |
157 | ||
9178ba29 | 158 | void (*pm_power_off)(void); |
03501dab PM |
159 | EXPORT_SYMBOL_GPL(pm_power_off); |
160 | ||
161 | void machine_halt(void) | |
162 | { | |
163 | machine_shutdown(); | |
b8e383d5 KG |
164 | if (ppc_md.halt) |
165 | ppc_md.halt(); | |
03501dab | 166 | smp_send_stop(); |
03501dab PM |
167 | printk(KERN_EMERG "System Halted, OK to turn off power\n"); |
168 | local_irq_disable(); | |
169 | while (1) ; | |
170 | } | |
171 | ||
172 | ||
173 | #ifdef CONFIG_TAU | |
174 | extern u32 cpu_temp(unsigned long cpu); | |
175 | extern u32 cpu_temp_both(unsigned long cpu); | |
176 | #endif /* CONFIG_TAU */ | |
177 | ||
178 | #ifdef CONFIG_SMP | |
6b7487fc | 179 | DEFINE_PER_CPU(unsigned int, cpu_pvr); |
03501dab PM |
180 | #endif |
181 | ||
2c2df038 AB |
182 | static void show_cpuinfo_summary(struct seq_file *m) |
183 | { | |
184 | struct device_node *root; | |
185 | const char *model = NULL; | |
186 | #if defined(CONFIG_SMP) && defined(CONFIG_PPC32) | |
187 | unsigned long bogosum = 0; | |
188 | int i; | |
189 | for_each_online_cpu(i) | |
190 | bogosum += loops_per_jiffy; | |
191 | seq_printf(m, "total bogomips\t: %lu.%02lu\n", | |
192 | bogosum/(500000/HZ), bogosum/(5000/HZ) % 100); | |
193 | #endif /* CONFIG_SMP && CONFIG_PPC32 */ | |
194 | seq_printf(m, "timebase\t: %lu\n", ppc_tb_freq); | |
195 | if (ppc_md.name) | |
196 | seq_printf(m, "platform\t: %s\n", ppc_md.name); | |
197 | root = of_find_node_by_path("/"); | |
198 | if (root) | |
199 | model = of_get_property(root, "model", NULL); | |
200 | if (model) | |
201 | seq_printf(m, "model\t\t: %s\n", model); | |
202 | of_node_put(root); | |
203 | ||
204 | if (ppc_md.show_cpuinfo != NULL) | |
205 | ppc_md.show_cpuinfo(m); | |
206 | ||
207 | #ifdef CONFIG_PPC32 | |
208 | /* Display the amount of memory */ | |
209 | seq_printf(m, "Memory\t\t: %d MB\n", | |
210 | (unsigned int)(total_memory / (1024 * 1024))); | |
211 | #endif | |
212 | } | |
213 | ||
03501dab PM |
214 | static int show_cpuinfo(struct seq_file *m, void *v) |
215 | { | |
216 | unsigned long cpu_id = (unsigned long)v - 1; | |
217 | unsigned int pvr; | |
2299d03a | 218 | unsigned long proc_freq; |
03501dab PM |
219 | unsigned short maj; |
220 | unsigned short min; | |
221 | ||
03501dab PM |
222 | /* We only show online cpus: disable preempt (overzealous, I |
223 | * knew) to prevent cpu going down. */ | |
224 | preempt_disable(); | |
225 | if (!cpu_online(cpu_id)) { | |
226 | preempt_enable(); | |
227 | return 0; | |
228 | } | |
229 | ||
230 | #ifdef CONFIG_SMP | |
6b7487fc | 231 | pvr = per_cpu(cpu_pvr, cpu_id); |
03501dab PM |
232 | #else |
233 | pvr = mfspr(SPRN_PVR); | |
234 | #endif | |
235 | maj = (pvr >> 8) & 0xFF; | |
236 | min = pvr & 0xFF; | |
237 | ||
238 | seq_printf(m, "processor\t: %lu\n", cpu_id); | |
239 | seq_printf(m, "cpu\t\t: "); | |
240 | ||
241 | if (cur_cpu_spec->pvr_mask) | |
242 | seq_printf(m, "%s", cur_cpu_spec->cpu_name); | |
243 | else | |
244 | seq_printf(m, "unknown (%08x)", pvr); | |
245 | ||
246 | #ifdef CONFIG_ALTIVEC | |
247 | if (cpu_has_feature(CPU_FTR_ALTIVEC)) | |
248 | seq_printf(m, ", altivec supported"); | |
249 | #endif /* CONFIG_ALTIVEC */ | |
250 | ||
251 | seq_printf(m, "\n"); | |
252 | ||
253 | #ifdef CONFIG_TAU | |
254 | if (cur_cpu_spec->cpu_features & CPU_FTR_TAU) { | |
255 | #ifdef CONFIG_TAU_AVERAGE | |
256 | /* more straightforward, but potentially misleading */ | |
257 | seq_printf(m, "temperature \t: %u C (uncalibrated)\n", | |
bccfd588 | 258 | cpu_temp(cpu_id)); |
03501dab PM |
259 | #else |
260 | /* show the actual temp sensor range */ | |
261 | u32 temp; | |
bccfd588 | 262 | temp = cpu_temp_both(cpu_id); |
03501dab PM |
263 | seq_printf(m, "temperature \t: %u-%u C (uncalibrated)\n", |
264 | temp & 0xff, temp >> 16); | |
265 | #endif | |
266 | } | |
267 | #endif /* CONFIG_TAU */ | |
268 | ||
269 | /* | |
2299d03a GS |
270 | * Platforms that have variable clock rates, should implement |
271 | * the method ppc_md.get_proc_freq() that reports the clock | |
272 | * rate of a given cpu. The rest can use ppc_proc_freq to | |
273 | * report the clock rate that is same across all cpus. | |
03501dab | 274 | */ |
2299d03a GS |
275 | if (ppc_md.get_proc_freq) |
276 | proc_freq = ppc_md.get_proc_freq(cpu_id); | |
277 | else | |
278 | proc_freq = ppc_proc_freq; | |
279 | ||
280 | if (proc_freq) | |
03501dab | 281 | seq_printf(m, "clock\t\t: %lu.%06luMHz\n", |
2299d03a | 282 | proc_freq / 1000000, proc_freq % 1000000); |
03501dab PM |
283 | |
284 | if (ppc_md.show_percpuinfo != NULL) | |
285 | ppc_md.show_percpuinfo(m, cpu_id); | |
286 | ||
287 | /* If we are a Freescale core do a simple check so | |
288 | * we dont have to keep adding cases in the future */ | |
289 | if (PVR_VER(pvr) & 0x8000) { | |
a501d8f3 ML |
290 | switch (PVR_VER(pvr)) { |
291 | case 0x8000: /* 7441/7450/7451, Voyager */ | |
292 | case 0x8001: /* 7445/7455, Apollo 6 */ | |
293 | case 0x8002: /* 7447/7457, Apollo 7 */ | |
294 | case 0x8003: /* 7447A, Apollo 7 PM */ | |
295 | case 0x8004: /* 7448, Apollo 8 */ | |
296 | case 0x800c: /* 7410, Nitro */ | |
297 | maj = ((pvr >> 8) & 0xF); | |
298 | min = PVR_MIN(pvr); | |
299 | break; | |
300 | default: /* e500/book-e */ | |
301 | maj = PVR_MAJ(pvr); | |
302 | min = PVR_MIN(pvr); | |
303 | break; | |
304 | } | |
03501dab PM |
305 | } else { |
306 | switch (PVR_VER(pvr)) { | |
307 | case 0x0020: /* 403 family */ | |
308 | maj = PVR_MAJ(pvr) + 1; | |
309 | min = PVR_MIN(pvr); | |
310 | break; | |
311 | case 0x1008: /* 740P/750P ?? */ | |
312 | maj = ((pvr >> 8) & 0xFF) - 1; | |
313 | min = pvr & 0xFF; | |
314 | break; | |
315 | default: | |
316 | maj = (pvr >> 8) & 0xFF; | |
317 | min = pvr & 0xFF; | |
318 | break; | |
319 | } | |
320 | } | |
321 | ||
322 | seq_printf(m, "revision\t: %hd.%hd (pvr %04x %04x)\n", | |
323 | maj, min, PVR_VER(pvr), PVR_REV(pvr)); | |
324 | ||
325 | #ifdef CONFIG_PPC32 | |
326 | seq_printf(m, "bogomips\t: %lu.%02lu\n", | |
327 | loops_per_jiffy / (500000/HZ), | |
328 | (loops_per_jiffy / (5000/HZ)) % 100); | |
329 | #endif | |
330 | ||
331 | #ifdef CONFIG_SMP | |
332 | seq_printf(m, "\n"); | |
333 | #endif | |
334 | ||
335 | preempt_enable(); | |
e6532c63 AB |
336 | |
337 | /* If this is the last cpu, print the summary */ | |
338 | if (cpumask_next(cpu_id, cpu_online_mask) >= nr_cpu_ids) | |
339 | show_cpuinfo_summary(m); | |
340 | ||
03501dab PM |
341 | return 0; |
342 | } | |
343 | ||
344 | static void *c_start(struct seq_file *m, loff_t *pos) | |
345 | { | |
e6532c63 AB |
346 | if (*pos == 0) /* just in case, cpu 0 is not the first */ |
347 | *pos = cpumask_first(cpu_online_mask); | |
348 | else | |
349 | *pos = cpumask_next(*pos - 1, cpu_online_mask); | |
350 | if ((*pos) < nr_cpu_ids) | |
351 | return (void *)(unsigned long)(*pos + 1); | |
352 | return NULL; | |
03501dab PM |
353 | } |
354 | ||
355 | static void *c_next(struct seq_file *m, void *v, loff_t *pos) | |
356 | { | |
e6532c63 | 357 | (*pos)++; |
03501dab PM |
358 | return c_start(m, pos); |
359 | } | |
360 | ||
361 | static void c_stop(struct seq_file *m, void *v) | |
362 | { | |
363 | } | |
364 | ||
88e9d34c | 365 | const struct seq_operations cpuinfo_op = { |
03501dab PM |
366 | .start =c_start, |
367 | .next = c_next, | |
368 | .stop = c_stop, | |
369 | .show = show_cpuinfo, | |
370 | }; | |
371 | ||
a82765b6 DW |
372 | void __init check_for_initrd(void) |
373 | { | |
374 | #ifdef CONFIG_BLK_DEV_INITRD | |
30437b3e DG |
375 | DBG(" -> check_for_initrd() initrd_start=0x%lx initrd_end=0x%lx\n", |
376 | initrd_start, initrd_end); | |
a82765b6 DW |
377 | |
378 | /* If we were passed an initrd, set the ROOT_DEV properly if the values | |
379 | * look sensible. If not, clear initrd reference. | |
380 | */ | |
51fae6de | 381 | if (is_kernel_addr(initrd_start) && is_kernel_addr(initrd_end) && |
a82765b6 DW |
382 | initrd_end > initrd_start) |
383 | ROOT_DEV = Root_RAM0; | |
6761c4a0 | 384 | else |
a82765b6 | 385 | initrd_start = initrd_end = 0; |
a82765b6 DW |
386 | |
387 | if (initrd_start) | |
a7696b36 | 388 | pr_info("Found initrd at 0x%lx:0x%lx\n", initrd_start, initrd_end); |
a82765b6 DW |
389 | |
390 | DBG(" <- check_for_initrd()\n"); | |
391 | #endif /* CONFIG_BLK_DEV_INITRD */ | |
392 | } | |
393 | ||
5ad57078 PM |
394 | #ifdef CONFIG_SMP |
395 | ||
5853aef1 | 396 | int threads_per_core, threads_per_subcore, threads_shift; |
8d089085 | 397 | cpumask_t threads_core_mask; |
de56a948 | 398 | EXPORT_SYMBOL_GPL(threads_per_core); |
5853aef1 | 399 | EXPORT_SYMBOL_GPL(threads_per_subcore); |
de56a948 PM |
400 | EXPORT_SYMBOL_GPL(threads_shift); |
401 | EXPORT_SYMBOL_GPL(threads_core_mask); | |
8d089085 BH |
402 | |
403 | static void __init cpu_init_thread_core_maps(int tpc) | |
404 | { | |
405 | int i; | |
406 | ||
407 | threads_per_core = tpc; | |
5853aef1 | 408 | threads_per_subcore = tpc; |
104699c0 | 409 | cpumask_clear(&threads_core_mask); |
8d089085 BH |
410 | |
411 | /* This implementation only supports power of 2 number of threads | |
412 | * for simplicity and performance | |
413 | */ | |
414 | threads_shift = ilog2(tpc); | |
415 | BUG_ON(tpc != (1 << threads_shift)); | |
416 | ||
417 | for (i = 0; i < tpc; i++) | |
104699c0 | 418 | cpumask_set_cpu(i, &threads_core_mask); |
8d089085 BH |
419 | |
420 | printk(KERN_INFO "CPU maps initialized for %d thread%s per core\n", | |
421 | tpc, tpc > 1 ? "s" : ""); | |
422 | printk(KERN_DEBUG " (thread shift is %d)\n", threads_shift); | |
423 | } | |
424 | ||
425 | ||
5ad57078 PM |
426 | /** |
427 | * setup_cpu_maps - initialize the following cpu maps: | |
828a6986 AB |
428 | * cpu_possible_mask |
429 | * cpu_present_mask | |
5ad57078 PM |
430 | * |
431 | * Having the possible map set up early allows us to restrict allocations | |
8657ae28 | 432 | * of things like irqstacks to nr_cpu_ids rather than NR_CPUS. |
5ad57078 PM |
433 | * |
434 | * We do not initialize the online map here; cpus set their own bits in | |
828a6986 | 435 | * cpu_online_mask as they come up. |
5ad57078 PM |
436 | * |
437 | * This function is valid only for Open Firmware systems. finish_device_tree | |
438 | * must be called before using this. | |
439 | * | |
440 | * While we're here, we may as well set the "physical" cpu ids in the paca. | |
4df20460 AB |
441 | * |
442 | * NOTE: This must match the parsing done in early_init_dt_scan_cpus. | |
5ad57078 PM |
443 | */ |
444 | void __init smp_setup_cpu_maps(void) | |
445 | { | |
446 | struct device_node *dn = NULL; | |
447 | int cpu = 0; | |
8d089085 BH |
448 | int nthreads = 1; |
449 | ||
450 | DBG("smp_setup_cpu_maps()\n"); | |
5ad57078 | 451 | |
8657ae28 | 452 | while ((dn = of_find_node_by_type(dn, "cpu")) && cpu < nr_cpu_ids) { |
ac13282d | 453 | const __be32 *intserv; |
43f88120 | 454 | __be32 cpu_be; |
8d089085 BH |
455 | int j, len; |
456 | ||
457 | DBG(" * %s...\n", dn->full_name); | |
5ad57078 | 458 | |
e2eb6392 SR |
459 | intserv = of_get_property(dn, "ibm,ppc-interrupt-server#s", |
460 | &len); | |
8d089085 | 461 | if (intserv) { |
8d089085 BH |
462 | DBG(" ibm,ppc-interrupt-server#s -> %d threads\n", |
463 | nthreads); | |
464 | } else { | |
465 | DBG(" no ibm,ppc-interrupt-server#s -> 1 thread\n"); | |
e16c8765 | 466 | intserv = of_get_property(dn, "reg", &len); |
43f88120 AP |
467 | if (!intserv) { |
468 | cpu_be = cpu_to_be32(cpu); | |
469 | intserv = &cpu_be; /* assume logical == phys */ | |
e16c8765 | 470 | len = 4; |
43f88120 | 471 | } |
5ad57078 PM |
472 | } |
473 | ||
e16c8765 AF |
474 | nthreads = len / sizeof(int); |
475 | ||
8657ae28 | 476 | for (j = 0; j < nthreads && cpu < nr_cpu_ids; j++) { |
6663a4fa SW |
477 | bool avail; |
478 | ||
8d089085 | 479 | DBG(" thread %d -> cpu %d (hard id %d)\n", |
ac13282d | 480 | j, cpu, be32_to_cpu(intserv[j])); |
6663a4fa SW |
481 | |
482 | avail = of_device_is_available(dn); | |
483 | if (!avail) | |
484 | avail = !of_property_match_string(dn, | |
485 | "enable-method", "spin-table"); | |
486 | ||
487 | set_cpu_present(cpu, avail); | |
ac13282d | 488 | set_hard_smp_processor_id(cpu, be32_to_cpu(intserv[j])); |
ea0f1cab | 489 | set_cpu_possible(cpu, true); |
5ad57078 PM |
490 | cpu++; |
491 | } | |
492 | } | |
493 | ||
8d089085 BH |
494 | /* If no SMT supported, nthreads is forced to 1 */ |
495 | if (!cpu_has_feature(CPU_FTR_SMT)) { | |
496 | DBG(" SMT disabled ! nthreads forced to 1\n"); | |
497 | nthreads = 1; | |
498 | } | |
499 | ||
5ad57078 PM |
500 | #ifdef CONFIG_PPC64 |
501 | /* | |
502 | * On pSeries LPAR, we need to know how many cpus | |
503 | * could possibly be added to this partition. | |
504 | */ | |
0f2b3442 | 505 | if (firmware_has_feature(FW_FEATURE_LPAR) && |
799d6046 | 506 | (dn = of_find_node_by_path("/rtas"))) { |
5ad57078 | 507 | int num_addr_cell, num_size_cell, maxcpus; |
01666c8e | 508 | const __be32 *ireg; |
5ad57078 | 509 | |
a8bda5dd | 510 | num_addr_cell = of_n_addr_cells(dn); |
9213feea | 511 | num_size_cell = of_n_size_cells(dn); |
5ad57078 | 512 | |
e2eb6392 | 513 | ireg = of_get_property(dn, "ibm,lrdr-capacity", NULL); |
5ad57078 PM |
514 | |
515 | if (!ireg) | |
516 | goto out; | |
517 | ||
01666c8e | 518 | maxcpus = be32_to_cpup(ireg + num_addr_cell + num_size_cell); |
5ad57078 PM |
519 | |
520 | /* Double maxcpus for processors which have SMT capability */ | |
521 | if (cpu_has_feature(CPU_FTR_SMT)) | |
8d089085 | 522 | maxcpus *= nthreads; |
5ad57078 | 523 | |
8657ae28 | 524 | if (maxcpus > nr_cpu_ids) { |
5ad57078 PM |
525 | printk(KERN_WARNING |
526 | "Partition configured for %d cpus, " | |
527 | "operating system maximum is %d.\n", | |
8657ae28 MM |
528 | maxcpus, nr_cpu_ids); |
529 | maxcpus = nr_cpu_ids; | |
5ad57078 PM |
530 | } else |
531 | printk(KERN_INFO "Partition configured for %d cpus.\n", | |
532 | maxcpus); | |
533 | ||
534 | for (cpu = 0; cpu < maxcpus; cpu++) | |
ea0f1cab | 535 | set_cpu_possible(cpu, true); |
5ad57078 PM |
536 | out: |
537 | of_node_put(dn); | |
538 | } | |
d5a7430d MT |
539 | vdso_data->processorCount = num_present_cpus(); |
540 | #endif /* CONFIG_PPC64 */ | |
8d089085 BH |
541 | |
542 | /* Initialize CPU <=> thread mapping/ | |
543 | * | |
544 | * WARNING: We assume that the number of threads is the same for | |
545 | * every CPU in the system. If that is not the case, then some code | |
546 | * here will have to be reworked | |
547 | */ | |
548 | cpu_init_thread_core_maps(nthreads); | |
1426d5a3 | 549 | |
c1854e00 | 550 | /* Now that possible cpus are set, set nr_cpu_ids for later use */ |
aa79bc21 | 551 | setup_nr_cpu_ids(); |
c1854e00 | 552 | |
1426d5a3 | 553 | free_unused_pacas(); |
d5a7430d | 554 | } |
5ad57078 | 555 | #endif /* CONFIG_SMP */ |
fca5dcd4 | 556 | |
d33b78df | 557 | #ifdef CONFIG_PCSPKR_PLATFORM |
e5c6c8e4 MN |
558 | static __init int add_pcspkr(void) |
559 | { | |
560 | struct device_node *np; | |
561 | struct platform_device *pd; | |
562 | int ret; | |
563 | ||
564 | np = of_find_compatible_node(NULL, NULL, "pnpPNP,100"); | |
565 | of_node_put(np); | |
566 | if (!np) | |
567 | return -ENODEV; | |
568 | ||
569 | pd = platform_device_alloc("pcspkr", -1); | |
570 | if (!pd) | |
571 | return -ENOMEM; | |
572 | ||
573 | ret = platform_device_add(pd); | |
574 | if (ret) | |
575 | platform_device_put(pd); | |
576 | ||
577 | return ret; | |
578 | } | |
579 | device_initcall(add_pcspkr); | |
d33b78df | 580 | #endif /* CONFIG_PCSPKR_PLATFORM */ |
95d465fd | 581 | |
e8222502 BH |
582 | void probe_machine(void) |
583 | { | |
584 | extern struct machdep_calls __machine_desc_start; | |
585 | extern struct machdep_calls __machine_desc_end; | |
84b62c72 | 586 | unsigned int i; |
e8222502 BH |
587 | |
588 | /* | |
589 | * Iterate all ppc_md structures until we find the proper | |
590 | * one for the current machine type | |
591 | */ | |
592 | DBG("Probing machine type ...\n"); | |
593 | ||
84b62c72 BH |
594 | /* |
595 | * Check ppc_md is empty, if not we have a bug, ie, we setup an | |
596 | * entry before probe_machine() which will be overwritten | |
597 | */ | |
598 | for (i = 0; i < (sizeof(ppc_md) / sizeof(void *)); i++) { | |
599 | if (((void **)&ppc_md)[i]) { | |
600 | printk(KERN_ERR "Entry %d in ppc_md non empty before" | |
601 | " machine probe !\n", i); | |
602 | } | |
603 | } | |
604 | ||
e8222502 BH |
605 | for (machine_id = &__machine_desc_start; |
606 | machine_id < &__machine_desc_end; | |
607 | machine_id++) { | |
608 | DBG(" %s ...", machine_id->name); | |
609 | memcpy(&ppc_md, machine_id, sizeof(struct machdep_calls)); | |
610 | if (ppc_md.probe()) { | |
611 | DBG(" match !\n"); | |
612 | break; | |
613 | } | |
614 | DBG("\n"); | |
615 | } | |
616 | /* What can we do if we didn't find ? */ | |
617 | if (machine_id >= &__machine_desc_end) { | |
618 | DBG("No suitable machine found !\n"); | |
619 | for (;;); | |
620 | } | |
621 | ||
622 | printk(KERN_INFO "Using %s machine description\n", ppc_md.name); | |
623 | } | |
1269277a | 624 | |
8d8a0241 | 625 | /* Match a class of boards, not a specific device configuration. */ |
1269277a DW |
626 | int check_legacy_ioport(unsigned long base_port) |
627 | { | |
8d8a0241 OH |
628 | struct device_node *parent, *np = NULL; |
629 | int ret = -ENODEV; | |
630 | ||
631 | switch(base_port) { | |
632 | case I8042_DATA_REG: | |
db0dbae9 WF |
633 | if (!(np = of_find_compatible_node(NULL, NULL, "pnpPNP,303"))) |
634 | np = of_find_compatible_node(NULL, NULL, "pnpPNP,f03"); | |
635 | if (np) { | |
636 | parent = of_get_parent(np); | |
540c6c39 MW |
637 | |
638 | of_i8042_kbd_irq = irq_of_parse_and_map(parent, 0); | |
639 | if (!of_i8042_kbd_irq) | |
640 | of_i8042_kbd_irq = 1; | |
641 | ||
642 | of_i8042_aux_irq = irq_of_parse_and_map(parent, 1); | |
643 | if (!of_i8042_aux_irq) | |
644 | of_i8042_aux_irq = 12; | |
645 | ||
db0dbae9 WF |
646 | of_node_put(np); |
647 | np = parent; | |
648 | break; | |
649 | } | |
8d8a0241 | 650 | np = of_find_node_by_type(NULL, "8042"); |
f5d834fc AC |
651 | /* Pegasos has no device_type on its 8042 node, look for the |
652 | * name instead */ | |
653 | if (!np) | |
654 | np = of_find_node_by_name(NULL, "8042"); | |
2c78027a GP |
655 | if (np) { |
656 | of_i8042_kbd_irq = 1; | |
657 | of_i8042_aux_irq = 12; | |
658 | } | |
8d8a0241 OH |
659 | break; |
660 | case FDC_BASE: /* FDC1 */ | |
661 | np = of_find_node_by_type(NULL, "fdc"); | |
662 | break; | |
8d8a0241 OH |
663 | default: |
664 | /* ipmi is supposed to fail here */ | |
665 | break; | |
666 | } | |
667 | if (!np) | |
668 | return ret; | |
669 | parent = of_get_parent(np); | |
670 | if (parent) { | |
671 | if (strcmp(parent->type, "isa") == 0) | |
672 | ret = 0; | |
673 | of_node_put(parent); | |
674 | } | |
675 | of_node_put(np); | |
676 | return ret; | |
1269277a DW |
677 | } |
678 | EXPORT_SYMBOL(check_legacy_ioport); | |
7e990266 KG |
679 | |
680 | static int ppc_panic_event(struct notifier_block *this, | |
681 | unsigned long event, void *ptr) | |
682 | { | |
ebaeb5ae MS |
683 | /* |
684 | * If firmware-assisted dump has been registered then trigger | |
685 | * firmware-assisted dump and let firmware handle everything else. | |
686 | */ | |
687 | crash_fadump(NULL, ptr); | |
7e990266 KG |
688 | ppc_md.panic(ptr); /* May not return */ |
689 | return NOTIFY_DONE; | |
690 | } | |
691 | ||
692 | static struct notifier_block ppc_panic_block = { | |
693 | .notifier_call = ppc_panic_event, | |
694 | .priority = INT_MIN /* may not return; must be done last */ | |
695 | }; | |
696 | ||
697 | void __init setup_panic(void) | |
698 | { | |
f7b9ebb7 BH |
699 | if (!ppc_md.panic) |
700 | return; | |
7e990266 KG |
701 | atomic_notifier_chain_register(&panic_notifier_list, &ppc_panic_block); |
702 | } | |
06cce43c DF |
703 | |
704 | #ifdef CONFIG_CHECK_CACHE_COHERENCY | |
705 | /* | |
706 | * For platforms that have configurable cache-coherency. This function | |
707 | * checks that the cache coherency setting of the kernel matches the setting | |
708 | * left by the firmware, as indicated in the device tree. Since a mismatch | |
709 | * will eventually result in DMA failures, we print * and error and call | |
710 | * BUG() in that case. | |
711 | */ | |
712 | ||
713 | #ifdef CONFIG_NOT_COHERENT_CACHE | |
714 | #define KERNEL_COHERENCY 0 | |
715 | #else | |
716 | #define KERNEL_COHERENCY 1 | |
717 | #endif | |
718 | ||
719 | static int __init check_cache_coherency(void) | |
720 | { | |
721 | struct device_node *np; | |
722 | const void *prop; | |
723 | int devtree_coherency; | |
724 | ||
725 | np = of_find_node_by_path("/"); | |
726 | prop = of_get_property(np, "coherency-off", NULL); | |
727 | of_node_put(np); | |
728 | ||
729 | devtree_coherency = prop ? 0 : 1; | |
730 | ||
731 | if (devtree_coherency != KERNEL_COHERENCY) { | |
732 | printk(KERN_ERR | |
733 | "kernel coherency:%s != device tree_coherency:%s\n", | |
734 | KERNEL_COHERENCY ? "on" : "off", | |
735 | devtree_coherency ? "on" : "off"); | |
736 | BUG(); | |
737 | } | |
738 | ||
739 | return 0; | |
740 | } | |
741 | ||
742 | late_initcall(check_cache_coherency); | |
743 | #endif /* CONFIG_CHECK_CACHE_COHERENCY */ | |
94a3807c ME |
744 | |
745 | #ifdef CONFIG_DEBUG_FS | |
746 | struct dentry *powerpc_debugfs_root; | |
907b1f45 | 747 | EXPORT_SYMBOL(powerpc_debugfs_root); |
94a3807c ME |
748 | |
749 | static int powerpc_debugfs_init(void) | |
750 | { | |
751 | powerpc_debugfs_root = debugfs_create_dir("powerpc", NULL); | |
752 | ||
753 | return powerpc_debugfs_root == NULL; | |
754 | } | |
755 | arch_initcall(powerpc_debugfs_init); | |
756 | #endif | |
d746286c | 757 | |
a9c0f41b | 758 | void ppc_printk_progress(char *s, unsigned short hex) |
d746286c | 759 | { |
a9c0f41b | 760 | pr_info("%s\n", s); |
d746286c KG |
761 | } |
762 | ||
314b02f5 | 763 | void arch_setup_pdev_archdata(struct platform_device *pdev) |
d746286c | 764 | { |
314b02f5 KG |
765 | pdev->archdata.dma_mask = DMA_BIT_MASK(32); |
766 | pdev->dev.dma_mask = &pdev->archdata.dma_mask; | |
767 | set_dma_ops(&pdev->dev, &dma_direct_ops); | |
d746286c | 768 | } |
b1923caa BH |
769 | |
770 | static __init void print_system_info(void) | |
771 | { | |
772 | pr_info("-----------------------------------------------------\n"); | |
773 | #ifdef CONFIG_PPC_STD_MMU_64 | |
774 | pr_info("ppc64_pft_size = 0x%llx\n", ppc64_pft_size); | |
775 | #endif | |
776 | #ifdef CONFIG_PPC_STD_MMU_32 | |
777 | pr_info("Hash_size = 0x%lx\n", Hash_size); | |
778 | #endif | |
779 | pr_info("phys_mem_size = 0x%llx\n", | |
780 | (unsigned long long)memblock_phys_mem_size()); | |
781 | ||
782 | pr_info("dcache_bsize = 0x%x\n", dcache_bsize); | |
783 | pr_info("icache_bsize = 0x%x\n", icache_bsize); | |
784 | if (ucache_bsize != 0) | |
785 | pr_info("ucache_bsize = 0x%x\n", ucache_bsize); | |
786 | ||
787 | pr_info("cpu_features = 0x%016lx\n", cur_cpu_spec->cpu_features); | |
788 | pr_info(" possible = 0x%016lx\n", | |
789 | (unsigned long)CPU_FTRS_POSSIBLE); | |
790 | pr_info(" always = 0x%016lx\n", | |
791 | (unsigned long)CPU_FTRS_ALWAYS); | |
792 | pr_info("cpu_user_features = 0x%08x 0x%08x\n", | |
793 | cur_cpu_spec->cpu_user_features, | |
794 | cur_cpu_spec->cpu_user_features2); | |
795 | pr_info("mmu_features = 0x%08x\n", cur_cpu_spec->mmu_features); | |
796 | #ifdef CONFIG_PPC64 | |
797 | pr_info("firmware_features = 0x%016lx\n", powerpc_firmware_features); | |
798 | #endif | |
799 | ||
800 | #ifdef CONFIG_PPC_STD_MMU_64 | |
801 | if (htab_address) | |
802 | pr_info("htab_address = 0x%p\n", htab_address); | |
803 | if (htab_hash_mask) | |
804 | pr_info("htab_hash_mask = 0x%lx\n", htab_hash_mask); | |
805 | #endif | |
806 | #ifdef CONFIG_PPC_STD_MMU_32 | |
807 | if (Hash) | |
808 | pr_info("Hash = 0x%p\n", Hash); | |
809 | if (Hash_mask) | |
810 | pr_info("Hash_mask = 0x%lx\n", Hash_mask); | |
811 | #endif | |
812 | ||
813 | if (PHYSICAL_START > 0) | |
814 | pr_info("physical_start = 0x%llx\n", | |
815 | (unsigned long long)PHYSICAL_START); | |
816 | pr_info("-----------------------------------------------------\n"); | |
817 | } | |
818 | ||
819 | /* | |
820 | * Called into from start_kernel this initializes memblock, which is used | |
821 | * to manage page allocation until mem_init is called. | |
822 | */ | |
823 | void __init setup_arch(char **cmdline_p) | |
824 | { | |
825 | *cmdline_p = boot_command_line; | |
826 | ||
827 | /* Set a half-reasonable default so udelay does something sensible */ | |
828 | loops_per_jiffy = 500000000 / HZ; | |
829 | ||
830 | /* Unflatten the device-tree passed by prom_init or kexec */ | |
831 | unflatten_device_tree(); | |
832 | ||
833 | /* | |
834 | * Initialize cache line/block info from device-tree (on ppc64) or | |
835 | * just cputable (on ppc32). | |
836 | */ | |
837 | initialize_cache_info(); | |
838 | ||
839 | /* Initialize RTAS if available. */ | |
840 | rtas_initialize(); | |
841 | ||
842 | /* Check if we have an initrd provided via the device-tree. */ | |
843 | check_for_initrd(); | |
844 | ||
845 | /* Probe the machine type, establish ppc_md. */ | |
846 | probe_machine(); | |
847 | ||
848 | /* Setup panic notifier if requested by the platform. */ | |
849 | setup_panic(); | |
850 | ||
851 | /* | |
852 | * Configure ppc_md.power_save (ppc32 only, 64-bit machines do | |
853 | * it from their respective probe() function. | |
854 | */ | |
855 | setup_power_save(); | |
856 | ||
857 | /* Discover standard serial ports. */ | |
858 | find_legacy_serial_ports(); | |
859 | ||
860 | /* Register early console with the printk subsystem. */ | |
861 | register_early_udbg_console(); | |
862 | ||
863 | /* Setup the various CPU maps based on the device-tree. */ | |
864 | smp_setup_cpu_maps(); | |
865 | ||
866 | /* Initialize xmon. */ | |
867 | xmon_setup(); | |
868 | ||
869 | /* Check the SMT related command line arguments (ppc64). */ | |
870 | check_smt_enabled(); | |
871 | ||
872 | /* On BookE, setup per-core TLB data structures. */ | |
873 | setup_tlb_core_data(); | |
874 | ||
875 | /* | |
876 | * Release secondary cpus out of their spinloops at 0x60 now that | |
877 | * we can map physical -> logical CPU ids. | |
878 | * | |
879 | * Freescale Book3e parts spin in a loop provided by firmware, | |
880 | * so smp_release_cpus() does nothing for them. | |
881 | */ | |
882 | #ifdef CONFIG_SMP | |
883 | smp_release_cpus(); | |
884 | #endif | |
885 | ||
886 | /* Print various info about the machine that has been gathered so far. */ | |
887 | print_system_info(); | |
888 | ||
889 | /* Reserve large chunks of memory for use by CMA for KVM. */ | |
890 | kvm_cma_reserve(); | |
891 | ||
892 | /* | |
893 | * Reserve any gigantic pages requested on the command line. | |
894 | * memblock needs to have been initialized by the time this is | |
895 | * called since this will reserve memory. | |
896 | */ | |
897 | reserve_hugetlb_gpages(); | |
898 | ||
899 | klp_init_thread_info(&init_thread_info); | |
900 | ||
901 | init_mm.start_code = (unsigned long)_stext; | |
902 | init_mm.end_code = (unsigned long) _etext; | |
903 | init_mm.end_data = (unsigned long) _edata; | |
904 | init_mm.brk = klimit; | |
905 | #ifdef CONFIG_PPC_64K_PAGES | |
906 | init_mm.context.pte_frag = NULL; | |
907 | #endif | |
908 | #ifdef CONFIG_SPAPR_TCE_IOMMU | |
909 | mm_iommu_init(&init_mm.context); | |
910 | #endif | |
911 | irqstack_early_init(); | |
912 | exc_lvl_early_init(); | |
913 | emergency_stack_init(); | |
914 | ||
915 | initmem_init(); | |
916 | ||
917 | #ifdef CONFIG_DUMMY_CONSOLE | |
918 | conswitchp = &dummy_con; | |
919 | #endif | |
920 | if (ppc_md.setup_arch) | |
921 | ppc_md.setup_arch(); | |
922 | ||
923 | paging_init(); | |
924 | ||
925 | /* Initialize the MMU context management stuff. */ | |
926 | mmu_context_init(); | |
927 | ||
928 | #ifdef CONFIG_PPC64 | |
929 | /* Interrupt code needs to be 64K-aligned. */ | |
930 | if ((unsigned long)_stext & 0xffff) | |
931 | panic("Kernelbase not 64K-aligned (0x%lx)!\n", | |
932 | (unsigned long)_stext); | |
933 | #endif | |
934 | } |