powerpc: Explicit alignment for .data.cacheline_aligned
[deliverable/linux.git] / arch / powerpc / kernel / lparcfg.c
1 /*
2 * PowerPC64 LPAR Configuration Information Driver
3 *
4 * Dave Engebretsen engebret@us.ibm.com
5 * Copyright (c) 2003 Dave Engebretsen
6 * Will Schmidt willschm@us.ibm.com
7 * SPLPAR updates, Copyright (c) 2003 Will Schmidt IBM Corporation.
8 * seq_file updates, Copyright (c) 2004 Will Schmidt IBM Corporation.
9 * Nathan Lynch nathanl@austin.ibm.com
10 * Added lparcfg_write, Copyright (C) 2004 Nathan Lynch IBM Corporation.
11 *
12 * This program is free software; you can redistribute it and/or
13 * modify it under the terms of the GNU General Public License
14 * as published by the Free Software Foundation; either version
15 * 2 of the License, or (at your option) any later version.
16 *
17 * This driver creates a proc file at /proc/ppc64/lparcfg which contains
18 * keyword - value pairs that specify the configuration of the partition.
19 */
20
21 #include <linux/module.h>
22 #include <linux/types.h>
23 #include <linux/errno.h>
24 #include <linux/proc_fs.h>
25 #include <linux/init.h>
26 #include <linux/seq_file.h>
27 #include <asm/uaccess.h>
28 #include <asm/iseries/hv_lp_config.h>
29 #include <asm/lppaca.h>
30 #include <asm/hvcall.h>
31 #include <asm/firmware.h>
32 #include <asm/rtas.h>
33 #include <asm/system.h>
34 #include <asm/time.h>
35 #include <asm/prom.h>
36 #include <asm/vdso_datapage.h>
37 #include <asm/vio.h>
38
39 #define MODULE_VERS "1.8"
40 #define MODULE_NAME "lparcfg"
41
42 /* #define LPARCFG_DEBUG */
43
44 static struct proc_dir_entry *proc_ppc64_lparcfg;
45
46 /*
47 * Track sum of all purrs across all processors. This is used to further
48 * calculate usage values by different applications
49 */
50 static unsigned long get_purr(void)
51 {
52 unsigned long sum_purr = 0;
53 int cpu;
54
55 for_each_possible_cpu(cpu) {
56 if (firmware_has_feature(FW_FEATURE_ISERIES))
57 sum_purr += lppaca[cpu].emulated_time_base;
58 else {
59 struct cpu_usage *cu;
60
61 cu = &per_cpu(cpu_usage_array, cpu);
62 sum_purr += cu->current_tb;
63 }
64 }
65 return sum_purr;
66 }
67
68 #ifdef CONFIG_PPC_ISERIES
69
70 /*
71 * Methods used to fetch LPAR data when running on an iSeries platform.
72 */
73 static int iseries_lparcfg_data(struct seq_file *m, void *v)
74 {
75 unsigned long pool_id;
76 int shared, entitled_capacity, max_entitled_capacity;
77 int processors, max_processors;
78 unsigned long purr = get_purr();
79
80 shared = (int)(local_paca->lppaca_ptr->shared_proc);
81
82 seq_printf(m, "system_active_processors=%d\n",
83 (int)HvLpConfig_getSystemPhysicalProcessors());
84
85 seq_printf(m, "system_potential_processors=%d\n",
86 (int)HvLpConfig_getSystemPhysicalProcessors());
87
88 processors = (int)HvLpConfig_getPhysicalProcessors();
89 seq_printf(m, "partition_active_processors=%d\n", processors);
90
91 max_processors = (int)HvLpConfig_getMaxPhysicalProcessors();
92 seq_printf(m, "partition_potential_processors=%d\n", max_processors);
93
94 if (shared) {
95 entitled_capacity = HvLpConfig_getSharedProcUnits();
96 max_entitled_capacity = HvLpConfig_getMaxSharedProcUnits();
97 } else {
98 entitled_capacity = processors * 100;
99 max_entitled_capacity = max_processors * 100;
100 }
101 seq_printf(m, "partition_entitled_capacity=%d\n", entitled_capacity);
102
103 seq_printf(m, "partition_max_entitled_capacity=%d\n",
104 max_entitled_capacity);
105
106 if (shared) {
107 pool_id = HvLpConfig_getSharedPoolIndex();
108 seq_printf(m, "pool=%d\n", (int)pool_id);
109 seq_printf(m, "pool_capacity=%d\n",
110 (int)(HvLpConfig_getNumProcsInSharedPool(pool_id) *
111 100));
112 seq_printf(m, "purr=%ld\n", purr);
113 }
114
115 seq_printf(m, "shared_processor_mode=%d\n", shared);
116
117 return 0;
118 }
119
120 #else /* CONFIG_PPC_ISERIES */
121
122 static int iseries_lparcfg_data(struct seq_file *m, void *v)
123 {
124 return 0;
125 }
126
127 #endif /* CONFIG_PPC_ISERIES */
128
129 #ifdef CONFIG_PPC_PSERIES
130 /*
131 * Methods used to fetch LPAR data when running on a pSeries platform.
132 */
133 /**
134 * h_get_mpp
135 * H_GET_MPP hcall returns info in 7 parms
136 */
137 int h_get_mpp(struct hvcall_mpp_data *mpp_data)
138 {
139 int rc;
140 unsigned long retbuf[PLPAR_HCALL9_BUFSIZE];
141
142 rc = plpar_hcall9(H_GET_MPP, retbuf);
143
144 mpp_data->entitled_mem = retbuf[0];
145 mpp_data->mapped_mem = retbuf[1];
146
147 mpp_data->group_num = (retbuf[2] >> 2 * 8) & 0xffff;
148 mpp_data->pool_num = retbuf[2] & 0xffff;
149
150 mpp_data->mem_weight = (retbuf[3] >> 7 * 8) & 0xff;
151 mpp_data->unallocated_mem_weight = (retbuf[3] >> 6 * 8) & 0xff;
152 mpp_data->unallocated_entitlement = retbuf[3] & 0xffffffffffff;
153
154 mpp_data->pool_size = retbuf[4];
155 mpp_data->loan_request = retbuf[5];
156 mpp_data->backing_mem = retbuf[6];
157
158 return rc;
159 }
160 EXPORT_SYMBOL(h_get_mpp);
161
162 struct hvcall_ppp_data {
163 u64 entitlement;
164 u64 unallocated_entitlement;
165 u16 group_num;
166 u16 pool_num;
167 u8 capped;
168 u8 weight;
169 u8 unallocated_weight;
170 u16 active_procs_in_pool;
171 u16 active_system_procs;
172 };
173
174 /*
175 * H_GET_PPP hcall returns info in 4 parms.
176 * entitled_capacity,unallocated_capacity,
177 * aggregation, resource_capability).
178 *
179 * R4 = Entitled Processor Capacity Percentage.
180 * R5 = Unallocated Processor Capacity Percentage.
181 * R6 (AABBCCDDEEFFGGHH).
182 * XXXX - reserved (0)
183 * XXXX - reserved (0)
184 * XXXX - Group Number
185 * XXXX - Pool Number.
186 * R7 (IIJJKKLLMMNNOOPP).
187 * XX - reserved. (0)
188 * XX - bit 0-6 reserved (0). bit 7 is Capped indicator.
189 * XX - variable processor Capacity Weight
190 * XX - Unallocated Variable Processor Capacity Weight.
191 * XXXX - Active processors in Physical Processor Pool.
192 * XXXX - Processors active on platform.
193 */
194 static unsigned int h_get_ppp(struct hvcall_ppp_data *ppp_data)
195 {
196 unsigned long rc;
197 unsigned long retbuf[PLPAR_HCALL_BUFSIZE];
198
199 rc = plpar_hcall(H_GET_PPP, retbuf);
200
201 ppp_data->entitlement = retbuf[0];
202 ppp_data->unallocated_entitlement = retbuf[1];
203
204 ppp_data->group_num = (retbuf[2] >> 2 * 8) & 0xffff;
205 ppp_data->pool_num = retbuf[2] & 0xffff;
206
207 ppp_data->capped = (retbuf[3] >> 6 * 8) & 0x01;
208 ppp_data->weight = (retbuf[3] >> 5 * 8) & 0xff;
209 ppp_data->unallocated_weight = (retbuf[3] >> 4 * 8) & 0xff;
210 ppp_data->active_procs_in_pool = (retbuf[3] >> 2 * 8) & 0xffff;
211 ppp_data->active_system_procs = retbuf[3] & 0xffff;
212
213 return rc;
214 }
215
216 static unsigned h_pic(unsigned long *pool_idle_time,
217 unsigned long *num_procs)
218 {
219 unsigned long rc;
220 unsigned long retbuf[PLPAR_HCALL_BUFSIZE];
221
222 rc = plpar_hcall(H_PIC, retbuf);
223
224 *pool_idle_time = retbuf[0];
225 *num_procs = retbuf[1];
226
227 return rc;
228 }
229
230 /*
231 * parse_ppp_data
232 * Parse out the data returned from h_get_ppp and h_pic
233 */
234 static void parse_ppp_data(struct seq_file *m)
235 {
236 struct hvcall_ppp_data ppp_data;
237 int rc;
238
239 rc = h_get_ppp(&ppp_data);
240 if (rc)
241 return;
242
243 seq_printf(m, "partition_entitled_capacity=%lld\n",
244 ppp_data.entitlement);
245 seq_printf(m, "group=%d\n", ppp_data.group_num);
246 seq_printf(m, "system_active_processors=%d\n",
247 ppp_data.active_system_procs);
248
249 /* pool related entries are apropriate for shared configs */
250 if (lppaca[0].shared_proc) {
251 unsigned long pool_idle_time, pool_procs;
252
253 seq_printf(m, "pool=%d\n", ppp_data.pool_num);
254
255 /* report pool_capacity in percentage */
256 seq_printf(m, "pool_capacity=%d\n",
257 ppp_data.active_procs_in_pool * 100);
258
259 h_pic(&pool_idle_time, &pool_procs);
260 seq_printf(m, "pool_idle_time=%ld\n", pool_idle_time);
261 seq_printf(m, "pool_num_procs=%ld\n", pool_procs);
262 }
263
264 seq_printf(m, "unallocated_capacity_weight=%d\n",
265 ppp_data.unallocated_weight);
266 seq_printf(m, "capacity_weight=%d\n", ppp_data.weight);
267 seq_printf(m, "capped=%d\n", ppp_data.capped);
268 seq_printf(m, "unallocated_capacity=%lld\n",
269 ppp_data.unallocated_entitlement);
270 }
271
272 /**
273 * parse_mpp_data
274 * Parse out data returned from h_get_mpp
275 */
276 static void parse_mpp_data(struct seq_file *m)
277 {
278 struct hvcall_mpp_data mpp_data;
279 int rc;
280
281 rc = h_get_mpp(&mpp_data);
282 if (rc)
283 return;
284
285 seq_printf(m, "entitled_memory=%ld\n", mpp_data.entitled_mem);
286
287 if (mpp_data.mapped_mem != -1)
288 seq_printf(m, "mapped_entitled_memory=%ld\n",
289 mpp_data.mapped_mem);
290
291 seq_printf(m, "entitled_memory_group_number=%d\n", mpp_data.group_num);
292 seq_printf(m, "entitled_memory_pool_number=%d\n", mpp_data.pool_num);
293
294 seq_printf(m, "entitled_memory_weight=%d\n", mpp_data.mem_weight);
295 seq_printf(m, "unallocated_entitled_memory_weight=%d\n",
296 mpp_data.unallocated_mem_weight);
297 seq_printf(m, "unallocated_io_mapping_entitlement=%ld\n",
298 mpp_data.unallocated_entitlement);
299
300 if (mpp_data.pool_size != -1)
301 seq_printf(m, "entitled_memory_pool_size=%ld bytes\n",
302 mpp_data.pool_size);
303
304 seq_printf(m, "entitled_memory_loan_request=%ld\n",
305 mpp_data.loan_request);
306
307 seq_printf(m, "backing_memory=%ld bytes\n", mpp_data.backing_mem);
308 }
309
310 #define SPLPAR_CHARACTERISTICS_TOKEN 20
311 #define SPLPAR_MAXLENGTH 1026*(sizeof(char))
312
313 /*
314 * parse_system_parameter_string()
315 * Retrieve the potential_processors, max_entitled_capacity and friends
316 * through the get-system-parameter rtas call. Replace keyword strings as
317 * necessary.
318 */
319 static void parse_system_parameter_string(struct seq_file *m)
320 {
321 int call_status;
322
323 unsigned char *local_buffer = kmalloc(SPLPAR_MAXLENGTH, GFP_KERNEL);
324 if (!local_buffer) {
325 printk(KERN_ERR "%s %s kmalloc failure at line %d \n",
326 __FILE__, __func__, __LINE__);
327 return;
328 }
329
330 spin_lock(&rtas_data_buf_lock);
331 memset(rtas_data_buf, 0, SPLPAR_MAXLENGTH);
332 call_status = rtas_call(rtas_token("ibm,get-system-parameter"), 3, 1,
333 NULL,
334 SPLPAR_CHARACTERISTICS_TOKEN,
335 __pa(rtas_data_buf),
336 RTAS_DATA_BUF_SIZE);
337 memcpy(local_buffer, rtas_data_buf, SPLPAR_MAXLENGTH);
338 spin_unlock(&rtas_data_buf_lock);
339
340 if (call_status != 0) {
341 printk(KERN_INFO
342 "%s %s Error calling get-system-parameter (0x%x)\n",
343 __FILE__, __func__, call_status);
344 } else {
345 int splpar_strlen;
346 int idx, w_idx;
347 char *workbuffer = kzalloc(SPLPAR_MAXLENGTH, GFP_KERNEL);
348 if (!workbuffer) {
349 printk(KERN_ERR "%s %s kmalloc failure at line %d \n",
350 __FILE__, __func__, __LINE__);
351 kfree(local_buffer);
352 return;
353 }
354 #ifdef LPARCFG_DEBUG
355 printk(KERN_INFO "success calling get-system-parameter \n");
356 #endif
357 splpar_strlen = local_buffer[0] * 256 + local_buffer[1];
358 local_buffer += 2; /* step over strlen value */
359
360 w_idx = 0;
361 idx = 0;
362 while ((*local_buffer) && (idx < splpar_strlen)) {
363 workbuffer[w_idx++] = local_buffer[idx++];
364 if ((local_buffer[idx] == ',')
365 || (local_buffer[idx] == '\0')) {
366 workbuffer[w_idx] = '\0';
367 if (w_idx) {
368 /* avoid the empty string */
369 seq_printf(m, "%s\n", workbuffer);
370 }
371 memset(workbuffer, 0, SPLPAR_MAXLENGTH);
372 idx++; /* skip the comma */
373 w_idx = 0;
374 } else if (local_buffer[idx] == '=') {
375 /* code here to replace workbuffer contents
376 with different keyword strings */
377 if (0 == strcmp(workbuffer, "MaxEntCap")) {
378 strcpy(workbuffer,
379 "partition_max_entitled_capacity");
380 w_idx = strlen(workbuffer);
381 }
382 if (0 == strcmp(workbuffer, "MaxPlatProcs")) {
383 strcpy(workbuffer,
384 "system_potential_processors");
385 w_idx = strlen(workbuffer);
386 }
387 }
388 }
389 kfree(workbuffer);
390 local_buffer -= 2; /* back up over strlen value */
391 }
392 kfree(local_buffer);
393 }
394
395 /* Return the number of processors in the system.
396 * This function reads through the device tree and counts
397 * the virtual processors, this does not include threads.
398 */
399 static int lparcfg_count_active_processors(void)
400 {
401 struct device_node *cpus_dn = NULL;
402 int count = 0;
403
404 while ((cpus_dn = of_find_node_by_type(cpus_dn, "cpu"))) {
405 #ifdef LPARCFG_DEBUG
406 printk(KERN_ERR "cpus_dn %p \n", cpus_dn);
407 #endif
408 count++;
409 }
410 return count;
411 }
412
413 static void pseries_cmo_data(struct seq_file *m)
414 {
415 int cpu;
416 unsigned long cmo_faults = 0;
417 unsigned long cmo_fault_time = 0;
418
419 seq_printf(m, "cmo_enabled=%d\n", firmware_has_feature(FW_FEATURE_CMO));
420
421 if (!firmware_has_feature(FW_FEATURE_CMO))
422 return;
423
424 for_each_possible_cpu(cpu) {
425 cmo_faults += lppaca[cpu].cmo_faults;
426 cmo_fault_time += lppaca[cpu].cmo_fault_time;
427 }
428
429 seq_printf(m, "cmo_faults=%lu\n", cmo_faults);
430 seq_printf(m, "cmo_fault_time_usec=%lu\n",
431 cmo_fault_time / tb_ticks_per_usec);
432 seq_printf(m, "cmo_primary_psp=%d\n", cmo_get_primary_psp());
433 seq_printf(m, "cmo_secondary_psp=%d\n", cmo_get_secondary_psp());
434 seq_printf(m, "cmo_page_size=%lu\n", cmo_get_page_size());
435 }
436
437 static void splpar_dispatch_data(struct seq_file *m)
438 {
439 int cpu;
440 unsigned long dispatches = 0;
441 unsigned long dispatch_dispersions = 0;
442
443 for_each_possible_cpu(cpu) {
444 dispatches += lppaca[cpu].yield_count;
445 dispatch_dispersions += lppaca[cpu].dispersion_count;
446 }
447
448 seq_printf(m, "dispatches=%lu\n", dispatches);
449 seq_printf(m, "dispatch_dispersions=%lu\n", dispatch_dispersions);
450 }
451
452 static int pseries_lparcfg_data(struct seq_file *m, void *v)
453 {
454 int partition_potential_processors;
455 int partition_active_processors;
456 struct device_node *rtas_node;
457 const int *lrdrp = NULL;
458
459 rtas_node = of_find_node_by_path("/rtas");
460 if (rtas_node)
461 lrdrp = of_get_property(rtas_node, "ibm,lrdr-capacity", NULL);
462
463 if (lrdrp == NULL) {
464 partition_potential_processors = vdso_data->processorCount;
465 } else {
466 partition_potential_processors = *(lrdrp + 4);
467 }
468 of_node_put(rtas_node);
469
470 partition_active_processors = lparcfg_count_active_processors();
471
472 if (firmware_has_feature(FW_FEATURE_SPLPAR)) {
473 /* this call handles the ibm,get-system-parameter contents */
474 parse_system_parameter_string(m);
475 parse_ppp_data(m);
476 parse_mpp_data(m);
477 pseries_cmo_data(m);
478 splpar_dispatch_data(m);
479
480 seq_printf(m, "purr=%ld\n", get_purr());
481 } else { /* non SPLPAR case */
482
483 seq_printf(m, "system_active_processors=%d\n",
484 partition_potential_processors);
485
486 seq_printf(m, "system_potential_processors=%d\n",
487 partition_potential_processors);
488
489 seq_printf(m, "partition_max_entitled_capacity=%d\n",
490 partition_potential_processors * 100);
491
492 seq_printf(m, "partition_entitled_capacity=%d\n",
493 partition_active_processors * 100);
494 }
495
496 seq_printf(m, "partition_active_processors=%d\n",
497 partition_active_processors);
498
499 seq_printf(m, "partition_potential_processors=%d\n",
500 partition_potential_processors);
501
502 seq_printf(m, "shared_processor_mode=%d\n", lppaca[0].shared_proc);
503
504 return 0;
505 }
506
507 static ssize_t update_ppp(u64 *entitlement, u8 *weight)
508 {
509 struct hvcall_ppp_data ppp_data;
510 u8 new_weight;
511 u64 new_entitled;
512 ssize_t retval;
513
514 /* Get our current parameters */
515 retval = h_get_ppp(&ppp_data);
516 if (retval)
517 return retval;
518
519 if (entitlement) {
520 new_weight = ppp_data.weight;
521 new_entitled = *entitlement;
522 } else if (weight) {
523 new_weight = *weight;
524 new_entitled = ppp_data.entitlement;
525 } else
526 return -EINVAL;
527
528 pr_debug("%s: current_entitled = %llu, current_weight = %u\n",
529 __func__, ppp_data.entitlement, ppp_data.weight);
530
531 pr_debug("%s: new_entitled = %llu, new_weight = %u\n",
532 __func__, new_entitled, new_weight);
533
534 retval = plpar_hcall_norets(H_SET_PPP, new_entitled, new_weight);
535 return retval;
536 }
537
538 /**
539 * update_mpp
540 *
541 * Update the memory entitlement and weight for the partition. Caller must
542 * specify either a new entitlement or weight, not both, to be updated
543 * since the h_set_mpp call takes both entitlement and weight as parameters.
544 */
545 static ssize_t update_mpp(u64 *entitlement, u8 *weight)
546 {
547 struct hvcall_mpp_data mpp_data;
548 u64 new_entitled;
549 u8 new_weight;
550 ssize_t rc;
551
552 if (entitlement) {
553 /* Check with vio to ensure the new memory entitlement
554 * can be handled.
555 */
556 rc = vio_cmo_entitlement_update(*entitlement);
557 if (rc)
558 return rc;
559 }
560
561 rc = h_get_mpp(&mpp_data);
562 if (rc)
563 return rc;
564
565 if (entitlement) {
566 new_weight = mpp_data.mem_weight;
567 new_entitled = *entitlement;
568 } else if (weight) {
569 new_weight = *weight;
570 new_entitled = mpp_data.entitled_mem;
571 } else
572 return -EINVAL;
573
574 pr_debug("%s: current_entitled = %lu, current_weight = %u\n",
575 __func__, mpp_data.entitled_mem, mpp_data.mem_weight);
576
577 pr_debug("%s: new_entitled = %llu, new_weight = %u\n",
578 __func__, new_entitled, new_weight);
579
580 rc = plpar_hcall_norets(H_SET_MPP, new_entitled, new_weight);
581 return rc;
582 }
583
584 /*
585 * Interface for changing system parameters (variable capacity weight
586 * and entitled capacity). Format of input is "param_name=value";
587 * anything after value is ignored. Valid parameters at this time are
588 * "partition_entitled_capacity" and "capacity_weight". We use
589 * H_SET_PPP to alter parameters.
590 *
591 * This function should be invoked only on systems with
592 * FW_FEATURE_SPLPAR.
593 */
594 static ssize_t lparcfg_write(struct file *file, const char __user * buf,
595 size_t count, loff_t * off)
596 {
597 int kbuf_sz = 64;
598 char kbuf[kbuf_sz];
599 char *tmp;
600 u64 new_entitled, *new_entitled_ptr = &new_entitled;
601 u8 new_weight, *new_weight_ptr = &new_weight;
602 ssize_t retval;
603
604 if (!firmware_has_feature(FW_FEATURE_SPLPAR) ||
605 firmware_has_feature(FW_FEATURE_ISERIES))
606 return -EINVAL;
607
608 if (count > kbuf_sz)
609 return -EINVAL;
610
611 if (copy_from_user(kbuf, buf, count))
612 return -EFAULT;
613
614 kbuf[count - 1] = '\0';
615 tmp = strchr(kbuf, '=');
616 if (!tmp)
617 return -EINVAL;
618
619 *tmp++ = '\0';
620
621 if (!strcmp(kbuf, "partition_entitled_capacity")) {
622 char *endp;
623 *new_entitled_ptr = (u64) simple_strtoul(tmp, &endp, 10);
624 if (endp == tmp)
625 return -EINVAL;
626
627 retval = update_ppp(new_entitled_ptr, NULL);
628 } else if (!strcmp(kbuf, "capacity_weight")) {
629 char *endp;
630 *new_weight_ptr = (u8) simple_strtoul(tmp, &endp, 10);
631 if (endp == tmp)
632 return -EINVAL;
633
634 retval = update_ppp(NULL, new_weight_ptr);
635 } else if (!strcmp(kbuf, "entitled_memory")) {
636 char *endp;
637 *new_entitled_ptr = (u64) simple_strtoul(tmp, &endp, 10);
638 if (endp == tmp)
639 return -EINVAL;
640
641 retval = update_mpp(new_entitled_ptr, NULL);
642 } else if (!strcmp(kbuf, "entitled_memory_weight")) {
643 char *endp;
644 *new_weight_ptr = (u8) simple_strtoul(tmp, &endp, 10);
645 if (endp == tmp)
646 return -EINVAL;
647
648 retval = update_mpp(NULL, new_weight_ptr);
649 } else
650 return -EINVAL;
651
652 if (retval == H_SUCCESS || retval == H_CONSTRAINED) {
653 retval = count;
654 } else if (retval == H_BUSY) {
655 retval = -EBUSY;
656 } else if (retval == H_HARDWARE) {
657 retval = -EIO;
658 } else if (retval == H_PARAMETER) {
659 retval = -EINVAL;
660 }
661
662 return retval;
663 }
664
665 #else /* CONFIG_PPC_PSERIES */
666
667 static int pseries_lparcfg_data(struct seq_file *m, void *v)
668 {
669 return 0;
670 }
671
672 static ssize_t lparcfg_write(struct file *file, const char __user * buf,
673 size_t count, loff_t * off)
674 {
675 return -EINVAL;
676 }
677
678 #endif /* CONFIG_PPC_PSERIES */
679
680 static int lparcfg_data(struct seq_file *m, void *v)
681 {
682 struct device_node *rootdn;
683 const char *model = "";
684 const char *system_id = "";
685 const char *tmp;
686 const unsigned int *lp_index_ptr;
687 unsigned int lp_index = 0;
688
689 seq_printf(m, "%s %s \n", MODULE_NAME, MODULE_VERS);
690
691 rootdn = of_find_node_by_path("/");
692 if (rootdn) {
693 tmp = of_get_property(rootdn, "model", NULL);
694 if (tmp) {
695 model = tmp;
696 /* Skip "IBM," - see platforms/iseries/dt.c */
697 if (firmware_has_feature(FW_FEATURE_ISERIES))
698 model += 4;
699 }
700 tmp = of_get_property(rootdn, "system-id", NULL);
701 if (tmp) {
702 system_id = tmp;
703 /* Skip "IBM," - see platforms/iseries/dt.c */
704 if (firmware_has_feature(FW_FEATURE_ISERIES))
705 system_id += 4;
706 }
707 lp_index_ptr = of_get_property(rootdn, "ibm,partition-no",
708 NULL);
709 if (lp_index_ptr)
710 lp_index = *lp_index_ptr;
711 of_node_put(rootdn);
712 }
713 seq_printf(m, "serial_number=%s\n", system_id);
714 seq_printf(m, "system_type=%s\n", model);
715 seq_printf(m, "partition_id=%d\n", (int)lp_index);
716
717 if (firmware_has_feature(FW_FEATURE_ISERIES))
718 return iseries_lparcfg_data(m, v);
719 return pseries_lparcfg_data(m, v);
720 }
721
722 static int lparcfg_open(struct inode *inode, struct file *file)
723 {
724 return single_open(file, lparcfg_data, NULL);
725 }
726
727 static const struct file_operations lparcfg_fops = {
728 .owner = THIS_MODULE,
729 .read = seq_read,
730 .write = lparcfg_write,
731 .open = lparcfg_open,
732 .release = single_release,
733 };
734
735 static int __init lparcfg_init(void)
736 {
737 struct proc_dir_entry *ent;
738 mode_t mode = S_IRUSR | S_IRGRP | S_IROTH;
739
740 /* Allow writing if we have FW_FEATURE_SPLPAR */
741 if (firmware_has_feature(FW_FEATURE_SPLPAR) &&
742 !firmware_has_feature(FW_FEATURE_ISERIES))
743 mode |= S_IWUSR;
744
745 ent = proc_create("ppc64/lparcfg", mode, NULL, &lparcfg_fops);
746 if (!ent) {
747 printk(KERN_ERR "Failed to create ppc64/lparcfg\n");
748 return -EIO;
749 }
750
751 proc_ppc64_lparcfg = ent;
752 return 0;
753 }
754
755 static void __exit lparcfg_cleanup(void)
756 {
757 if (proc_ppc64_lparcfg)
758 remove_proc_entry("lparcfg", proc_ppc64_lparcfg->parent);
759 }
760
761 module_init(lparcfg_init);
762 module_exit(lparcfg_cleanup);
763 MODULE_DESCRIPTION("Interface for LPAR configuration data");
764 MODULE_AUTHOR("Dave Engebretsen");
765 MODULE_LICENSE("GPL");
This page took 0.059696 seconds and 5 git commands to generate.