Commit | Line | Data |
---|---|---|
1da177e4 LT |
1 | /* |
2 | * drivers/s390/cio/device_ops.c | |
3 | * | |
1da177e4 LT |
4 | * Copyright (C) 2002 IBM Deutschland Entwicklung GmbH, |
5 | * IBM Corporation | |
6 | * Author(s): Martin Schwidefsky (schwidefsky@de.ibm.com) | |
4ce3b30c | 7 | * Cornelia Huck (cornelia.huck@de.ibm.com) |
1da177e4 | 8 | */ |
1da177e4 LT |
9 | #include <linux/module.h> |
10 | #include <linux/init.h> | |
11 | #include <linux/errno.h> | |
12 | #include <linux/slab.h> | |
13 | #include <linux/list.h> | |
14 | #include <linux/device.h> | |
15 | #include <linux/delay.h> | |
16 | ||
17 | #include <asm/ccwdev.h> | |
18 | #include <asm/idals.h> | |
1da177e4 LT |
19 | |
20 | #include "cio.h" | |
21 | #include "cio_debug.h" | |
22 | #include "css.h" | |
23 | #include "chsc.h" | |
24 | #include "device.h" | |
1da177e4 LT |
25 | |
26 | int | |
27 | ccw_device_set_options(struct ccw_device *cdev, unsigned long flags) | |
28 | { | |
29 | /* | |
30 | * The flag usage is mutal exclusive ... | |
31 | */ | |
32 | if ((flags & CCWDEV_EARLY_NOTIFICATION) && | |
33 | (flags & CCWDEV_REPORT_ALL)) | |
34 | return -EINVAL; | |
35 | cdev->private->options.fast = (flags & CCWDEV_EARLY_NOTIFICATION) != 0; | |
36 | cdev->private->options.repall = (flags & CCWDEV_REPORT_ALL) != 0; | |
37 | cdev->private->options.pgroup = (flags & CCWDEV_DO_PATHGROUP) != 0; | |
38 | cdev->private->options.force = (flags & CCWDEV_ALLOW_FORCE) != 0; | |
39 | return 0; | |
40 | } | |
41 | ||
42 | int | |
43 | ccw_device_clear(struct ccw_device *cdev, unsigned long intparm) | |
44 | { | |
45 | struct subchannel *sch; | |
46 | int ret; | |
47 | ||
48 | if (!cdev) | |
49 | return -ENODEV; | |
50 | if (cdev->private->state == DEV_STATE_NOT_OPER) | |
51 | return -ENODEV; | |
52 | if (cdev->private->state != DEV_STATE_ONLINE && | |
53 | cdev->private->state != DEV_STATE_WAIT4IO && | |
54 | cdev->private->state != DEV_STATE_W4SENSE) | |
55 | return -EINVAL; | |
56 | sch = to_subchannel(cdev->dev.parent); | |
57 | if (!sch) | |
58 | return -ENODEV; | |
59 | ret = cio_clear(sch); | |
60 | if (ret == 0) | |
61 | cdev->private->intparm = intparm; | |
62 | return ret; | |
63 | } | |
64 | ||
65 | int | |
66 | ccw_device_start_key(struct ccw_device *cdev, struct ccw1 *cpa, | |
67 | unsigned long intparm, __u8 lpm, __u8 key, | |
68 | unsigned long flags) | |
69 | { | |
70 | struct subchannel *sch; | |
71 | int ret; | |
72 | ||
73 | if (!cdev) | |
74 | return -ENODEV; | |
75 | sch = to_subchannel(cdev->dev.parent); | |
76 | if (!sch) | |
77 | return -ENODEV; | |
78 | if (cdev->private->state == DEV_STATE_NOT_OPER) | |
79 | return -ENODEV; | |
b4f7b1ee CH |
80 | if (cdev->private->state == DEV_STATE_VERIFY || |
81 | cdev->private->state == DEV_STATE_CLEAR_VERIFY) { | |
1da177e4 LT |
82 | /* Remember to fake irb when finished. */ |
83 | if (!cdev->private->flags.fake_irb) { | |
84 | cdev->private->flags.fake_irb = 1; | |
85 | cdev->private->intparm = intparm; | |
86 | return 0; | |
87 | } else | |
88 | /* There's already a fake I/O around. */ | |
89 | return -EBUSY; | |
90 | } | |
91 | if (cdev->private->state != DEV_STATE_ONLINE || | |
92 | ((sch->schib.scsw.stctl & SCSW_STCTL_PRIM_STATUS) && | |
93 | !(sch->schib.scsw.stctl & SCSW_STCTL_SEC_STATUS)) || | |
94 | cdev->private->flags.doverify) | |
95 | return -EBUSY; | |
96 | ret = cio_set_options (sch, flags); | |
97 | if (ret) | |
98 | return ret; | |
99 | ret = cio_start_key (sch, cpa, lpm, key); | |
100 | if (ret == 0) | |
101 | cdev->private->intparm = intparm; | |
102 | return ret; | |
103 | } | |
104 | ||
105 | ||
106 | int | |
107 | ccw_device_start_timeout_key(struct ccw_device *cdev, struct ccw1 *cpa, | |
108 | unsigned long intparm, __u8 lpm, __u8 key, | |
109 | unsigned long flags, int expires) | |
110 | { | |
111 | int ret; | |
112 | ||
113 | if (!cdev) | |
114 | return -ENODEV; | |
115 | ccw_device_set_timeout(cdev, expires); | |
116 | ret = ccw_device_start_key(cdev, cpa, intparm, lpm, key, flags); | |
117 | if (ret != 0) | |
118 | ccw_device_set_timeout(cdev, 0); | |
119 | return ret; | |
120 | } | |
121 | ||
122 | int | |
123 | ccw_device_start(struct ccw_device *cdev, struct ccw1 *cpa, | |
124 | unsigned long intparm, __u8 lpm, unsigned long flags) | |
125 | { | |
126 | return ccw_device_start_key(cdev, cpa, intparm, lpm, | |
0b642ede | 127 | PAGE_DEFAULT_KEY, flags); |
1da177e4 LT |
128 | } |
129 | ||
130 | int | |
131 | ccw_device_start_timeout(struct ccw_device *cdev, struct ccw1 *cpa, | |
132 | unsigned long intparm, __u8 lpm, unsigned long flags, | |
133 | int expires) | |
134 | { | |
135 | return ccw_device_start_timeout_key(cdev, cpa, intparm, lpm, | |
0b642ede | 136 | PAGE_DEFAULT_KEY, flags, |
1da177e4 LT |
137 | expires); |
138 | } | |
139 | ||
140 | ||
141 | int | |
142 | ccw_device_halt(struct ccw_device *cdev, unsigned long intparm) | |
143 | { | |
144 | struct subchannel *sch; | |
145 | int ret; | |
146 | ||
147 | if (!cdev) | |
148 | return -ENODEV; | |
149 | if (cdev->private->state == DEV_STATE_NOT_OPER) | |
150 | return -ENODEV; | |
151 | if (cdev->private->state != DEV_STATE_ONLINE && | |
152 | cdev->private->state != DEV_STATE_WAIT4IO && | |
153 | cdev->private->state != DEV_STATE_W4SENSE) | |
154 | return -EINVAL; | |
155 | sch = to_subchannel(cdev->dev.parent); | |
156 | if (!sch) | |
157 | return -ENODEV; | |
158 | ret = cio_halt(sch); | |
159 | if (ret == 0) | |
160 | cdev->private->intparm = intparm; | |
161 | return ret; | |
162 | } | |
163 | ||
164 | int | |
165 | ccw_device_resume(struct ccw_device *cdev) | |
166 | { | |
167 | struct subchannel *sch; | |
168 | ||
169 | if (!cdev) | |
170 | return -ENODEV; | |
171 | sch = to_subchannel(cdev->dev.parent); | |
172 | if (!sch) | |
173 | return -ENODEV; | |
174 | if (cdev->private->state == DEV_STATE_NOT_OPER) | |
175 | return -ENODEV; | |
176 | if (cdev->private->state != DEV_STATE_ONLINE || | |
177 | !(sch->schib.scsw.actl & SCSW_ACTL_SUSPENDED)) | |
178 | return -EINVAL; | |
179 | return cio_resume(sch); | |
180 | } | |
181 | ||
182 | /* | |
183 | * Pass interrupt to device driver. | |
184 | */ | |
185 | int | |
186 | ccw_device_call_handler(struct ccw_device *cdev) | |
187 | { | |
188 | struct subchannel *sch; | |
189 | unsigned int stctl; | |
190 | int ending_status; | |
191 | ||
192 | sch = to_subchannel(cdev->dev.parent); | |
193 | ||
194 | /* | |
195 | * we allow for the device action handler if . | |
196 | * - we received ending status | |
197 | * - the action handler requested to see all interrupts | |
198 | * - we received an intermediate status | |
199 | * - fast notification was requested (primary status) | |
200 | * - unsolicited interrupts | |
201 | */ | |
202 | stctl = cdev->private->irb.scsw.stctl; | |
203 | ending_status = (stctl & SCSW_STCTL_SEC_STATUS) || | |
204 | (stctl == (SCSW_STCTL_ALERT_STATUS | SCSW_STCTL_STATUS_PEND)) || | |
205 | (stctl == SCSW_STCTL_STATUS_PEND); | |
206 | if (!ending_status && | |
207 | !cdev->private->options.repall && | |
208 | !(stctl & SCSW_STCTL_INTER_STATUS) && | |
209 | !(cdev->private->options.fast && | |
210 | (stctl & SCSW_STCTL_PRIM_STATUS))) | |
211 | return 0; | |
212 | ||
213 | /* | |
214 | * Now we are ready to call the device driver interrupt handler. | |
215 | */ | |
216 | if (cdev->handler) | |
217 | cdev->handler(cdev, cdev->private->intparm, | |
218 | &cdev->private->irb); | |
219 | ||
220 | /* | |
221 | * Clear the old and now useless interrupt response block. | |
222 | */ | |
223 | memset(&cdev->private->irb, 0, sizeof(struct irb)); | |
224 | ||
225 | return 1; | |
226 | } | |
227 | ||
228 | /* | |
229 | * Search for CIW command in extended sense data. | |
230 | */ | |
231 | struct ciw * | |
232 | ccw_device_get_ciw(struct ccw_device *cdev, __u32 ct) | |
233 | { | |
234 | int ciw_cnt; | |
235 | ||
236 | if (cdev->private->flags.esid == 0) | |
237 | return NULL; | |
238 | for (ciw_cnt = 0; ciw_cnt < MAX_CIWS; ciw_cnt++) | |
239 | if (cdev->private->senseid.ciw[ciw_cnt].ct == ct) | |
240 | return cdev->private->senseid.ciw + ciw_cnt; | |
241 | return NULL; | |
242 | } | |
243 | ||
244 | __u8 | |
245 | ccw_device_get_path_mask(struct ccw_device *cdev) | |
246 | { | |
247 | struct subchannel *sch; | |
248 | ||
249 | sch = to_subchannel(cdev->dev.parent); | |
250 | if (!sch) | |
251 | return 0; | |
252 | else | |
253 | return sch->vpm; | |
254 | } | |
255 | ||
256 | static void | |
257 | ccw_device_wake_up(struct ccw_device *cdev, unsigned long ip, struct irb *irb) | |
258 | { | |
259 | if (!ip) | |
260 | /* unsolicited interrupt */ | |
261 | return; | |
262 | ||
263 | /* Abuse intparm for error reporting. */ | |
264 | if (IS_ERR(irb)) | |
265 | cdev->private->intparm = -EIO; | |
4bc0c4f9 CH |
266 | else if (irb->scsw.cc == 1) |
267 | /* Retry for deferred condition code. */ | |
268 | cdev->private->intparm = -EAGAIN; | |
1da177e4 LT |
269 | else if ((irb->scsw.dstat != |
270 | (DEV_STAT_CHN_END|DEV_STAT_DEV_END)) || | |
271 | (irb->scsw.cstat != 0)) { | |
272 | /* | |
273 | * We didn't get channel end / device end. Check if path | |
274 | * verification has been started; we can retry after it has | |
275 | * finished. We also retry unit checks except for command reject | |
da039bd1 PO |
276 | * or intervention required. Also check for long busy |
277 | * conditions. | |
1da177e4 LT |
278 | */ |
279 | if (cdev->private->flags.doverify || | |
280 | cdev->private->state == DEV_STATE_VERIFY) | |
281 | cdev->private->intparm = -EAGAIN; | |
282 | if ((irb->scsw.dstat & DEV_STAT_UNIT_CHECK) && | |
283 | !(irb->ecw[0] & | |
284 | (SNS0_CMD_REJECT | SNS0_INTERVENTION_REQ))) | |
285 | cdev->private->intparm = -EAGAIN; | |
da039bd1 PO |
286 | else if ((irb->scsw.dstat & DEV_STAT_ATTENTION) && |
287 | (irb->scsw.dstat & DEV_STAT_DEV_END) && | |
288 | (irb->scsw.dstat & DEV_STAT_UNIT_EXCEP)) | |
289 | cdev->private->intparm = -EAGAIN; | |
1da177e4 LT |
290 | else |
291 | cdev->private->intparm = -EIO; | |
292 | ||
293 | } else | |
294 | cdev->private->intparm = 0; | |
295 | wake_up(&cdev->private->wait_q); | |
296 | } | |
297 | ||
298 | static inline int | |
299 | __ccw_device_retry_loop(struct ccw_device *cdev, struct ccw1 *ccw, long magic, __u8 lpm) | |
300 | { | |
301 | int ret; | |
302 | struct subchannel *sch; | |
303 | ||
304 | sch = to_subchannel(cdev->dev.parent); | |
305 | do { | |
306 | ret = cio_start (sch, ccw, lpm); | |
307 | if ((ret == -EBUSY) || (ret == -EACCES)) { | |
308 | /* Try again later. */ | |
309 | spin_unlock_irq(&sch->lock); | |
310 | msleep(10); | |
311 | spin_lock_irq(&sch->lock); | |
312 | continue; | |
313 | } | |
314 | if (ret != 0) | |
315 | /* Non-retryable error. */ | |
316 | break; | |
317 | /* Wait for end of request. */ | |
318 | cdev->private->intparm = magic; | |
319 | spin_unlock_irq(&sch->lock); | |
320 | wait_event(cdev->private->wait_q, | |
321 | (cdev->private->intparm == -EIO) || | |
322 | (cdev->private->intparm == -EAGAIN) || | |
323 | (cdev->private->intparm == 0)); | |
324 | spin_lock_irq(&sch->lock); | |
325 | /* Check at least for channel end / device end */ | |
326 | if (cdev->private->intparm == -EIO) { | |
327 | /* Non-retryable error. */ | |
328 | ret = -EIO; | |
329 | break; | |
330 | } | |
331 | if (cdev->private->intparm == 0) | |
332 | /* Success. */ | |
333 | break; | |
334 | /* Try again later. */ | |
335 | spin_unlock_irq(&sch->lock); | |
336 | msleep(10); | |
337 | spin_lock_irq(&sch->lock); | |
338 | } while (1); | |
339 | ||
340 | return ret; | |
341 | } | |
342 | ||
343 | /** | |
344 | * read_dev_chars() - read device characteristics | |
345 | * @param cdev target ccw device | |
346 | * @param buffer pointer to buffer for rdc data | |
347 | * @param length size of rdc data | |
348 | * @returns 0 for success, negative error value on failure | |
349 | * | |
350 | * Context: | |
351 | * called for online device, lock not held | |
352 | **/ | |
353 | int | |
354 | read_dev_chars (struct ccw_device *cdev, void **buffer, int length) | |
355 | { | |
356 | void (*handler)(struct ccw_device *, unsigned long, struct irb *); | |
357 | struct subchannel *sch; | |
358 | int ret; | |
359 | struct ccw1 *rdc_ccw; | |
360 | ||
361 | if (!cdev) | |
362 | return -ENODEV; | |
363 | if (!buffer || !length) | |
364 | return -EINVAL; | |
365 | sch = to_subchannel(cdev->dev.parent); | |
366 | ||
367 | CIO_TRACE_EVENT (4, "rddevch"); | |
368 | CIO_TRACE_EVENT (4, sch->dev.bus_id); | |
369 | ||
88abaab4 | 370 | rdc_ccw = kzalloc(sizeof(struct ccw1), GFP_KERNEL | GFP_DMA); |
1da177e4 LT |
371 | if (!rdc_ccw) |
372 | return -ENOMEM; | |
1da177e4 LT |
373 | rdc_ccw->cmd_code = CCW_CMD_RDC; |
374 | rdc_ccw->count = length; | |
375 | rdc_ccw->flags = CCW_FLAG_SLI; | |
376 | ret = set_normalized_cda (rdc_ccw, (*buffer)); | |
377 | if (ret != 0) { | |
378 | kfree(rdc_ccw); | |
379 | return ret; | |
380 | } | |
381 | ||
382 | spin_lock_irq(&sch->lock); | |
383 | /* Save interrupt handler. */ | |
384 | handler = cdev->handler; | |
385 | /* Temporarily install own handler. */ | |
386 | cdev->handler = ccw_device_wake_up; | |
387 | if (cdev->private->state != DEV_STATE_ONLINE) | |
388 | ret = -ENODEV; | |
389 | else if (((sch->schib.scsw.stctl & SCSW_STCTL_PRIM_STATUS) && | |
390 | !(sch->schib.scsw.stctl & SCSW_STCTL_SEC_STATUS)) || | |
391 | cdev->private->flags.doverify) | |
392 | ret = -EBUSY; | |
393 | else | |
394 | /* 0x00D9C4C3 == ebcdic "RDC" */ | |
395 | ret = __ccw_device_retry_loop(cdev, rdc_ccw, 0x00D9C4C3, 0); | |
396 | ||
397 | /* Restore interrupt handler. */ | |
398 | cdev->handler = handler; | |
399 | spin_unlock_irq(&sch->lock); | |
400 | ||
401 | clear_normalized_cda (rdc_ccw); | |
402 | kfree(rdc_ccw); | |
403 | ||
404 | return ret; | |
405 | } | |
406 | ||
407 | /* | |
408 | * Read Configuration data using path mask | |
409 | */ | |
410 | int | |
411 | read_conf_data_lpm (struct ccw_device *cdev, void **buffer, int *length, __u8 lpm) | |
412 | { | |
413 | void (*handler)(struct ccw_device *, unsigned long, struct irb *); | |
414 | struct subchannel *sch; | |
415 | struct ciw *ciw; | |
416 | char *rcd_buf; | |
417 | int ret; | |
418 | struct ccw1 *rcd_ccw; | |
419 | ||
420 | if (!cdev) | |
421 | return -ENODEV; | |
422 | if (!buffer || !length) | |
423 | return -EINVAL; | |
424 | sch = to_subchannel(cdev->dev.parent); | |
425 | ||
426 | CIO_TRACE_EVENT (4, "rdconf"); | |
427 | CIO_TRACE_EVENT (4, sch->dev.bus_id); | |
428 | ||
429 | /* | |
430 | * scan for RCD command in extended SenseID data | |
431 | */ | |
432 | ciw = ccw_device_get_ciw(cdev, CIW_TYPE_RCD); | |
433 | if (!ciw || ciw->cmd == 0) | |
434 | return -EOPNOTSUPP; | |
435 | ||
88abaab4 | 436 | rcd_ccw = kzalloc(sizeof(struct ccw1), GFP_KERNEL | GFP_DMA); |
1da177e4 LT |
437 | if (!rcd_ccw) |
438 | return -ENOMEM; | |
88abaab4 | 439 | rcd_buf = kzalloc(ciw->count, GFP_KERNEL | GFP_DMA); |
1da177e4 LT |
440 | if (!rcd_buf) { |
441 | kfree(rcd_ccw); | |
442 | return -ENOMEM; | |
443 | } | |
1da177e4 LT |
444 | rcd_ccw->cmd_code = ciw->cmd; |
445 | rcd_ccw->cda = (__u32) __pa (rcd_buf); | |
446 | rcd_ccw->count = ciw->count; | |
447 | rcd_ccw->flags = CCW_FLAG_SLI; | |
448 | ||
449 | spin_lock_irq(&sch->lock); | |
450 | /* Save interrupt handler. */ | |
451 | handler = cdev->handler; | |
452 | /* Temporarily install own handler. */ | |
453 | cdev->handler = ccw_device_wake_up; | |
454 | if (cdev->private->state != DEV_STATE_ONLINE) | |
455 | ret = -ENODEV; | |
456 | else if (((sch->schib.scsw.stctl & SCSW_STCTL_PRIM_STATUS) && | |
457 | !(sch->schib.scsw.stctl & SCSW_STCTL_SEC_STATUS)) || | |
458 | cdev->private->flags.doverify) | |
459 | ret = -EBUSY; | |
460 | else | |
461 | /* 0x00D9C3C4 == ebcdic "RCD" */ | |
462 | ret = __ccw_device_retry_loop(cdev, rcd_ccw, 0x00D9C3C4, lpm); | |
463 | ||
464 | /* Restore interrupt handler. */ | |
465 | cdev->handler = handler; | |
466 | spin_unlock_irq(&sch->lock); | |
467 | ||
468 | /* | |
469 | * on success we update the user input parms | |
470 | */ | |
471 | if (ret) { | |
472 | kfree (rcd_buf); | |
473 | *buffer = NULL; | |
474 | *length = 0; | |
475 | } else { | |
476 | *length = ciw->count; | |
477 | *buffer = rcd_buf; | |
478 | } | |
479 | kfree(rcd_ccw); | |
480 | ||
481 | return ret; | |
482 | } | |
483 | ||
484 | /* | |
485 | * Read Configuration data | |
486 | */ | |
487 | int | |
488 | read_conf_data (struct ccw_device *cdev, void **buffer, int *length) | |
489 | { | |
490 | return read_conf_data_lpm (cdev, buffer, length, 0); | |
491 | } | |
492 | ||
493 | /* | |
494 | * Try to break the lock on a boxed device. | |
495 | */ | |
496 | int | |
497 | ccw_device_stlck(struct ccw_device *cdev) | |
498 | { | |
499 | void *buf, *buf2; | |
500 | unsigned long flags; | |
501 | struct subchannel *sch; | |
502 | int ret; | |
503 | ||
504 | if (!cdev) | |
505 | return -ENODEV; | |
506 | ||
507 | if (cdev->drv && !cdev->private->options.force) | |
508 | return -EINVAL; | |
509 | ||
510 | sch = to_subchannel(cdev->dev.parent); | |
511 | ||
512 | CIO_TRACE_EVENT(2, "stl lock"); | |
513 | CIO_TRACE_EVENT(2, cdev->dev.bus_id); | |
514 | ||
515 | buf = kmalloc(32*sizeof(char), GFP_DMA|GFP_KERNEL); | |
516 | if (!buf) | |
517 | return -ENOMEM; | |
518 | buf2 = kmalloc(32*sizeof(char), GFP_DMA|GFP_KERNEL); | |
519 | if (!buf2) { | |
520 | kfree(buf); | |
521 | return -ENOMEM; | |
522 | } | |
523 | spin_lock_irqsave(&sch->lock, flags); | |
524 | ret = cio_enable_subchannel(sch, 3); | |
525 | if (ret) | |
526 | goto out_unlock; | |
527 | /* | |
528 | * Setup ccw. We chain an unconditional reserve and a release so we | |
529 | * only break the lock. | |
530 | */ | |
531 | cdev->private->iccws[0].cmd_code = CCW_CMD_STLCK; | |
532 | cdev->private->iccws[0].cda = (__u32) __pa(buf); | |
533 | cdev->private->iccws[0].count = 32; | |
534 | cdev->private->iccws[0].flags = CCW_FLAG_CC; | |
535 | cdev->private->iccws[1].cmd_code = CCW_CMD_RELEASE; | |
536 | cdev->private->iccws[1].cda = (__u32) __pa(buf2); | |
537 | cdev->private->iccws[1].count = 32; | |
538 | cdev->private->iccws[1].flags = 0; | |
539 | ret = cio_start(sch, cdev->private->iccws, 0); | |
540 | if (ret) { | |
541 | cio_disable_subchannel(sch); //FIXME: return code? | |
542 | goto out_unlock; | |
543 | } | |
544 | cdev->private->irb.scsw.actl |= SCSW_ACTL_START_PEND; | |
545 | spin_unlock_irqrestore(&sch->lock, flags); | |
546 | wait_event(cdev->private->wait_q, cdev->private->irb.scsw.actl == 0); | |
547 | spin_lock_irqsave(&sch->lock, flags); | |
548 | cio_disable_subchannel(sch); //FIXME: return code? | |
549 | if ((cdev->private->irb.scsw.dstat != | |
550 | (DEV_STAT_CHN_END|DEV_STAT_DEV_END)) || | |
551 | (cdev->private->irb.scsw.cstat != 0)) | |
552 | ret = -EIO; | |
553 | /* Clear irb. */ | |
554 | memset(&cdev->private->irb, 0, sizeof(struct irb)); | |
555 | out_unlock: | |
17fd682e JJ |
556 | kfree(buf); |
557 | kfree(buf2); | |
1da177e4 LT |
558 | spin_unlock_irqrestore(&sch->lock, flags); |
559 | return ret; | |
560 | } | |
561 | ||
562 | void * | |
563 | ccw_device_get_chp_desc(struct ccw_device *cdev, int chp_no) | |
564 | { | |
565 | struct subchannel *sch; | |
566 | ||
567 | sch = to_subchannel(cdev->dev.parent); | |
568 | return chsc_get_chp_desc(sch, chp_no); | |
569 | } | |
570 | ||
571 | // FIXME: these have to go: | |
572 | ||
573 | int | |
574 | _ccw_device_get_subchannel_number(struct ccw_device *cdev) | |
575 | { | |
a8237fc4 | 576 | return cdev->private->sch_no; |
1da177e4 LT |
577 | } |
578 | ||
579 | int | |
580 | _ccw_device_get_device_number(struct ccw_device *cdev) | |
581 | { | |
582 | return cdev->private->devno; | |
583 | } | |
584 | ||
585 | ||
586 | MODULE_LICENSE("GPL"); | |
587 | EXPORT_SYMBOL(ccw_device_set_options); | |
588 | EXPORT_SYMBOL(ccw_device_clear); | |
589 | EXPORT_SYMBOL(ccw_device_halt); | |
590 | EXPORT_SYMBOL(ccw_device_resume); | |
591 | EXPORT_SYMBOL(ccw_device_start_timeout); | |
592 | EXPORT_SYMBOL(ccw_device_start); | |
593 | EXPORT_SYMBOL(ccw_device_start_timeout_key); | |
594 | EXPORT_SYMBOL(ccw_device_start_key); | |
595 | EXPORT_SYMBOL(ccw_device_get_ciw); | |
596 | EXPORT_SYMBOL(ccw_device_get_path_mask); | |
597 | EXPORT_SYMBOL(read_conf_data); | |
598 | EXPORT_SYMBOL(read_dev_chars); | |
599 | EXPORT_SYMBOL(_ccw_device_get_subchannel_number); | |
600 | EXPORT_SYMBOL(_ccw_device_get_device_number); | |
601 | EXPORT_SYMBOL_GPL(ccw_device_get_chp_desc); | |
602 | EXPORT_SYMBOL_GPL(read_conf_data_lpm); |