tpm: two-phase chip management functions
[deliverable/linux.git] / drivers / char / tpm / xen-tpmfront.c
1 /*
2 * Implementation of the Xen vTPM device frontend
3 *
4 * Author: Daniel De Graaf <dgdegra@tycho.nsa.gov>
5 *
6 * This program is free software; you can redistribute it and/or modify
7 * it under the terms of the GNU General Public License version 2,
8 * as published by the Free Software Foundation.
9 */
10 #include <linux/errno.h>
11 #include <linux/err.h>
12 #include <linux/interrupt.h>
13 #include <xen/xen.h>
14 #include <xen/events.h>
15 #include <xen/interface/io/tpmif.h>
16 #include <xen/grant_table.h>
17 #include <xen/xenbus.h>
18 #include <xen/page.h>
19 #include "tpm.h"
20 #include <xen/platform_pci.h>
21
22 struct tpm_private {
23 struct tpm_chip *chip;
24 struct xenbus_device *dev;
25
26 struct vtpm_shared_page *shr;
27
28 unsigned int evtchn;
29 int ring_ref;
30 domid_t backend_id;
31 };
32
33 enum status_bits {
34 VTPM_STATUS_RUNNING = 0x1,
35 VTPM_STATUS_IDLE = 0x2,
36 VTPM_STATUS_RESULT = 0x4,
37 VTPM_STATUS_CANCELED = 0x8,
38 };
39
40 static u8 vtpm_status(struct tpm_chip *chip)
41 {
42 struct tpm_private *priv = TPM_VPRIV(chip);
43 switch (priv->shr->state) {
44 case VTPM_STATE_IDLE:
45 return VTPM_STATUS_IDLE | VTPM_STATUS_CANCELED;
46 case VTPM_STATE_FINISH:
47 return VTPM_STATUS_IDLE | VTPM_STATUS_RESULT;
48 case VTPM_STATE_SUBMIT:
49 case VTPM_STATE_CANCEL: /* cancel requested, not yet canceled */
50 return VTPM_STATUS_RUNNING;
51 default:
52 return 0;
53 }
54 }
55
56 static bool vtpm_req_canceled(struct tpm_chip *chip, u8 status)
57 {
58 return status & VTPM_STATUS_CANCELED;
59 }
60
61 static void vtpm_cancel(struct tpm_chip *chip)
62 {
63 struct tpm_private *priv = TPM_VPRIV(chip);
64 priv->shr->state = VTPM_STATE_CANCEL;
65 wmb();
66 notify_remote_via_evtchn(priv->evtchn);
67 }
68
69 static unsigned int shr_data_offset(struct vtpm_shared_page *shr)
70 {
71 return sizeof(*shr) + sizeof(u32) * shr->nr_extra_pages;
72 }
73
74 static int vtpm_send(struct tpm_chip *chip, u8 *buf, size_t count)
75 {
76 struct tpm_private *priv = TPM_VPRIV(chip);
77 struct vtpm_shared_page *shr = priv->shr;
78 unsigned int offset = shr_data_offset(shr);
79
80 u32 ordinal;
81 unsigned long duration;
82
83 if (offset > PAGE_SIZE)
84 return -EINVAL;
85
86 if (offset + count > PAGE_SIZE)
87 return -EINVAL;
88
89 /* Wait for completion of any existing command or cancellation */
90 if (wait_for_tpm_stat(chip, VTPM_STATUS_IDLE, chip->vendor.timeout_c,
91 &chip->vendor.read_queue, true) < 0) {
92 vtpm_cancel(chip);
93 return -ETIME;
94 }
95
96 memcpy(offset + (u8 *)shr, buf, count);
97 shr->length = count;
98 barrier();
99 shr->state = VTPM_STATE_SUBMIT;
100 wmb();
101 notify_remote_via_evtchn(priv->evtchn);
102
103 ordinal = be32_to_cpu(((struct tpm_input_header*)buf)->ordinal);
104 duration = tpm_calc_ordinal_duration(chip, ordinal);
105
106 if (wait_for_tpm_stat(chip, VTPM_STATUS_IDLE, duration,
107 &chip->vendor.read_queue, true) < 0) {
108 /* got a signal or timeout, try to cancel */
109 vtpm_cancel(chip);
110 return -ETIME;
111 }
112
113 return count;
114 }
115
116 static int vtpm_recv(struct tpm_chip *chip, u8 *buf, size_t count)
117 {
118 struct tpm_private *priv = TPM_VPRIV(chip);
119 struct vtpm_shared_page *shr = priv->shr;
120 unsigned int offset = shr_data_offset(shr);
121 size_t length = shr->length;
122
123 if (shr->state == VTPM_STATE_IDLE)
124 return -ECANCELED;
125
126 /* In theory the wait at the end of _send makes this one unnecessary */
127 if (wait_for_tpm_stat(chip, VTPM_STATUS_RESULT, chip->vendor.timeout_c,
128 &chip->vendor.read_queue, true) < 0) {
129 vtpm_cancel(chip);
130 return -ETIME;
131 }
132
133 if (offset > PAGE_SIZE)
134 return -EIO;
135
136 if (offset + length > PAGE_SIZE)
137 length = PAGE_SIZE - offset;
138
139 if (length > count)
140 length = count;
141
142 memcpy(buf, offset + (u8 *)shr, length);
143
144 return length;
145 }
146
147 static const struct tpm_class_ops tpm_vtpm = {
148 .status = vtpm_status,
149 .recv = vtpm_recv,
150 .send = vtpm_send,
151 .cancel = vtpm_cancel,
152 .req_complete_mask = VTPM_STATUS_IDLE | VTPM_STATUS_RESULT,
153 .req_complete_val = VTPM_STATUS_IDLE | VTPM_STATUS_RESULT,
154 .req_canceled = vtpm_req_canceled,
155 };
156
157 static irqreturn_t tpmif_interrupt(int dummy, void *dev_id)
158 {
159 struct tpm_private *priv = dev_id;
160
161 switch (priv->shr->state) {
162 case VTPM_STATE_IDLE:
163 case VTPM_STATE_FINISH:
164 wake_up_interruptible(&priv->chip->vendor.read_queue);
165 break;
166 case VTPM_STATE_SUBMIT:
167 case VTPM_STATE_CANCEL:
168 default:
169 break;
170 }
171 return IRQ_HANDLED;
172 }
173
174 static int setup_chip(struct device *dev, struct tpm_private *priv)
175 {
176 struct tpm_chip *chip;
177
178 chip = tpmm_chip_alloc(dev, &tpm_vtpm);
179 if (IS_ERR(chip))
180 return PTR_ERR(chip);
181
182 init_waitqueue_head(&chip->vendor.read_queue);
183
184 priv->chip = chip;
185 TPM_VPRIV(chip) = priv;
186
187 return 0;
188 }
189
190 /* caller must clean up in case of errors */
191 static int setup_ring(struct xenbus_device *dev, struct tpm_private *priv)
192 {
193 struct xenbus_transaction xbt;
194 const char *message = NULL;
195 int rv;
196
197 priv->shr = (void *)__get_free_page(GFP_KERNEL|__GFP_ZERO);
198 if (!priv->shr) {
199 xenbus_dev_fatal(dev, -ENOMEM, "allocating shared ring");
200 return -ENOMEM;
201 }
202
203 rv = xenbus_grant_ring(dev, virt_to_mfn(priv->shr));
204 if (rv < 0)
205 return rv;
206
207 priv->ring_ref = rv;
208
209 rv = xenbus_alloc_evtchn(dev, &priv->evtchn);
210 if (rv)
211 return rv;
212
213 rv = bind_evtchn_to_irqhandler(priv->evtchn, tpmif_interrupt, 0,
214 "tpmif", priv);
215 if (rv <= 0) {
216 xenbus_dev_fatal(dev, rv, "allocating TPM irq");
217 return rv;
218 }
219 priv->chip->vendor.irq = rv;
220
221 again:
222 rv = xenbus_transaction_start(&xbt);
223 if (rv) {
224 xenbus_dev_fatal(dev, rv, "starting transaction");
225 return rv;
226 }
227
228 rv = xenbus_printf(xbt, dev->nodename,
229 "ring-ref", "%u", priv->ring_ref);
230 if (rv) {
231 message = "writing ring-ref";
232 goto abort_transaction;
233 }
234
235 rv = xenbus_printf(xbt, dev->nodename, "event-channel", "%u",
236 priv->evtchn);
237 if (rv) {
238 message = "writing event-channel";
239 goto abort_transaction;
240 }
241
242 rv = xenbus_printf(xbt, dev->nodename, "feature-protocol-v2", "1");
243 if (rv) {
244 message = "writing feature-protocol-v2";
245 goto abort_transaction;
246 }
247
248 rv = xenbus_transaction_end(xbt, 0);
249 if (rv == -EAGAIN)
250 goto again;
251 if (rv) {
252 xenbus_dev_fatal(dev, rv, "completing transaction");
253 return rv;
254 }
255
256 xenbus_switch_state(dev, XenbusStateInitialised);
257
258 return 0;
259
260 abort_transaction:
261 xenbus_transaction_end(xbt, 1);
262 if (message)
263 xenbus_dev_error(dev, rv, "%s", message);
264
265 return rv;
266 }
267
268 static void ring_free(struct tpm_private *priv)
269 {
270 if (!priv)
271 return;
272
273 if (priv->ring_ref)
274 gnttab_end_foreign_access(priv->ring_ref, 0,
275 (unsigned long)priv->shr);
276 else
277 free_page((unsigned long)priv->shr);
278
279 if (priv->chip && priv->chip->vendor.irq)
280 unbind_from_irqhandler(priv->chip->vendor.irq, priv);
281
282 kfree(priv);
283 }
284
285 static int tpmfront_probe(struct xenbus_device *dev,
286 const struct xenbus_device_id *id)
287 {
288 struct tpm_private *priv;
289 struct tpm_chip *chip;
290 int rv;
291
292 priv = kzalloc(sizeof(*priv), GFP_KERNEL);
293 if (!priv) {
294 xenbus_dev_fatal(dev, -ENOMEM, "allocating priv structure");
295 return -ENOMEM;
296 }
297
298 rv = setup_chip(&dev->dev, priv);
299 if (rv) {
300 kfree(priv);
301 return rv;
302 }
303
304 rv = setup_ring(dev, priv);
305 if (rv) {
306 chip = dev_get_drvdata(&dev->dev);
307 tpm_chip_unregister(chip);
308 ring_free(priv);
309 return rv;
310 }
311
312 tpm_get_timeouts(priv->chip);
313
314 return tpm_chip_register(priv->chip);
315 }
316
317 static int tpmfront_remove(struct xenbus_device *dev)
318 {
319 struct tpm_chip *chip = dev_get_drvdata(&dev->dev);
320 struct tpm_private *priv = TPM_VPRIV(chip);
321 tpm_chip_unregister(chip);
322 ring_free(priv);
323 TPM_VPRIV(chip) = NULL;
324 return 0;
325 }
326
327 static int tpmfront_resume(struct xenbus_device *dev)
328 {
329 /* A suspend/resume/migrate will interrupt a vTPM anyway */
330 tpmfront_remove(dev);
331 return tpmfront_probe(dev, NULL);
332 }
333
334 static void backend_changed(struct xenbus_device *dev,
335 enum xenbus_state backend_state)
336 {
337 int val;
338
339 switch (backend_state) {
340 case XenbusStateInitialised:
341 case XenbusStateConnected:
342 if (dev->state == XenbusStateConnected)
343 break;
344
345 if (xenbus_scanf(XBT_NIL, dev->otherend,
346 "feature-protocol-v2", "%d", &val) < 0)
347 val = 0;
348 if (!val) {
349 xenbus_dev_fatal(dev, -EINVAL,
350 "vTPM protocol 2 required");
351 return;
352 }
353 xenbus_switch_state(dev, XenbusStateConnected);
354 break;
355
356 case XenbusStateClosing:
357 case XenbusStateClosed:
358 device_unregister(&dev->dev);
359 xenbus_frontend_closed(dev);
360 break;
361 default:
362 break;
363 }
364 }
365
366 static const struct xenbus_device_id tpmfront_ids[] = {
367 { "vtpm" },
368 { "" }
369 };
370 MODULE_ALIAS("xen:vtpm");
371
372 static struct xenbus_driver tpmfront_driver = {
373 .ids = tpmfront_ids,
374 .probe = tpmfront_probe,
375 .remove = tpmfront_remove,
376 .resume = tpmfront_resume,
377 .otherend_changed = backend_changed,
378 };
379
380 static int __init xen_tpmfront_init(void)
381 {
382 if (!xen_domain())
383 return -ENODEV;
384
385 if (!xen_has_pv_devices())
386 return -ENODEV;
387
388 return xenbus_register_frontend(&tpmfront_driver);
389 }
390 module_init(xen_tpmfront_init);
391
392 static void __exit xen_tpmfront_exit(void)
393 {
394 xenbus_unregister_driver(&tpmfront_driver);
395 }
396 module_exit(xen_tpmfront_exit);
397
398 MODULE_AUTHOR("Daniel De Graaf <dgdegra@tycho.nsa.gov>");
399 MODULE_DESCRIPTION("Xen vTPM Driver");
400 MODULE_LICENSE("GPL");
This page took 0.039208 seconds and 5 git commands to generate.