root/drivers/char/tpm/xen-tpmfront.c

/* [<][>][^][v][top][bottom][index][help] */

DEFINITIONS

This source file includes following definitions.
  1. wait_for_tpm_stat_cond
  2. wait_for_tpm_stat
  3. vtpm_status
  4. vtpm_req_canceled
  5. vtpm_cancel
  6. shr_data_offset
  7. vtpm_send
  8. vtpm_recv
  9. tpmif_interrupt
  10. setup_chip
  11. setup_ring
  12. ring_free
  13. tpmfront_probe
  14. tpmfront_remove
  15. tpmfront_resume
  16. backend_changed
  17. xen_tpmfront_init
  18. xen_tpmfront_exit

   1 // SPDX-License-Identifier: GPL-2.0-only
   2 /*
   3  * Implementation of the Xen vTPM device frontend
   4  *
   5  * Author:  Daniel De Graaf <dgdegra@tycho.nsa.gov>
   6  */
   7 #include <linux/errno.h>
   8 #include <linux/err.h>
   9 #include <linux/interrupt.h>
  10 #include <linux/freezer.h>
  11 #include <xen/xen.h>
  12 #include <xen/events.h>
  13 #include <xen/interface/io/tpmif.h>
  14 #include <xen/grant_table.h>
  15 #include <xen/xenbus.h>
  16 #include <xen/page.h>
  17 #include "tpm.h"
  18 #include <xen/platform_pci.h>
  19 
  20 struct tpm_private {
  21         struct tpm_chip *chip;
  22         struct xenbus_device *dev;
  23 
  24         struct vtpm_shared_page *shr;
  25 
  26         unsigned int evtchn;
  27         int ring_ref;
  28         domid_t backend_id;
  29         int irq;
  30         wait_queue_head_t read_queue;
  31 };
  32 
  33 enum status_bits {
  34         VTPM_STATUS_RUNNING  = 0x1,
  35         VTPM_STATUS_IDLE     = 0x2,
  36         VTPM_STATUS_RESULT   = 0x4,
  37         VTPM_STATUS_CANCELED = 0x8,
  38 };
  39 
  40 static bool wait_for_tpm_stat_cond(struct tpm_chip *chip, u8 mask,
  41                                         bool check_cancel, bool *canceled)
  42 {
  43         u8 status = chip->ops->status(chip);
  44 
  45         *canceled = false;
  46         if ((status & mask) == mask)
  47                 return true;
  48         if (check_cancel && chip->ops->req_canceled(chip, status)) {
  49                 *canceled = true;
  50                 return true;
  51         }
  52         return false;
  53 }
  54 
  55 static int wait_for_tpm_stat(struct tpm_chip *chip, u8 mask,
  56                 unsigned long timeout, wait_queue_head_t *queue,
  57                 bool check_cancel)
  58 {
  59         unsigned long stop;
  60         long rc;
  61         u8 status;
  62         bool canceled = false;
  63 
  64         /* check current status */
  65         status = chip->ops->status(chip);
  66         if ((status & mask) == mask)
  67                 return 0;
  68 
  69         stop = jiffies + timeout;
  70 
  71         if (chip->flags & TPM_CHIP_FLAG_IRQ) {
  72 again:
  73                 timeout = stop - jiffies;
  74                 if ((long)timeout <= 0)
  75                         return -ETIME;
  76                 rc = wait_event_interruptible_timeout(*queue,
  77                         wait_for_tpm_stat_cond(chip, mask, check_cancel,
  78                                                &canceled),
  79                         timeout);
  80                 if (rc > 0) {
  81                         if (canceled)
  82                                 return -ECANCELED;
  83                         return 0;
  84                 }
  85                 if (rc == -ERESTARTSYS && freezing(current)) {
  86                         clear_thread_flag(TIF_SIGPENDING);
  87                         goto again;
  88                 }
  89         } else {
  90                 do {
  91                         tpm_msleep(TPM_TIMEOUT);
  92                         status = chip->ops->status(chip);
  93                         if ((status & mask) == mask)
  94                                 return 0;
  95                 } while (time_before(jiffies, stop));
  96         }
  97         return -ETIME;
  98 }
  99 
 100 static u8 vtpm_status(struct tpm_chip *chip)
 101 {
 102         struct tpm_private *priv = dev_get_drvdata(&chip->dev);
 103         switch (priv->shr->state) {
 104         case VTPM_STATE_IDLE:
 105                 return VTPM_STATUS_IDLE | VTPM_STATUS_CANCELED;
 106         case VTPM_STATE_FINISH:
 107                 return VTPM_STATUS_IDLE | VTPM_STATUS_RESULT;
 108         case VTPM_STATE_SUBMIT:
 109         case VTPM_STATE_CANCEL: /* cancel requested, not yet canceled */
 110                 return VTPM_STATUS_RUNNING;
 111         default:
 112                 return 0;
 113         }
 114 }
 115 
 116 static bool vtpm_req_canceled(struct tpm_chip *chip, u8 status)
 117 {
 118         return status & VTPM_STATUS_CANCELED;
 119 }
 120 
 121 static void vtpm_cancel(struct tpm_chip *chip)
 122 {
 123         struct tpm_private *priv = dev_get_drvdata(&chip->dev);
 124         priv->shr->state = VTPM_STATE_CANCEL;
 125         wmb();
 126         notify_remote_via_evtchn(priv->evtchn);
 127 }
 128 
 129 static unsigned int shr_data_offset(struct vtpm_shared_page *shr)
 130 {
 131         return sizeof(*shr) + sizeof(u32) * shr->nr_extra_pages;
 132 }
 133 
 134 static int vtpm_send(struct tpm_chip *chip, u8 *buf, size_t count)
 135 {
 136         struct tpm_private *priv = dev_get_drvdata(&chip->dev);
 137         struct vtpm_shared_page *shr = priv->shr;
 138         unsigned int offset = shr_data_offset(shr);
 139 
 140         u32 ordinal;
 141         unsigned long duration;
 142 
 143         if (offset > PAGE_SIZE)
 144                 return -EINVAL;
 145 
 146         if (offset + count > PAGE_SIZE)
 147                 return -EINVAL;
 148 
 149         /* Wait for completion of any existing command or cancellation */
 150         if (wait_for_tpm_stat(chip, VTPM_STATUS_IDLE, chip->timeout_c,
 151                         &priv->read_queue, true) < 0) {
 152                 vtpm_cancel(chip);
 153                 return -ETIME;
 154         }
 155 
 156         memcpy(offset + (u8 *)shr, buf, count);
 157         shr->length = count;
 158         barrier();
 159         shr->state = VTPM_STATE_SUBMIT;
 160         wmb();
 161         notify_remote_via_evtchn(priv->evtchn);
 162 
 163         ordinal = be32_to_cpu(((struct tpm_header *)buf)->ordinal);
 164         duration = tpm_calc_ordinal_duration(chip, ordinal);
 165 
 166         if (wait_for_tpm_stat(chip, VTPM_STATUS_IDLE, duration,
 167                         &priv->read_queue, true) < 0) {
 168                 /* got a signal or timeout, try to cancel */
 169                 vtpm_cancel(chip);
 170                 return -ETIME;
 171         }
 172 
 173         return 0;
 174 }
 175 
 176 static int vtpm_recv(struct tpm_chip *chip, u8 *buf, size_t count)
 177 {
 178         struct tpm_private *priv = dev_get_drvdata(&chip->dev);
 179         struct vtpm_shared_page *shr = priv->shr;
 180         unsigned int offset = shr_data_offset(shr);
 181         size_t length = shr->length;
 182 
 183         if (shr->state == VTPM_STATE_IDLE)
 184                 return -ECANCELED;
 185 
 186         /* In theory the wait at the end of _send makes this one unnecessary */
 187         if (wait_for_tpm_stat(chip, VTPM_STATUS_RESULT, chip->timeout_c,
 188                         &priv->read_queue, true) < 0) {
 189                 vtpm_cancel(chip);
 190                 return -ETIME;
 191         }
 192 
 193         if (offset > PAGE_SIZE)
 194                 return -EIO;
 195 
 196         if (offset + length > PAGE_SIZE)
 197                 length = PAGE_SIZE - offset;
 198 
 199         if (length > count)
 200                 length = count;
 201 
 202         memcpy(buf, offset + (u8 *)shr, length);
 203 
 204         return length;
 205 }
 206 
 207 static const struct tpm_class_ops tpm_vtpm = {
 208         .status = vtpm_status,
 209         .recv = vtpm_recv,
 210         .send = vtpm_send,
 211         .cancel = vtpm_cancel,
 212         .req_complete_mask = VTPM_STATUS_IDLE | VTPM_STATUS_RESULT,
 213         .req_complete_val  = VTPM_STATUS_IDLE | VTPM_STATUS_RESULT,
 214         .req_canceled      = vtpm_req_canceled,
 215 };
 216 
 217 static irqreturn_t tpmif_interrupt(int dummy, void *dev_id)
 218 {
 219         struct tpm_private *priv = dev_id;
 220 
 221         switch (priv->shr->state) {
 222         case VTPM_STATE_IDLE:
 223         case VTPM_STATE_FINISH:
 224                 wake_up_interruptible(&priv->read_queue);
 225                 break;
 226         case VTPM_STATE_SUBMIT:
 227         case VTPM_STATE_CANCEL:
 228         default:
 229                 break;
 230         }
 231         return IRQ_HANDLED;
 232 }
 233 
 234 static int setup_chip(struct device *dev, struct tpm_private *priv)
 235 {
 236         struct tpm_chip *chip;
 237 
 238         chip = tpmm_chip_alloc(dev, &tpm_vtpm);
 239         if (IS_ERR(chip))
 240                 return PTR_ERR(chip);
 241 
 242         init_waitqueue_head(&priv->read_queue);
 243 
 244         priv->chip = chip;
 245         dev_set_drvdata(&chip->dev, priv);
 246 
 247         return 0;
 248 }
 249 
 250 /* caller must clean up in case of errors */
 251 static int setup_ring(struct xenbus_device *dev, struct tpm_private *priv)
 252 {
 253         struct xenbus_transaction xbt;
 254         const char *message = NULL;
 255         int rv;
 256         grant_ref_t gref;
 257 
 258         priv->shr = (void *)__get_free_page(GFP_KERNEL|__GFP_ZERO);
 259         if (!priv->shr) {
 260                 xenbus_dev_fatal(dev, -ENOMEM, "allocating shared ring");
 261                 return -ENOMEM;
 262         }
 263 
 264         rv = xenbus_grant_ring(dev, priv->shr, 1, &gref);
 265         if (rv < 0)
 266                 return rv;
 267 
 268         priv->ring_ref = gref;
 269 
 270         rv = xenbus_alloc_evtchn(dev, &priv->evtchn);
 271         if (rv)
 272                 return rv;
 273 
 274         rv = bind_evtchn_to_irqhandler(priv->evtchn, tpmif_interrupt, 0,
 275                                        "tpmif", priv);
 276         if (rv <= 0) {
 277                 xenbus_dev_fatal(dev, rv, "allocating TPM irq");
 278                 return rv;
 279         }
 280         priv->irq = rv;
 281 
 282  again:
 283         rv = xenbus_transaction_start(&xbt);
 284         if (rv) {
 285                 xenbus_dev_fatal(dev, rv, "starting transaction");
 286                 return rv;
 287         }
 288 
 289         rv = xenbus_printf(xbt, dev->nodename,
 290                         "ring-ref", "%u", priv->ring_ref);
 291         if (rv) {
 292                 message = "writing ring-ref";
 293                 goto abort_transaction;
 294         }
 295 
 296         rv = xenbus_printf(xbt, dev->nodename, "event-channel", "%u",
 297                         priv->evtchn);
 298         if (rv) {
 299                 message = "writing event-channel";
 300                 goto abort_transaction;
 301         }
 302 
 303         rv = xenbus_printf(xbt, dev->nodename, "feature-protocol-v2", "1");
 304         if (rv) {
 305                 message = "writing feature-protocol-v2";
 306                 goto abort_transaction;
 307         }
 308 
 309         rv = xenbus_transaction_end(xbt, 0);
 310         if (rv == -EAGAIN)
 311                 goto again;
 312         if (rv) {
 313                 xenbus_dev_fatal(dev, rv, "completing transaction");
 314                 return rv;
 315         }
 316 
 317         xenbus_switch_state(dev, XenbusStateInitialised);
 318 
 319         return 0;
 320 
 321  abort_transaction:
 322         xenbus_transaction_end(xbt, 1);
 323         if (message)
 324                 xenbus_dev_error(dev, rv, "%s", message);
 325 
 326         return rv;
 327 }
 328 
 329 static void ring_free(struct tpm_private *priv)
 330 {
 331         if (!priv)
 332                 return;
 333 
 334         if (priv->ring_ref)
 335                 gnttab_end_foreign_access(priv->ring_ref, 0,
 336                                 (unsigned long)priv->shr);
 337         else
 338                 free_page((unsigned long)priv->shr);
 339 
 340         if (priv->irq)
 341                 unbind_from_irqhandler(priv->irq, priv);
 342 
 343         kfree(priv);
 344 }
 345 
 346 static int tpmfront_probe(struct xenbus_device *dev,
 347                 const struct xenbus_device_id *id)
 348 {
 349         struct tpm_private *priv;
 350         int rv;
 351 
 352         priv = kzalloc(sizeof(*priv), GFP_KERNEL);
 353         if (!priv) {
 354                 xenbus_dev_fatal(dev, -ENOMEM, "allocating priv structure");
 355                 return -ENOMEM;
 356         }
 357 
 358         rv = setup_chip(&dev->dev, priv);
 359         if (rv) {
 360                 kfree(priv);
 361                 return rv;
 362         }
 363 
 364         rv = setup_ring(dev, priv);
 365         if (rv) {
 366                 ring_free(priv);
 367                 return rv;
 368         }
 369 
 370         tpm_get_timeouts(priv->chip);
 371 
 372         return tpm_chip_register(priv->chip);
 373 }
 374 
 375 static int tpmfront_remove(struct xenbus_device *dev)
 376 {
 377         struct tpm_chip *chip = dev_get_drvdata(&dev->dev);
 378         struct tpm_private *priv = dev_get_drvdata(&chip->dev);
 379         tpm_chip_unregister(chip);
 380         ring_free(priv);
 381         dev_set_drvdata(&chip->dev, NULL);
 382         return 0;
 383 }
 384 
 385 static int tpmfront_resume(struct xenbus_device *dev)
 386 {
 387         /* A suspend/resume/migrate will interrupt a vTPM anyway */
 388         tpmfront_remove(dev);
 389         return tpmfront_probe(dev, NULL);
 390 }
 391 
 392 static void backend_changed(struct xenbus_device *dev,
 393                 enum xenbus_state backend_state)
 394 {
 395         switch (backend_state) {
 396         case XenbusStateInitialised:
 397         case XenbusStateConnected:
 398                 if (dev->state == XenbusStateConnected)
 399                         break;
 400 
 401                 if (!xenbus_read_unsigned(dev->otherend, "feature-protocol-v2",
 402                                           0)) {
 403                         xenbus_dev_fatal(dev, -EINVAL,
 404                                         "vTPM protocol 2 required");
 405                         return;
 406                 }
 407                 xenbus_switch_state(dev, XenbusStateConnected);
 408                 break;
 409 
 410         case XenbusStateClosing:
 411         case XenbusStateClosed:
 412                 device_unregister(&dev->dev);
 413                 xenbus_frontend_closed(dev);
 414                 break;
 415         default:
 416                 break;
 417         }
 418 }
 419 
 420 static const struct xenbus_device_id tpmfront_ids[] = {
 421         { "vtpm" },
 422         { "" }
 423 };
 424 MODULE_ALIAS("xen:vtpm");
 425 
 426 static struct xenbus_driver tpmfront_driver = {
 427         .ids = tpmfront_ids,
 428         .probe = tpmfront_probe,
 429         .remove = tpmfront_remove,
 430         .resume = tpmfront_resume,
 431         .otherend_changed = backend_changed,
 432 };
 433 
 434 static int __init xen_tpmfront_init(void)
 435 {
 436         if (!xen_domain())
 437                 return -ENODEV;
 438 
 439         if (!xen_has_pv_devices())
 440                 return -ENODEV;
 441 
 442         return xenbus_register_frontend(&tpmfront_driver);
 443 }
 444 module_init(xen_tpmfront_init);
 445 
 446 static void __exit xen_tpmfront_exit(void)
 447 {
 448         xenbus_unregister_driver(&tpmfront_driver);
 449 }
 450 module_exit(xen_tpmfront_exit);
 451 
 452 MODULE_AUTHOR("Daniel De Graaf <dgdegra@tycho.nsa.gov>");
 453 MODULE_DESCRIPTION("Xen vTPM Driver");
 454 MODULE_LICENSE("GPL");

/* [<][>][^][v][top][bottom][index][help] */