root/drivers/crypto/ccree/cc_driver.c

/* [<][>][^][v][top][bottom][index][help] */

DEFINITIONS

This source file includes following definitions.
  1. cc_read_idr
  2. __dump_byte_array
  3. cc_isr
  4. cc_wait_for_reset_completion
  5. init_cc_regs
  6. init_cc_resources
  7. fini_cc_regs
  8. cleanup_cc_resources
  9. cc_clk_on
  10. cc_get_default_hash_len
  11. cc_clk_off
  12. ccree_probe
  13. ccree_remove
  14. ccree_init
  15. ccree_exit

   1 // SPDX-License-Identifier: GPL-2.0
   2 /* Copyright (C) 2012-2019 ARM Limited or its affiliates. */
   3 
   4 #include <linux/kernel.h>
   5 #include <linux/module.h>
   6 
   7 #include <linux/crypto.h>
   8 #include <linux/moduleparam.h>
   9 #include <linux/types.h>
  10 #include <linux/interrupt.h>
  11 #include <linux/platform_device.h>
  12 #include <linux/slab.h>
  13 #include <linux/spinlock.h>
  14 #include <linux/of.h>
  15 #include <linux/clk.h>
  16 #include <linux/of_address.h>
  17 
  18 #include "cc_driver.h"
  19 #include "cc_request_mgr.h"
  20 #include "cc_buffer_mgr.h"
  21 #include "cc_debugfs.h"
  22 #include "cc_cipher.h"
  23 #include "cc_aead.h"
  24 #include "cc_hash.h"
  25 #include "cc_sram_mgr.h"
  26 #include "cc_pm.h"
  27 #include "cc_fips.h"
  28 
  29 bool cc_dump_desc;
  30 module_param_named(dump_desc, cc_dump_desc, bool, 0600);
  31 MODULE_PARM_DESC(cc_dump_desc, "Dump descriptors to kernel log as debugging aid");
  32 bool cc_dump_bytes;
  33 module_param_named(dump_bytes, cc_dump_bytes, bool, 0600);
  34 MODULE_PARM_DESC(cc_dump_bytes, "Dump buffers to kernel log as debugging aid");
  35 
  36 static bool cc_sec_disable;
  37 module_param_named(sec_disable, cc_sec_disable, bool, 0600);
  38 MODULE_PARM_DESC(cc_sec_disable, "Disable security functions");
  39 
  40 struct cc_hw_data {
  41         char *name;
  42         enum cc_hw_rev rev;
  43         u32 sig;
  44         u32 cidr_0123;
  45         u32 pidr_0124;
  46         int std_bodies;
  47 };
  48 
  49 #define CC_NUM_IDRS 4
  50 #define CC_HW_RESET_LOOP_COUNT 10
  51 
  52 /* Note: PIDR3 holds CMOD/Rev so ignored for HW identification purposes */
  53 static const u32 pidr_0124_offsets[CC_NUM_IDRS] = {
  54         CC_REG(PERIPHERAL_ID_0), CC_REG(PERIPHERAL_ID_1),
  55         CC_REG(PERIPHERAL_ID_2), CC_REG(PERIPHERAL_ID_4)
  56 };
  57 
  58 static const u32 cidr_0123_offsets[CC_NUM_IDRS] = {
  59         CC_REG(COMPONENT_ID_0), CC_REG(COMPONENT_ID_1),
  60         CC_REG(COMPONENT_ID_2), CC_REG(COMPONENT_ID_3)
  61 };
  62 
  63 /* Hardware revisions defs. */
  64 
  65 /* The 703 is a OSCCA only variant of the 713 */
  66 static const struct cc_hw_data cc703_hw = {
  67         .name = "703", .rev = CC_HW_REV_713, .cidr_0123 = 0xB105F00DU,
  68         .pidr_0124 = 0x040BB0D0U, .std_bodies = CC_STD_OSCCA
  69 };
  70 
  71 static const struct cc_hw_data cc713_hw = {
  72         .name = "713", .rev = CC_HW_REV_713, .cidr_0123 = 0xB105F00DU,
  73         .pidr_0124 = 0x040BB0D0U, .std_bodies = CC_STD_ALL
  74 };
  75 
  76 static const struct cc_hw_data cc712_hw = {
  77         .name = "712", .rev = CC_HW_REV_712, .sig =  0xDCC71200U,
  78         .std_bodies = CC_STD_ALL
  79 };
  80 
  81 static const struct cc_hw_data cc710_hw = {
  82         .name = "710", .rev = CC_HW_REV_710, .sig =  0xDCC63200U,
  83         .std_bodies = CC_STD_ALL
  84 };
  85 
  86 static const struct cc_hw_data cc630p_hw = {
  87         .name = "630P", .rev = CC_HW_REV_630, .sig = 0xDCC63000U,
  88         .std_bodies = CC_STD_ALL
  89 };
  90 
  91 static const struct of_device_id arm_ccree_dev_of_match[] = {
  92         { .compatible = "arm,cryptocell-703-ree", .data = &cc703_hw },
  93         { .compatible = "arm,cryptocell-713-ree", .data = &cc713_hw },
  94         { .compatible = "arm,cryptocell-712-ree", .data = &cc712_hw },
  95         { .compatible = "arm,cryptocell-710-ree", .data = &cc710_hw },
  96         { .compatible = "arm,cryptocell-630p-ree", .data = &cc630p_hw },
  97         {}
  98 };
  99 MODULE_DEVICE_TABLE(of, arm_ccree_dev_of_match);
 100 
 101 static u32 cc_read_idr(struct cc_drvdata *drvdata, const u32 *idr_offsets)
 102 {
 103         int i;
 104         union {
 105                 u8 regs[CC_NUM_IDRS];
 106                 __le32 val;
 107         } idr;
 108 
 109         for (i = 0; i < CC_NUM_IDRS; ++i)
 110                 idr.regs[i] = cc_ioread(drvdata, idr_offsets[i]);
 111 
 112         return le32_to_cpu(idr.val);
 113 }
 114 
 115 void __dump_byte_array(const char *name, const u8 *buf, size_t len)
 116 {
 117         char prefix[64];
 118 
 119         if (!buf)
 120                 return;
 121 
 122         snprintf(prefix, sizeof(prefix), "%s[%zu]: ", name, len);
 123 
 124         print_hex_dump(KERN_DEBUG, prefix, DUMP_PREFIX_ADDRESS, 16, 1, buf,
 125                        len, false);
 126 }
 127 
 128 static irqreturn_t cc_isr(int irq, void *dev_id)
 129 {
 130         struct cc_drvdata *drvdata = (struct cc_drvdata *)dev_id;
 131         struct device *dev = drvdata_to_dev(drvdata);
 132         u32 irr;
 133         u32 imr;
 134 
 135         /* STAT_OP_TYPE_GENERIC STAT_PHASE_0: Interrupt */
 136         /* if driver suspended return, probebly shared interrupt */
 137         if (cc_pm_is_dev_suspended(dev))
 138                 return IRQ_NONE;
 139 
 140         /* read the interrupt status */
 141         irr = cc_ioread(drvdata, CC_REG(HOST_IRR));
 142         dev_dbg(dev, "Got IRR=0x%08X\n", irr);
 143 
 144         if (irr == 0) /* Probably shared interrupt line */
 145                 return IRQ_NONE;
 146 
 147         imr = cc_ioread(drvdata, CC_REG(HOST_IMR));
 148 
 149         /* clear interrupt - must be before processing events */
 150         cc_iowrite(drvdata, CC_REG(HOST_ICR), irr);
 151 
 152         drvdata->irq = irr;
 153         /* Completion interrupt - most probable */
 154         if (irr & drvdata->comp_mask) {
 155                 /* Mask all completion interrupts - will be unmasked in
 156                  * deferred service handler
 157                  */
 158                 cc_iowrite(drvdata, CC_REG(HOST_IMR), imr | drvdata->comp_mask);
 159                 irr &= ~drvdata->comp_mask;
 160                 complete_request(drvdata);
 161         }
 162 #ifdef CONFIG_CRYPTO_FIPS
 163         /* TEE FIPS interrupt */
 164         if (irr & CC_GPR0_IRQ_MASK) {
 165                 /* Mask interrupt - will be unmasked in Deferred service
 166                  * handler
 167                  */
 168                 cc_iowrite(drvdata, CC_REG(HOST_IMR), imr | CC_GPR0_IRQ_MASK);
 169                 irr &= ~CC_GPR0_IRQ_MASK;
 170                 fips_handler(drvdata);
 171         }
 172 #endif
 173         /* AXI error interrupt */
 174         if (irr & CC_AXI_ERR_IRQ_MASK) {
 175                 u32 axi_err;
 176 
 177                 /* Read the AXI error ID */
 178                 axi_err = cc_ioread(drvdata, CC_REG(AXIM_MON_ERR));
 179                 dev_dbg(dev, "AXI completion error: axim_mon_err=0x%08X\n",
 180                         axi_err);
 181 
 182                 irr &= ~CC_AXI_ERR_IRQ_MASK;
 183         }
 184 
 185         if (irr) {
 186                 dev_dbg_ratelimited(dev, "IRR includes unknown cause bits (0x%08X)\n",
 187                                     irr);
 188                 /* Just warning */
 189         }
 190 
 191         return IRQ_HANDLED;
 192 }
 193 
 194 bool cc_wait_for_reset_completion(struct cc_drvdata *drvdata)
 195 {
 196         unsigned int val;
 197         unsigned int i;
 198 
 199         /* 712/710/63 has no reset completion indication, always return true */
 200         if (drvdata->hw_rev <= CC_HW_REV_712)
 201                 return true;
 202 
 203         for (i = 0; i < CC_HW_RESET_LOOP_COUNT; i++) {
 204                 /* in cc7x3 NVM_IS_IDLE indicates that CC reset is
 205                  *  completed and device is fully functional
 206                  */
 207                 val = cc_ioread(drvdata, CC_REG(NVM_IS_IDLE));
 208                 if (val & CC_NVM_IS_IDLE_MASK) {
 209                         /* hw indicate reset completed */
 210                         return true;
 211                 }
 212                 /* allow scheduling other process on the processor */
 213                 schedule();
 214         }
 215         /* reset not completed */
 216         return false;
 217 }
 218 
 219 int init_cc_regs(struct cc_drvdata *drvdata, bool is_probe)
 220 {
 221         unsigned int val, cache_params;
 222         struct device *dev = drvdata_to_dev(drvdata);
 223 
 224         /* Unmask all AXI interrupt sources AXI_CFG1 register   */
 225         /* AXI interrupt config are obsoleted startign at cc7x3 */
 226         if (drvdata->hw_rev <= CC_HW_REV_712) {
 227                 val = cc_ioread(drvdata, CC_REG(AXIM_CFG));
 228                 cc_iowrite(drvdata, CC_REG(AXIM_CFG), val & ~CC_AXI_IRQ_MASK);
 229                 dev_dbg(dev, "AXIM_CFG=0x%08X\n",
 230                         cc_ioread(drvdata, CC_REG(AXIM_CFG)));
 231         }
 232 
 233         /* Clear all pending interrupts */
 234         val = cc_ioread(drvdata, CC_REG(HOST_IRR));
 235         dev_dbg(dev, "IRR=0x%08X\n", val);
 236         cc_iowrite(drvdata, CC_REG(HOST_ICR), val);
 237 
 238         /* Unmask relevant interrupt cause */
 239         val = drvdata->comp_mask | CC_AXI_ERR_IRQ_MASK;
 240 
 241         if (drvdata->hw_rev >= CC_HW_REV_712)
 242                 val |= CC_GPR0_IRQ_MASK;
 243 
 244         cc_iowrite(drvdata, CC_REG(HOST_IMR), ~val);
 245 
 246         cache_params = (drvdata->coherent ? CC_COHERENT_CACHE_PARAMS : 0x0);
 247 
 248         val = cc_ioread(drvdata, CC_REG(AXIM_CACHE_PARAMS));
 249 
 250         if (is_probe)
 251                 dev_dbg(dev, "Cache params previous: 0x%08X\n", val);
 252 
 253         cc_iowrite(drvdata, CC_REG(AXIM_CACHE_PARAMS), cache_params);
 254         val = cc_ioread(drvdata, CC_REG(AXIM_CACHE_PARAMS));
 255 
 256         if (is_probe)
 257                 dev_dbg(dev, "Cache params current: 0x%08X (expect: 0x%08X)\n",
 258                         val, cache_params);
 259 
 260         return 0;
 261 }
 262 
 263 static int init_cc_resources(struct platform_device *plat_dev)
 264 {
 265         struct resource *req_mem_cc_regs = NULL;
 266         struct cc_drvdata *new_drvdata;
 267         struct device *dev = &plat_dev->dev;
 268         struct device_node *np = dev->of_node;
 269         u32 val, hw_rev_pidr, sig_cidr;
 270         u64 dma_mask;
 271         const struct cc_hw_data *hw_rev;
 272         const struct of_device_id *dev_id;
 273         struct clk *clk;
 274         int rc = 0;
 275 
 276         new_drvdata = devm_kzalloc(dev, sizeof(*new_drvdata), GFP_KERNEL);
 277         if (!new_drvdata)
 278                 return -ENOMEM;
 279 
 280         dev_id = of_match_node(arm_ccree_dev_of_match, np);
 281         if (!dev_id)
 282                 return -ENODEV;
 283 
 284         hw_rev = (struct cc_hw_data *)dev_id->data;
 285         new_drvdata->hw_rev_name = hw_rev->name;
 286         new_drvdata->hw_rev = hw_rev->rev;
 287         new_drvdata->std_bodies = hw_rev->std_bodies;
 288 
 289         if (hw_rev->rev >= CC_HW_REV_712) {
 290                 new_drvdata->axim_mon_offset = CC_REG(AXIM_MON_COMP);
 291                 new_drvdata->sig_offset = CC_REG(HOST_SIGNATURE_712);
 292                 new_drvdata->ver_offset = CC_REG(HOST_VERSION_712);
 293         } else {
 294                 new_drvdata->axim_mon_offset = CC_REG(AXIM_MON_COMP8);
 295                 new_drvdata->sig_offset = CC_REG(HOST_SIGNATURE_630);
 296                 new_drvdata->ver_offset = CC_REG(HOST_VERSION_630);
 297         }
 298 
 299         new_drvdata->comp_mask = CC_COMP_IRQ_MASK;
 300 
 301         platform_set_drvdata(plat_dev, new_drvdata);
 302         new_drvdata->plat_dev = plat_dev;
 303 
 304         clk = devm_clk_get(dev, NULL);
 305         if (IS_ERR(clk))
 306                 switch (PTR_ERR(clk)) {
 307                 /* Clock is optional so this might be fine */
 308                 case -ENOENT:
 309                         break;
 310 
 311                 /* Clock not available, let's try again soon */
 312                 case -EPROBE_DEFER:
 313                         return -EPROBE_DEFER;
 314 
 315                 default:
 316                         dev_err(dev, "Error getting clock: %ld\n",
 317                                 PTR_ERR(clk));
 318                         return PTR_ERR(clk);
 319                 }
 320         new_drvdata->clk = clk;
 321 
 322         new_drvdata->coherent = of_dma_is_coherent(np);
 323 
 324         /* Get device resources */
 325         /* First CC registers space */
 326         req_mem_cc_regs = platform_get_resource(plat_dev, IORESOURCE_MEM, 0);
 327         /* Map registers space */
 328         new_drvdata->cc_base = devm_ioremap_resource(dev, req_mem_cc_regs);
 329         if (IS_ERR(new_drvdata->cc_base)) {
 330                 dev_err(dev, "Failed to ioremap registers");
 331                 return PTR_ERR(new_drvdata->cc_base);
 332         }
 333 
 334         dev_dbg(dev, "Got MEM resource (%s): %pR\n", req_mem_cc_regs->name,
 335                 req_mem_cc_regs);
 336         dev_dbg(dev, "CC registers mapped from %pa to 0x%p\n",
 337                 &req_mem_cc_regs->start, new_drvdata->cc_base);
 338 
 339         /* Then IRQ */
 340         new_drvdata->irq = platform_get_irq(plat_dev, 0);
 341         if (new_drvdata->irq < 0)
 342                 return new_drvdata->irq;
 343 
 344         init_completion(&new_drvdata->hw_queue_avail);
 345 
 346         if (!plat_dev->dev.dma_mask)
 347                 plat_dev->dev.dma_mask = &plat_dev->dev.coherent_dma_mask;
 348 
 349         dma_mask = DMA_BIT_MASK(DMA_BIT_MASK_LEN);
 350         while (dma_mask > 0x7fffffffUL) {
 351                 if (dma_supported(&plat_dev->dev, dma_mask)) {
 352                         rc = dma_set_coherent_mask(&plat_dev->dev, dma_mask);
 353                         if (!rc)
 354                                 break;
 355                 }
 356                 dma_mask >>= 1;
 357         }
 358 
 359         if (rc) {
 360                 dev_err(dev, "Failed in dma_set_mask, mask=%llx\n", dma_mask);
 361                 return rc;
 362         }
 363 
 364         rc = cc_clk_on(new_drvdata);
 365         if (rc) {
 366                 dev_err(dev, "Failed to enable clock");
 367                 return rc;
 368         }
 369 
 370         new_drvdata->sec_disabled = cc_sec_disable;
 371 
 372         /* wait for Crytpcell reset completion */
 373         if (!cc_wait_for_reset_completion(new_drvdata)) {
 374                 dev_err(dev, "Cryptocell reset not completed");
 375         }
 376 
 377         if (hw_rev->rev <= CC_HW_REV_712) {
 378                 /* Verify correct mapping */
 379                 val = cc_ioread(new_drvdata, new_drvdata->sig_offset);
 380                 if (val != hw_rev->sig) {
 381                         dev_err(dev, "Invalid CC signature: SIGNATURE=0x%08X != expected=0x%08X\n",
 382                                 val, hw_rev->sig);
 383                         rc = -EINVAL;
 384                         goto post_clk_err;
 385                 }
 386                 sig_cidr = val;
 387                 hw_rev_pidr = cc_ioread(new_drvdata, new_drvdata->ver_offset);
 388         } else {
 389                 /* Verify correct mapping */
 390                 val = cc_read_idr(new_drvdata, pidr_0124_offsets);
 391                 if (val != hw_rev->pidr_0124) {
 392                         dev_err(dev, "Invalid CC PIDR: PIDR0124=0x%08X != expected=0x%08X\n",
 393                                 val,  hw_rev->pidr_0124);
 394                         rc = -EINVAL;
 395                         goto post_clk_err;
 396                 }
 397                 hw_rev_pidr = val;
 398 
 399                 val = cc_read_idr(new_drvdata, cidr_0123_offsets);
 400                 if (val != hw_rev->cidr_0123) {
 401                         dev_err(dev, "Invalid CC CIDR: CIDR0123=0x%08X != expected=0x%08X\n",
 402                         val,  hw_rev->cidr_0123);
 403                         rc = -EINVAL;
 404                         goto post_clk_err;
 405                 }
 406                 sig_cidr = val;
 407 
 408                 /* Check HW engine configuration */
 409                 val = cc_ioread(new_drvdata, CC_REG(HOST_REMOVE_INPUT_PINS));
 410                 switch (val) {
 411                 case CC_PINS_FULL:
 412                         /* This is fine */
 413                         break;
 414                 case CC_PINS_SLIM:
 415                         if (new_drvdata->std_bodies & CC_STD_NIST) {
 416                                 dev_warn(dev, "703 mode forced due to HW configuration.\n");
 417                                 new_drvdata->std_bodies = CC_STD_OSCCA;
 418                         }
 419                         break;
 420                 default:
 421                         dev_err(dev, "Unsupported engines configuration.\n");
 422                         rc = -EINVAL;
 423                         goto post_clk_err;
 424                 }
 425 
 426                 /* Check security disable state */
 427                 val = cc_ioread(new_drvdata, CC_REG(SECURITY_DISABLED));
 428                 val &= CC_SECURITY_DISABLED_MASK;
 429                 new_drvdata->sec_disabled |= !!val;
 430 
 431                 if (!new_drvdata->sec_disabled) {
 432                         new_drvdata->comp_mask |= CC_CPP_SM4_ABORT_MASK;
 433                         if (new_drvdata->std_bodies & CC_STD_NIST)
 434                                 new_drvdata->comp_mask |= CC_CPP_AES_ABORT_MASK;
 435                 }
 436         }
 437 
 438         if (new_drvdata->sec_disabled)
 439                 dev_info(dev, "Security Disabled mode is in effect. Security functions disabled.\n");
 440 
 441         /* Display HW versions */
 442         dev_info(dev, "ARM CryptoCell %s Driver: HW version 0x%08X/0x%8X, Driver version %s\n",
 443                  hw_rev->name, hw_rev_pidr, sig_cidr, DRV_MODULE_VERSION);
 444         /* register the driver isr function */
 445         rc = devm_request_irq(dev, new_drvdata->irq, cc_isr,
 446                               IRQF_SHARED, "ccree", new_drvdata);
 447         if (rc) {
 448                 dev_err(dev, "Could not register to interrupt %d\n",
 449                         new_drvdata->irq);
 450                 goto post_clk_err;
 451         }
 452         dev_dbg(dev, "Registered to IRQ: %d\n", new_drvdata->irq);
 453 
 454         rc = init_cc_regs(new_drvdata, true);
 455         if (rc) {
 456                 dev_err(dev, "init_cc_regs failed\n");
 457                 goto post_clk_err;
 458         }
 459 
 460         rc = cc_debugfs_init(new_drvdata);
 461         if (rc) {
 462                 dev_err(dev, "Failed registering debugfs interface\n");
 463                 goto post_regs_err;
 464         }
 465 
 466         rc = cc_fips_init(new_drvdata);
 467         if (rc) {
 468                 dev_err(dev, "CC_FIPS_INIT failed 0x%x\n", rc);
 469                 goto post_debugfs_err;
 470         }
 471         rc = cc_sram_mgr_init(new_drvdata);
 472         if (rc) {
 473                 dev_err(dev, "cc_sram_mgr_init failed\n");
 474                 goto post_fips_init_err;
 475         }
 476 
 477         new_drvdata->mlli_sram_addr =
 478                 cc_sram_alloc(new_drvdata, MAX_MLLI_BUFF_SIZE);
 479         if (new_drvdata->mlli_sram_addr == NULL_SRAM_ADDR) {
 480                 dev_err(dev, "Failed to alloc MLLI Sram buffer\n");
 481                 rc = -ENOMEM;
 482                 goto post_sram_mgr_err;
 483         }
 484 
 485         rc = cc_req_mgr_init(new_drvdata);
 486         if (rc) {
 487                 dev_err(dev, "cc_req_mgr_init failed\n");
 488                 goto post_sram_mgr_err;
 489         }
 490 
 491         rc = cc_buffer_mgr_init(new_drvdata);
 492         if (rc) {
 493                 dev_err(dev, "buffer_mgr_init failed\n");
 494                 goto post_req_mgr_err;
 495         }
 496 
 497         rc = cc_pm_init(new_drvdata);
 498         if (rc) {
 499                 dev_err(dev, "ssi_power_mgr_init failed\n");
 500                 goto post_buf_mgr_err;
 501         }
 502 
 503         /* Allocate crypto algs */
 504         rc = cc_cipher_alloc(new_drvdata);
 505         if (rc) {
 506                 dev_err(dev, "cc_cipher_alloc failed\n");
 507                 goto post_buf_mgr_err;
 508         }
 509 
 510         /* hash must be allocated before aead since hash exports APIs */
 511         rc = cc_hash_alloc(new_drvdata);
 512         if (rc) {
 513                 dev_err(dev, "cc_hash_alloc failed\n");
 514                 goto post_cipher_err;
 515         }
 516 
 517         rc = cc_aead_alloc(new_drvdata);
 518         if (rc) {
 519                 dev_err(dev, "cc_aead_alloc failed\n");
 520                 goto post_hash_err;
 521         }
 522 
 523         /* All set, we can allow autosuspend */
 524         cc_pm_go(new_drvdata);
 525 
 526         /* If we got here and FIPS mode is enabled
 527          * it means all FIPS test passed, so let TEE
 528          * know we're good.
 529          */
 530         cc_set_ree_fips_status(new_drvdata, true);
 531 
 532         return 0;
 533 
 534 post_hash_err:
 535         cc_hash_free(new_drvdata);
 536 post_cipher_err:
 537         cc_cipher_free(new_drvdata);
 538 post_buf_mgr_err:
 539          cc_buffer_mgr_fini(new_drvdata);
 540 post_req_mgr_err:
 541         cc_req_mgr_fini(new_drvdata);
 542 post_sram_mgr_err:
 543         cc_sram_mgr_fini(new_drvdata);
 544 post_fips_init_err:
 545         cc_fips_fini(new_drvdata);
 546 post_debugfs_err:
 547         cc_debugfs_fini(new_drvdata);
 548 post_regs_err:
 549         fini_cc_regs(new_drvdata);
 550 post_clk_err:
 551         cc_clk_off(new_drvdata);
 552         return rc;
 553 }
 554 
 555 void fini_cc_regs(struct cc_drvdata *drvdata)
 556 {
 557         /* Mask all interrupts */
 558         cc_iowrite(drvdata, CC_REG(HOST_IMR), 0xFFFFFFFF);
 559 }
 560 
 561 static void cleanup_cc_resources(struct platform_device *plat_dev)
 562 {
 563         struct cc_drvdata *drvdata =
 564                 (struct cc_drvdata *)platform_get_drvdata(plat_dev);
 565 
 566         cc_aead_free(drvdata);
 567         cc_hash_free(drvdata);
 568         cc_cipher_free(drvdata);
 569         cc_pm_fini(drvdata);
 570         cc_buffer_mgr_fini(drvdata);
 571         cc_req_mgr_fini(drvdata);
 572         cc_sram_mgr_fini(drvdata);
 573         cc_fips_fini(drvdata);
 574         cc_debugfs_fini(drvdata);
 575         fini_cc_regs(drvdata);
 576         cc_clk_off(drvdata);
 577 }
 578 
 579 int cc_clk_on(struct cc_drvdata *drvdata)
 580 {
 581         struct clk *clk = drvdata->clk;
 582         int rc;
 583 
 584         if (IS_ERR(clk))
 585                 /* Not all devices have a clock associated with CCREE  */
 586                 return 0;
 587 
 588         rc = clk_prepare_enable(clk);
 589         if (rc)
 590                 return rc;
 591 
 592         return 0;
 593 }
 594 
 595 unsigned int cc_get_default_hash_len(struct cc_drvdata *drvdata)
 596 {
 597         if (drvdata->hw_rev >= CC_HW_REV_712)
 598                 return HASH_LEN_SIZE_712;
 599         else
 600                 return HASH_LEN_SIZE_630;
 601 }
 602 
 603 void cc_clk_off(struct cc_drvdata *drvdata)
 604 {
 605         struct clk *clk = drvdata->clk;
 606 
 607         if (IS_ERR(clk))
 608                 /* Not all devices have a clock associated with CCREE */
 609                 return;
 610 
 611         clk_disable_unprepare(clk);
 612 }
 613 
 614 static int ccree_probe(struct platform_device *plat_dev)
 615 {
 616         int rc;
 617         struct device *dev = &plat_dev->dev;
 618 
 619         /* Map registers space */
 620         rc = init_cc_resources(plat_dev);
 621         if (rc)
 622                 return rc;
 623 
 624         dev_info(dev, "ARM ccree device initialized\n");
 625 
 626         return 0;
 627 }
 628 
 629 static int ccree_remove(struct platform_device *plat_dev)
 630 {
 631         struct device *dev = &plat_dev->dev;
 632 
 633         dev_dbg(dev, "Releasing ccree resources...\n");
 634 
 635         cleanup_cc_resources(plat_dev);
 636 
 637         dev_info(dev, "ARM ccree device terminated\n");
 638 
 639         return 0;
 640 }
 641 
 642 static struct platform_driver ccree_driver = {
 643         .driver = {
 644                    .name = "ccree",
 645                    .of_match_table = arm_ccree_dev_of_match,
 646 #ifdef CONFIG_PM
 647                    .pm = &ccree_pm,
 648 #endif
 649         },
 650         .probe = ccree_probe,
 651         .remove = ccree_remove,
 652 };
 653 
 654 static int __init ccree_init(void)
 655 {
 656         cc_hash_global_init();
 657         cc_debugfs_global_init();
 658 
 659         return platform_driver_register(&ccree_driver);
 660 }
 661 module_init(ccree_init);
 662 
 663 static void __exit ccree_exit(void)
 664 {
 665         platform_driver_unregister(&ccree_driver);
 666         cc_debugfs_global_fini();
 667 }
 668 module_exit(ccree_exit);
 669 
 670 /* Module description */
 671 MODULE_DESCRIPTION("ARM TrustZone CryptoCell REE Driver");
 672 MODULE_VERSION(DRV_MODULE_VERSION);
 673 MODULE_AUTHOR("ARM");
 674 MODULE_LICENSE("GPL v2");

/* [<][>][^][v][top][bottom][index][help] */