tmpl 3860 arch/x86/kvm/vmx/vmx.c unsigned long tmpl; tmpl 3902 arch/x86/kvm/vmx/vmx.c rdmsrl(MSR_IA32_SYSENTER_EIP, tmpl); tmpl 3903 arch/x86/kvm/vmx/vmx.c vmcs_writel(HOST_IA32_SYSENTER_EIP, tmpl); /* 22.2.3 */ tmpl 500 crypto/adiantum.c static int adiantum_create(struct crypto_template *tmpl, struct rtattr **tb) tmpl 624 crypto/adiantum.c err = skcipher_register_instance(tmpl, inst); tmpl 194 crypto/aead.c inst->tmpl->free(inst); tmpl 231 crypto/aead.c struct aead_instance *aead_geniv_alloc(struct crypto_template *tmpl, tmpl 279 crypto/aead.c "%s(%s)", tmpl->name, alg->base.cra_name) >= tmpl 283 crypto/aead.c "%s(%s)", tmpl->name, alg->base.cra_driver_name) >= tmpl 451 crypto/aead.c int aead_register_instance(struct crypto_template *tmpl, tmpl 460 crypto/aead.c return crypto_register_instance(tmpl, aead_crypto_instance(inst)); tmpl 636 crypto/ahash.c int ahash_register_instance(struct crypto_template *tmpl, tmpl 645 crypto/ahash.c return crypto_register_instance(tmpl, ahash_crypto_instance(inst)); tmpl 146 crypto/akcipher.c int akcipher_register_instance(struct crypto_template *tmpl, tmpl 150 crypto/akcipher.c return crypto_register_instance(tmpl, akcipher_crypto_instance(inst)); tmpl 69 crypto/algapi.c inst->tmpl->free(inst); tmpl 79 crypto/algapi.c struct crypto_template *tmpl = inst->tmpl; tmpl 82 crypto/algapi.c crypto_tmpl_put(tmpl); tmpl 110 crypto/algapi.c struct crypto_template *tmpl = inst->tmpl; tmpl 119 crypto/algapi.c if (!tmpl || !crypto_tmpl_get(tmpl)) tmpl 474 crypto/algapi.c int crypto_register_template(struct crypto_template *tmpl) tmpl 481 crypto/algapi.c crypto_check_module_sig(tmpl->module); tmpl 484 crypto/algapi.c if (q == tmpl) tmpl 488 crypto/algapi.c list_add(&tmpl->list, &crypto_template_list); tmpl 514 crypto/algapi.c void crypto_unregister_template(struct crypto_template *tmpl) tmpl 523 crypto/algapi.c BUG_ON(list_empty(&tmpl->list)); tmpl 524 crypto/algapi.c list_del_init(&tmpl->list); tmpl 526 crypto/algapi.c list = &tmpl->instances; tmpl 554 crypto/algapi.c struct crypto_template *q, *tmpl = NULL; tmpl 563 crypto/algapi.c tmpl = q; tmpl 568 crypto/algapi.c return tmpl; tmpl 578 crypto/algapi.c int crypto_register_instance(struct crypto_template *tmpl, tmpl 588 crypto/algapi.c inst->alg.cra_module = tmpl->module; tmpl 597 crypto/algapi.c hlist_add_head(&inst->list, &tmpl->instances); tmpl 598 crypto/algapi.c inst->tmpl = tmpl; tmpl 60 crypto/algboss.c struct crypto_template *tmpl; tmpl 64 crypto/algboss.c tmpl = crypto_lookup_template(param->template); tmpl 65 crypto/algboss.c if (!tmpl) tmpl 69 crypto/algboss.c if (tmpl->create) { tmpl 70 crypto/algboss.c err = tmpl->create(tmpl, param->tb); tmpl 74 crypto/algboss.c inst = tmpl->alloc(param->tb); tmpl 77 crypto/algboss.c else if ((err = crypto_register_instance(tmpl, inst))) tmpl 78 crypto/algboss.c tmpl->free(inst); tmpl 81 crypto/algboss.c crypto_tmpl_put(tmpl); tmpl 382 crypto/authenc.c static int crypto_authenc_create(struct crypto_template *tmpl, tmpl 472 crypto/authenc.c err = aead_register_instance(tmpl, inst); tmpl 400 crypto/authencesn.c static int crypto_authenc_esn_create(struct crypto_template *tmpl, tmpl 487 crypto/authencesn.c err = aead_register_instance(tmpl, inst); tmpl 51 crypto/cbc.c static int crypto_cbc_create(struct crypto_template *tmpl, struct rtattr **tb) tmpl 57 crypto/cbc.c inst = skcipher_alloc_instance_simple(tmpl, tb, &alg); tmpl 68 crypto/cbc.c err = skcipher_register_instance(tmpl, inst); tmpl 454 crypto/ccm.c static int crypto_ccm_create_common(struct crypto_template *tmpl, tmpl 547 crypto/ccm.c err = aead_register_instance(tmpl, inst); tmpl 564 crypto/ccm.c static int crypto_ccm_create(struct crypto_template *tmpl, struct rtattr **tb) tmpl 582 crypto/ccm.c return crypto_ccm_create_common(tmpl, tb, ctr_name, mac_name); tmpl 585 crypto/ccm.c static int crypto_ccm_base_create(struct crypto_template *tmpl, tmpl 599 crypto/ccm.c return crypto_ccm_create_common(tmpl, tb, ctr_name, mac_name); tmpl 744 crypto/ccm.c static int crypto_rfc4309_create(struct crypto_template *tmpl, tmpl 818 crypto/ccm.c err = aead_register_instance(tmpl, inst); tmpl 917 crypto/ccm.c static int cbcmac_create(struct crypto_template *tmpl, struct rtattr **tb) tmpl 960 crypto/ccm.c err = shash_register_instance(tmpl, inst); tmpl 200 crypto/cfb.c static int crypto_cfb_create(struct crypto_template *tmpl, struct rtattr **tb) tmpl 206 crypto/cfb.c inst = skcipher_alloc_instance_simple(tmpl, tb, &alg); tmpl 222 crypto/cfb.c err = skcipher_register_instance(tmpl, inst); tmpl 562 crypto/chacha20poly1305.c static int chachapoly_create(struct crypto_template *tmpl, struct rtattr **tb, tmpl 664 crypto/chacha20poly1305.c err = aead_register_instance(tmpl, inst); tmpl 681 crypto/chacha20poly1305.c static int rfc7539_create(struct crypto_template *tmpl, struct rtattr **tb) tmpl 683 crypto/chacha20poly1305.c return chachapoly_create(tmpl, tb, "rfc7539", 12); tmpl 686 crypto/chacha20poly1305.c static int rfc7539esp_create(struct crypto_template *tmpl, struct rtattr **tb) tmpl 688 crypto/chacha20poly1305.c return chachapoly_create(tmpl, tb, "rfc7539esp", 8); tmpl 222 crypto/cmac.c static int cmac_create(struct crypto_template *tmpl, struct rtattr **tb) tmpl 283 crypto/cmac.c err = shash_register_instance(tmpl, inst); tmpl 396 crypto/cryptd.c static int cryptd_create_skcipher(struct crypto_template *tmpl, tmpl 453 crypto/cryptd.c err = skcipher_register_instance(tmpl, inst); tmpl 669 crypto/cryptd.c static int cryptd_create_hash(struct crypto_template *tmpl, struct rtattr **tb, tmpl 722 crypto/cryptd.c err = ahash_register_instance(tmpl, inst); tmpl 852 crypto/cryptd.c static int cryptd_create_aead(struct crypto_template *tmpl, tmpl 901 crypto/cryptd.c err = aead_register_instance(tmpl, inst); tmpl 913 crypto/cryptd.c static int cryptd_create(struct crypto_template *tmpl, struct rtattr **tb) tmpl 923 crypto/cryptd.c return cryptd_create_skcipher(tmpl, tb, &queue); tmpl 925 crypto/cryptd.c return cryptd_create_hash(tmpl, tb, &queue); tmpl 927 crypto/cryptd.c return cryptd_create_aead(tmpl, tb, &queue); tmpl 126 crypto/ctr.c static int crypto_ctr_create(struct crypto_template *tmpl, struct rtattr **tb) tmpl 132 crypto/ctr.c inst = skcipher_alloc_instance_simple(tmpl, tb, &alg); tmpl 157 crypto/ctr.c err = skcipher_register_instance(tmpl, inst); tmpl 262 crypto/ctr.c static int crypto_rfc3686_create(struct crypto_template *tmpl, tmpl 344 crypto/ctr.c err = skcipher_register_instance(tmpl, inst); tmpl 328 crypto/cts.c static int crypto_cts_create(struct crypto_template *tmpl, struct rtattr **tb) tmpl 396 crypto/cts.c err = skcipher_register_instance(tmpl, inst); tmpl 61 crypto/ecb.c static int crypto_ecb_create(struct crypto_template *tmpl, struct rtattr **tb) tmpl 67 crypto/ecb.c inst = skcipher_alloc_instance_simple(tmpl, tb, &alg); tmpl 76 crypto/ecb.c err = skcipher_register_instance(tmpl, inst); tmpl 112 crypto/echainiv.c static int echainiv_aead_create(struct crypto_template *tmpl, tmpl 118 crypto/echainiv.c inst = aead_geniv_alloc(tmpl, tb, 0, 0); tmpl 138 crypto/echainiv.c err = aead_register_instance(tmpl, inst); tmpl 456 crypto/essiv.c static int essiv_create(struct crypto_template *tmpl, struct rtattr **tb) tmpl 604 crypto/essiv.c err = skcipher_register_instance(tmpl, skcipher_inst); tmpl 619 crypto/essiv.c err = aead_register_instance(tmpl, aead_inst); tmpl 582 crypto/gcm.c static int crypto_gcm_create_common(struct crypto_template *tmpl, tmpl 675 crypto/gcm.c err = aead_register_instance(tmpl, inst); tmpl 692 crypto/gcm.c static int crypto_gcm_create(struct crypto_template *tmpl, struct rtattr **tb) tmpl 705 crypto/gcm.c return crypto_gcm_create_common(tmpl, tb, ctr_name, "ghash"); tmpl 708 crypto/gcm.c static int crypto_gcm_base_create(struct crypto_template *tmpl, tmpl 722 crypto/gcm.c return crypto_gcm_create_common(tmpl, tb, ctr_name, ghash_name); tmpl 866 crypto/gcm.c static int crypto_rfc4106_create(struct crypto_template *tmpl, tmpl 940 crypto/gcm.c err = aead_register_instance(tmpl, inst); tmpl 1102 crypto/gcm.c static int crypto_rfc4543_create(struct crypto_template *tmpl, tmpl 1178 crypto/gcm.c err = aead_register_instance(tmpl, inst); tmpl 170 crypto/hmac.c static int hmac_create(struct crypto_template *tmpl, struct rtattr **tb) tmpl 231 crypto/hmac.c err = shash_register_instance(tmpl, inst); tmpl 100 crypto/internal.h static inline int crypto_tmpl_get(struct crypto_template *tmpl) tmpl 102 crypto/internal.h return try_module_get(tmpl->module); tmpl 105 crypto/internal.h static inline void crypto_tmpl_put(struct crypto_template *tmpl) tmpl 107 crypto/internal.h module_put(tmpl->module); tmpl 263 crypto/keywrap.c static int crypto_kw_create(struct crypto_template *tmpl, struct rtattr **tb) tmpl 269 crypto/keywrap.c inst = skcipher_alloc_instance_simple(tmpl, tb, &alg); tmpl 285 crypto/keywrap.c err = skcipher_register_instance(tmpl, inst); tmpl 298 crypto/lrw.c static int create(struct crypto_template *tmpl, struct rtattr **tb) tmpl 406 crypto/lrw.c err = skcipher_register_instance(tmpl, inst); tmpl 52 crypto/ofb.c static int crypto_ofb_create(struct crypto_template *tmpl, struct rtattr **tb) tmpl 58 crypto/ofb.c inst = skcipher_alloc_instance_simple(tmpl, tb, &alg); tmpl 74 crypto/ofb.c err = skcipher_register_instance(tmpl, inst); tmpl 153 crypto/pcbc.c static int crypto_pcbc_create(struct crypto_template *tmpl, struct rtattr **tb) tmpl 159 crypto/pcbc.c inst = skcipher_alloc_instance_simple(tmpl, tb, &alg); tmpl 166 crypto/pcbc.c err = skcipher_register_instance(tmpl, inst); tmpl 229 crypto/pcrypt.c static int pcrypt_create_aead(struct crypto_template *tmpl, struct rtattr **tb, tmpl 290 crypto/pcrypt.c err = aead_register_instance(tmpl, inst); tmpl 308 crypto/pcrypt.c static int pcrypt_create(struct crypto_template *tmpl, struct rtattr **tb) tmpl 318 crypto/pcrypt.c return pcrypt_create_aead(tmpl, tb, algt->type, algt->mask); tmpl 597 crypto/rsa-pkcs1pad.c static int pkcs1pad_create(struct crypto_template *tmpl, struct rtattr **tb) tmpl 691 crypto/rsa-pkcs1pad.c err = akcipher_register_instance(tmpl, inst); tmpl 138 crypto/seqiv.c static int seqiv_aead_create(struct crypto_template *tmpl, struct rtattr **tb) tmpl 143 crypto/seqiv.c inst = aead_geniv_alloc(tmpl, tb, 0, 0); tmpl 161 crypto/seqiv.c err = aead_register_instance(tmpl, inst); tmpl 173 crypto/seqiv.c static int seqiv_create(struct crypto_template *tmpl, struct rtattr **tb) tmpl 184 crypto/seqiv.c return seqiv_aead_create(tmpl, tb); tmpl 540 crypto/shash.c int shash_register_instance(struct crypto_template *tmpl, tmpl 549 crypto/shash.c return crypto_register_instance(tmpl, shash_crypto_instance(inst)); tmpl 1095 crypto/skcipher.c int skcipher_register_instance(struct crypto_template *tmpl, tmpl 1104 crypto/skcipher.c return crypto_register_instance(tmpl, skcipher_crypto_instance(inst)); tmpl 1170 crypto/skcipher.c skcipher_alloc_instance_simple(struct crypto_template *tmpl, struct rtattr **tb, tmpl 1202 crypto/skcipher.c err = crypto_inst_setname(skcipher_crypto_instance(inst), tmpl->name, tmpl 620 crypto/vmac.c static int vmac_create(struct crypto_template *tmpl, struct rtattr **tb) tmpl 639 crypto/vmac.c inst = shash_alloc_instance(tmpl->name, alg); tmpl 665 crypto/vmac.c err = shash_register_instance(tmpl, inst); tmpl 188 crypto/xcbc.c static int xcbc_create(struct crypto_template *tmpl, struct rtattr **tb) tmpl 245 crypto/xcbc.c err = shash_register_instance(tmpl, inst); tmpl 60 crypto/xor.c do_xor_speed(struct xor_block_template *tmpl, void *b1, void *b2) tmpl 66 crypto/xor.c tmpl->next = template_list; tmpl 67 crypto/xor.c template_list = tmpl; tmpl 84 crypto/xor.c tmpl->do_2(BENCH_SIZE, b1, b2); tmpl 96 crypto/xor.c tmpl->speed = speed; tmpl 98 crypto/xor.c printk(KERN_INFO " %-10s: %5d.%03d MB/sec\n", tmpl->name, tmpl 337 crypto/xts.c static int create(struct crypto_template *tmpl, struct rtattr **tb) tmpl 444 crypto/xts.c err = skcipher_register_instance(tmpl, inst); tmpl 65 drivers/ata/libata-transport.c #define to_ata_internal(tmpl) container_of(tmpl, struct ata_internal, t) tmpl 2141 drivers/atm/idt77252.c unsigned long tmpl, modl; tmpl 2162 drivers/atm/idt77252.c tmpl = (unsigned long) tcra * ((unsigned long) card->tst_size - 2); tmpl 2163 drivers/atm/idt77252.c modl = tmpl % (unsigned long)card->utopia_pcr; tmpl 2165 drivers/atm/idt77252.c tst_entries = (int) (tmpl / card->utopia_pcr); tmpl 3252 drivers/atm/idt77252.c unsigned long tmpl, modl; tmpl 3476 drivers/atm/idt77252.c tmpl = (unsigned long) rsvdcr * ((unsigned long) card->tst_size - 2); tmpl 3477 drivers/atm/idt77252.c modl = tmpl % (unsigned long)card->utopia_pcr; tmpl 3478 drivers/atm/idt77252.c tst_entries = (int) (tmpl / (unsigned long)card->utopia_pcr); tmpl 1232 drivers/atm/nicstar.c unsigned long tmpl, modl; tmpl 1294 drivers/atm/nicstar.c tmpl = tmpl 1297 drivers/atm/nicstar.c modl = tmpl % card->max_pcr; tmpl 1299 drivers/atm/nicstar.c n = (int)(tmpl / card->max_pcr); tmpl 2619 drivers/crypto/ccree/cc_aead.c static struct cc_crypto_alg *cc_create_aead_alg(struct cc_alg_template *tmpl, tmpl 2629 drivers/crypto/ccree/cc_aead.c alg = &tmpl->template_aead; tmpl 2631 drivers/crypto/ccree/cc_aead.c snprintf(alg->base.cra_name, CRYPTO_MAX_ALG_NAME, "%s", tmpl->name); tmpl 2633 drivers/crypto/ccree/cc_aead.c tmpl->driver_name); tmpl 2644 drivers/crypto/ccree/cc_aead.c t_alg->cipher_mode = tmpl->cipher_mode; tmpl 2645 drivers/crypto/ccree/cc_aead.c t_alg->flow_mode = tmpl->flow_mode; tmpl 2646 drivers/crypto/ccree/cc_aead.c t_alg->auth_mode = tmpl->auth_mode; tmpl 1635 drivers/crypto/ccree/cc_cipher.c static struct cc_crypto_alg *cc_create_alg(const struct cc_alg_template *tmpl, tmpl 1647 drivers/crypto/ccree/cc_cipher.c memcpy(alg, &tmpl->template_skcipher, sizeof(*alg)); tmpl 1649 drivers/crypto/ccree/cc_cipher.c snprintf(alg->base.cra_name, CRYPTO_MAX_ALG_NAME, "%s", tmpl->name); tmpl 1651 drivers/crypto/ccree/cc_cipher.c tmpl->driver_name); tmpl 1654 drivers/crypto/ccree/cc_cipher.c alg->base.cra_blocksize = tmpl->blocksize; tmpl 1662 drivers/crypto/ccree/cc_cipher.c t_alg->cipher_mode = tmpl->cipher_mode; tmpl 1663 drivers/crypto/ccree/cc_cipher.c t_alg->flow_mode = tmpl->flow_mode; tmpl 1664 drivers/crypto/ccree/cc_cipher.c t_alg->data_unit = tmpl->data_unit; tmpl 1065 drivers/crypto/inside-secure/safexcel_cipher.c struct safexcel_alg_template *tmpl = tmpl 1072 drivers/crypto/inside-secure/safexcel_cipher.c ctx->priv = tmpl->priv; tmpl 1554 drivers/crypto/inside-secure/safexcel_cipher.c struct safexcel_alg_template *tmpl = tmpl 1561 drivers/crypto/inside-secure/safexcel_cipher.c ctx->priv = tmpl->priv; tmpl 788 drivers/crypto/inside-secure/safexcel_hash.c struct safexcel_alg_template *tmpl = tmpl 792 drivers/crypto/inside-secure/safexcel_hash.c ctx->priv = tmpl->priv; tmpl 411 drivers/crypto/marvell/cipher.c struct mv_cesa_op_ctx *tmpl) tmpl 432 drivers/crypto/marvell/cipher.c mv_cesa_update_op_cfg(tmpl, CESA_SA_DESC_CFG_OP_CRYPT_ONLY, tmpl 436 drivers/crypto/marvell/cipher.c ret = mv_cesa_skcipher_dma_req_init(req, tmpl); tmpl 438 drivers/crypto/marvell/cipher.c ret = mv_cesa_skcipher_std_req_init(req, tmpl); tmpl 444 drivers/crypto/marvell/cipher.c struct mv_cesa_op_ctx *tmpl) tmpl 450 drivers/crypto/marvell/cipher.c ret = mv_cesa_skcipher_req_init(req, tmpl); tmpl 466 drivers/crypto/marvell/cipher.c struct mv_cesa_op_ctx *tmpl) tmpl 470 drivers/crypto/marvell/cipher.c mv_cesa_update_op_cfg(tmpl, CESA_SA_DESC_CFG_CRYPTM_DES, tmpl 473 drivers/crypto/marvell/cipher.c memcpy(tmpl->ctx.blkcipher.key, ctx->key, DES_KEY_SIZE); tmpl 475 drivers/crypto/marvell/cipher.c return mv_cesa_skcipher_queue_req(req, tmpl); tmpl 480 drivers/crypto/marvell/cipher.c struct mv_cesa_op_ctx tmpl; tmpl 482 drivers/crypto/marvell/cipher.c mv_cesa_set_op_cfg(&tmpl, tmpl 486 drivers/crypto/marvell/cipher.c return mv_cesa_des_op(req, &tmpl); tmpl 491 drivers/crypto/marvell/cipher.c struct mv_cesa_op_ctx tmpl; tmpl 493 drivers/crypto/marvell/cipher.c mv_cesa_set_op_cfg(&tmpl, tmpl 497 drivers/crypto/marvell/cipher.c return mv_cesa_des_op(req, &tmpl); tmpl 521 drivers/crypto/marvell/cipher.c struct mv_cesa_op_ctx *tmpl) tmpl 523 drivers/crypto/marvell/cipher.c mv_cesa_update_op_cfg(tmpl, CESA_SA_DESC_CFG_CRYPTCM_CBC, tmpl 526 drivers/crypto/marvell/cipher.c memcpy(tmpl->ctx.blkcipher.iv, req->iv, DES_BLOCK_SIZE); tmpl 528 drivers/crypto/marvell/cipher.c return mv_cesa_des_op(req, tmpl); tmpl 533 drivers/crypto/marvell/cipher.c struct mv_cesa_op_ctx tmpl; tmpl 535 drivers/crypto/marvell/cipher.c mv_cesa_set_op_cfg(&tmpl, CESA_SA_DESC_CFG_DIR_ENC); tmpl 537 drivers/crypto/marvell/cipher.c return mv_cesa_cbc_des_op(req, &tmpl); tmpl 542 drivers/crypto/marvell/cipher.c struct mv_cesa_op_ctx tmpl; tmpl 544 drivers/crypto/marvell/cipher.c mv_cesa_set_op_cfg(&tmpl, CESA_SA_DESC_CFG_DIR_DEC); tmpl 546 drivers/crypto/marvell/cipher.c return mv_cesa_cbc_des_op(req, &tmpl); tmpl 571 drivers/crypto/marvell/cipher.c struct mv_cesa_op_ctx *tmpl) tmpl 575 drivers/crypto/marvell/cipher.c mv_cesa_update_op_cfg(tmpl, CESA_SA_DESC_CFG_CRYPTM_3DES, tmpl 578 drivers/crypto/marvell/cipher.c memcpy(tmpl->ctx.blkcipher.key, ctx->key, DES3_EDE_KEY_SIZE); tmpl 580 drivers/crypto/marvell/cipher.c return mv_cesa_skcipher_queue_req(req, tmpl); tmpl 585 drivers/crypto/marvell/cipher.c struct mv_cesa_op_ctx tmpl; tmpl 587 drivers/crypto/marvell/cipher.c mv_cesa_set_op_cfg(&tmpl, tmpl 592 drivers/crypto/marvell/cipher.c return mv_cesa_des3_op(req, &tmpl); tmpl 597 drivers/crypto/marvell/cipher.c struct mv_cesa_op_ctx tmpl; tmpl 599 drivers/crypto/marvell/cipher.c mv_cesa_set_op_cfg(&tmpl, tmpl 604 drivers/crypto/marvell/cipher.c return mv_cesa_des3_op(req, &tmpl); tmpl 629 drivers/crypto/marvell/cipher.c struct mv_cesa_op_ctx *tmpl) tmpl 631 drivers/crypto/marvell/cipher.c memcpy(tmpl->ctx.blkcipher.iv, req->iv, DES3_EDE_BLOCK_SIZE); tmpl 633 drivers/crypto/marvell/cipher.c return mv_cesa_des3_op(req, tmpl); tmpl 638 drivers/crypto/marvell/cipher.c struct mv_cesa_op_ctx tmpl; tmpl 640 drivers/crypto/marvell/cipher.c mv_cesa_set_op_cfg(&tmpl, tmpl 645 drivers/crypto/marvell/cipher.c return mv_cesa_cbc_des3_op(req, &tmpl); tmpl 650 drivers/crypto/marvell/cipher.c struct mv_cesa_op_ctx tmpl; tmpl 652 drivers/crypto/marvell/cipher.c mv_cesa_set_op_cfg(&tmpl, tmpl 657 drivers/crypto/marvell/cipher.c return mv_cesa_cbc_des3_op(req, &tmpl); tmpl 682 drivers/crypto/marvell/cipher.c struct mv_cesa_op_ctx *tmpl) tmpl 691 drivers/crypto/marvell/cipher.c if (mv_cesa_get_op_cfg(tmpl) & CESA_SA_DESC_CFG_DIR_DEC) tmpl 697 drivers/crypto/marvell/cipher.c tmpl->ctx.blkcipher.key[i] = cpu_to_le32(key[i]); tmpl 704 drivers/crypto/marvell/cipher.c mv_cesa_update_op_cfg(tmpl, cfg, tmpl 708 drivers/crypto/marvell/cipher.c return mv_cesa_skcipher_queue_req(req, tmpl); tmpl 713 drivers/crypto/marvell/cipher.c struct mv_cesa_op_ctx tmpl; tmpl 715 drivers/crypto/marvell/cipher.c mv_cesa_set_op_cfg(&tmpl, tmpl 719 drivers/crypto/marvell/cipher.c return mv_cesa_aes_op(req, &tmpl); tmpl 724 drivers/crypto/marvell/cipher.c struct mv_cesa_op_ctx tmpl; tmpl 726 drivers/crypto/marvell/cipher.c mv_cesa_set_op_cfg(&tmpl, tmpl 730 drivers/crypto/marvell/cipher.c return mv_cesa_aes_op(req, &tmpl); tmpl 754 drivers/crypto/marvell/cipher.c struct mv_cesa_op_ctx *tmpl) tmpl 756 drivers/crypto/marvell/cipher.c mv_cesa_update_op_cfg(tmpl, CESA_SA_DESC_CFG_CRYPTCM_CBC, tmpl 758 drivers/crypto/marvell/cipher.c memcpy(tmpl->ctx.blkcipher.iv, req->iv, AES_BLOCK_SIZE); tmpl 760 drivers/crypto/marvell/cipher.c return mv_cesa_aes_op(req, tmpl); tmpl 765 drivers/crypto/marvell/cipher.c struct mv_cesa_op_ctx tmpl; tmpl 767 drivers/crypto/marvell/cipher.c mv_cesa_set_op_cfg(&tmpl, CESA_SA_DESC_CFG_DIR_ENC); tmpl 769 drivers/crypto/marvell/cipher.c return mv_cesa_cbc_aes_op(req, &tmpl); tmpl 774 drivers/crypto/marvell/cipher.c struct mv_cesa_op_ctx tmpl; tmpl 776 drivers/crypto/marvell/cipher.c mv_cesa_set_op_cfg(&tmpl, CESA_SA_DESC_CFG_DIR_DEC); tmpl 778 drivers/crypto/marvell/cipher.c return mv_cesa_cbc_aes_op(req, &tmpl); tmpl 409 drivers/crypto/marvell/hash.c struct mv_cesa_op_ctx *tmpl, bool algo_le) tmpl 414 drivers/crypto/marvell/hash.c mv_cesa_update_op_cfg(tmpl, tmpl 419 drivers/crypto/marvell/hash.c mv_cesa_set_mac_op_total_len(tmpl, 0); tmpl 420 drivers/crypto/marvell/hash.c mv_cesa_set_mac_op_frag_len(tmpl, 0); tmpl 421 drivers/crypto/marvell/hash.c creq->op_tmpl = *tmpl; tmpl 460 drivers/crypto/marvell/hash.c struct mv_cesa_op_ctx *tmpl, unsigned int frag_len, tmpl 466 drivers/crypto/marvell/hash.c op = mv_cesa_dma_add_op(chain, tmpl, false, flags); tmpl 478 drivers/crypto/marvell/hash.c if (mv_cesa_mac_op_is_first_frag(tmpl)) tmpl 479 drivers/crypto/marvell/hash.c mv_cesa_update_op_cfg(tmpl, tmpl 785 drivers/crypto/marvell/hash.c struct mv_cesa_op_ctx *tmpl = &creq->op_tmpl; tmpl 787 drivers/crypto/marvell/hash.c mv_cesa_set_mac_op_total_len(tmpl, creq->len); tmpl 797 drivers/crypto/marvell/hash.c struct mv_cesa_op_ctx *tmpl = &creq->op_tmpl; tmpl 800 drivers/crypto/marvell/hash.c mv_cesa_set_mac_op_total_len(tmpl, creq->len); tmpl 861 drivers/crypto/marvell/hash.c struct mv_cesa_op_ctx tmpl = { }; tmpl 863 drivers/crypto/marvell/hash.c mv_cesa_set_op_cfg(&tmpl, CESA_SA_DESC_CFG_MACM_MD5); tmpl 865 drivers/crypto/marvell/hash.c mv_cesa_ahash_init(req, &tmpl, true); tmpl 930 drivers/crypto/marvell/hash.c struct mv_cesa_op_ctx tmpl = { }; tmpl 932 drivers/crypto/marvell/hash.c mv_cesa_set_op_cfg(&tmpl, CESA_SA_DESC_CFG_MACM_SHA1); tmpl 934 drivers/crypto/marvell/hash.c mv_cesa_ahash_init(req, &tmpl, false); tmpl 1000 drivers/crypto/marvell/hash.c struct mv_cesa_op_ctx tmpl = { }; tmpl 1002 drivers/crypto/marvell/hash.c mv_cesa_set_op_cfg(&tmpl, CESA_SA_DESC_CFG_MACM_SHA256); tmpl 1004 drivers/crypto/marvell/hash.c mv_cesa_ahash_init(req, &tmpl, false); tmpl 1237 drivers/crypto/marvell/hash.c struct mv_cesa_op_ctx tmpl = { }; tmpl 1239 drivers/crypto/marvell/hash.c mv_cesa_set_op_cfg(&tmpl, CESA_SA_DESC_CFG_MACM_HMAC_MD5); tmpl 1240 drivers/crypto/marvell/hash.c memcpy(tmpl.ctx.hash.iv, ctx->iv, sizeof(ctx->iv)); tmpl 1242 drivers/crypto/marvell/hash.c mv_cesa_ahash_init(req, &tmpl, true); tmpl 1307 drivers/crypto/marvell/hash.c struct mv_cesa_op_ctx tmpl = { }; tmpl 1309 drivers/crypto/marvell/hash.c mv_cesa_set_op_cfg(&tmpl, CESA_SA_DESC_CFG_MACM_HMAC_SHA1); tmpl 1310 drivers/crypto/marvell/hash.c memcpy(tmpl.ctx.hash.iv, ctx->iv, sizeof(ctx->iv)); tmpl 1312 drivers/crypto/marvell/hash.c mv_cesa_ahash_init(req, &tmpl, false); tmpl 1397 drivers/crypto/marvell/hash.c struct mv_cesa_op_ctx tmpl = { }; tmpl 1399 drivers/crypto/marvell/hash.c mv_cesa_set_op_cfg(&tmpl, CESA_SA_DESC_CFG_MACM_HMAC_SHA256); tmpl 1400 drivers/crypto/marvell/hash.c memcpy(tmpl.ctx.hash.iv, ctx->iv, sizeof(ctx->iv)); tmpl 1402 drivers/crypto/marvell/hash.c mv_cesa_ahash_init(req, &tmpl, false); tmpl 1374 drivers/crypto/n2_core.c static int __n2_register_one_cipher(const struct n2_cipher_tmpl *tmpl) tmpl 1385 drivers/crypto/n2_core.c snprintf(alg->cra_name, CRYPTO_MAX_ALG_NAME, "%s", tmpl->name); tmpl 1386 drivers/crypto/n2_core.c snprintf(alg->cra_driver_name, CRYPTO_MAX_ALG_NAME, "%s-n2", tmpl->drv_name); tmpl 1390 drivers/crypto/n2_core.c alg->cra_blocksize = tmpl->block_size; tmpl 1391 drivers/crypto/n2_core.c p->enc_type = tmpl->enc_type; tmpl 1394 drivers/crypto/n2_core.c alg->cra_u.ablkcipher = tmpl->ablkcipher; tmpl 1448 drivers/crypto/n2_core.c static int __n2_register_one_ahash(const struct n2_hash_tmpl *tmpl) tmpl 1459 drivers/crypto/n2_core.c p->hash_zero = tmpl->hash_zero; tmpl 1460 drivers/crypto/n2_core.c p->hash_init = tmpl->hash_init; tmpl 1461 drivers/crypto/n2_core.c p->auth_type = tmpl->auth_type; tmpl 1462 drivers/crypto/n2_core.c p->hmac_type = tmpl->hmac_type; tmpl 1463 drivers/crypto/n2_core.c p->hw_op_hashsz = tmpl->hw_op_hashsz; tmpl 1464 drivers/crypto/n2_core.c p->digest_size = tmpl->digest_size; tmpl 1476 drivers/crypto/n2_core.c halg->digestsize = tmpl->digest_size; tmpl 1479 drivers/crypto/n2_core.c snprintf(base->cra_name, CRYPTO_MAX_ALG_NAME, "%s", tmpl->name); tmpl 1480 drivers/crypto/n2_core.c snprintf(base->cra_driver_name, CRYPTO_MAX_ALG_NAME, "%s-n2", tmpl->name); tmpl 1484 drivers/crypto/n2_core.c base->cra_blocksize = tmpl->block_size; tmpl 22 drivers/crypto/qce/ablkcipher.c struct qce_alg_template *tmpl = to_cipher_tmpl(async_req->tfm); tmpl 23 drivers/crypto/qce/ablkcipher.c struct qce_device *qce = tmpl->qce; tmpl 48 drivers/crypto/qce/ablkcipher.c qce->async_req_done(tmpl->qce, error); tmpl 57 drivers/crypto/qce/ablkcipher.c struct qce_alg_template *tmpl = to_cipher_tmpl(async_req->tfm); tmpl 58 drivers/crypto/qce/ablkcipher.c struct qce_device *qce = tmpl->qce; tmpl 134 drivers/crypto/qce/ablkcipher.c ret = qce_start(async_req, tmpl->crypto_alg_type, req->nbytes, 0); tmpl 216 drivers/crypto/qce/ablkcipher.c struct qce_alg_template *tmpl = to_cipher_tmpl(tfm); tmpl 219 drivers/crypto/qce/ablkcipher.c rctx->flags = tmpl->alg_flags; tmpl 237 drivers/crypto/qce/ablkcipher.c return tmpl->qce->async_req_enqueue(tmpl->qce, &req->base); tmpl 357 drivers/crypto/qce/ablkcipher.c struct qce_alg_template *tmpl; tmpl 361 drivers/crypto/qce/ablkcipher.c tmpl = kzalloc(sizeof(*tmpl), GFP_KERNEL); tmpl 362 drivers/crypto/qce/ablkcipher.c if (!tmpl) tmpl 365 drivers/crypto/qce/ablkcipher.c alg = &tmpl->alg.crypto; tmpl 391 drivers/crypto/qce/ablkcipher.c INIT_LIST_HEAD(&tmpl->entry); tmpl 392 drivers/crypto/qce/ablkcipher.c tmpl->crypto_alg_type = CRYPTO_ALG_TYPE_ABLKCIPHER; tmpl 393 drivers/crypto/qce/ablkcipher.c tmpl->alg_flags = def->flags; tmpl 394 drivers/crypto/qce/ablkcipher.c tmpl->qce = qce; tmpl 398 drivers/crypto/qce/ablkcipher.c kfree(tmpl); tmpl 403 drivers/crypto/qce/ablkcipher.c list_add_tail(&tmpl->entry, &ablkcipher_algs); tmpl 410 drivers/crypto/qce/ablkcipher.c struct qce_alg_template *tmpl, *n; tmpl 412 drivers/crypto/qce/ablkcipher.c list_for_each_entry_safe(tmpl, n, &ablkcipher_algs, entry) { tmpl 413 drivers/crypto/qce/ablkcipher.c crypto_unregister_alg(&tmpl->alg.crypto); tmpl 414 drivers/crypto/qce/ablkcipher.c list_del(&tmpl->entry); tmpl 415 drivers/crypto/qce/ablkcipher.c kfree(tmpl); tmpl 228 drivers/crypto/qce/common.c struct qce_alg_template *tmpl = to_ahash_tmpl(async_req->tfm); tmpl 229 drivers/crypto/qce/common.c struct qce_device *qce = tmpl->qce; tmpl 313 drivers/crypto/qce/common.c struct qce_alg_template *tmpl = to_cipher_tmpl(async_req->tfm); tmpl 314 drivers/crypto/qce/common.c struct qce_device *qce = tmpl->qce; tmpl 35 drivers/crypto/qce/sha.c struct qce_alg_template *tmpl = to_ahash_tmpl(async_req->tfm); tmpl 36 drivers/crypto/qce/sha.c struct qce_device *qce = tmpl->qce; tmpl 65 drivers/crypto/qce/sha.c qce->async_req_done(tmpl->qce, error); tmpl 73 drivers/crypto/qce/sha.c struct qce_alg_template *tmpl = to_ahash_tmpl(async_req->tfm); tmpl 74 drivers/crypto/qce/sha.c struct qce_device *qce = tmpl->qce; tmpl 109 drivers/crypto/qce/sha.c ret = qce_start(async_req, tmpl->crypto_alg_type, 0, 0); tmpl 127 drivers/crypto/qce/sha.c struct qce_alg_template *tmpl = to_ahash_tmpl(req->base.tfm); tmpl 128 drivers/crypto/qce/sha.c const u32 *std_iv = tmpl->std_iv; tmpl 133 drivers/crypto/qce/sha.c rctx->flags = tmpl->alg_flags; tmpl 230 drivers/crypto/qce/sha.c struct qce_alg_template *tmpl = to_ahash_tmpl(req->base.tfm); tmpl 231 drivers/crypto/qce/sha.c struct qce_device *qce = tmpl->qce; tmpl 299 drivers/crypto/qce/sha.c return qce->async_req_enqueue(tmpl->qce, &req->base); tmpl 305 drivers/crypto/qce/sha.c struct qce_alg_template *tmpl = to_ahash_tmpl(req->base.tfm); tmpl 306 drivers/crypto/qce/sha.c struct qce_device *qce = tmpl->qce; tmpl 322 drivers/crypto/qce/sha.c return qce->async_req_enqueue(tmpl->qce, &req->base); tmpl 328 drivers/crypto/qce/sha.c struct qce_alg_template *tmpl = to_ahash_tmpl(req->base.tfm); tmpl 329 drivers/crypto/qce/sha.c struct qce_device *qce = tmpl->qce; tmpl 341 drivers/crypto/qce/sha.c return qce->async_req_enqueue(tmpl->qce, &req->base); tmpl 472 drivers/crypto/qce/sha.c struct qce_alg_template *tmpl; tmpl 477 drivers/crypto/qce/sha.c tmpl = kzalloc(sizeof(*tmpl), GFP_KERNEL); tmpl 478 drivers/crypto/qce/sha.c if (!tmpl) tmpl 481 drivers/crypto/qce/sha.c tmpl->std_iv = def->std_iv; tmpl 483 drivers/crypto/qce/sha.c alg = &tmpl->alg.ahash; tmpl 508 drivers/crypto/qce/sha.c INIT_LIST_HEAD(&tmpl->entry); tmpl 509 drivers/crypto/qce/sha.c tmpl->crypto_alg_type = CRYPTO_ALG_TYPE_AHASH; tmpl 510 drivers/crypto/qce/sha.c tmpl->alg_flags = def->flags; tmpl 511 drivers/crypto/qce/sha.c tmpl->qce = qce; tmpl 515 drivers/crypto/qce/sha.c kfree(tmpl); tmpl 520 drivers/crypto/qce/sha.c list_add_tail(&tmpl->entry, &ahash_algs); tmpl 527 drivers/crypto/qce/sha.c struct qce_alg_template *tmpl, *n; tmpl 529 drivers/crypto/qce/sha.c list_for_each_entry_safe(tmpl, n, &ahash_algs, entry) { tmpl 530 drivers/crypto/qce/sha.c crypto_unregister_ahash(&tmpl->alg.ahash); tmpl 531 drivers/crypto/qce/sha.c list_del(&tmpl->entry); tmpl 532 drivers/crypto/qce/sha.c kfree(tmpl); tmpl 264 drivers/media/dvb-frontends/mn88443x.c u32 cpmon, tmpu, tmpl, flg; tmpl 298 drivers/media/dvb-frontends/mn88443x.c regmap_read(r_s, CNRDXL_S, &tmpl); tmpl 299 drivers/media/dvb-frontends/mn88443x.c x = (tmpu << 8) | tmpl; tmpl 301 drivers/media/dvb-frontends/mn88443x.c regmap_read(r_s, CNRDYL_S, &tmpl); tmpl 302 drivers/media/dvb-frontends/mn88443x.c y = (tmpu << 8) | tmpl; tmpl 340 drivers/media/dvb-frontends/mn88443x.c regmap_read(r_s, BERVRDL_S, &tmpl); tmpl 341 drivers/media/dvb-frontends/mn88443x.c bit_err = (tmpu << 8) | tmpl; tmpl 466 drivers/media/dvb-frontends/mn88443x.c u32 seqrd, st, flg, tmpu, tmpm, tmpl; tmpl 486 drivers/media/dvb-frontends/mn88443x.c regmap_read(r_t, AGCRDL_T, &tmpl); tmpl 487 drivers/media/dvb-frontends/mn88443x.c agc = (tmpu << 8) | tmpl; tmpl 501 drivers/media/dvb-frontends/mn88443x.c regmap_read(r_t, CNRDL_T, &tmpl); tmpl 503 drivers/media/dvb-frontends/mn88443x.c if (tmpu || tmpl) { tmpl 506 drivers/media/dvb-frontends/mn88443x.c tmp = (u64)80807124 - intlog10((tmpu << 8) | tmpl) tmpl 529 drivers/media/dvb-frontends/mn88443x.c regmap_read(r_t, BERRDL_T, &tmpl); tmpl 530 drivers/media/dvb-frontends/mn88443x.c bit_err = (tmpu << 16) | (tmpm << 8) | tmpl; tmpl 533 drivers/media/dvb-frontends/mn88443x.c regmap_read(r_t, BERLENRDL_T, &tmpl); tmpl 534 drivers/media/dvb-frontends/mn88443x.c bit_cnt = ((tmpu << 8) | tmpl) * 203 * 8; tmpl 1828 drivers/net/bonding/bond_3ad.c static const struct port_params tmpl = { tmpl 1859 drivers/net/bonding/bond_3ad.c memcpy(&port->partner_admin, &tmpl, sizeof(tmpl)); tmpl 1860 drivers/net/bonding/bond_3ad.c memcpy(&port->partner_oper, &tmpl, sizeof(tmpl)); tmpl 1189 drivers/net/ethernet/mellanox/mlxsw/spectrum_fid.c const struct mlxsw_sp_fid_family *tmpl) tmpl 1191 drivers/net/ethernet/mellanox/mlxsw/spectrum_fid.c u16 nr_fids = tmpl->end_index - tmpl->start_index + 1; tmpl 1195 drivers/net/ethernet/mellanox/mlxsw/spectrum_fid.c fid_family = kmemdup(tmpl, sizeof(*fid_family), GFP_KERNEL); tmpl 1213 drivers/net/ethernet/mellanox/mlxsw/spectrum_fid.c mlxsw_sp->fid_core->fid_family_arr[tmpl->type] = fid_family; tmpl 1220 drivers/net/wireless/ti/wlcore/cmd.c struct wl12xx_arp_rsp_template *tmpl; tmpl 1224 drivers/net/wireless/ti/wlcore/cmd.c skb = dev_alloc_skb(sizeof(*hdr) + sizeof(__le16) + sizeof(*tmpl) + tmpl 1233 drivers/net/wireless/ti/wlcore/cmd.c tmpl = skb_put_zero(skb, sizeof(*tmpl)); tmpl 1236 drivers/net/wireless/ti/wlcore/cmd.c memcpy(tmpl->llc_hdr, rfc1042_header, sizeof(rfc1042_header)); tmpl 1237 drivers/net/wireless/ti/wlcore/cmd.c tmpl->llc_type = cpu_to_be16(ETH_P_ARP); tmpl 1240 drivers/net/wireless/ti/wlcore/cmd.c arp_hdr = &tmpl->arp_hdr; tmpl 1248 drivers/net/wireless/ti/wlcore/cmd.c memcpy(tmpl->sender_hw, vif->addr, ETH_ALEN); tmpl 1249 drivers/net/wireless/ti/wlcore/cmd.c tmpl->sender_ip = wlvif->ip_addr; tmpl 147 drivers/net/wireless/ti/wlcore/init.c struct wl12xx_disconn_template *tmpl; tmpl 151 drivers/net/wireless/ti/wlcore/init.c tmpl = kzalloc(sizeof(*tmpl), GFP_KERNEL); tmpl 152 drivers/net/wireless/ti/wlcore/init.c if (!tmpl) { tmpl 157 drivers/net/wireless/ti/wlcore/init.c tmpl->header.frame_ctl = cpu_to_le16(IEEE80211_FTYPE_MGMT | tmpl 163 drivers/net/wireless/ti/wlcore/init.c tmpl, sizeof(*tmpl), 0, rate); tmpl 166 drivers/net/wireless/ti/wlcore/init.c kfree(tmpl); tmpl 18 drivers/s390/net/fsm.c int nr_events, const fsm_node *tmpl, int tmpl_len, gfp_t order) tmpl 57 drivers/s390/net/fsm.c if ((tmpl[i].cond_state >= nr_states) || tmpl 58 drivers/s390/net/fsm.c (tmpl[i].cond_event >= nr_events) ) { tmpl 61 drivers/s390/net/fsm.c name, i, (long)tmpl[i].cond_state, (long)f->nr_states, tmpl 62 drivers/s390/net/fsm.c (long)tmpl[i].cond_event, (long)f->nr_events); tmpl 66 drivers/s390/net/fsm.c m[nr_states * tmpl[i].cond_event + tmpl[i].cond_state] = tmpl 67 drivers/s390/net/fsm.c tmpl[i].function; tmpl 112 drivers/s390/net/fsm.h int nr_states, int nr_events, const fsm_node *tmpl, tmpl 28 drivers/scsi/fnic/cq_exch_desc.h u8 tmpl; tmpl 62 drivers/scsi/fnic/cq_exch_desc.h u16 tmpl; tmpl 95 drivers/scsi/fnic/cq_exch_desc.h u16 *tmpl, tmpl 116 drivers/scsi/fnic/cq_exch_desc.h *tmpl = desc_ptr->tmpl & CQ_FCP_RQ_DESC_TMPL_MASK; tmpl 141 drivers/scsi/fnic/cq_exch_desc.h u16 tmpl; tmpl 169 drivers/scsi/fnic/cq_exch_desc.h u16 *tmpl, tmpl 178 drivers/scsi/fnic/cq_exch_desc.h *tmpl = desc_ptr->tmpl & CQ_SGL_TMPL_MASK; tmpl 849 drivers/scsi/fnic/fnic_fcs.c u16 exchange_id, tmpl; tmpl 866 drivers/scsi/fnic/fnic_fcs.c &tmpl, &fcp_bytes_written, &sof, &eof, tmpl 677 drivers/scsi/qla2xxx/qla_nvme.c struct nvme_fc_port_template *tmpl; tmpl 686 drivers/scsi/qla2xxx/qla_nvme.c tmpl = &qla_nvme_fc_transport; tmpl 705 drivers/scsi/qla2xxx/qla_nvme.c ret = nvme_fc_register_localport(&pinfo, tmpl, tmpl 38 drivers/scsi/raid_class.c #define to_raid_internal(tmpl) container_of(tmpl, struct raid_internal, r) tmpl 41 drivers/scsi/scsi_sas_internal.h #define to_sas_internal(tmpl) container_of(tmpl, struct sas_internal, t) tmpl 343 drivers/scsi/scsi_transport_fc.c #define to_fc_internal(tmpl) container_of(tmpl, struct fc_internal, t) tmpl 101 drivers/scsi/scsi_transport_iscsi.c #define to_iscsi_internal(tmpl) \ tmpl 102 drivers/scsi/scsi_transport_iscsi.c container_of(tmpl, struct iscsi_internal, t) tmpl 64 drivers/scsi/scsi_transport_spi.c #define to_spi_internal(tmpl) container_of(tmpl, struct spi_internal, t) tmpl 42 drivers/scsi/scsi_transport_srp.c #define to_srp_internal(tmpl) container_of(tmpl, struct srp_internal, t) tmpl 49 include/crypto/algapi.h struct crypto_template *tmpl; tmpl 62 include/crypto/algapi.h int (*create)(struct crypto_template *tmpl, struct rtattr **tb); tmpl 140 include/crypto/algapi.h int crypto_register_template(struct crypto_template *tmpl); tmpl 142 include/crypto/algapi.h void crypto_unregister_template(struct crypto_template *tmpl); tmpl 146 include/crypto/algapi.h int crypto_register_instance(struct crypto_template *tmpl, tmpl 182 include/crypto/internal/aead.h int aead_register_instance(struct crypto_template *tmpl, tmpl 137 include/crypto/internal/akcipher.h int akcipher_register_instance(struct crypto_template *tmpl, tmpl 22 include/crypto/internal/geniv.h struct aead_instance *aead_geniv_alloc(struct crypto_template *tmpl, tmpl 76 include/crypto/internal/hash.h int ahash_register_instance(struct crypto_template *tmpl, tmpl 105 include/crypto/internal/hash.h int shash_register_instance(struct crypto_template *tmpl, tmpl 133 include/crypto/internal/skcipher.h int skcipher_register_instance(struct crypto_template *tmpl, tmpl 250 include/crypto/internal/skcipher.h skcipher_alloc_instance_simple(struct crypto_template *tmpl, struct rtattr **tb, tmpl 221 include/linux/sunrpc/cache.h extern struct cache_detail *cache_create_net(const struct cache_detail *tmpl, struct net *net); tmpl 47 include/net/fib_notifier.h fib_notifier_ops_register(const struct fib_notifier_ops *tmpl, struct net *net); tmpl 312 include/net/netfilter/nf_conntrack.h void nf_ct_tmpl_free(struct nf_conn *tmpl); tmpl 115 include/net/netfilter/nf_conntrack_helper.h int __nf_ct_try_assign_helper(struct nf_conn *ct, struct nf_conn *tmpl, tmpl 78 include/net/netfilter/nf_conntrack_l4proto.h int nf_conntrack_inet_error(struct nf_conn *tmpl, struct sk_buff *skb, tmpl 84 include/net/netfilter/nf_conntrack_l4proto.h int nf_conntrack_icmpv4_error(struct nf_conn *tmpl, tmpl 89 include/net/netfilter/nf_conntrack_l4proto.h int nf_conntrack_icmpv6_error(struct nf_conn *tmpl, tmpl 33 include/net/netfilter/nf_conntrack_synproxy.h const struct nf_conn *tmpl) tmpl 36 include/net/netfilter/nf_conntrack_synproxy.h if (tmpl && nfct_synproxy(tmpl)) { tmpl 29 include/net/netfilter/nf_conntrack_zones.h nf_ct_zone_tmpl(const struct nf_conn *tmpl, const struct sk_buff *skb, tmpl 33 include/net/netfilter/nf_conntrack_zones.h if (!tmpl) tmpl 36 include/net/netfilter/nf_conntrack_zones.h if (tmpl->zone.flags & NF_CT_FLAG_MARK) tmpl 37 include/net/netfilter/nf_conntrack_zones.h return nf_ct_zone_init(tmp, skb->mark, tmpl->zone.dir, 0); tmpl 39 include/net/netfilter/nf_conntrack_zones.h return nf_ct_zone(tmpl); tmpl 23 include/net/netfilter/nf_synproxy.h struct nf_conn *tmpl; tmpl 563 include/net/netfilter/nf_tables.h static inline void nft_set_ext_prepare(struct nft_set_ext_tmpl *tmpl) tmpl 565 include/net/netfilter/nf_tables.h memset(tmpl, 0, sizeof(*tmpl)); tmpl 566 include/net/netfilter/nf_tables.h tmpl->len = sizeof(struct nft_set_ext); tmpl 569 include/net/netfilter/nf_tables.h static inline void nft_set_ext_add_length(struct nft_set_ext_tmpl *tmpl, u8 id, tmpl 572 include/net/netfilter/nf_tables.h tmpl->len = ALIGN(tmpl->len, nft_set_ext_types[id].align); tmpl 573 include/net/netfilter/nf_tables.h BUG_ON(tmpl->len > U8_MAX); tmpl 574 include/net/netfilter/nf_tables.h tmpl->offset[id] = tmpl->len; tmpl 575 include/net/netfilter/nf_tables.h tmpl->len += nft_set_ext_types[id].len + len; tmpl 578 include/net/netfilter/nf_tables.h static inline void nft_set_ext_add(struct nft_set_ext_tmpl *tmpl, u8 id) tmpl 580 include/net/netfilter/nf_tables.h nft_set_ext_add_length(tmpl, id, 0); tmpl 584 include/net/netfilter/nf_tables.h const struct nft_set_ext_tmpl *tmpl) tmpl 586 include/net/netfilter/nf_tables.h memcpy(ext->offset, tmpl->offset, sizeof(ext->offset)); tmpl 657 include/net/netfilter/nf_tables.h const struct nft_set_ext_tmpl *tmpl, tmpl 13 include/net/tc_act/tc_ct.h struct nf_conn *tmpl; tmpl 1060 include/net/xfrm.h __xfrm4_state_addr_cmp(const struct xfrm_tmpl *tmpl, const struct xfrm_state *x) tmpl 1062 include/net/xfrm.h return (tmpl->saddr.a4 && tmpl 1063 include/net/xfrm.h tmpl->saddr.a4 != x->props.saddr.a4); tmpl 1067 include/net/xfrm.h __xfrm6_state_addr_cmp(const struct xfrm_tmpl *tmpl, const struct xfrm_state *x) tmpl 1069 include/net/xfrm.h return (!ipv6_addr_any((struct in6_addr*)&tmpl->saddr) && tmpl 1070 include/net/xfrm.h !ipv6_addr_equal((struct in6_addr *)&tmpl->saddr, (struct in6_addr*)&x->props.saddr)); tmpl 1074 include/net/xfrm.h xfrm_state_addr_cmp(const struct xfrm_tmpl *tmpl, const struct xfrm_state *x, unsigned short family) tmpl 1078 include/net/xfrm.h return __xfrm4_state_addr_cmp(tmpl, x); tmpl 1080 include/net/xfrm.h return __xfrm6_state_addr_cmp(tmpl, x); tmpl 1481 include/net/xfrm.h struct xfrm_tmpl *tmpl, tmpl 151 net/core/fib_notifier.c fib_notifier_ops_register(const struct fib_notifier_ops *tmpl, struct net *net) tmpl 156 net/core/fib_notifier.c ops = kmemdup(tmpl, sizeof(*ops), GFP_KERNEL); tmpl 150 net/core/fib_rules.c fib_rules_register(const struct fib_rules_ops *tmpl, struct net *net) tmpl 155 net/core/fib_rules.c ops = kmemdup(tmpl, sizeof(*ops), GFP_KERNEL); tmpl 545 net/netfilter/nf_conntrack_core.c struct nf_conn *tmpl, *p; tmpl 548 net/netfilter/nf_conntrack_core.c tmpl = kzalloc(sizeof(*tmpl) + NFCT_INFOMASK, flags); tmpl 549 net/netfilter/nf_conntrack_core.c if (!tmpl) tmpl 552 net/netfilter/nf_conntrack_core.c p = tmpl; tmpl 553 net/netfilter/nf_conntrack_core.c tmpl = (struct nf_conn *)NFCT_ALIGN((unsigned long)p); tmpl 554 net/netfilter/nf_conntrack_core.c if (tmpl != p) { tmpl 555 net/netfilter/nf_conntrack_core.c tmpl = (struct nf_conn *)NFCT_ALIGN((unsigned long)p); tmpl 556 net/netfilter/nf_conntrack_core.c tmpl->proto.tmpl_padto = (char *)tmpl - (char *)p; tmpl 559 net/netfilter/nf_conntrack_core.c tmpl = kzalloc(sizeof(*tmpl), flags); tmpl 560 net/netfilter/nf_conntrack_core.c if (!tmpl) tmpl 564 net/netfilter/nf_conntrack_core.c tmpl->status = IPS_TEMPLATE; tmpl 565 net/netfilter/nf_conntrack_core.c write_pnet(&tmpl->ct_net, net); tmpl 566 net/netfilter/nf_conntrack_core.c nf_ct_zone_add(tmpl, zone); tmpl 567 net/netfilter/nf_conntrack_core.c atomic_set(&tmpl->ct_general.use, 0); tmpl 569 net/netfilter/nf_conntrack_core.c return tmpl; tmpl 573 net/netfilter/nf_conntrack_core.c void nf_ct_tmpl_free(struct nf_conn *tmpl) tmpl 575 net/netfilter/nf_conntrack_core.c nf_ct_ext_destroy(tmpl); tmpl 576 net/netfilter/nf_conntrack_core.c nf_ct_ext_free(tmpl); tmpl 579 net/netfilter/nf_conntrack_core.c kfree((char *)tmpl - tmpl->proto.tmpl_padto); tmpl 581 net/netfilter/nf_conntrack_core.c kfree(tmpl); tmpl 1431 net/netfilter/nf_conntrack_core.c init_conntrack(struct net *net, struct nf_conn *tmpl, tmpl 1450 net/netfilter/nf_conntrack_core.c zone = nf_ct_zone_tmpl(tmpl, skb, &tmp); tmpl 1456 net/netfilter/nf_conntrack_core.c if (!nf_ct_add_synproxy(ct, tmpl)) { tmpl 1461 net/netfilter/nf_conntrack_core.c timeout_ext = tmpl ? nf_ct_timeout_find(tmpl) : NULL; tmpl 1471 net/netfilter/nf_conntrack_core.c ecache = tmpl ? nf_ct_ecache_find(tmpl) : NULL; tmpl 1504 net/netfilter/nf_conntrack_core.c __nf_ct_try_assign_helper(ct, tmpl, GFP_ATOMIC); tmpl 1523 net/netfilter/nf_conntrack_core.c resolve_normal_ct(struct nf_conn *tmpl, tmpl 1545 net/netfilter/nf_conntrack_core.c zone = nf_ct_zone_tmpl(tmpl, skb, &tmp); tmpl 1549 net/netfilter/nf_conntrack_core.c h = init_conntrack(state->net, tmpl, &tuple, tmpl 1586 net/netfilter/nf_conntrack_core.c nf_conntrack_handle_icmp(struct nf_conn *tmpl, tmpl 1595 net/netfilter/nf_conntrack_core.c ret = nf_conntrack_icmpv4_error(tmpl, skb, dataoff, state); tmpl 1598 net/netfilter/nf_conntrack_core.c ret = nf_conntrack_icmpv6_error(tmpl, skb, dataoff, state); tmpl 1672 net/netfilter/nf_conntrack_core.c struct nf_conn *ct, *tmpl; tmpl 1676 net/netfilter/nf_conntrack_core.c tmpl = nf_ct_get(skb, &ctinfo); tmpl 1677 net/netfilter/nf_conntrack_core.c if (tmpl || ctinfo == IP_CT_UNTRACKED) { tmpl 1679 net/netfilter/nf_conntrack_core.c if ((tmpl && !nf_ct_is_template(tmpl)) || tmpl 1698 net/netfilter/nf_conntrack_core.c ret = nf_conntrack_handle_icmp(tmpl, skb, dataoff, tmpl 1709 net/netfilter/nf_conntrack_core.c ret = resolve_normal_ct(tmpl, skb, dataoff, tmpl 1750 net/netfilter/nf_conntrack_core.c if (tmpl) tmpl 1751 net/netfilter/nf_conntrack_core.c nf_ct_put(tmpl); tmpl 232 net/netfilter/nf_conntrack_helper.c int __nf_ct_try_assign_helper(struct nf_conn *ct, struct nf_conn *tmpl, tmpl 248 net/netfilter/nf_conntrack_helper.c if (tmpl != NULL) { tmpl 249 net/netfilter/nf_conntrack_helper.c help = nfct_help(tmpl); tmpl 187 net/netfilter/nf_conntrack_proto.c struct nf_conn *tmpl; tmpl 189 net/netfilter/nf_conntrack_proto.c tmpl = nf_ct_get(skb, &ctinfo); tmpl 190 net/netfilter/nf_conntrack_proto.c if (tmpl && nf_ct_is_template(tmpl)) { tmpl 195 net/netfilter/nf_conntrack_proto.c nf_ct_put(tmpl); tmpl 104 net/netfilter/nf_conntrack_proto_icmp.c int nf_conntrack_inet_error(struct nf_conn *tmpl, struct sk_buff *skb, tmpl 119 net/netfilter/nf_conntrack_proto_icmp.c zone = nf_ct_zone_tmpl(tmpl, skb, &tmp); tmpl 203 net/netfilter/nf_conntrack_proto_icmp.c int nf_conntrack_icmpv4_error(struct nf_conn *tmpl, tmpl 249 net/netfilter/nf_conntrack_proto_icmp.c return nf_conntrack_inet_error(tmpl, skb, dataoff, state, tmpl 131 net/netfilter/nf_conntrack_proto_icmpv6.c int nf_conntrack_icmpv6_error(struct nf_conn *tmpl, tmpl 168 net/netfilter/nf_conntrack_proto_icmpv6.c return nf_conntrack_inet_error(tmpl, skb, dataoff, state, tmpl 348 net/netfilter/nf_synproxy_core.c snet->tmpl = ct; tmpl 372 net/netfilter/nf_synproxy_core.c nf_ct_put(snet->tmpl); tmpl 548 net/netfilter/nf_synproxy_core.c synproxy_send_tcp(net, skb, nskb, &snet->tmpl->ct_general, IP_CT_NEW, tmpl 963 net/netfilter/nf_synproxy_core.c synproxy_send_tcp_ipv6(net, skb, nskb, &snet->tmpl->ct_general, tmpl 4422 net/netfilter/nf_tables_api.c const struct nft_set_ext_tmpl *tmpl, tmpl 4429 net/netfilter/nf_tables_api.c elem = kzalloc(set->ops->elemsize + tmpl->len, gfp); tmpl 4434 net/netfilter/nf_tables_api.c nft_set_ext_init(ext, tmpl); tmpl 4497 net/netfilter/nf_tables_api.c struct nft_set_ext_tmpl tmpl; tmpl 4520 net/netfilter/nf_tables_api.c nft_set_ext_prepare(&tmpl); tmpl 4526 net/netfilter/nf_tables_api.c nft_set_ext_add(&tmpl, NFT_SET_EXT_FLAGS); tmpl 4576 net/netfilter/nf_tables_api.c nft_set_ext_add_length(&tmpl, NFT_SET_EXT_KEY, d1.len); tmpl 4578 net/netfilter/nf_tables_api.c nft_set_ext_add(&tmpl, NFT_SET_EXT_EXPIRATION); tmpl 4580 net/netfilter/nf_tables_api.c nft_set_ext_add(&tmpl, NFT_SET_EXT_TIMEOUT); tmpl 4595 net/netfilter/nf_tables_api.c nft_set_ext_add(&tmpl, NFT_SET_EXT_OBJREF); tmpl 4633 net/netfilter/nf_tables_api.c nft_set_ext_add_length(&tmpl, NFT_SET_EXT_DATA, d2.len); tmpl 4644 net/netfilter/nf_tables_api.c nft_set_ext_add_length(&tmpl, NFT_SET_EXT_USERDATA, tmpl 4649 net/netfilter/nf_tables_api.c elem.priv = nft_set_elem_init(set, &tmpl, elem.key.val.data, data.data, tmpl 4813 net/netfilter/nf_tables_api.c struct nft_set_ext_tmpl tmpl; tmpl 4831 net/netfilter/nf_tables_api.c nft_set_ext_prepare(&tmpl); tmpl 4837 net/netfilter/nf_tables_api.c nft_set_ext_add(&tmpl, NFT_SET_EXT_FLAGS); tmpl 4848 net/netfilter/nf_tables_api.c nft_set_ext_add_length(&tmpl, NFT_SET_EXT_KEY, desc.len); tmpl 4851 net/netfilter/nf_tables_api.c elem.priv = nft_set_elem_init(set, &tmpl, elem.key.val.data, NULL, 0, tmpl 17 net/netfilter/nft_dynset.c struct nft_set_ext_tmpl tmpl; tmpl 56 net/netfilter/nft_dynset.c elem = nft_set_elem_init(set, &priv->tmpl, tmpl 226 net/netfilter/nft_dynset.c nft_set_ext_prepare(&priv->tmpl); tmpl 227 net/netfilter/nft_dynset.c nft_set_ext_add_length(&priv->tmpl, NFT_SET_EXT_KEY, set->klen); tmpl 229 net/netfilter/nft_dynset.c nft_set_ext_add_length(&priv->tmpl, NFT_SET_EXT_DATA, set->dlen); tmpl 231 net/netfilter/nft_dynset.c nft_set_ext_add_length(&priv->tmpl, NFT_SET_EXT_EXPR, tmpl 235 net/netfilter/nft_dynset.c nft_set_ext_add(&priv->tmpl, NFT_SET_EXT_EXPIRATION); tmpl 958 net/openvswitch/conntrack.c struct nf_conn *tmpl = info->ct; tmpl 962 net/openvswitch/conntrack.c if (tmpl) { tmpl 965 net/openvswitch/conntrack.c nf_conntrack_get(&tmpl->ct_general); tmpl 966 net/openvswitch/conntrack.c nf_ct_set(skb, tmpl, IP_CT_NEW); tmpl 210 net/sched/act_ct.c if (params->tmpl) tmpl 211 net/sched/act_ct.c nf_conntrack_put(¶ms->tmpl->ct_general); tmpl 386 net/sched/act_ct.c struct nf_conn *tmpl = NULL; tmpl 399 net/sched/act_ct.c tmpl = p->tmpl; tmpl 440 net/sched/act_ct.c if (tmpl) { tmpl 444 net/sched/act_ct.c nf_conntrack_get(&tmpl->ct_general); tmpl 445 net/sched/act_ct.c nf_ct_set(skb, tmpl, IP_CT_NEW); tmpl 595 net/sched/act_ct.c struct nf_conn *tmpl; tmpl 655 net/sched/act_ct.c tmpl = nf_ct_tmpl_alloc(net, &zone, GFP_KERNEL); tmpl 656 net/sched/act_ct.c if (!tmpl) { tmpl 660 net/sched/act_ct.c __set_bit(IPS_CONFIRMED_BIT, &tmpl->status); tmpl 661 net/sched/act_ct.c nf_conntrack_get(&tmpl->ct_general); tmpl 662 net/sched/act_ct.c p->tmpl = tmpl; tmpl 1717 net/sunrpc/cache.c struct cache_detail *cache_create_net(const struct cache_detail *tmpl, struct net *net) tmpl 1722 net/sunrpc/cache.c cd = kmemdup(tmpl, sizeof(struct cache_detail), GFP_KERNEL); tmpl 2372 net/xfrm/xfrm_policy.c struct xfrm_tmpl *tmpl = &policy->xfrm_vec[i]; tmpl 2374 net/xfrm/xfrm_policy.c if (tmpl->mode == XFRM_MODE_TUNNEL || tmpl 2375 net/xfrm/xfrm_policy.c tmpl->mode == XFRM_MODE_BEET) { tmpl 2376 net/xfrm/xfrm_policy.c remote = &tmpl->id.daddr; tmpl 2377 net/xfrm/xfrm_policy.c local = &tmpl->saddr; tmpl 2378 net/xfrm/xfrm_policy.c if (xfrm_addr_any(local, tmpl->encap_family)) { tmpl 2381 net/xfrm/xfrm_policy.c tmpl->encap_family, 0); tmpl 2388 net/xfrm/xfrm_policy.c x = xfrm_state_find(remote, local, fl, tmpl, policy, &error, tmpl 2405 net/xfrm/xfrm_policy.c if (!tmpl->optional) tmpl 3217 net/xfrm/xfrm_policy.c xfrm_state_ok(const struct xfrm_tmpl *tmpl, const struct xfrm_state *x, tmpl 3221 net/xfrm/xfrm_policy.c return tmpl->optional && !xfrm_state_addr_cmp(tmpl, x, tmpl->encap_family); tmpl 3222 net/xfrm/xfrm_policy.c return x->id.proto == tmpl->id.proto && tmpl 3223 net/xfrm/xfrm_policy.c (x->id.spi == tmpl->id.spi || !tmpl->id.spi) && tmpl 3224 net/xfrm/xfrm_policy.c (x->props.reqid == tmpl->reqid || !tmpl->reqid) && tmpl 3225 net/xfrm/xfrm_policy.c x->props.mode == tmpl->mode && tmpl 3226 net/xfrm/xfrm_policy.c (tmpl->allalgs || (tmpl->aalgos & (1<<x->props.aalgo)) || tmpl 3227 net/xfrm/xfrm_policy.c !(xfrm_id_proto_match(tmpl->id.proto, IPSEC_PROTO_ANY))) && tmpl 3229 net/xfrm/xfrm_policy.c xfrm_state_addr_cmp(tmpl, x, family)); tmpl 3240 net/xfrm/xfrm_policy.c xfrm_policy_ok(const struct xfrm_tmpl *tmpl, const struct sec_path *sp, int start, tmpl 3245 net/xfrm/xfrm_policy.c if (tmpl->optional) { tmpl 3246 net/xfrm/xfrm_policy.c if (tmpl->mode == XFRM_MODE_TRANSPORT) tmpl 3251 net/xfrm/xfrm_policy.c if (xfrm_state_ok(tmpl, sp->xvec[idx], family)) tmpl 890 net/xfrm/xfrm_state.c const struct xfrm_tmpl *tmpl, tmpl 903 net/xfrm/xfrm_state.c x->id = tmpl->id; tmpl 905 net/xfrm/xfrm_state.c switch (tmpl->encap_family) { tmpl 909 net/xfrm/xfrm_state.c x->props.saddr = tmpl->saddr; tmpl 916 net/xfrm/xfrm_state.c memcpy(&x->props.saddr, &tmpl->saddr, sizeof(x->props.saddr)); tmpl 922 net/xfrm/xfrm_state.c x->props.mode = tmpl->mode; tmpl 923 net/xfrm/xfrm_state.c x->props.reqid = tmpl->reqid; tmpl 924 net/xfrm/xfrm_state.c x->props.family = tmpl->encap_family; tmpl 1040 net/xfrm/xfrm_state.c const struct flowi *fl, struct xfrm_tmpl *tmpl, tmpl 1052 net/xfrm/xfrm_state.c unsigned short encap_family = tmpl->encap_family; tmpl 1061 net/xfrm/xfrm_state.c h = xfrm_dst_hash(net, daddr, saddr, tmpl->reqid, encap_family); tmpl 1064 net/xfrm/xfrm_state.c x->props.reqid == tmpl->reqid && tmpl 1069 net/xfrm/xfrm_state.c tmpl->mode == x->props.mode && tmpl 1070 net/xfrm/xfrm_state.c tmpl->id.proto == x->id.proto && tmpl 1071 net/xfrm/xfrm_state.c (tmpl->id.spi == x->id.spi || !tmpl->id.spi)) tmpl 1078 net/xfrm/xfrm_state.c h_wildcard = xfrm_dst_hash(net, daddr, &saddr_wildcard, tmpl->reqid, encap_family); tmpl 1081 net/xfrm/xfrm_state.c x->props.reqid == tmpl->reqid && tmpl 1086 net/xfrm/xfrm_state.c tmpl->mode == x->props.mode && tmpl 1087 net/xfrm/xfrm_state.c tmpl->id.proto == x->id.proto && tmpl 1088 net/xfrm/xfrm_state.c (tmpl->id.spi == x->id.spi || !tmpl->id.spi)) tmpl 1096 net/xfrm/xfrm_state.c if (tmpl->id.spi && tmpl 1097 net/xfrm/xfrm_state.c (x0 = __xfrm_state_lookup(net, mark, daddr, tmpl->id.spi, tmpl 1098 net/xfrm/xfrm_state.c tmpl->id.proto, encap_family)) != NULL) { tmpl 1121 net/xfrm/xfrm_state.c xfrm_init_tempstate(x, fl, tmpl, daddr, saddr, family); tmpl 1133 net/xfrm/xfrm_state.c if (km_query(x, tmpl, pol) == 0) {