tmpl             3860 arch/x86/kvm/vmx/vmx.c 	unsigned long tmpl;
tmpl             3902 arch/x86/kvm/vmx/vmx.c 	rdmsrl(MSR_IA32_SYSENTER_EIP, tmpl);
tmpl             3903 arch/x86/kvm/vmx/vmx.c 	vmcs_writel(HOST_IA32_SYSENTER_EIP, tmpl);   /* 22.2.3 */
tmpl              500 crypto/adiantum.c static int adiantum_create(struct crypto_template *tmpl, struct rtattr **tb)
tmpl              624 crypto/adiantum.c 	err = skcipher_register_instance(tmpl, inst);
tmpl              194 crypto/aead.c  		inst->tmpl->free(inst);
tmpl              231 crypto/aead.c  struct aead_instance *aead_geniv_alloc(struct crypto_template *tmpl,
tmpl              279 crypto/aead.c  		     "%s(%s)", tmpl->name, alg->base.cra_name) >=
tmpl              283 crypto/aead.c  		     "%s(%s)", tmpl->name, alg->base.cra_driver_name) >=
tmpl              451 crypto/aead.c  int aead_register_instance(struct crypto_template *tmpl,
tmpl              460 crypto/aead.c  	return crypto_register_instance(tmpl, aead_crypto_instance(inst));
tmpl              636 crypto/ahash.c int ahash_register_instance(struct crypto_template *tmpl,
tmpl              645 crypto/ahash.c 	return crypto_register_instance(tmpl, ahash_crypto_instance(inst));
tmpl              146 crypto/akcipher.c int akcipher_register_instance(struct crypto_template *tmpl,
tmpl              150 crypto/akcipher.c 	return crypto_register_instance(tmpl, akcipher_crypto_instance(inst));
tmpl               69 crypto/algapi.c 		inst->tmpl->free(inst);
tmpl               79 crypto/algapi.c 	struct crypto_template *tmpl = inst->tmpl;
tmpl               82 crypto/algapi.c 	crypto_tmpl_put(tmpl);
tmpl              110 crypto/algapi.c 	struct crypto_template *tmpl = inst->tmpl;
tmpl              119 crypto/algapi.c 	if (!tmpl || !crypto_tmpl_get(tmpl))
tmpl              474 crypto/algapi.c int crypto_register_template(struct crypto_template *tmpl)
tmpl              481 crypto/algapi.c 	crypto_check_module_sig(tmpl->module);
tmpl              484 crypto/algapi.c 		if (q == tmpl)
tmpl              488 crypto/algapi.c 	list_add(&tmpl->list, &crypto_template_list);
tmpl              514 crypto/algapi.c void crypto_unregister_template(struct crypto_template *tmpl)
tmpl              523 crypto/algapi.c 	BUG_ON(list_empty(&tmpl->list));
tmpl              524 crypto/algapi.c 	list_del_init(&tmpl->list);
tmpl              526 crypto/algapi.c 	list = &tmpl->instances;
tmpl              554 crypto/algapi.c 	struct crypto_template *q, *tmpl = NULL;
tmpl              563 crypto/algapi.c 		tmpl = q;
tmpl              568 crypto/algapi.c 	return tmpl;
tmpl              578 crypto/algapi.c int crypto_register_instance(struct crypto_template *tmpl,
tmpl              588 crypto/algapi.c 	inst->alg.cra_module = tmpl->module;
tmpl              597 crypto/algapi.c 	hlist_add_head(&inst->list, &tmpl->instances);
tmpl              598 crypto/algapi.c 	inst->tmpl = tmpl;
tmpl               60 crypto/algboss.c 	struct crypto_template *tmpl;
tmpl               64 crypto/algboss.c 	tmpl = crypto_lookup_template(param->template);
tmpl               65 crypto/algboss.c 	if (!tmpl)
tmpl               69 crypto/algboss.c 		if (tmpl->create) {
tmpl               70 crypto/algboss.c 			err = tmpl->create(tmpl, param->tb);
tmpl               74 crypto/algboss.c 		inst = tmpl->alloc(param->tb);
tmpl               77 crypto/algboss.c 		else if ((err = crypto_register_instance(tmpl, inst)))
tmpl               78 crypto/algboss.c 			tmpl->free(inst);
tmpl               81 crypto/algboss.c 	crypto_tmpl_put(tmpl);
tmpl              382 crypto/authenc.c static int crypto_authenc_create(struct crypto_template *tmpl,
tmpl              472 crypto/authenc.c 	err = aead_register_instance(tmpl, inst);
tmpl              400 crypto/authencesn.c static int crypto_authenc_esn_create(struct crypto_template *tmpl,
tmpl              487 crypto/authencesn.c 	err = aead_register_instance(tmpl, inst);
tmpl               51 crypto/cbc.c   static int crypto_cbc_create(struct crypto_template *tmpl, struct rtattr **tb)
tmpl               57 crypto/cbc.c   	inst = skcipher_alloc_instance_simple(tmpl, tb, &alg);
tmpl               68 crypto/cbc.c   	err = skcipher_register_instance(tmpl, inst);
tmpl              454 crypto/ccm.c   static int crypto_ccm_create_common(struct crypto_template *tmpl,
tmpl              547 crypto/ccm.c   	err = aead_register_instance(tmpl, inst);
tmpl              564 crypto/ccm.c   static int crypto_ccm_create(struct crypto_template *tmpl, struct rtattr **tb)
tmpl              582 crypto/ccm.c   	return crypto_ccm_create_common(tmpl, tb, ctr_name, mac_name);
tmpl              585 crypto/ccm.c   static int crypto_ccm_base_create(struct crypto_template *tmpl,
tmpl              599 crypto/ccm.c   	return crypto_ccm_create_common(tmpl, tb, ctr_name, mac_name);
tmpl              744 crypto/ccm.c   static int crypto_rfc4309_create(struct crypto_template *tmpl,
tmpl              818 crypto/ccm.c   	err = aead_register_instance(tmpl, inst);
tmpl              917 crypto/ccm.c   static int cbcmac_create(struct crypto_template *tmpl, struct rtattr **tb)
tmpl              960 crypto/ccm.c   	err = shash_register_instance(tmpl, inst);
tmpl              200 crypto/cfb.c   static int crypto_cfb_create(struct crypto_template *tmpl, struct rtattr **tb)
tmpl              206 crypto/cfb.c   	inst = skcipher_alloc_instance_simple(tmpl, tb, &alg);
tmpl              222 crypto/cfb.c   	err = skcipher_register_instance(tmpl, inst);
tmpl              562 crypto/chacha20poly1305.c static int chachapoly_create(struct crypto_template *tmpl, struct rtattr **tb,
tmpl              664 crypto/chacha20poly1305.c 	err = aead_register_instance(tmpl, inst);
tmpl              681 crypto/chacha20poly1305.c static int rfc7539_create(struct crypto_template *tmpl, struct rtattr **tb)
tmpl              683 crypto/chacha20poly1305.c 	return chachapoly_create(tmpl, tb, "rfc7539", 12);
tmpl              686 crypto/chacha20poly1305.c static int rfc7539esp_create(struct crypto_template *tmpl, struct rtattr **tb)
tmpl              688 crypto/chacha20poly1305.c 	return chachapoly_create(tmpl, tb, "rfc7539esp", 8);
tmpl              222 crypto/cmac.c  static int cmac_create(struct crypto_template *tmpl, struct rtattr **tb)
tmpl              283 crypto/cmac.c  	err = shash_register_instance(tmpl, inst);
tmpl              396 crypto/cryptd.c static int cryptd_create_skcipher(struct crypto_template *tmpl,
tmpl              453 crypto/cryptd.c 	err = skcipher_register_instance(tmpl, inst);
tmpl              669 crypto/cryptd.c static int cryptd_create_hash(struct crypto_template *tmpl, struct rtattr **tb,
tmpl              722 crypto/cryptd.c 	err = ahash_register_instance(tmpl, inst);
tmpl              852 crypto/cryptd.c static int cryptd_create_aead(struct crypto_template *tmpl,
tmpl              901 crypto/cryptd.c 	err = aead_register_instance(tmpl, inst);
tmpl              913 crypto/cryptd.c static int cryptd_create(struct crypto_template *tmpl, struct rtattr **tb)
tmpl              923 crypto/cryptd.c 		return cryptd_create_skcipher(tmpl, tb, &queue);
tmpl              925 crypto/cryptd.c 		return cryptd_create_hash(tmpl, tb, &queue);
tmpl              927 crypto/cryptd.c 		return cryptd_create_aead(tmpl, tb, &queue);
tmpl              126 crypto/ctr.c   static int crypto_ctr_create(struct crypto_template *tmpl, struct rtattr **tb)
tmpl              132 crypto/ctr.c   	inst = skcipher_alloc_instance_simple(tmpl, tb, &alg);
tmpl              157 crypto/ctr.c   	err = skcipher_register_instance(tmpl, inst);
tmpl              262 crypto/ctr.c   static int crypto_rfc3686_create(struct crypto_template *tmpl,
tmpl              344 crypto/ctr.c   	err = skcipher_register_instance(tmpl, inst);
tmpl              328 crypto/cts.c   static int crypto_cts_create(struct crypto_template *tmpl, struct rtattr **tb)
tmpl              396 crypto/cts.c   	err = skcipher_register_instance(tmpl, inst);
tmpl               61 crypto/ecb.c   static int crypto_ecb_create(struct crypto_template *tmpl, struct rtattr **tb)
tmpl               67 crypto/ecb.c   	inst = skcipher_alloc_instance_simple(tmpl, tb, &alg);
tmpl               76 crypto/ecb.c   	err = skcipher_register_instance(tmpl, inst);
tmpl              112 crypto/echainiv.c static int echainiv_aead_create(struct crypto_template *tmpl,
tmpl              118 crypto/echainiv.c 	inst = aead_geniv_alloc(tmpl, tb, 0, 0);
tmpl              138 crypto/echainiv.c 	err = aead_register_instance(tmpl, inst);
tmpl              456 crypto/essiv.c static int essiv_create(struct crypto_template *tmpl, struct rtattr **tb)
tmpl              604 crypto/essiv.c 		err = skcipher_register_instance(tmpl, skcipher_inst);
tmpl              619 crypto/essiv.c 		err = aead_register_instance(tmpl, aead_inst);
tmpl              582 crypto/gcm.c   static int crypto_gcm_create_common(struct crypto_template *tmpl,
tmpl              675 crypto/gcm.c   	err = aead_register_instance(tmpl, inst);
tmpl              692 crypto/gcm.c   static int crypto_gcm_create(struct crypto_template *tmpl, struct rtattr **tb)
tmpl              705 crypto/gcm.c   	return crypto_gcm_create_common(tmpl, tb, ctr_name, "ghash");
tmpl              708 crypto/gcm.c   static int crypto_gcm_base_create(struct crypto_template *tmpl,
tmpl              722 crypto/gcm.c   	return crypto_gcm_create_common(tmpl, tb, ctr_name, ghash_name);
tmpl              866 crypto/gcm.c   static int crypto_rfc4106_create(struct crypto_template *tmpl,
tmpl              940 crypto/gcm.c   	err = aead_register_instance(tmpl, inst);
tmpl             1102 crypto/gcm.c   static int crypto_rfc4543_create(struct crypto_template *tmpl,
tmpl             1178 crypto/gcm.c   	err = aead_register_instance(tmpl, inst);
tmpl              170 crypto/hmac.c  static int hmac_create(struct crypto_template *tmpl, struct rtattr **tb)
tmpl              231 crypto/hmac.c  	err = shash_register_instance(tmpl, inst);
tmpl              100 crypto/internal.h static inline int crypto_tmpl_get(struct crypto_template *tmpl)
tmpl              102 crypto/internal.h 	return try_module_get(tmpl->module);
tmpl              105 crypto/internal.h static inline void crypto_tmpl_put(struct crypto_template *tmpl)
tmpl              107 crypto/internal.h 	module_put(tmpl->module);
tmpl              263 crypto/keywrap.c static int crypto_kw_create(struct crypto_template *tmpl, struct rtattr **tb)
tmpl              269 crypto/keywrap.c 	inst = skcipher_alloc_instance_simple(tmpl, tb, &alg);
tmpl              285 crypto/keywrap.c 	err = skcipher_register_instance(tmpl, inst);
tmpl              298 crypto/lrw.c   static int create(struct crypto_template *tmpl, struct rtattr **tb)
tmpl              406 crypto/lrw.c   	err = skcipher_register_instance(tmpl, inst);
tmpl               52 crypto/ofb.c   static int crypto_ofb_create(struct crypto_template *tmpl, struct rtattr **tb)
tmpl               58 crypto/ofb.c   	inst = skcipher_alloc_instance_simple(tmpl, tb, &alg);
tmpl               74 crypto/ofb.c   	err = skcipher_register_instance(tmpl, inst);
tmpl              153 crypto/pcbc.c  static int crypto_pcbc_create(struct crypto_template *tmpl, struct rtattr **tb)
tmpl              159 crypto/pcbc.c  	inst = skcipher_alloc_instance_simple(tmpl, tb, &alg);
tmpl              166 crypto/pcbc.c  	err = skcipher_register_instance(tmpl, inst);
tmpl              229 crypto/pcrypt.c static int pcrypt_create_aead(struct crypto_template *tmpl, struct rtattr **tb,
tmpl              290 crypto/pcrypt.c 	err = aead_register_instance(tmpl, inst);
tmpl              308 crypto/pcrypt.c static int pcrypt_create(struct crypto_template *tmpl, struct rtattr **tb)
tmpl              318 crypto/pcrypt.c 		return pcrypt_create_aead(tmpl, tb, algt->type, algt->mask);
tmpl              597 crypto/rsa-pkcs1pad.c static int pkcs1pad_create(struct crypto_template *tmpl, struct rtattr **tb)
tmpl              691 crypto/rsa-pkcs1pad.c 	err = akcipher_register_instance(tmpl, inst);
tmpl              138 crypto/seqiv.c static int seqiv_aead_create(struct crypto_template *tmpl, struct rtattr **tb)
tmpl              143 crypto/seqiv.c 	inst = aead_geniv_alloc(tmpl, tb, 0, 0);
tmpl              161 crypto/seqiv.c 	err = aead_register_instance(tmpl, inst);
tmpl              173 crypto/seqiv.c static int seqiv_create(struct crypto_template *tmpl, struct rtattr **tb)
tmpl              184 crypto/seqiv.c 	return seqiv_aead_create(tmpl, tb);
tmpl              540 crypto/shash.c int shash_register_instance(struct crypto_template *tmpl,
tmpl              549 crypto/shash.c 	return crypto_register_instance(tmpl, shash_crypto_instance(inst));
tmpl             1095 crypto/skcipher.c int skcipher_register_instance(struct crypto_template *tmpl,
tmpl             1104 crypto/skcipher.c 	return crypto_register_instance(tmpl, skcipher_crypto_instance(inst));
tmpl             1170 crypto/skcipher.c skcipher_alloc_instance_simple(struct crypto_template *tmpl, struct rtattr **tb,
tmpl             1202 crypto/skcipher.c 	err = crypto_inst_setname(skcipher_crypto_instance(inst), tmpl->name,
tmpl              620 crypto/vmac.c  static int vmac_create(struct crypto_template *tmpl, struct rtattr **tb)
tmpl              639 crypto/vmac.c  	inst = shash_alloc_instance(tmpl->name, alg);
tmpl              665 crypto/vmac.c  	err = shash_register_instance(tmpl, inst);
tmpl              188 crypto/xcbc.c  static int xcbc_create(struct crypto_template *tmpl, struct rtattr **tb)
tmpl              245 crypto/xcbc.c  	err = shash_register_instance(tmpl, inst);
tmpl               60 crypto/xor.c   do_xor_speed(struct xor_block_template *tmpl, void *b1, void *b2)
tmpl               66 crypto/xor.c   	tmpl->next = template_list;
tmpl               67 crypto/xor.c   	template_list = tmpl;
tmpl               84 crypto/xor.c   			tmpl->do_2(BENCH_SIZE, b1, b2);
tmpl               96 crypto/xor.c   	tmpl->speed = speed;
tmpl               98 crypto/xor.c   	printk(KERN_INFO "   %-10s: %5d.%03d MB/sec\n", tmpl->name,
tmpl              337 crypto/xts.c   static int create(struct crypto_template *tmpl, struct rtattr **tb)
tmpl              444 crypto/xts.c   	err = skcipher_register_instance(tmpl, inst);
tmpl               65 drivers/ata/libata-transport.c #define to_ata_internal(tmpl)	container_of(tmpl, struct ata_internal, t)
tmpl             2141 drivers/atm/idt77252.c 	unsigned long tmpl, modl;
tmpl             2162 drivers/atm/idt77252.c 	tmpl = (unsigned long) tcra * ((unsigned long) card->tst_size - 2);
tmpl             2163 drivers/atm/idt77252.c 	modl = tmpl % (unsigned long)card->utopia_pcr;
tmpl             2165 drivers/atm/idt77252.c 	tst_entries = (int) (tmpl / card->utopia_pcr);
tmpl             3252 drivers/atm/idt77252.c 	unsigned long tmpl, modl;
tmpl             3476 drivers/atm/idt77252.c 	tmpl = (unsigned long) rsvdcr * ((unsigned long) card->tst_size - 2);
tmpl             3477 drivers/atm/idt77252.c 	modl = tmpl % (unsigned long)card->utopia_pcr;
tmpl             3478 drivers/atm/idt77252.c 	tst_entries = (int) (tmpl / (unsigned long)card->utopia_pcr);
tmpl             1232 drivers/atm/nicstar.c 	unsigned long tmpl, modl;
tmpl             1294 drivers/atm/nicstar.c 			tmpl =
tmpl             1297 drivers/atm/nicstar.c 			modl = tmpl % card->max_pcr;
tmpl             1299 drivers/atm/nicstar.c 			n = (int)(tmpl / card->max_pcr);
tmpl             2619 drivers/crypto/ccree/cc_aead.c static struct cc_crypto_alg *cc_create_aead_alg(struct cc_alg_template *tmpl,
tmpl             2629 drivers/crypto/ccree/cc_aead.c 	alg = &tmpl->template_aead;
tmpl             2631 drivers/crypto/ccree/cc_aead.c 	snprintf(alg->base.cra_name, CRYPTO_MAX_ALG_NAME, "%s", tmpl->name);
tmpl             2633 drivers/crypto/ccree/cc_aead.c 		 tmpl->driver_name);
tmpl             2644 drivers/crypto/ccree/cc_aead.c 	t_alg->cipher_mode = tmpl->cipher_mode;
tmpl             2645 drivers/crypto/ccree/cc_aead.c 	t_alg->flow_mode = tmpl->flow_mode;
tmpl             2646 drivers/crypto/ccree/cc_aead.c 	t_alg->auth_mode = tmpl->auth_mode;
tmpl             1635 drivers/crypto/ccree/cc_cipher.c static struct cc_crypto_alg *cc_create_alg(const struct cc_alg_template *tmpl,
tmpl             1647 drivers/crypto/ccree/cc_cipher.c 	memcpy(alg, &tmpl->template_skcipher, sizeof(*alg));
tmpl             1649 drivers/crypto/ccree/cc_cipher.c 	snprintf(alg->base.cra_name, CRYPTO_MAX_ALG_NAME, "%s", tmpl->name);
tmpl             1651 drivers/crypto/ccree/cc_cipher.c 		 tmpl->driver_name);
tmpl             1654 drivers/crypto/ccree/cc_cipher.c 	alg->base.cra_blocksize = tmpl->blocksize;
tmpl             1662 drivers/crypto/ccree/cc_cipher.c 	t_alg->cipher_mode = tmpl->cipher_mode;
tmpl             1663 drivers/crypto/ccree/cc_cipher.c 	t_alg->flow_mode = tmpl->flow_mode;
tmpl             1664 drivers/crypto/ccree/cc_cipher.c 	t_alg->data_unit = tmpl->data_unit;
tmpl             1065 drivers/crypto/inside-secure/safexcel_cipher.c 	struct safexcel_alg_template *tmpl =
tmpl             1072 drivers/crypto/inside-secure/safexcel_cipher.c 	ctx->priv = tmpl->priv;
tmpl             1554 drivers/crypto/inside-secure/safexcel_cipher.c 	struct safexcel_alg_template *tmpl =
tmpl             1561 drivers/crypto/inside-secure/safexcel_cipher.c 	ctx->priv = tmpl->priv;
tmpl              788 drivers/crypto/inside-secure/safexcel_hash.c 	struct safexcel_alg_template *tmpl =
tmpl              792 drivers/crypto/inside-secure/safexcel_hash.c 	ctx->priv = tmpl->priv;
tmpl              411 drivers/crypto/marvell/cipher.c 				     struct mv_cesa_op_ctx *tmpl)
tmpl              432 drivers/crypto/marvell/cipher.c 	mv_cesa_update_op_cfg(tmpl, CESA_SA_DESC_CFG_OP_CRYPT_ONLY,
tmpl              436 drivers/crypto/marvell/cipher.c 		ret = mv_cesa_skcipher_dma_req_init(req, tmpl);
tmpl              438 drivers/crypto/marvell/cipher.c 		ret = mv_cesa_skcipher_std_req_init(req, tmpl);
tmpl              444 drivers/crypto/marvell/cipher.c 				      struct mv_cesa_op_ctx *tmpl)
tmpl              450 drivers/crypto/marvell/cipher.c 	ret = mv_cesa_skcipher_req_init(req, tmpl);
tmpl              466 drivers/crypto/marvell/cipher.c 			  struct mv_cesa_op_ctx *tmpl)
tmpl              470 drivers/crypto/marvell/cipher.c 	mv_cesa_update_op_cfg(tmpl, CESA_SA_DESC_CFG_CRYPTM_DES,
tmpl              473 drivers/crypto/marvell/cipher.c 	memcpy(tmpl->ctx.blkcipher.key, ctx->key, DES_KEY_SIZE);
tmpl              475 drivers/crypto/marvell/cipher.c 	return mv_cesa_skcipher_queue_req(req, tmpl);
tmpl              480 drivers/crypto/marvell/cipher.c 	struct mv_cesa_op_ctx tmpl;
tmpl              482 drivers/crypto/marvell/cipher.c 	mv_cesa_set_op_cfg(&tmpl,
tmpl              486 drivers/crypto/marvell/cipher.c 	return mv_cesa_des_op(req, &tmpl);
tmpl              491 drivers/crypto/marvell/cipher.c 	struct mv_cesa_op_ctx tmpl;
tmpl              493 drivers/crypto/marvell/cipher.c 	mv_cesa_set_op_cfg(&tmpl,
tmpl              497 drivers/crypto/marvell/cipher.c 	return mv_cesa_des_op(req, &tmpl);
tmpl              521 drivers/crypto/marvell/cipher.c 			      struct mv_cesa_op_ctx *tmpl)
tmpl              523 drivers/crypto/marvell/cipher.c 	mv_cesa_update_op_cfg(tmpl, CESA_SA_DESC_CFG_CRYPTCM_CBC,
tmpl              526 drivers/crypto/marvell/cipher.c 	memcpy(tmpl->ctx.blkcipher.iv, req->iv, DES_BLOCK_SIZE);
tmpl              528 drivers/crypto/marvell/cipher.c 	return mv_cesa_des_op(req, tmpl);
tmpl              533 drivers/crypto/marvell/cipher.c 	struct mv_cesa_op_ctx tmpl;
tmpl              535 drivers/crypto/marvell/cipher.c 	mv_cesa_set_op_cfg(&tmpl, CESA_SA_DESC_CFG_DIR_ENC);
tmpl              537 drivers/crypto/marvell/cipher.c 	return mv_cesa_cbc_des_op(req, &tmpl);
tmpl              542 drivers/crypto/marvell/cipher.c 	struct mv_cesa_op_ctx tmpl;
tmpl              544 drivers/crypto/marvell/cipher.c 	mv_cesa_set_op_cfg(&tmpl, CESA_SA_DESC_CFG_DIR_DEC);
tmpl              546 drivers/crypto/marvell/cipher.c 	return mv_cesa_cbc_des_op(req, &tmpl);
tmpl              571 drivers/crypto/marvell/cipher.c 			   struct mv_cesa_op_ctx *tmpl)
tmpl              575 drivers/crypto/marvell/cipher.c 	mv_cesa_update_op_cfg(tmpl, CESA_SA_DESC_CFG_CRYPTM_3DES,
tmpl              578 drivers/crypto/marvell/cipher.c 	memcpy(tmpl->ctx.blkcipher.key, ctx->key, DES3_EDE_KEY_SIZE);
tmpl              580 drivers/crypto/marvell/cipher.c 	return mv_cesa_skcipher_queue_req(req, tmpl);
tmpl              585 drivers/crypto/marvell/cipher.c 	struct mv_cesa_op_ctx tmpl;
tmpl              587 drivers/crypto/marvell/cipher.c 	mv_cesa_set_op_cfg(&tmpl,
tmpl              592 drivers/crypto/marvell/cipher.c 	return mv_cesa_des3_op(req, &tmpl);
tmpl              597 drivers/crypto/marvell/cipher.c 	struct mv_cesa_op_ctx tmpl;
tmpl              599 drivers/crypto/marvell/cipher.c 	mv_cesa_set_op_cfg(&tmpl,
tmpl              604 drivers/crypto/marvell/cipher.c 	return mv_cesa_des3_op(req, &tmpl);
tmpl              629 drivers/crypto/marvell/cipher.c 			       struct mv_cesa_op_ctx *tmpl)
tmpl              631 drivers/crypto/marvell/cipher.c 	memcpy(tmpl->ctx.blkcipher.iv, req->iv, DES3_EDE_BLOCK_SIZE);
tmpl              633 drivers/crypto/marvell/cipher.c 	return mv_cesa_des3_op(req, tmpl);
tmpl              638 drivers/crypto/marvell/cipher.c 	struct mv_cesa_op_ctx tmpl;
tmpl              640 drivers/crypto/marvell/cipher.c 	mv_cesa_set_op_cfg(&tmpl,
tmpl              645 drivers/crypto/marvell/cipher.c 	return mv_cesa_cbc_des3_op(req, &tmpl);
tmpl              650 drivers/crypto/marvell/cipher.c 	struct mv_cesa_op_ctx tmpl;
tmpl              652 drivers/crypto/marvell/cipher.c 	mv_cesa_set_op_cfg(&tmpl,
tmpl              657 drivers/crypto/marvell/cipher.c 	return mv_cesa_cbc_des3_op(req, &tmpl);
tmpl              682 drivers/crypto/marvell/cipher.c 			  struct mv_cesa_op_ctx *tmpl)
tmpl              691 drivers/crypto/marvell/cipher.c 	if (mv_cesa_get_op_cfg(tmpl) & CESA_SA_DESC_CFG_DIR_DEC)
tmpl              697 drivers/crypto/marvell/cipher.c 		tmpl->ctx.blkcipher.key[i] = cpu_to_le32(key[i]);
tmpl              704 drivers/crypto/marvell/cipher.c 	mv_cesa_update_op_cfg(tmpl, cfg,
tmpl              708 drivers/crypto/marvell/cipher.c 	return mv_cesa_skcipher_queue_req(req, tmpl);
tmpl              713 drivers/crypto/marvell/cipher.c 	struct mv_cesa_op_ctx tmpl;
tmpl              715 drivers/crypto/marvell/cipher.c 	mv_cesa_set_op_cfg(&tmpl,
tmpl              719 drivers/crypto/marvell/cipher.c 	return mv_cesa_aes_op(req, &tmpl);
tmpl              724 drivers/crypto/marvell/cipher.c 	struct mv_cesa_op_ctx tmpl;
tmpl              726 drivers/crypto/marvell/cipher.c 	mv_cesa_set_op_cfg(&tmpl,
tmpl              730 drivers/crypto/marvell/cipher.c 	return mv_cesa_aes_op(req, &tmpl);
tmpl              754 drivers/crypto/marvell/cipher.c 			      struct mv_cesa_op_ctx *tmpl)
tmpl              756 drivers/crypto/marvell/cipher.c 	mv_cesa_update_op_cfg(tmpl, CESA_SA_DESC_CFG_CRYPTCM_CBC,
tmpl              758 drivers/crypto/marvell/cipher.c 	memcpy(tmpl->ctx.blkcipher.iv, req->iv, AES_BLOCK_SIZE);
tmpl              760 drivers/crypto/marvell/cipher.c 	return mv_cesa_aes_op(req, tmpl);
tmpl              765 drivers/crypto/marvell/cipher.c 	struct mv_cesa_op_ctx tmpl;
tmpl              767 drivers/crypto/marvell/cipher.c 	mv_cesa_set_op_cfg(&tmpl, CESA_SA_DESC_CFG_DIR_ENC);
tmpl              769 drivers/crypto/marvell/cipher.c 	return mv_cesa_cbc_aes_op(req, &tmpl);
tmpl              774 drivers/crypto/marvell/cipher.c 	struct mv_cesa_op_ctx tmpl;
tmpl              776 drivers/crypto/marvell/cipher.c 	mv_cesa_set_op_cfg(&tmpl, CESA_SA_DESC_CFG_DIR_DEC);
tmpl              778 drivers/crypto/marvell/cipher.c 	return mv_cesa_cbc_aes_op(req, &tmpl);
tmpl              409 drivers/crypto/marvell/hash.c 			      struct mv_cesa_op_ctx *tmpl, bool algo_le)
tmpl              414 drivers/crypto/marvell/hash.c 	mv_cesa_update_op_cfg(tmpl,
tmpl              419 drivers/crypto/marvell/hash.c 	mv_cesa_set_mac_op_total_len(tmpl, 0);
tmpl              420 drivers/crypto/marvell/hash.c 	mv_cesa_set_mac_op_frag_len(tmpl, 0);
tmpl              421 drivers/crypto/marvell/hash.c 	creq->op_tmpl = *tmpl;
tmpl              460 drivers/crypto/marvell/hash.c 		     struct mv_cesa_op_ctx *tmpl, unsigned int frag_len,
tmpl              466 drivers/crypto/marvell/hash.c 	op = mv_cesa_dma_add_op(chain, tmpl, false, flags);
tmpl              478 drivers/crypto/marvell/hash.c 	if (mv_cesa_mac_op_is_first_frag(tmpl))
tmpl              479 drivers/crypto/marvell/hash.c 		mv_cesa_update_op_cfg(tmpl,
tmpl              785 drivers/crypto/marvell/hash.c 	struct mv_cesa_op_ctx *tmpl = &creq->op_tmpl;
tmpl              787 drivers/crypto/marvell/hash.c 	mv_cesa_set_mac_op_total_len(tmpl, creq->len);
tmpl              797 drivers/crypto/marvell/hash.c 	struct mv_cesa_op_ctx *tmpl = &creq->op_tmpl;
tmpl              800 drivers/crypto/marvell/hash.c 	mv_cesa_set_mac_op_total_len(tmpl, creq->len);
tmpl              861 drivers/crypto/marvell/hash.c 	struct mv_cesa_op_ctx tmpl = { };
tmpl              863 drivers/crypto/marvell/hash.c 	mv_cesa_set_op_cfg(&tmpl, CESA_SA_DESC_CFG_MACM_MD5);
tmpl              865 drivers/crypto/marvell/hash.c 	mv_cesa_ahash_init(req, &tmpl, true);
tmpl              930 drivers/crypto/marvell/hash.c 	struct mv_cesa_op_ctx tmpl = { };
tmpl              932 drivers/crypto/marvell/hash.c 	mv_cesa_set_op_cfg(&tmpl, CESA_SA_DESC_CFG_MACM_SHA1);
tmpl              934 drivers/crypto/marvell/hash.c 	mv_cesa_ahash_init(req, &tmpl, false);
tmpl             1000 drivers/crypto/marvell/hash.c 	struct mv_cesa_op_ctx tmpl = { };
tmpl             1002 drivers/crypto/marvell/hash.c 	mv_cesa_set_op_cfg(&tmpl, CESA_SA_DESC_CFG_MACM_SHA256);
tmpl             1004 drivers/crypto/marvell/hash.c 	mv_cesa_ahash_init(req, &tmpl, false);
tmpl             1237 drivers/crypto/marvell/hash.c 	struct mv_cesa_op_ctx tmpl = { };
tmpl             1239 drivers/crypto/marvell/hash.c 	mv_cesa_set_op_cfg(&tmpl, CESA_SA_DESC_CFG_MACM_HMAC_MD5);
tmpl             1240 drivers/crypto/marvell/hash.c 	memcpy(tmpl.ctx.hash.iv, ctx->iv, sizeof(ctx->iv));
tmpl             1242 drivers/crypto/marvell/hash.c 	mv_cesa_ahash_init(req, &tmpl, true);
tmpl             1307 drivers/crypto/marvell/hash.c 	struct mv_cesa_op_ctx tmpl = { };
tmpl             1309 drivers/crypto/marvell/hash.c 	mv_cesa_set_op_cfg(&tmpl, CESA_SA_DESC_CFG_MACM_HMAC_SHA1);
tmpl             1310 drivers/crypto/marvell/hash.c 	memcpy(tmpl.ctx.hash.iv, ctx->iv, sizeof(ctx->iv));
tmpl             1312 drivers/crypto/marvell/hash.c 	mv_cesa_ahash_init(req, &tmpl, false);
tmpl             1397 drivers/crypto/marvell/hash.c 	struct mv_cesa_op_ctx tmpl = { };
tmpl             1399 drivers/crypto/marvell/hash.c 	mv_cesa_set_op_cfg(&tmpl, CESA_SA_DESC_CFG_MACM_HMAC_SHA256);
tmpl             1400 drivers/crypto/marvell/hash.c 	memcpy(tmpl.ctx.hash.iv, ctx->iv, sizeof(ctx->iv));
tmpl             1402 drivers/crypto/marvell/hash.c 	mv_cesa_ahash_init(req, &tmpl, false);
tmpl             1374 drivers/crypto/n2_core.c static int __n2_register_one_cipher(const struct n2_cipher_tmpl *tmpl)
tmpl             1385 drivers/crypto/n2_core.c 	snprintf(alg->cra_name, CRYPTO_MAX_ALG_NAME, "%s", tmpl->name);
tmpl             1386 drivers/crypto/n2_core.c 	snprintf(alg->cra_driver_name, CRYPTO_MAX_ALG_NAME, "%s-n2", tmpl->drv_name);
tmpl             1390 drivers/crypto/n2_core.c 	alg->cra_blocksize = tmpl->block_size;
tmpl             1391 drivers/crypto/n2_core.c 	p->enc_type = tmpl->enc_type;
tmpl             1394 drivers/crypto/n2_core.c 	alg->cra_u.ablkcipher = tmpl->ablkcipher;
tmpl             1448 drivers/crypto/n2_core.c static int __n2_register_one_ahash(const struct n2_hash_tmpl *tmpl)
tmpl             1459 drivers/crypto/n2_core.c 	p->hash_zero = tmpl->hash_zero;
tmpl             1460 drivers/crypto/n2_core.c 	p->hash_init = tmpl->hash_init;
tmpl             1461 drivers/crypto/n2_core.c 	p->auth_type = tmpl->auth_type;
tmpl             1462 drivers/crypto/n2_core.c 	p->hmac_type = tmpl->hmac_type;
tmpl             1463 drivers/crypto/n2_core.c 	p->hw_op_hashsz = tmpl->hw_op_hashsz;
tmpl             1464 drivers/crypto/n2_core.c 	p->digest_size = tmpl->digest_size;
tmpl             1476 drivers/crypto/n2_core.c 	halg->digestsize = tmpl->digest_size;
tmpl             1479 drivers/crypto/n2_core.c 	snprintf(base->cra_name, CRYPTO_MAX_ALG_NAME, "%s", tmpl->name);
tmpl             1480 drivers/crypto/n2_core.c 	snprintf(base->cra_driver_name, CRYPTO_MAX_ALG_NAME, "%s-n2", tmpl->name);
tmpl             1484 drivers/crypto/n2_core.c 	base->cra_blocksize = tmpl->block_size;
tmpl               22 drivers/crypto/qce/ablkcipher.c 	struct qce_alg_template *tmpl = to_cipher_tmpl(async_req->tfm);
tmpl               23 drivers/crypto/qce/ablkcipher.c 	struct qce_device *qce = tmpl->qce;
tmpl               48 drivers/crypto/qce/ablkcipher.c 	qce->async_req_done(tmpl->qce, error);
tmpl               57 drivers/crypto/qce/ablkcipher.c 	struct qce_alg_template *tmpl = to_cipher_tmpl(async_req->tfm);
tmpl               58 drivers/crypto/qce/ablkcipher.c 	struct qce_device *qce = tmpl->qce;
tmpl              134 drivers/crypto/qce/ablkcipher.c 	ret = qce_start(async_req, tmpl->crypto_alg_type, req->nbytes, 0);
tmpl              216 drivers/crypto/qce/ablkcipher.c 	struct qce_alg_template *tmpl = to_cipher_tmpl(tfm);
tmpl              219 drivers/crypto/qce/ablkcipher.c 	rctx->flags = tmpl->alg_flags;
tmpl              237 drivers/crypto/qce/ablkcipher.c 	return tmpl->qce->async_req_enqueue(tmpl->qce, &req->base);
tmpl              357 drivers/crypto/qce/ablkcipher.c 	struct qce_alg_template *tmpl;
tmpl              361 drivers/crypto/qce/ablkcipher.c 	tmpl = kzalloc(sizeof(*tmpl), GFP_KERNEL);
tmpl              362 drivers/crypto/qce/ablkcipher.c 	if (!tmpl)
tmpl              365 drivers/crypto/qce/ablkcipher.c 	alg = &tmpl->alg.crypto;
tmpl              391 drivers/crypto/qce/ablkcipher.c 	INIT_LIST_HEAD(&tmpl->entry);
tmpl              392 drivers/crypto/qce/ablkcipher.c 	tmpl->crypto_alg_type = CRYPTO_ALG_TYPE_ABLKCIPHER;
tmpl              393 drivers/crypto/qce/ablkcipher.c 	tmpl->alg_flags = def->flags;
tmpl              394 drivers/crypto/qce/ablkcipher.c 	tmpl->qce = qce;
tmpl              398 drivers/crypto/qce/ablkcipher.c 		kfree(tmpl);
tmpl              403 drivers/crypto/qce/ablkcipher.c 	list_add_tail(&tmpl->entry, &ablkcipher_algs);
tmpl              410 drivers/crypto/qce/ablkcipher.c 	struct qce_alg_template *tmpl, *n;
tmpl              412 drivers/crypto/qce/ablkcipher.c 	list_for_each_entry_safe(tmpl, n, &ablkcipher_algs, entry) {
tmpl              413 drivers/crypto/qce/ablkcipher.c 		crypto_unregister_alg(&tmpl->alg.crypto);
tmpl              414 drivers/crypto/qce/ablkcipher.c 		list_del(&tmpl->entry);
tmpl              415 drivers/crypto/qce/ablkcipher.c 		kfree(tmpl);
tmpl              228 drivers/crypto/qce/common.c 	struct qce_alg_template *tmpl = to_ahash_tmpl(async_req->tfm);
tmpl              229 drivers/crypto/qce/common.c 	struct qce_device *qce = tmpl->qce;
tmpl              313 drivers/crypto/qce/common.c 	struct qce_alg_template *tmpl = to_cipher_tmpl(async_req->tfm);
tmpl              314 drivers/crypto/qce/common.c 	struct qce_device *qce = tmpl->qce;
tmpl               35 drivers/crypto/qce/sha.c 	struct qce_alg_template *tmpl = to_ahash_tmpl(async_req->tfm);
tmpl               36 drivers/crypto/qce/sha.c 	struct qce_device *qce = tmpl->qce;
tmpl               65 drivers/crypto/qce/sha.c 	qce->async_req_done(tmpl->qce, error);
tmpl               73 drivers/crypto/qce/sha.c 	struct qce_alg_template *tmpl = to_ahash_tmpl(async_req->tfm);
tmpl               74 drivers/crypto/qce/sha.c 	struct qce_device *qce = tmpl->qce;
tmpl              109 drivers/crypto/qce/sha.c 	ret = qce_start(async_req, tmpl->crypto_alg_type, 0, 0);
tmpl              127 drivers/crypto/qce/sha.c 	struct qce_alg_template *tmpl = to_ahash_tmpl(req->base.tfm);
tmpl              128 drivers/crypto/qce/sha.c 	const u32 *std_iv = tmpl->std_iv;
tmpl              133 drivers/crypto/qce/sha.c 	rctx->flags = tmpl->alg_flags;
tmpl              230 drivers/crypto/qce/sha.c 	struct qce_alg_template *tmpl = to_ahash_tmpl(req->base.tfm);
tmpl              231 drivers/crypto/qce/sha.c 	struct qce_device *qce = tmpl->qce;
tmpl              299 drivers/crypto/qce/sha.c 	return qce->async_req_enqueue(tmpl->qce, &req->base);
tmpl              305 drivers/crypto/qce/sha.c 	struct qce_alg_template *tmpl = to_ahash_tmpl(req->base.tfm);
tmpl              306 drivers/crypto/qce/sha.c 	struct qce_device *qce = tmpl->qce;
tmpl              322 drivers/crypto/qce/sha.c 	return qce->async_req_enqueue(tmpl->qce, &req->base);
tmpl              328 drivers/crypto/qce/sha.c 	struct qce_alg_template *tmpl = to_ahash_tmpl(req->base.tfm);
tmpl              329 drivers/crypto/qce/sha.c 	struct qce_device *qce = tmpl->qce;
tmpl              341 drivers/crypto/qce/sha.c 	return qce->async_req_enqueue(tmpl->qce, &req->base);
tmpl              472 drivers/crypto/qce/sha.c 	struct qce_alg_template *tmpl;
tmpl              477 drivers/crypto/qce/sha.c 	tmpl = kzalloc(sizeof(*tmpl), GFP_KERNEL);
tmpl              478 drivers/crypto/qce/sha.c 	if (!tmpl)
tmpl              481 drivers/crypto/qce/sha.c 	tmpl->std_iv = def->std_iv;
tmpl              483 drivers/crypto/qce/sha.c 	alg = &tmpl->alg.ahash;
tmpl              508 drivers/crypto/qce/sha.c 	INIT_LIST_HEAD(&tmpl->entry);
tmpl              509 drivers/crypto/qce/sha.c 	tmpl->crypto_alg_type = CRYPTO_ALG_TYPE_AHASH;
tmpl              510 drivers/crypto/qce/sha.c 	tmpl->alg_flags = def->flags;
tmpl              511 drivers/crypto/qce/sha.c 	tmpl->qce = qce;
tmpl              515 drivers/crypto/qce/sha.c 		kfree(tmpl);
tmpl              520 drivers/crypto/qce/sha.c 	list_add_tail(&tmpl->entry, &ahash_algs);
tmpl              527 drivers/crypto/qce/sha.c 	struct qce_alg_template *tmpl, *n;
tmpl              529 drivers/crypto/qce/sha.c 	list_for_each_entry_safe(tmpl, n, &ahash_algs, entry) {
tmpl              530 drivers/crypto/qce/sha.c 		crypto_unregister_ahash(&tmpl->alg.ahash);
tmpl              531 drivers/crypto/qce/sha.c 		list_del(&tmpl->entry);
tmpl              532 drivers/crypto/qce/sha.c 		kfree(tmpl);
tmpl              264 drivers/media/dvb-frontends/mn88443x.c 	u32 cpmon, tmpu, tmpl, flg;
tmpl              298 drivers/media/dvb-frontends/mn88443x.c 		regmap_read(r_s, CNRDXL_S, &tmpl);
tmpl              299 drivers/media/dvb-frontends/mn88443x.c 		x = (tmpu << 8) | tmpl;
tmpl              301 drivers/media/dvb-frontends/mn88443x.c 		regmap_read(r_s, CNRDYL_S, &tmpl);
tmpl              302 drivers/media/dvb-frontends/mn88443x.c 		y = (tmpu << 8) | tmpl;
tmpl              340 drivers/media/dvb-frontends/mn88443x.c 		regmap_read(r_s, BERVRDL_S, &tmpl);
tmpl              341 drivers/media/dvb-frontends/mn88443x.c 		bit_err = (tmpu << 8) | tmpl;
tmpl              466 drivers/media/dvb-frontends/mn88443x.c 	u32 seqrd, st, flg, tmpu, tmpm, tmpl;
tmpl              486 drivers/media/dvb-frontends/mn88443x.c 		regmap_read(r_t, AGCRDL_T, &tmpl);
tmpl              487 drivers/media/dvb-frontends/mn88443x.c 		agc = (tmpu << 8) | tmpl;
tmpl              501 drivers/media/dvb-frontends/mn88443x.c 		regmap_read(r_t, CNRDL_T, &tmpl);
tmpl              503 drivers/media/dvb-frontends/mn88443x.c 		if (tmpu || tmpl) {
tmpl              506 drivers/media/dvb-frontends/mn88443x.c 			tmp = (u64)80807124 - intlog10((tmpu << 8) | tmpl)
tmpl              529 drivers/media/dvb-frontends/mn88443x.c 		regmap_read(r_t, BERRDL_T, &tmpl);
tmpl              530 drivers/media/dvb-frontends/mn88443x.c 		bit_err = (tmpu << 16) | (tmpm << 8) | tmpl;
tmpl              533 drivers/media/dvb-frontends/mn88443x.c 		regmap_read(r_t, BERLENRDL_T, &tmpl);
tmpl              534 drivers/media/dvb-frontends/mn88443x.c 		bit_cnt = ((tmpu << 8) | tmpl) * 203 * 8;
tmpl             1828 drivers/net/bonding/bond_3ad.c 	static const struct port_params tmpl = {
tmpl             1859 drivers/net/bonding/bond_3ad.c 		memcpy(&port->partner_admin, &tmpl, sizeof(tmpl));
tmpl             1860 drivers/net/bonding/bond_3ad.c 		memcpy(&port->partner_oper, &tmpl, sizeof(tmpl));
tmpl             1189 drivers/net/ethernet/mellanox/mlxsw/spectrum_fid.c 					const struct mlxsw_sp_fid_family *tmpl)
tmpl             1191 drivers/net/ethernet/mellanox/mlxsw/spectrum_fid.c 	u16 nr_fids = tmpl->end_index - tmpl->start_index + 1;
tmpl             1195 drivers/net/ethernet/mellanox/mlxsw/spectrum_fid.c 	fid_family = kmemdup(tmpl, sizeof(*fid_family), GFP_KERNEL);
tmpl             1213 drivers/net/ethernet/mellanox/mlxsw/spectrum_fid.c 	mlxsw_sp->fid_core->fid_family_arr[tmpl->type] = fid_family;
tmpl             1220 drivers/net/wireless/ti/wlcore/cmd.c 	struct wl12xx_arp_rsp_template *tmpl;
tmpl             1224 drivers/net/wireless/ti/wlcore/cmd.c 	skb = dev_alloc_skb(sizeof(*hdr) + sizeof(__le16) + sizeof(*tmpl) +
tmpl             1233 drivers/net/wireless/ti/wlcore/cmd.c 	tmpl = skb_put_zero(skb, sizeof(*tmpl));
tmpl             1236 drivers/net/wireless/ti/wlcore/cmd.c 	memcpy(tmpl->llc_hdr, rfc1042_header, sizeof(rfc1042_header));
tmpl             1237 drivers/net/wireless/ti/wlcore/cmd.c 	tmpl->llc_type = cpu_to_be16(ETH_P_ARP);
tmpl             1240 drivers/net/wireless/ti/wlcore/cmd.c 	arp_hdr = &tmpl->arp_hdr;
tmpl             1248 drivers/net/wireless/ti/wlcore/cmd.c 	memcpy(tmpl->sender_hw, vif->addr, ETH_ALEN);
tmpl             1249 drivers/net/wireless/ti/wlcore/cmd.c 	tmpl->sender_ip = wlvif->ip_addr;
tmpl              147 drivers/net/wireless/ti/wlcore/init.c 	struct wl12xx_disconn_template *tmpl;
tmpl              151 drivers/net/wireless/ti/wlcore/init.c 	tmpl = kzalloc(sizeof(*tmpl), GFP_KERNEL);
tmpl              152 drivers/net/wireless/ti/wlcore/init.c 	if (!tmpl) {
tmpl              157 drivers/net/wireless/ti/wlcore/init.c 	tmpl->header.frame_ctl = cpu_to_le16(IEEE80211_FTYPE_MGMT |
tmpl              163 drivers/net/wireless/ti/wlcore/init.c 				      tmpl, sizeof(*tmpl), 0, rate);
tmpl              166 drivers/net/wireless/ti/wlcore/init.c 	kfree(tmpl);
tmpl               18 drivers/s390/net/fsm.c 		int nr_events, const fsm_node *tmpl, int tmpl_len, gfp_t order)
tmpl               57 drivers/s390/net/fsm.c 		if ((tmpl[i].cond_state >= nr_states) ||
tmpl               58 drivers/s390/net/fsm.c 		    (tmpl[i].cond_event >= nr_events)   ) {
tmpl               61 drivers/s390/net/fsm.c 				name, i, (long)tmpl[i].cond_state, (long)f->nr_states,
tmpl               62 drivers/s390/net/fsm.c 				(long)tmpl[i].cond_event, (long)f->nr_events);
tmpl               66 drivers/s390/net/fsm.c 			m[nr_states * tmpl[i].cond_event + tmpl[i].cond_state] =
tmpl               67 drivers/s390/net/fsm.c 				tmpl[i].function;
tmpl              112 drivers/s390/net/fsm.h 	 int nr_states, int nr_events, const fsm_node *tmpl,
tmpl               28 drivers/scsi/fnic/cq_exch_desc.h 	u8  tmpl;
tmpl               62 drivers/scsi/fnic/cq_exch_desc.h 	u16 tmpl;
tmpl               95 drivers/scsi/fnic/cq_exch_desc.h 				      u16 *tmpl,
tmpl              116 drivers/scsi/fnic/cq_exch_desc.h 	*tmpl = desc_ptr->tmpl & CQ_FCP_RQ_DESC_TMPL_MASK;
tmpl              141 drivers/scsi/fnic/cq_exch_desc.h 	u16 tmpl;
tmpl              169 drivers/scsi/fnic/cq_exch_desc.h 				   u16 *tmpl,
tmpl              178 drivers/scsi/fnic/cq_exch_desc.h 	*tmpl = desc_ptr->tmpl & CQ_SGL_TMPL_MASK;
tmpl              849 drivers/scsi/fnic/fnic_fcs.c 	u16 exchange_id, tmpl;
tmpl              866 drivers/scsi/fnic/fnic_fcs.c 				   &tmpl, &fcp_bytes_written, &sof, &eof,
tmpl              677 drivers/scsi/qla2xxx/qla_nvme.c 	struct nvme_fc_port_template *tmpl;
tmpl              686 drivers/scsi/qla2xxx/qla_nvme.c 	tmpl = &qla_nvme_fc_transport;
tmpl              705 drivers/scsi/qla2xxx/qla_nvme.c 	ret = nvme_fc_register_localport(&pinfo, tmpl,
tmpl               38 drivers/scsi/raid_class.c #define to_raid_internal(tmpl)	container_of(tmpl, struct raid_internal, r)
tmpl               41 drivers/scsi/scsi_sas_internal.h #define to_sas_internal(tmpl)	container_of(tmpl, struct sas_internal, t)
tmpl              343 drivers/scsi/scsi_transport_fc.c #define to_fc_internal(tmpl)	container_of(tmpl, struct fc_internal, t)
tmpl              101 drivers/scsi/scsi_transport_iscsi.c #define to_iscsi_internal(tmpl) \
tmpl              102 drivers/scsi/scsi_transport_iscsi.c 	container_of(tmpl, struct iscsi_internal, t)
tmpl               64 drivers/scsi/scsi_transport_spi.c #define to_spi_internal(tmpl)	container_of(tmpl, struct spi_internal, t)
tmpl               42 drivers/scsi/scsi_transport_srp.c #define to_srp_internal(tmpl) container_of(tmpl, struct srp_internal, t)
tmpl               49 include/crypto/algapi.h 	struct crypto_template *tmpl;
tmpl               62 include/crypto/algapi.h 	int (*create)(struct crypto_template *tmpl, struct rtattr **tb);
tmpl              140 include/crypto/algapi.h int crypto_register_template(struct crypto_template *tmpl);
tmpl              142 include/crypto/algapi.h void crypto_unregister_template(struct crypto_template *tmpl);
tmpl              146 include/crypto/algapi.h int crypto_register_instance(struct crypto_template *tmpl,
tmpl              182 include/crypto/internal/aead.h int aead_register_instance(struct crypto_template *tmpl,
tmpl              137 include/crypto/internal/akcipher.h int akcipher_register_instance(struct crypto_template *tmpl,
tmpl               22 include/crypto/internal/geniv.h struct aead_instance *aead_geniv_alloc(struct crypto_template *tmpl,
tmpl               76 include/crypto/internal/hash.h int ahash_register_instance(struct crypto_template *tmpl,
tmpl              105 include/crypto/internal/hash.h int shash_register_instance(struct crypto_template *tmpl,
tmpl              133 include/crypto/internal/skcipher.h int skcipher_register_instance(struct crypto_template *tmpl,
tmpl              250 include/crypto/internal/skcipher.h skcipher_alloc_instance_simple(struct crypto_template *tmpl, struct rtattr **tb,
tmpl              221 include/linux/sunrpc/cache.h extern struct cache_detail *cache_create_net(const struct cache_detail *tmpl, struct net *net);
tmpl               47 include/net/fib_notifier.h fib_notifier_ops_register(const struct fib_notifier_ops *tmpl, struct net *net);
tmpl              312 include/net/netfilter/nf_conntrack.h void nf_ct_tmpl_free(struct nf_conn *tmpl);
tmpl              115 include/net/netfilter/nf_conntrack_helper.h int __nf_ct_try_assign_helper(struct nf_conn *ct, struct nf_conn *tmpl,
tmpl               78 include/net/netfilter/nf_conntrack_l4proto.h int nf_conntrack_inet_error(struct nf_conn *tmpl, struct sk_buff *skb,
tmpl               84 include/net/netfilter/nf_conntrack_l4proto.h int nf_conntrack_icmpv4_error(struct nf_conn *tmpl,
tmpl               89 include/net/netfilter/nf_conntrack_l4proto.h int nf_conntrack_icmpv6_error(struct nf_conn *tmpl,
tmpl               33 include/net/netfilter/nf_conntrack_synproxy.h 				      const struct nf_conn *tmpl)
tmpl               36 include/net/netfilter/nf_conntrack_synproxy.h 	if (tmpl && nfct_synproxy(tmpl)) {
tmpl               29 include/net/netfilter/nf_conntrack_zones.h nf_ct_zone_tmpl(const struct nf_conn *tmpl, const struct sk_buff *skb,
tmpl               33 include/net/netfilter/nf_conntrack_zones.h 	if (!tmpl)
tmpl               36 include/net/netfilter/nf_conntrack_zones.h 	if (tmpl->zone.flags & NF_CT_FLAG_MARK)
tmpl               37 include/net/netfilter/nf_conntrack_zones.h 		return nf_ct_zone_init(tmp, skb->mark, tmpl->zone.dir, 0);
tmpl               39 include/net/netfilter/nf_conntrack_zones.h 	return nf_ct_zone(tmpl);
tmpl               23 include/net/netfilter/nf_synproxy.h 	struct nf_conn			*tmpl;
tmpl              563 include/net/netfilter/nf_tables.h static inline void nft_set_ext_prepare(struct nft_set_ext_tmpl *tmpl)
tmpl              565 include/net/netfilter/nf_tables.h 	memset(tmpl, 0, sizeof(*tmpl));
tmpl              566 include/net/netfilter/nf_tables.h 	tmpl->len = sizeof(struct nft_set_ext);
tmpl              569 include/net/netfilter/nf_tables.h static inline void nft_set_ext_add_length(struct nft_set_ext_tmpl *tmpl, u8 id,
tmpl              572 include/net/netfilter/nf_tables.h 	tmpl->len	 = ALIGN(tmpl->len, nft_set_ext_types[id].align);
tmpl              573 include/net/netfilter/nf_tables.h 	BUG_ON(tmpl->len > U8_MAX);
tmpl              574 include/net/netfilter/nf_tables.h 	tmpl->offset[id] = tmpl->len;
tmpl              575 include/net/netfilter/nf_tables.h 	tmpl->len	+= nft_set_ext_types[id].len + len;
tmpl              578 include/net/netfilter/nf_tables.h static inline void nft_set_ext_add(struct nft_set_ext_tmpl *tmpl, u8 id)
tmpl              580 include/net/netfilter/nf_tables.h 	nft_set_ext_add_length(tmpl, id, 0);
tmpl              584 include/net/netfilter/nf_tables.h 				    const struct nft_set_ext_tmpl *tmpl)
tmpl              586 include/net/netfilter/nf_tables.h 	memcpy(ext->offset, tmpl->offset, sizeof(ext->offset));
tmpl              657 include/net/netfilter/nf_tables.h 			const struct nft_set_ext_tmpl *tmpl,
tmpl               13 include/net/tc_act/tc_ct.h 	struct nf_conn *tmpl;
tmpl             1060 include/net/xfrm.h __xfrm4_state_addr_cmp(const struct xfrm_tmpl *tmpl, const struct xfrm_state *x)
tmpl             1062 include/net/xfrm.h 	return	(tmpl->saddr.a4 &&
tmpl             1063 include/net/xfrm.h 		 tmpl->saddr.a4 != x->props.saddr.a4);
tmpl             1067 include/net/xfrm.h __xfrm6_state_addr_cmp(const struct xfrm_tmpl *tmpl, const struct xfrm_state *x)
tmpl             1069 include/net/xfrm.h 	return	(!ipv6_addr_any((struct in6_addr*)&tmpl->saddr) &&
tmpl             1070 include/net/xfrm.h 		 !ipv6_addr_equal((struct in6_addr *)&tmpl->saddr, (struct in6_addr*)&x->props.saddr));
tmpl             1074 include/net/xfrm.h xfrm_state_addr_cmp(const struct xfrm_tmpl *tmpl, const struct xfrm_state *x, unsigned short family)
tmpl             1078 include/net/xfrm.h 		return __xfrm4_state_addr_cmp(tmpl, x);
tmpl             1080 include/net/xfrm.h 		return __xfrm6_state_addr_cmp(tmpl, x);
tmpl             1481 include/net/xfrm.h 				   struct xfrm_tmpl *tmpl,
tmpl              151 net/core/fib_notifier.c fib_notifier_ops_register(const struct fib_notifier_ops *tmpl, struct net *net)
tmpl              156 net/core/fib_notifier.c 	ops = kmemdup(tmpl, sizeof(*ops), GFP_KERNEL);
tmpl              150 net/core/fib_rules.c fib_rules_register(const struct fib_rules_ops *tmpl, struct net *net)
tmpl              155 net/core/fib_rules.c 	ops = kmemdup(tmpl, sizeof(*ops), GFP_KERNEL);
tmpl              545 net/netfilter/nf_conntrack_core.c 	struct nf_conn *tmpl, *p;
tmpl              548 net/netfilter/nf_conntrack_core.c 		tmpl = kzalloc(sizeof(*tmpl) + NFCT_INFOMASK, flags);
tmpl              549 net/netfilter/nf_conntrack_core.c 		if (!tmpl)
tmpl              552 net/netfilter/nf_conntrack_core.c 		p = tmpl;
tmpl              553 net/netfilter/nf_conntrack_core.c 		tmpl = (struct nf_conn *)NFCT_ALIGN((unsigned long)p);
tmpl              554 net/netfilter/nf_conntrack_core.c 		if (tmpl != p) {
tmpl              555 net/netfilter/nf_conntrack_core.c 			tmpl = (struct nf_conn *)NFCT_ALIGN((unsigned long)p);
tmpl              556 net/netfilter/nf_conntrack_core.c 			tmpl->proto.tmpl_padto = (char *)tmpl - (char *)p;
tmpl              559 net/netfilter/nf_conntrack_core.c 		tmpl = kzalloc(sizeof(*tmpl), flags);
tmpl              560 net/netfilter/nf_conntrack_core.c 		if (!tmpl)
tmpl              564 net/netfilter/nf_conntrack_core.c 	tmpl->status = IPS_TEMPLATE;
tmpl              565 net/netfilter/nf_conntrack_core.c 	write_pnet(&tmpl->ct_net, net);
tmpl              566 net/netfilter/nf_conntrack_core.c 	nf_ct_zone_add(tmpl, zone);
tmpl              567 net/netfilter/nf_conntrack_core.c 	atomic_set(&tmpl->ct_general.use, 0);
tmpl              569 net/netfilter/nf_conntrack_core.c 	return tmpl;
tmpl              573 net/netfilter/nf_conntrack_core.c void nf_ct_tmpl_free(struct nf_conn *tmpl)
tmpl              575 net/netfilter/nf_conntrack_core.c 	nf_ct_ext_destroy(tmpl);
tmpl              576 net/netfilter/nf_conntrack_core.c 	nf_ct_ext_free(tmpl);
tmpl              579 net/netfilter/nf_conntrack_core.c 		kfree((char *)tmpl - tmpl->proto.tmpl_padto);
tmpl              581 net/netfilter/nf_conntrack_core.c 		kfree(tmpl);
tmpl             1431 net/netfilter/nf_conntrack_core.c init_conntrack(struct net *net, struct nf_conn *tmpl,
tmpl             1450 net/netfilter/nf_conntrack_core.c 	zone = nf_ct_zone_tmpl(tmpl, skb, &tmp);
tmpl             1456 net/netfilter/nf_conntrack_core.c 	if (!nf_ct_add_synproxy(ct, tmpl)) {
tmpl             1461 net/netfilter/nf_conntrack_core.c 	timeout_ext = tmpl ? nf_ct_timeout_find(tmpl) : NULL;
tmpl             1471 net/netfilter/nf_conntrack_core.c 	ecache = tmpl ? nf_ct_ecache_find(tmpl) : NULL;
tmpl             1504 net/netfilter/nf_conntrack_core.c 		__nf_ct_try_assign_helper(ct, tmpl, GFP_ATOMIC);
tmpl             1523 net/netfilter/nf_conntrack_core.c resolve_normal_ct(struct nf_conn *tmpl,
tmpl             1545 net/netfilter/nf_conntrack_core.c 	zone = nf_ct_zone_tmpl(tmpl, skb, &tmp);
tmpl             1549 net/netfilter/nf_conntrack_core.c 		h = init_conntrack(state->net, tmpl, &tuple,
tmpl             1586 net/netfilter/nf_conntrack_core.c nf_conntrack_handle_icmp(struct nf_conn *tmpl,
tmpl             1595 net/netfilter/nf_conntrack_core.c 		ret = nf_conntrack_icmpv4_error(tmpl, skb, dataoff, state);
tmpl             1598 net/netfilter/nf_conntrack_core.c 		ret = nf_conntrack_icmpv6_error(tmpl, skb, dataoff, state);
tmpl             1672 net/netfilter/nf_conntrack_core.c 	struct nf_conn *ct, *tmpl;
tmpl             1676 net/netfilter/nf_conntrack_core.c 	tmpl = nf_ct_get(skb, &ctinfo);
tmpl             1677 net/netfilter/nf_conntrack_core.c 	if (tmpl || ctinfo == IP_CT_UNTRACKED) {
tmpl             1679 net/netfilter/nf_conntrack_core.c 		if ((tmpl && !nf_ct_is_template(tmpl)) ||
tmpl             1698 net/netfilter/nf_conntrack_core.c 		ret = nf_conntrack_handle_icmp(tmpl, skb, dataoff,
tmpl             1709 net/netfilter/nf_conntrack_core.c 	ret = resolve_normal_ct(tmpl, skb, dataoff,
tmpl             1750 net/netfilter/nf_conntrack_core.c 	if (tmpl)
tmpl             1751 net/netfilter/nf_conntrack_core.c 		nf_ct_put(tmpl);
tmpl              232 net/netfilter/nf_conntrack_helper.c int __nf_ct_try_assign_helper(struct nf_conn *ct, struct nf_conn *tmpl,
tmpl              248 net/netfilter/nf_conntrack_helper.c 	if (tmpl != NULL) {
tmpl              249 net/netfilter/nf_conntrack_helper.c 		help = nfct_help(tmpl);
tmpl              187 net/netfilter/nf_conntrack_proto.c 		struct nf_conn *tmpl;
tmpl              189 net/netfilter/nf_conntrack_proto.c 		tmpl = nf_ct_get(skb, &ctinfo);
tmpl              190 net/netfilter/nf_conntrack_proto.c 		if (tmpl && nf_ct_is_template(tmpl)) {
tmpl              195 net/netfilter/nf_conntrack_proto.c 			nf_ct_put(tmpl);
tmpl              104 net/netfilter/nf_conntrack_proto_icmp.c int nf_conntrack_inet_error(struct nf_conn *tmpl, struct sk_buff *skb,
tmpl              119 net/netfilter/nf_conntrack_proto_icmp.c 	zone = nf_ct_zone_tmpl(tmpl, skb, &tmp);
tmpl              203 net/netfilter/nf_conntrack_proto_icmp.c int nf_conntrack_icmpv4_error(struct nf_conn *tmpl,
tmpl              249 net/netfilter/nf_conntrack_proto_icmp.c 	return nf_conntrack_inet_error(tmpl, skb, dataoff, state,
tmpl              131 net/netfilter/nf_conntrack_proto_icmpv6.c int nf_conntrack_icmpv6_error(struct nf_conn *tmpl,
tmpl              168 net/netfilter/nf_conntrack_proto_icmpv6.c 	return nf_conntrack_inet_error(tmpl, skb, dataoff, state,
tmpl              348 net/netfilter/nf_synproxy_core.c 	snet->tmpl = ct;
tmpl              372 net/netfilter/nf_synproxy_core.c 	nf_ct_put(snet->tmpl);
tmpl              548 net/netfilter/nf_synproxy_core.c 	synproxy_send_tcp(net, skb, nskb, &snet->tmpl->ct_general, IP_CT_NEW,
tmpl              963 net/netfilter/nf_synproxy_core.c 	synproxy_send_tcp_ipv6(net, skb, nskb, &snet->tmpl->ct_general,
tmpl             4422 net/netfilter/nf_tables_api.c 			const struct nft_set_ext_tmpl *tmpl,
tmpl             4429 net/netfilter/nf_tables_api.c 	elem = kzalloc(set->ops->elemsize + tmpl->len, gfp);
tmpl             4434 net/netfilter/nf_tables_api.c 	nft_set_ext_init(ext, tmpl);
tmpl             4497 net/netfilter/nf_tables_api.c 	struct nft_set_ext_tmpl tmpl;
tmpl             4520 net/netfilter/nf_tables_api.c 	nft_set_ext_prepare(&tmpl);
tmpl             4526 net/netfilter/nf_tables_api.c 		nft_set_ext_add(&tmpl, NFT_SET_EXT_FLAGS);
tmpl             4576 net/netfilter/nf_tables_api.c 	nft_set_ext_add_length(&tmpl, NFT_SET_EXT_KEY, d1.len);
tmpl             4578 net/netfilter/nf_tables_api.c 		nft_set_ext_add(&tmpl, NFT_SET_EXT_EXPIRATION);
tmpl             4580 net/netfilter/nf_tables_api.c 			nft_set_ext_add(&tmpl, NFT_SET_EXT_TIMEOUT);
tmpl             4595 net/netfilter/nf_tables_api.c 		nft_set_ext_add(&tmpl, NFT_SET_EXT_OBJREF);
tmpl             4633 net/netfilter/nf_tables_api.c 		nft_set_ext_add_length(&tmpl, NFT_SET_EXT_DATA, d2.len);
tmpl             4644 net/netfilter/nf_tables_api.c 			nft_set_ext_add_length(&tmpl, NFT_SET_EXT_USERDATA,
tmpl             4649 net/netfilter/nf_tables_api.c 	elem.priv = nft_set_elem_init(set, &tmpl, elem.key.val.data, data.data,
tmpl             4813 net/netfilter/nf_tables_api.c 	struct nft_set_ext_tmpl tmpl;
tmpl             4831 net/netfilter/nf_tables_api.c 	nft_set_ext_prepare(&tmpl);
tmpl             4837 net/netfilter/nf_tables_api.c 		nft_set_ext_add(&tmpl, NFT_SET_EXT_FLAGS);
tmpl             4848 net/netfilter/nf_tables_api.c 	nft_set_ext_add_length(&tmpl, NFT_SET_EXT_KEY, desc.len);
tmpl             4851 net/netfilter/nf_tables_api.c 	elem.priv = nft_set_elem_init(set, &tmpl, elem.key.val.data, NULL, 0,
tmpl               17 net/netfilter/nft_dynset.c 	struct nft_set_ext_tmpl		tmpl;
tmpl               56 net/netfilter/nft_dynset.c 	elem = nft_set_elem_init(set, &priv->tmpl,
tmpl              226 net/netfilter/nft_dynset.c 	nft_set_ext_prepare(&priv->tmpl);
tmpl              227 net/netfilter/nft_dynset.c 	nft_set_ext_add_length(&priv->tmpl, NFT_SET_EXT_KEY, set->klen);
tmpl              229 net/netfilter/nft_dynset.c 		nft_set_ext_add_length(&priv->tmpl, NFT_SET_EXT_DATA, set->dlen);
tmpl              231 net/netfilter/nft_dynset.c 		nft_set_ext_add_length(&priv->tmpl, NFT_SET_EXT_EXPR,
tmpl              235 net/netfilter/nft_dynset.c 			nft_set_ext_add(&priv->tmpl, NFT_SET_EXT_EXPIRATION);
tmpl              958 net/openvswitch/conntrack.c 		struct nf_conn *tmpl = info->ct;
tmpl              962 net/openvswitch/conntrack.c 		if (tmpl) {
tmpl              965 net/openvswitch/conntrack.c 			nf_conntrack_get(&tmpl->ct_general);
tmpl              966 net/openvswitch/conntrack.c 			nf_ct_set(skb, tmpl, IP_CT_NEW);
tmpl              210 net/sched/act_ct.c 	if (params->tmpl)
tmpl              211 net/sched/act_ct.c 		nf_conntrack_put(&params->tmpl->ct_general);
tmpl              386 net/sched/act_ct.c 	struct nf_conn *tmpl = NULL;
tmpl              399 net/sched/act_ct.c 	tmpl = p->tmpl;
tmpl              440 net/sched/act_ct.c 		if (tmpl) {
tmpl              444 net/sched/act_ct.c 			nf_conntrack_get(&tmpl->ct_general);
tmpl              445 net/sched/act_ct.c 			nf_ct_set(skb, tmpl, IP_CT_NEW);
tmpl              595 net/sched/act_ct.c 	struct nf_conn *tmpl;
tmpl              655 net/sched/act_ct.c 	tmpl = nf_ct_tmpl_alloc(net, &zone, GFP_KERNEL);
tmpl              656 net/sched/act_ct.c 	if (!tmpl) {
tmpl              660 net/sched/act_ct.c 	__set_bit(IPS_CONFIRMED_BIT, &tmpl->status);
tmpl              661 net/sched/act_ct.c 	nf_conntrack_get(&tmpl->ct_general);
tmpl              662 net/sched/act_ct.c 	p->tmpl = tmpl;
tmpl             1717 net/sunrpc/cache.c struct cache_detail *cache_create_net(const struct cache_detail *tmpl, struct net *net)
tmpl             1722 net/sunrpc/cache.c 	cd = kmemdup(tmpl, sizeof(struct cache_detail), GFP_KERNEL);
tmpl             2372 net/xfrm/xfrm_policy.c 		struct xfrm_tmpl *tmpl = &policy->xfrm_vec[i];
tmpl             2374 net/xfrm/xfrm_policy.c 		if (tmpl->mode == XFRM_MODE_TUNNEL ||
tmpl             2375 net/xfrm/xfrm_policy.c 		    tmpl->mode == XFRM_MODE_BEET) {
tmpl             2376 net/xfrm/xfrm_policy.c 			remote = &tmpl->id.daddr;
tmpl             2377 net/xfrm/xfrm_policy.c 			local = &tmpl->saddr;
tmpl             2378 net/xfrm/xfrm_policy.c 			if (xfrm_addr_any(local, tmpl->encap_family)) {
tmpl             2381 net/xfrm/xfrm_policy.c 						       tmpl->encap_family, 0);
tmpl             2388 net/xfrm/xfrm_policy.c 		x = xfrm_state_find(remote, local, fl, tmpl, policy, &error,
tmpl             2405 net/xfrm/xfrm_policy.c 		if (!tmpl->optional)
tmpl             3217 net/xfrm/xfrm_policy.c xfrm_state_ok(const struct xfrm_tmpl *tmpl, const struct xfrm_state *x,
tmpl             3221 net/xfrm/xfrm_policy.c 		return tmpl->optional && !xfrm_state_addr_cmp(tmpl, x, tmpl->encap_family);
tmpl             3222 net/xfrm/xfrm_policy.c 	return	x->id.proto == tmpl->id.proto &&
tmpl             3223 net/xfrm/xfrm_policy.c 		(x->id.spi == tmpl->id.spi || !tmpl->id.spi) &&
tmpl             3224 net/xfrm/xfrm_policy.c 		(x->props.reqid == tmpl->reqid || !tmpl->reqid) &&
tmpl             3225 net/xfrm/xfrm_policy.c 		x->props.mode == tmpl->mode &&
tmpl             3226 net/xfrm/xfrm_policy.c 		(tmpl->allalgs || (tmpl->aalgos & (1<<x->props.aalgo)) ||
tmpl             3227 net/xfrm/xfrm_policy.c 		 !(xfrm_id_proto_match(tmpl->id.proto, IPSEC_PROTO_ANY))) &&
tmpl             3229 net/xfrm/xfrm_policy.c 		  xfrm_state_addr_cmp(tmpl, x, family));
tmpl             3240 net/xfrm/xfrm_policy.c xfrm_policy_ok(const struct xfrm_tmpl *tmpl, const struct sec_path *sp, int start,
tmpl             3245 net/xfrm/xfrm_policy.c 	if (tmpl->optional) {
tmpl             3246 net/xfrm/xfrm_policy.c 		if (tmpl->mode == XFRM_MODE_TRANSPORT)
tmpl             3251 net/xfrm/xfrm_policy.c 		if (xfrm_state_ok(tmpl, sp->xvec[idx], family))
tmpl              890 net/xfrm/xfrm_state.c 		    const struct xfrm_tmpl *tmpl,
tmpl              903 net/xfrm/xfrm_state.c 	x->id = tmpl->id;
tmpl              905 net/xfrm/xfrm_state.c 	switch (tmpl->encap_family) {
tmpl              909 net/xfrm/xfrm_state.c 		x->props.saddr = tmpl->saddr;
tmpl              916 net/xfrm/xfrm_state.c 		memcpy(&x->props.saddr, &tmpl->saddr, sizeof(x->props.saddr));
tmpl              922 net/xfrm/xfrm_state.c 	x->props.mode = tmpl->mode;
tmpl              923 net/xfrm/xfrm_state.c 	x->props.reqid = tmpl->reqid;
tmpl              924 net/xfrm/xfrm_state.c 	x->props.family = tmpl->encap_family;
tmpl             1040 net/xfrm/xfrm_state.c 		const struct flowi *fl, struct xfrm_tmpl *tmpl,
tmpl             1052 net/xfrm/xfrm_state.c 	unsigned short encap_family = tmpl->encap_family;
tmpl             1061 net/xfrm/xfrm_state.c 	h = xfrm_dst_hash(net, daddr, saddr, tmpl->reqid, encap_family);
tmpl             1064 net/xfrm/xfrm_state.c 		    x->props.reqid == tmpl->reqid &&
tmpl             1069 net/xfrm/xfrm_state.c 		    tmpl->mode == x->props.mode &&
tmpl             1070 net/xfrm/xfrm_state.c 		    tmpl->id.proto == x->id.proto &&
tmpl             1071 net/xfrm/xfrm_state.c 		    (tmpl->id.spi == x->id.spi || !tmpl->id.spi))
tmpl             1078 net/xfrm/xfrm_state.c 	h_wildcard = xfrm_dst_hash(net, daddr, &saddr_wildcard, tmpl->reqid, encap_family);
tmpl             1081 net/xfrm/xfrm_state.c 		    x->props.reqid == tmpl->reqid &&
tmpl             1086 net/xfrm/xfrm_state.c 		    tmpl->mode == x->props.mode &&
tmpl             1087 net/xfrm/xfrm_state.c 		    tmpl->id.proto == x->id.proto &&
tmpl             1088 net/xfrm/xfrm_state.c 		    (tmpl->id.spi == x->id.spi || !tmpl->id.spi))
tmpl             1096 net/xfrm/xfrm_state.c 		if (tmpl->id.spi &&
tmpl             1097 net/xfrm/xfrm_state.c 		    (x0 = __xfrm_state_lookup(net, mark, daddr, tmpl->id.spi,
tmpl             1098 net/xfrm/xfrm_state.c 					      tmpl->id.proto, encap_family)) != NULL) {
tmpl             1121 net/xfrm/xfrm_state.c 		xfrm_init_tempstate(x, fl, tmpl, daddr, saddr, family);
tmpl             1133 net/xfrm/xfrm_state.c 		if (km_query(x, tmpl, pol) == 0) {