Home
last modified time | relevance | path

Searched refs:rctx (Results 1 – 38 of 38) sorted by relevance

/linux-4.1.27/drivers/crypto/ccp/
Dccp-crypto-sha.c30 struct ccp_sha_req_ctx *rctx = ahash_request_ctx(req); in ccp_sha_complete() local
36 if (rctx->hash_rem) { in ccp_sha_complete()
38 unsigned int offset = rctx->nbytes - rctx->hash_rem; in ccp_sha_complete()
40 scatterwalk_map_and_copy(rctx->buf, rctx->src, in ccp_sha_complete()
41 offset, rctx->hash_rem, 0); in ccp_sha_complete()
42 rctx->buf_count = rctx->hash_rem; in ccp_sha_complete()
44 rctx->buf_count = 0; in ccp_sha_complete()
49 memcpy(req->result, rctx->ctx, digest_size); in ccp_sha_complete()
52 sg_free_table(&rctx->data_sg); in ccp_sha_complete()
62 struct ccp_sha_req_ctx *rctx = ahash_request_ctx(req); in ccp_do_sha_update() local
[all …]
Dccp-crypto-aes-cmac.c31 struct ccp_aes_cmac_req_ctx *rctx = ahash_request_ctx(req); in ccp_aes_cmac_complete() local
37 if (rctx->hash_rem) { in ccp_aes_cmac_complete()
39 unsigned int offset = rctx->nbytes - rctx->hash_rem; in ccp_aes_cmac_complete()
41 scatterwalk_map_and_copy(rctx->buf, rctx->src, in ccp_aes_cmac_complete()
42 offset, rctx->hash_rem, 0); in ccp_aes_cmac_complete()
43 rctx->buf_count = rctx->hash_rem; in ccp_aes_cmac_complete()
45 rctx->buf_count = 0; in ccp_aes_cmac_complete()
50 memcpy(req->result, rctx->iv, digest_size); in ccp_aes_cmac_complete()
53 sg_free_table(&rctx->data_sg); in ccp_aes_cmac_complete()
63 struct ccp_aes_cmac_req_ctx *rctx = ahash_request_ctx(req); in ccp_do_cmac_update() local
[all …]
Dccp-crypto-aes.c29 struct ccp_aes_req_ctx *rctx = ablkcipher_request_ctx(req); in ccp_aes_complete() local
35 memcpy(req->info, rctx->iv, AES_BLOCK_SIZE); in ccp_aes_complete()
73 struct ccp_aes_req_ctx *rctx = ablkcipher_request_ctx(req); in ccp_aes_crypt() local
91 memcpy(rctx->iv, req->info, AES_BLOCK_SIZE); in ccp_aes_crypt()
92 iv_sg = &rctx->iv_sg; in ccp_aes_crypt()
94 sg_init_one(iv_sg, rctx->iv, iv_len); in ccp_aes_crypt()
97 memset(&rctx->cmd, 0, sizeof(rctx->cmd)); in ccp_aes_crypt()
98 INIT_LIST_HEAD(&rctx->cmd.entry); in ccp_aes_crypt()
99 rctx->cmd.engine = CCP_ENGINE_AES; in ccp_aes_crypt()
100 rctx->cmd.u.aes.type = ctx->u.aes.type; in ccp_aes_crypt()
[all …]
Dccp-crypto-aes-xts.c87 struct ccp_aes_req_ctx *rctx = ablkcipher_request_ctx(req); in ccp_aes_xts_complete() local
92 memcpy(req->info, rctx->iv, AES_BLOCK_SIZE); in ccp_aes_xts_complete()
123 struct ccp_aes_req_ctx *rctx = ablkcipher_request_ctx(req); in ccp_aes_xts_crypt() local
160 memcpy(rctx->iv, req->info, AES_BLOCK_SIZE); in ccp_aes_xts_crypt()
161 sg_init_one(&rctx->iv_sg, rctx->iv, AES_BLOCK_SIZE); in ccp_aes_xts_crypt()
163 memset(&rctx->cmd, 0, sizeof(rctx->cmd)); in ccp_aes_xts_crypt()
164 INIT_LIST_HEAD(&rctx->cmd.entry); in ccp_aes_xts_crypt()
165 rctx->cmd.engine = CCP_ENGINE_XTS_AES_128; in ccp_aes_xts_crypt()
166 rctx->cmd.u.xts.action = (encrypt) ? CCP_AES_ACTION_ENCRYPT in ccp_aes_xts_crypt()
168 rctx->cmd.u.xts.unit_size = unit_size; in ccp_aes_xts_crypt()
[all …]
/linux-4.1.27/drivers/crypto/qce/
Dsha.c42 struct qce_sha_reqctx *rctx = ahash_request_ctx(req); in qce_ahash_done() local
54 qce_unmapsg(qce->dev, req->src, rctx->src_nents, DMA_TO_DEVICE, in qce_ahash_done()
55 rctx->src_chained); in qce_ahash_done()
56 qce_unmapsg(qce->dev, &rctx->result_sg, 1, DMA_FROM_DEVICE, 0); in qce_ahash_done()
58 memcpy(rctx->digest, result->auth_iv, digestsize); in qce_ahash_done()
62 rctx->byte_count[0] = cpu_to_be32(result->auth_byte_count[0]); in qce_ahash_done()
63 rctx->byte_count[1] = cpu_to_be32(result->auth_byte_count[1]); in qce_ahash_done()
69 req->src = rctx->src_orig; in qce_ahash_done()
70 req->nbytes = rctx->nbytes_orig; in qce_ahash_done()
71 rctx->last_blk = false; in qce_ahash_done()
[all …]
Dablkcipher.c29 struct qce_cipher_reqctx *rctx = ablkcipher_request_ctx(req); in qce_ablkcipher_done() local
47 qce_unmapsg(qce->dev, rctx->src_sg, rctx->src_nents, dir_src, in qce_ablkcipher_done()
48 rctx->dst_chained); in qce_ablkcipher_done()
49 qce_unmapsg(qce->dev, rctx->dst_sg, rctx->dst_nents, dir_dst, in qce_ablkcipher_done()
50 rctx->dst_chained); in qce_ablkcipher_done()
52 sg_free_table(&rctx->dst_tbl); in qce_ablkcipher_done()
65 struct qce_cipher_reqctx *rctx = ablkcipher_request_ctx(req); in qce_ablkcipher_async_req_handle() local
75 rctx->iv = req->info; in qce_ablkcipher_async_req_handle()
76 rctx->ivsize = crypto_ablkcipher_ivsize(ablkcipher); in qce_ablkcipher_async_req_handle()
77 rctx->cryptlen = req->nbytes; in qce_ablkcipher_async_req_handle()
[all …]
Dcommon.c235 struct qce_sha_reqctx *rctx = ahash_request_ctx(req); in qce_setup_regs_ahash() local
246 if (!rctx->last_blk && req->nbytes % blocksize) in qce_setup_regs_ahash()
251 if (IS_CMAC(rctx->flags)) { in qce_setup_regs_ahash()
259 auth_cfg = qce_auth_cfg(rctx->flags, rctx->authklen); in qce_setup_regs_ahash()
262 if (IS_SHA_HMAC(rctx->flags) || IS_CMAC(rctx->flags)) { in qce_setup_regs_ahash()
263 u32 authkey_words = rctx->authklen / sizeof(u32); in qce_setup_regs_ahash()
265 qce_cpu_to_be32p_array(mackey, rctx->authkey, rctx->authklen); in qce_setup_regs_ahash()
270 if (IS_CMAC(rctx->flags)) in qce_setup_regs_ahash()
273 if (rctx->first_blk) in qce_setup_regs_ahash()
274 memcpy(auth, rctx->digest, digestsize); in qce_setup_regs_ahash()
[all …]
/linux-4.1.27/arch/x86/crypto/sha-mb/
Dsha1_mb.c96 static void req_ctx_init(struct mcryptd_hash_request_ctx *rctx, in req_ctx_init() argument
99 rctx->flag = HASH_UPDATE; in req_ctx_init()
366 static int sha1_mb_set_results(struct mcryptd_hash_request_ctx *rctx) in sha1_mb_set_results() argument
369 struct sha1_hash_ctx *sctx = shash_desc_ctx(&rctx->desc); in sha1_mb_set_results()
370 __be32 *dst = (__be32 *) rctx->out; in sha1_mb_set_results()
383 struct mcryptd_hash_request_ctx *rctx = *ret_rctx; in sha_finish_walk() local
387 while (!(rctx->flag & HASH_DONE)) { in sha_finish_walk()
388 nbytes = crypto_ahash_walk_done(&rctx->walk, 0); in sha_finish_walk()
394 if (crypto_ahash_walk_last(&rctx->walk)) { in sha_finish_walk()
395 rctx->flag |= HASH_DONE; in sha_finish_walk()
[all …]
/linux-4.1.27/drivers/crypto/
Dsahara.c581 struct sahara_aes_reqctx *rctx; in sahara_aes_process() local
595 rctx = ablkcipher_request_ctx(req); in sahara_aes_process()
597 rctx->mode &= FLAGS_MODE_MASK; in sahara_aes_process()
598 dev->flags = (dev->flags & ~FLAGS_MODE_MASK) | rctx->mode; in sahara_aes_process()
666 struct sahara_aes_reqctx *rctx = ablkcipher_request_ctx(req); in sahara_aes_crypt() local
679 rctx->mode = mode; in sahara_aes_crypt()
789 struct sahara_sha_reqctx *rctx) in sahara_sha_init_hdr() argument
793 hdr = rctx->mode; in sahara_sha_init_hdr()
795 if (rctx->first) { in sahara_sha_init_hdr()
802 if (rctx->last) in sahara_sha_init_hdr()
[all …]
Dn2_core.c305 struct n2_hash_req_ctx *rctx = ahash_request_ctx(req); in n2_hash_async_init() local
309 ahash_request_set_tfm(&rctx->fallback_req, ctx->fallback_tfm); in n2_hash_async_init()
310 rctx->fallback_req.base.flags = req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP; in n2_hash_async_init()
312 return crypto_ahash_init(&rctx->fallback_req); in n2_hash_async_init()
317 struct n2_hash_req_ctx *rctx = ahash_request_ctx(req); in n2_hash_async_update() local
321 ahash_request_set_tfm(&rctx->fallback_req, ctx->fallback_tfm); in n2_hash_async_update()
322 rctx->fallback_req.base.flags = req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP; in n2_hash_async_update()
323 rctx->fallback_req.nbytes = req->nbytes; in n2_hash_async_update()
324 rctx->fallback_req.src = req->src; in n2_hash_async_update()
326 return crypto_ahash_update(&rctx->fallback_req); in n2_hash_async_update()
[all …]
Dmxs-dcp.c201 struct dcp_aes_req_ctx *rctx = ablkcipher_request_ctx(req); in mxs_dcp_run_aes() local
220 if (rctx->enc) in mxs_dcp_run_aes()
227 if (rctx->ecb) in mxs_dcp_run_aes()
255 struct dcp_aes_req_ctx *rctx = ablkcipher_request_ctx(req); in mxs_dcp_aes_block_crypt() local
280 if (!rctx->ecb) { in mxs_dcp_aes_block_crypt()
400 struct dcp_aes_req_ctx *rctx = ablkcipher_request_ctx(req); in mxs_dcp_aes_enqueue() local
406 rctx->enc = enc; in mxs_dcp_aes_enqueue()
407 rctx->ecb = ecb; in mxs_dcp_aes_enqueue()
516 struct dcp_sha_req_ctx *rctx = ahash_request_ctx(req); in mxs_dcp_run_sha() local
529 if (rctx->init) in mxs_dcp_run_sha()
[all …]
Dimg-hash.c484 struct img_hash_request_ctx *rctx = ahash_request_ctx(req); in img_hash_init() local
487 ahash_request_set_tfm(&rctx->fallback_req, ctx->fallback); in img_hash_init()
488 rctx->fallback_req.base.flags = req->base.flags in img_hash_init()
491 return crypto_ahash_init(&rctx->fallback_req); in img_hash_init()
547 struct img_hash_request_ctx *rctx = ahash_request_ctx(req); in img_hash_update() local
551 ahash_request_set_tfm(&rctx->fallback_req, ctx->fallback); in img_hash_update()
552 rctx->fallback_req.base.flags = req->base.flags in img_hash_update()
554 rctx->fallback_req.nbytes = req->nbytes; in img_hash_update()
555 rctx->fallback_req.src = req->src; in img_hash_update()
557 return crypto_ahash_update(&rctx->fallback_req); in img_hash_update()
[all …]
Dhifn_795x.c1166 struct hifn_context *ctx, struct hifn_request_context *rctx, in hifn_setup_cmd_desc() argument
1178 switch (rctx->op) { in hifn_setup_cmd_desc()
1195 if (rctx->op == ACRYPTO_OP_ENCRYPT || rctx->op == ACRYPTO_OP_DECRYPT) { in hifn_setup_cmd_desc()
1200 if (rctx->iv && rctx->mode != ACRYPTO_MODE_ECB) in hifn_setup_cmd_desc()
1203 switch (rctx->mode) { in hifn_setup_cmd_desc()
1220 switch (rctx->type) { in hifn_setup_cmd_desc()
1255 rctx->iv, rctx->ivsize, md); in hifn_setup_cmd_desc()
1371 struct hifn_context *ctx, struct hifn_request_context *rctx, in hifn_setup_dma() argument
1392 t = &rctx->walk.cache[0]; in hifn_setup_dma()
1395 if (t->length && rctx->walk.flags & ASYNC_FLAGS_MISALIGNED) { in hifn_setup_dma()
[all …]
Domap-des.c594 struct omap_des_reqctx *rctx; in omap_des_handle_queue() local
639 rctx = ablkcipher_request_ctx(req); in omap_des_handle_queue()
641 rctx->mode &= FLAGS_MODE_MASK; in omap_des_handle_queue()
642 dd->flags = (dd->flags & ~FLAGS_MODE_MASK) | rctx->mode; in omap_des_handle_queue()
704 struct omap_des_reqctx *rctx = ablkcipher_request_ctx(req); in omap_des_crypt() local
720 rctx->mode = mode; in omap_des_crypt()
Domap-aes.c612 struct omap_aes_reqctx *rctx; in omap_aes_handle_queue() local
657 rctx = ablkcipher_request_ctx(req); in omap_aes_handle_queue()
659 rctx->mode &= FLAGS_MODE_MASK; in omap_aes_handle_queue()
660 dd->flags = (dd->flags & ~FLAGS_MODE_MASK) | rctx->mode; in omap_aes_handle_queue()
722 struct omap_aes_reqctx *rctx = ablkcipher_request_ctx(req); in omap_aes_crypt() local
738 rctx->mode = mode; in omap_aes_crypt()
Datmel-tdes.c593 struct atmel_tdes_reqctx *rctx; in atmel_tdes_handle_queue() local
626 rctx = ablkcipher_request_ctx(req); in atmel_tdes_handle_queue()
628 rctx->mode &= TDES_FLAGS_MODE_MASK; in atmel_tdes_handle_queue()
629 dd->flags = (dd->flags & ~TDES_FLAGS_MODE_MASK) | rctx->mode; in atmel_tdes_handle_queue()
675 struct atmel_tdes_reqctx *rctx = ablkcipher_request_ctx(req); in atmel_tdes_crypt() local
703 rctx->mode = mode; in atmel_tdes_crypt()
Datmel-aes.c567 struct atmel_aes_reqctx *rctx; in atmel_aes_handle_queue() local
600 rctx = ablkcipher_request_ctx(req); in atmel_aes_handle_queue()
602 rctx->mode &= AES_FLAGS_MODE_MASK; in atmel_aes_handle_queue()
603 dd->flags = (dd->flags & ~AES_FLAGS_MODE_MASK) | rctx->mode; in atmel_aes_handle_queue()
710 struct atmel_aes_reqctx *rctx = ablkcipher_request_ctx(req); in atmel_aes_crypt() local
749 rctx->mode = mode; in atmel_aes_crypt()
/linux-4.1.27/crypto/
Drmd256.c237 struct rmd256_ctx *rctx = shash_desc_ctx(desc); in rmd256_init() local
239 rctx->byte_count = 0; in rmd256_init()
241 rctx->state[0] = RMD_H0; in rmd256_init()
242 rctx->state[1] = RMD_H1; in rmd256_init()
243 rctx->state[2] = RMD_H2; in rmd256_init()
244 rctx->state[3] = RMD_H3; in rmd256_init()
245 rctx->state[4] = RMD_H5; in rmd256_init()
246 rctx->state[5] = RMD_H6; in rmd256_init()
247 rctx->state[6] = RMD_H7; in rmd256_init()
248 rctx->state[7] = RMD_H8; in rmd256_init()
[all …]
Dmcryptd.c101 struct mcryptd_hash_request_ctx *rctx) in mcryptd_enqueue_request() argument
108 rctx->tag.cpu = cpu; in mcryptd_enqueue_request()
322 struct mcryptd_hash_request_ctx *rctx = ahash_request_ctx(req); in mcryptd_hash_enqueue() local
327 rctx->complete = req->base.complete; in mcryptd_hash_enqueue()
330 ret = mcryptd_enqueue_request(queue, &req->base, rctx); in mcryptd_hash_enqueue()
340 struct mcryptd_hash_request_ctx *rctx = ahash_request_ctx(req); in mcryptd_hash_init() local
341 struct shash_desc *desc = &rctx->desc; in mcryptd_hash_init()
351 req->base.complete = rctx->complete; in mcryptd_hash_init()
355 rctx->complete(&req->base, err); in mcryptd_hash_init()
367 struct mcryptd_hash_request_ctx *rctx = ahash_request_ctx(req); in mcryptd_hash_update() local
[all …]
Drmd128.c222 struct rmd128_ctx *rctx = shash_desc_ctx(desc); in rmd128_init() local
224 rctx->byte_count = 0; in rmd128_init()
226 rctx->state[0] = RMD_H0; in rmd128_init()
227 rctx->state[1] = RMD_H1; in rmd128_init()
228 rctx->state[2] = RMD_H2; in rmd128_init()
229 rctx->state[3] = RMD_H3; in rmd128_init()
231 memset(rctx->buffer, 0, sizeof(rctx->buffer)); in rmd128_init()
239 struct rmd128_ctx *rctx = shash_desc_ctx(desc); in rmd128_update() local
240 const u32 avail = sizeof(rctx->buffer) - (rctx->byte_count & 0x3f); in rmd128_update()
242 rctx->byte_count += len; in rmd128_update()
[all …]
Dcryptd.c209 struct cryptd_blkcipher_request_ctx *rctx; in cryptd_blkcipher_crypt() local
212 rctx = ablkcipher_request_ctx(req); in cryptd_blkcipher_crypt()
223 req->base.complete = rctx->complete; in cryptd_blkcipher_crypt()
227 rctx->complete(&req->base, err); in cryptd_blkcipher_crypt()
252 struct cryptd_blkcipher_request_ctx *rctx = ablkcipher_request_ctx(req); in cryptd_blkcipher_enqueue() local
257 rctx->complete = req->base.complete; in cryptd_blkcipher_enqueue()
439 struct cryptd_hash_request_ctx *rctx = ahash_request_ctx(req); in cryptd_hash_enqueue() local
444 rctx->complete = req->base.complete; in cryptd_hash_enqueue()
455 struct cryptd_hash_request_ctx *rctx = ahash_request_ctx(req); in cryptd_hash_init() local
456 struct shash_desc *desc = &rctx->desc; in cryptd_hash_init()
[all …]
Drmd320.c284 struct rmd320_ctx *rctx = shash_desc_ctx(desc); in rmd320_init() local
286 rctx->byte_count = 0; in rmd320_init()
288 rctx->state[0] = RMD_H0; in rmd320_init()
289 rctx->state[1] = RMD_H1; in rmd320_init()
290 rctx->state[2] = RMD_H2; in rmd320_init()
291 rctx->state[3] = RMD_H3; in rmd320_init()
292 rctx->state[4] = RMD_H4; in rmd320_init()
293 rctx->state[5] = RMD_H5; in rmd320_init()
294 rctx->state[6] = RMD_H6; in rmd320_init()
295 rctx->state[7] = RMD_H7; in rmd320_init()
[all …]
Drmd160.c265 struct rmd160_ctx *rctx = shash_desc_ctx(desc); in rmd160_init() local
267 rctx->byte_count = 0; in rmd160_init()
269 rctx->state[0] = RMD_H0; in rmd160_init()
270 rctx->state[1] = RMD_H1; in rmd160_init()
271 rctx->state[2] = RMD_H2; in rmd160_init()
272 rctx->state[3] = RMD_H3; in rmd160_init()
273 rctx->state[4] = RMD_H4; in rmd160_init()
275 memset(rctx->buffer, 0, sizeof(rctx->buffer)); in rmd160_init()
283 struct rmd160_ctx *rctx = shash_desc_ctx(desc); in rmd160_update() local
284 const u32 avail = sizeof(rctx->buffer) - (rctx->byte_count & 0x3f); in rmd160_update()
[all …]
Dgcm.c1107 struct crypto_rfc4543_req_ctx *rctx = crypto_rfc4543_reqctx(req); in crypto_rfc4543_done() local
1110 scatterwalk_map_and_copy(rctx->auth_tag, req->dst, in crypto_rfc4543_done()
1123 struct crypto_rfc4543_req_ctx *rctx = crypto_rfc4543_reqctx(req); in crypto_rfc4543_crypt() local
1124 struct aead_request *subreq = &rctx->subreq; in crypto_rfc4543_crypt()
1126 struct scatterlist *cipher = rctx->cipher; in crypto_rfc4543_crypt()
1127 struct scatterlist *payload = rctx->payload; in crypto_rfc4543_crypt()
1128 struct scatterlist *assoc = rctx->assoc; in crypto_rfc4543_crypt()
1133 u8 *iv = PTR_ALIGN((u8 *)(rctx + 1) + crypto_aead_reqsize(ctx->child), in crypto_rfc4543_crypt()
1141 memset(rctx->auth_tag, 0, authsize); in crypto_rfc4543_crypt()
1143 scatterwalk_map_and_copy(rctx->auth_tag, src, in crypto_rfc4543_crypt()
[all …]
Dctr.c284 struct crypto_rfc3686_req_ctx *rctx = in crypto_rfc3686_crypt() local
286 struct ablkcipher_request *subreq = &rctx->subreq; in crypto_rfc3686_crypt()
287 u8 *iv = rctx->iv; in crypto_rfc3686_crypt()
/linux-4.1.27/kernel/events/
Dcallchain.c135 static struct perf_callchain_entry *get_callchain_entry(int *rctx) in get_callchain_entry() argument
140 *rctx = get_recursion_context(this_cpu_ptr(callchain_recursion)); in get_callchain_entry()
141 if (*rctx == -1) in get_callchain_entry()
150 return &entries->cpu_entries[cpu][*rctx]; in get_callchain_entry()
154 put_callchain_entry(int rctx) in put_callchain_entry() argument
156 put_recursion_context(this_cpu_ptr(callchain_recursion), rctx); in put_callchain_entry()
162 int rctx; in perf_callchain() local
171 entry = get_callchain_entry(&rctx); in perf_callchain()
172 if (rctx == -1) in perf_callchain()
206 put_callchain_entry(rctx); in perf_callchain()
Dinternal.h199 int rctx; in get_recursion_context() local
202 rctx = 3; in get_recursion_context()
204 rctx = 2; in get_recursion_context()
206 rctx = 1; in get_recursion_context()
208 rctx = 0; in get_recursion_context()
210 if (recursion[rctx]) in get_recursion_context()
213 recursion[rctx]++; in get_recursion_context()
216 return rctx; in get_recursion_context()
219 static inline void put_recursion_context(int *recursion, int rctx) in put_recursion_context() argument
222 recursion[rctx]--; in put_recursion_context()
Dcore.c4818 int rctx; in perf_pending_event() local
4820 rctx = perf_swevent_get_recursion_context(); in perf_pending_event()
4836 if (rctx >= 0) in perf_pending_event()
4837 perf_swevent_put_recursion_context(rctx); in perf_pending_event()
6389 inline void perf_swevent_put_recursion_context(int rctx) in perf_swevent_put_recursion_context() argument
6393 put_recursion_context(swhash->recursion, rctx); in perf_swevent_put_recursion_context()
6409 int rctx; in __perf_sw_event() local
6412 rctx = perf_swevent_get_recursion_context(); in __perf_sw_event()
6413 if (unlikely(rctx < 0)) in __perf_sw_event()
6418 perf_swevent_put_recursion_context(rctx); in __perf_sw_event()
[all …]
/linux-4.1.27/drivers/crypto/nx/
Dnx-aes-gcm.c333 struct nx_gcm_rctx *rctx = aead_request_ctx(req); in gcm_aes_nx_crypt() local
343 desc.info = rctx->iv; in gcm_aes_nx_crypt()
438 struct nx_gcm_rctx *rctx = aead_request_ctx(req); in gcm_aes_nx_encrypt() local
439 char *iv = rctx->iv; in gcm_aes_nx_encrypt()
448 struct nx_gcm_rctx *rctx = aead_request_ctx(req); in gcm_aes_nx_decrypt() local
449 char *iv = rctx->iv; in gcm_aes_nx_decrypt()
459 struct nx_gcm_rctx *rctx = aead_request_ctx(req); in gcm4106_aes_nx_encrypt() local
460 char *iv = rctx->iv; in gcm4106_aes_nx_encrypt()
472 struct nx_gcm_rctx *rctx = aead_request_ctx(req); in gcm4106_aes_nx_decrypt() local
473 char *iv = rctx->iv; in gcm4106_aes_nx_decrypt()
Dnx-aes-ccm.c497 struct nx_gcm_rctx *rctx = aead_request_ctx(req); in ccm4309_aes_nx_encrypt() local
499 u8 *iv = rctx->iv; in ccm4309_aes_nx_encrypt()
529 struct nx_gcm_rctx *rctx = aead_request_ctx(req); in ccm4309_aes_nx_decrypt() local
531 u8 *iv = rctx->iv; in ccm4309_aes_nx_decrypt()
/linux-4.1.27/kernel/trace/
Dtrace_syscalls.c554 int rctx; in perf_syscall_enter() local
577 sys_data->enter_event->event.type, NULL, &rctx); in perf_syscall_enter()
584 perf_trace_buf_submit(rec, size, rctx, 0, 1, regs, head, NULL); in perf_syscall_enter()
628 int rctx; in perf_syscall_exit() local
650 sys_data->exit_event->event.type, NULL, &rctx); in perf_syscall_exit()
656 perf_trace_buf_submit(rec, size, rctx, 0, 1, regs, head, NULL); in perf_syscall_exit()
Dtrace_event_perf.c308 int rctx; in perf_ftrace_function_call() local
321 entry = perf_trace_buf_prepare(ENTRY_SIZE, TRACE_FN, NULL, &rctx); in perf_ftrace_function_call()
327 perf_trace_buf_submit(entry, ENTRY_SIZE, rctx, 0, in perf_ftrace_function_call()
Dtrace_kprobe.c1142 int rctx; in kprobe_perf_func() local
1156 entry = perf_trace_buf_prepare(size, call->event.type, NULL, &rctx); in kprobe_perf_func()
1163 perf_trace_buf_submit(entry, size, rctx, 0, 1, regs, head, NULL); in kprobe_perf_func()
1177 int rctx; in kretprobe_perf_func() local
1191 entry = perf_trace_buf_prepare(size, call->event.type, NULL, &rctx); in kretprobe_perf_func()
1198 perf_trace_buf_submit(entry, size, rctx, 0, 1, regs, head, NULL); in kretprobe_perf_func()
Dtrace_uprobe.c1101 int rctx; in __uprobe_perf_func() local
1115 entry = perf_trace_buf_prepare(size, call->event.type, NULL, &rctx); in __uprobe_perf_func()
1136 perf_trace_buf_submit(entry, size, rctx, 0, 1, regs, head, NULL); in __uprobe_perf_func()
/linux-4.1.27/drivers/crypto/amcc/
Dcrypto4xx_core.h172 struct crypto4xx_ctx *rctx);
173 extern void crypto4xx_free_sa_rctx(struct crypto4xx_ctx *rctx);
/linux-4.1.27/include/linux/
Dperf_event.h872 struct hlist_head *head, int rctx,
905 extern void perf_swevent_put_recursion_context(int rctx);
959 static inline void perf_swevent_put_recursion_context(int rctx) { } in perf_swevent_put_recursion_context() argument
Dftrace_event.h619 perf_trace_buf_submit(void *raw_data, int size, int rctx, u64 addr, in perf_trace_buf_submit() argument
623 perf_tp_event(addr, count, raw_data, size, regs, head, rctx, task); in perf_trace_buf_submit()
/linux-4.1.27/include/trace/
Dftrace.h812 int rctx; \
826 event_call->event.type, &__regs, &rctx); \
836 perf_trace_buf_submit(entry, __entry_size, rctx, __addr, \