c_ctx             761 drivers/crypto/chelsio/chcr_algo.c 	struct ablk_ctx *ablkctx = ABLK_CTX(c_ctx(tfm));
c_ctx             774 drivers/crypto/chelsio/chcr_algo.c 	struct adapter *adap = padap(c_ctx(tfm)->dev);
c_ctx             794 drivers/crypto/chelsio/chcr_algo.c 		FILL_SEC_CPL_OP_IVINSR(c_ctx(tfm)->tx_chan_id, 2, 1);
c_ctx             838 drivers/crypto/chelsio/chcr_algo.c 	create_wreq(c_ctx(tfm), chcr_req, &(wrparam->req->base), reqctx->imm, 0,
c_ctx             874 drivers/crypto/chelsio/chcr_algo.c 	struct ablk_ctx *ablkctx = ABLK_CTX(c_ctx(cipher));
c_ctx             893 drivers/crypto/chelsio/chcr_algo.c 	struct ablk_ctx *ablkctx = ABLK_CTX(c_ctx(cipher));
c_ctx             925 drivers/crypto/chelsio/chcr_algo.c 	struct ablk_ctx *ablkctx = ABLK_CTX(c_ctx(cipher));
c_ctx             956 drivers/crypto/chelsio/chcr_algo.c 	struct ablk_ctx *ablkctx = ABLK_CTX(c_ctx(cipher));
c_ctx            1024 drivers/crypto/chelsio/chcr_algo.c 	struct ablk_ctx *ablkctx = ABLK_CTX(c_ctx(tfm));
c_ctx            1115 drivers/crypto/chelsio/chcr_algo.c 	struct uld_ctx *u_ctx = ULD_CTX(c_ctx(tfm));
c_ctx            1116 drivers/crypto/chelsio/chcr_algo.c 	struct ablk_ctx *ablkctx = ABLK_CTX(c_ctx(tfm));
c_ctx            1121 drivers/crypto/chelsio/chcr_algo.c 	struct chcr_dev *dev = c_ctx(tfm)->dev;
c_ctx            1127 drivers/crypto/chelsio/chcr_algo.c 		chcr_cipher_dma_unmap(&ULD_CTX(c_ctx(tfm))->lldi.pdev->dev,
c_ctx            1150 drivers/crypto/chelsio/chcr_algo.c 		chcr_cipher_dma_unmap(&ULD_CTX(c_ctx(tfm))->lldi.pdev->dev,
c_ctx            1165 drivers/crypto/chelsio/chcr_algo.c 	wrparam.qid = u_ctx->lldi.rxq_ids[c_ctx(tfm)->rx_qidx];
c_ctx            1175 drivers/crypto/chelsio/chcr_algo.c 	set_wr_txq(skb, CPL_PRIORITY_DATA, c_ctx(tfm)->tx_qidx);
c_ctx            1181 drivers/crypto/chelsio/chcr_algo.c 	chcr_cipher_dma_unmap(&ULD_CTX(c_ctx(tfm))->lldi.pdev->dev, req);
c_ctx            1196 drivers/crypto/chelsio/chcr_algo.c 	struct ablk_ctx *ablkctx = ABLK_CTX(c_ctx(tfm));
c_ctx            1211 drivers/crypto/chelsio/chcr_algo.c 	err = chcr_cipher_dma_map(&ULD_CTX(c_ctx(tfm))->lldi.pdev->dev, req);
c_ctx            1265 drivers/crypto/chelsio/chcr_algo.c 		chcr_cipher_dma_unmap(&ULD_CTX(c_ctx(tfm))->lldi.pdev->dev,
c_ctx            1294 drivers/crypto/chelsio/chcr_algo.c 	chcr_cipher_dma_unmap(&ULD_CTX(c_ctx(tfm))->lldi.pdev->dev, req);
c_ctx            1302 drivers/crypto/chelsio/chcr_algo.c 	struct chcr_dev *dev = c_ctx(tfm)->dev;
c_ctx            1305 drivers/crypto/chelsio/chcr_algo.c 	struct uld_ctx *u_ctx = ULD_CTX(c_ctx(tfm));
c_ctx            1311 drivers/crypto/chelsio/chcr_algo.c 					    c_ctx(tfm)->tx_qidx))) {
c_ctx            1319 drivers/crypto/chelsio/chcr_algo.c 	err = process_cipher(req, u_ctx->lldi.rxq_ids[c_ctx(tfm)->rx_qidx],
c_ctx            1324 drivers/crypto/chelsio/chcr_algo.c 	set_wr_txq(skb, CPL_PRIORITY_DATA, c_ctx(tfm)->tx_qidx);
c_ctx            1335 drivers/crypto/chelsio/chcr_algo.c 	struct uld_ctx *u_ctx = ULD_CTX(c_ctx(tfm));
c_ctx            1336 drivers/crypto/chelsio/chcr_algo.c 	struct chcr_dev *dev = c_ctx(tfm)->dev;
c_ctx            1345 drivers/crypto/chelsio/chcr_algo.c 					    c_ctx(tfm)->tx_qidx))) {
c_ctx            1351 drivers/crypto/chelsio/chcr_algo.c 	err = process_cipher(req, u_ctx->lldi.rxq_ids[c_ctx(tfm)->rx_qidx],
c_ctx            1356 drivers/crypto/chelsio/chcr_algo.c 	set_wr_txq(skb, CPL_PRIORITY_DATA, c_ctx(tfm)->tx_qidx);
c_ctx            2154 drivers/crypto/chelsio/chcr_algo.c 	struct ablk_ctx *ablkctx = ABLK_CTX(c_ctx(cipher));
c_ctx            2609 drivers/crypto/chelsio/chcr_algo.c 	struct chcr_context *ctx = c_ctx(tfm);
c_ctx             556 drivers/lightnvm/pblk-rb.c 	struct pblk_c_ctx *c_ctx = nvm_rq_to_pdu(rqd);
c_ctx             572 drivers/lightnvm/pblk-rb.c 	c_ctx->sentry = pos;
c_ctx             573 drivers/lightnvm/pblk-rb.c 	c_ctx->nr_valid = to_read;
c_ctx             574 drivers/lightnvm/pblk-rb.c 	c_ctx->nr_padded = pad;
c_ctx              23 drivers/lightnvm/pblk-write.c 				    struct pblk_c_ctx *c_ctx)
c_ctx              30 drivers/lightnvm/pblk-write.c 	for (i = 0; i < c_ctx->nr_valid; i++) {
c_ctx              32 drivers/lightnvm/pblk-write.c 		int pos = c_ctx->sentry + i;
c_ctx              52 drivers/lightnvm/pblk-write.c 	if (c_ctx->nr_padded)
c_ctx              53 drivers/lightnvm/pblk-write.c 		pblk_bio_free_pages(pblk, rqd->bio, c_ctx->nr_valid,
c_ctx              54 drivers/lightnvm/pblk-write.c 							c_ctx->nr_padded);
c_ctx              60 drivers/lightnvm/pblk-write.c 	ret = pblk_rb_sync_advance(&pblk->rwb, c_ctx->nr_valid);
c_ctx              70 drivers/lightnvm/pblk-write.c 					   struct pblk_c_ctx *c_ctx)
c_ctx              72 drivers/lightnvm/pblk-write.c 	list_del(&c_ctx->list);
c_ctx              73 drivers/lightnvm/pblk-write.c 	return pblk_end_w_bio(pblk, rqd, c_ctx);
c_ctx              77 drivers/lightnvm/pblk-write.c 				struct pblk_c_ctx *c_ctx)
c_ctx              84 drivers/lightnvm/pblk-write.c 	atomic_long_sub(c_ctx->nr_valid, &pblk->inflight_writes);
c_ctx              86 drivers/lightnvm/pblk-write.c 	pblk_up_rq(pblk, c_ctx->lun_bitmap);
c_ctx              89 drivers/lightnvm/pblk-write.c 	if (pos == c_ctx->sentry) {
c_ctx              90 drivers/lightnvm/pblk-write.c 		pos = pblk_end_w_bio(pblk, rqd, c_ctx);
c_ctx             101 drivers/lightnvm/pblk-write.c 		WARN_ON(nvm_rq_from_c_ctx(c_ctx) != rqd);
c_ctx             102 drivers/lightnvm/pblk-write.c 		list_add_tail(&c_ctx->list, &pblk->compl_list);
c_ctx             186 drivers/lightnvm/pblk-write.c static void pblk_queue_resubmit(struct pblk *pblk, struct pblk_c_ctx *c_ctx)
c_ctx             195 drivers/lightnvm/pblk-write.c 	r_ctx->sentry = c_ctx->sentry;
c_ctx             196 drivers/lightnvm/pblk-write.c 	r_ctx->nr_valid = c_ctx->nr_valid;
c_ctx             197 drivers/lightnvm/pblk-write.c 	r_ctx->nr_padded = c_ctx->nr_padded;
c_ctx             204 drivers/lightnvm/pblk-write.c 	atomic_long_add(c_ctx->nr_valid, &pblk->recov_writes);
c_ctx             214 drivers/lightnvm/pblk-write.c 	struct pblk_c_ctx *c_ctx = nvm_rq_to_pdu(rqd);
c_ctx             220 drivers/lightnvm/pblk-write.c 	pblk_queue_resubmit(pblk, c_ctx);
c_ctx             222 drivers/lightnvm/pblk-write.c 	pblk_up_rq(pblk, c_ctx->lun_bitmap);
c_ctx             223 drivers/lightnvm/pblk-write.c 	if (c_ctx->nr_padded)
c_ctx             224 drivers/lightnvm/pblk-write.c 		pblk_bio_free_pages(pblk, rqd->bio, c_ctx->nr_valid,
c_ctx             225 drivers/lightnvm/pblk-write.c 							c_ctx->nr_padded);
c_ctx             255 drivers/lightnvm/pblk-write.c 	struct pblk_c_ctx *c_ctx = nvm_rq_to_pdu(rqd);
c_ctx             268 drivers/lightnvm/pblk-write.c 	pblk_complete_write(pblk, rqd, c_ctx);
c_ctx             320 drivers/lightnvm/pblk-write.c 	struct pblk_c_ctx *c_ctx = nvm_rq_to_pdu(rqd);
c_ctx             321 drivers/lightnvm/pblk-write.c 	unsigned int valid = c_ctx->nr_valid;
c_ctx             322 drivers/lightnvm/pblk-write.c 	unsigned int padded = c_ctx->nr_padded;
c_ctx             330 drivers/lightnvm/pblk-write.c 	c_ctx->lun_bitmap = lun_bitmap;
c_ctx             339 drivers/lightnvm/pblk-write.c 		ret = pblk_map_rq(pblk, rqd, c_ctx->sentry, lun_bitmap,
c_ctx             342 drivers/lightnvm/pblk-write.c 		ret = pblk_map_erase_rq(pblk, rqd, c_ctx->sentry, lun_bitmap,
c_ctx             547 drivers/lightnvm/pblk-write.c 	struct pblk_c_ctx *c_ctx = nvm_rq_to_pdu(rqd);
c_ctx             550 drivers/lightnvm/pblk-write.c 	if (c_ctx->nr_padded)
c_ctx             551 drivers/lightnvm/pblk-write.c 		pblk_bio_free_pages(pblk, bio, c_ctx->nr_valid,
c_ctx             552 drivers/lightnvm/pblk-write.c 							c_ctx->nr_padded);
c_ctx             770 drivers/lightnvm/pblk.h 			struct pblk_c_ctx *c_ctx);
c_ctx             929 drivers/lightnvm/pblk.h static inline struct nvm_rq *nvm_rq_from_c_ctx(void *c_ctx)
c_ctx             931 drivers/lightnvm/pblk.h 	return c_ctx - sizeof(struct nvm_rq);