c_ctx 761 drivers/crypto/chelsio/chcr_algo.c struct ablk_ctx *ablkctx = ABLK_CTX(c_ctx(tfm)); c_ctx 774 drivers/crypto/chelsio/chcr_algo.c struct adapter *adap = padap(c_ctx(tfm)->dev); c_ctx 794 drivers/crypto/chelsio/chcr_algo.c FILL_SEC_CPL_OP_IVINSR(c_ctx(tfm)->tx_chan_id, 2, 1); c_ctx 838 drivers/crypto/chelsio/chcr_algo.c create_wreq(c_ctx(tfm), chcr_req, &(wrparam->req->base), reqctx->imm, 0, c_ctx 874 drivers/crypto/chelsio/chcr_algo.c struct ablk_ctx *ablkctx = ABLK_CTX(c_ctx(cipher)); c_ctx 893 drivers/crypto/chelsio/chcr_algo.c struct ablk_ctx *ablkctx = ABLK_CTX(c_ctx(cipher)); c_ctx 925 drivers/crypto/chelsio/chcr_algo.c struct ablk_ctx *ablkctx = ABLK_CTX(c_ctx(cipher)); c_ctx 956 drivers/crypto/chelsio/chcr_algo.c struct ablk_ctx *ablkctx = ABLK_CTX(c_ctx(cipher)); c_ctx 1024 drivers/crypto/chelsio/chcr_algo.c struct ablk_ctx *ablkctx = ABLK_CTX(c_ctx(tfm)); c_ctx 1115 drivers/crypto/chelsio/chcr_algo.c struct uld_ctx *u_ctx = ULD_CTX(c_ctx(tfm)); c_ctx 1116 drivers/crypto/chelsio/chcr_algo.c struct ablk_ctx *ablkctx = ABLK_CTX(c_ctx(tfm)); c_ctx 1121 drivers/crypto/chelsio/chcr_algo.c struct chcr_dev *dev = c_ctx(tfm)->dev; c_ctx 1127 drivers/crypto/chelsio/chcr_algo.c chcr_cipher_dma_unmap(&ULD_CTX(c_ctx(tfm))->lldi.pdev->dev, c_ctx 1150 drivers/crypto/chelsio/chcr_algo.c chcr_cipher_dma_unmap(&ULD_CTX(c_ctx(tfm))->lldi.pdev->dev, c_ctx 1165 drivers/crypto/chelsio/chcr_algo.c wrparam.qid = u_ctx->lldi.rxq_ids[c_ctx(tfm)->rx_qidx]; c_ctx 1175 drivers/crypto/chelsio/chcr_algo.c set_wr_txq(skb, CPL_PRIORITY_DATA, c_ctx(tfm)->tx_qidx); c_ctx 1181 drivers/crypto/chelsio/chcr_algo.c chcr_cipher_dma_unmap(&ULD_CTX(c_ctx(tfm))->lldi.pdev->dev, req); c_ctx 1196 drivers/crypto/chelsio/chcr_algo.c struct ablk_ctx *ablkctx = ABLK_CTX(c_ctx(tfm)); c_ctx 1211 drivers/crypto/chelsio/chcr_algo.c err = chcr_cipher_dma_map(&ULD_CTX(c_ctx(tfm))->lldi.pdev->dev, req); c_ctx 1265 drivers/crypto/chelsio/chcr_algo.c chcr_cipher_dma_unmap(&ULD_CTX(c_ctx(tfm))->lldi.pdev->dev, c_ctx 1294 drivers/crypto/chelsio/chcr_algo.c chcr_cipher_dma_unmap(&ULD_CTX(c_ctx(tfm))->lldi.pdev->dev, req); c_ctx 1302 drivers/crypto/chelsio/chcr_algo.c struct chcr_dev *dev = c_ctx(tfm)->dev; c_ctx 1305 drivers/crypto/chelsio/chcr_algo.c struct uld_ctx *u_ctx = ULD_CTX(c_ctx(tfm)); c_ctx 1311 drivers/crypto/chelsio/chcr_algo.c c_ctx(tfm)->tx_qidx))) { c_ctx 1319 drivers/crypto/chelsio/chcr_algo.c err = process_cipher(req, u_ctx->lldi.rxq_ids[c_ctx(tfm)->rx_qidx], c_ctx 1324 drivers/crypto/chelsio/chcr_algo.c set_wr_txq(skb, CPL_PRIORITY_DATA, c_ctx(tfm)->tx_qidx); c_ctx 1335 drivers/crypto/chelsio/chcr_algo.c struct uld_ctx *u_ctx = ULD_CTX(c_ctx(tfm)); c_ctx 1336 drivers/crypto/chelsio/chcr_algo.c struct chcr_dev *dev = c_ctx(tfm)->dev; c_ctx 1345 drivers/crypto/chelsio/chcr_algo.c c_ctx(tfm)->tx_qidx))) { c_ctx 1351 drivers/crypto/chelsio/chcr_algo.c err = process_cipher(req, u_ctx->lldi.rxq_ids[c_ctx(tfm)->rx_qidx], c_ctx 1356 drivers/crypto/chelsio/chcr_algo.c set_wr_txq(skb, CPL_PRIORITY_DATA, c_ctx(tfm)->tx_qidx); c_ctx 2154 drivers/crypto/chelsio/chcr_algo.c struct ablk_ctx *ablkctx = ABLK_CTX(c_ctx(cipher)); c_ctx 2609 drivers/crypto/chelsio/chcr_algo.c struct chcr_context *ctx = c_ctx(tfm); c_ctx 556 drivers/lightnvm/pblk-rb.c struct pblk_c_ctx *c_ctx = nvm_rq_to_pdu(rqd); c_ctx 572 drivers/lightnvm/pblk-rb.c c_ctx->sentry = pos; c_ctx 573 drivers/lightnvm/pblk-rb.c c_ctx->nr_valid = to_read; c_ctx 574 drivers/lightnvm/pblk-rb.c c_ctx->nr_padded = pad; c_ctx 23 drivers/lightnvm/pblk-write.c struct pblk_c_ctx *c_ctx) c_ctx 30 drivers/lightnvm/pblk-write.c for (i = 0; i < c_ctx->nr_valid; i++) { c_ctx 32 drivers/lightnvm/pblk-write.c int pos = c_ctx->sentry + i; c_ctx 52 drivers/lightnvm/pblk-write.c if (c_ctx->nr_padded) c_ctx 53 drivers/lightnvm/pblk-write.c pblk_bio_free_pages(pblk, rqd->bio, c_ctx->nr_valid, c_ctx 54 drivers/lightnvm/pblk-write.c c_ctx->nr_padded); c_ctx 60 drivers/lightnvm/pblk-write.c ret = pblk_rb_sync_advance(&pblk->rwb, c_ctx->nr_valid); c_ctx 70 drivers/lightnvm/pblk-write.c struct pblk_c_ctx *c_ctx) c_ctx 72 drivers/lightnvm/pblk-write.c list_del(&c_ctx->list); c_ctx 73 drivers/lightnvm/pblk-write.c return pblk_end_w_bio(pblk, rqd, c_ctx); c_ctx 77 drivers/lightnvm/pblk-write.c struct pblk_c_ctx *c_ctx) c_ctx 84 drivers/lightnvm/pblk-write.c atomic_long_sub(c_ctx->nr_valid, &pblk->inflight_writes); c_ctx 86 drivers/lightnvm/pblk-write.c pblk_up_rq(pblk, c_ctx->lun_bitmap); c_ctx 89 drivers/lightnvm/pblk-write.c if (pos == c_ctx->sentry) { c_ctx 90 drivers/lightnvm/pblk-write.c pos = pblk_end_w_bio(pblk, rqd, c_ctx); c_ctx 101 drivers/lightnvm/pblk-write.c WARN_ON(nvm_rq_from_c_ctx(c_ctx) != rqd); c_ctx 102 drivers/lightnvm/pblk-write.c list_add_tail(&c_ctx->list, &pblk->compl_list); c_ctx 186 drivers/lightnvm/pblk-write.c static void pblk_queue_resubmit(struct pblk *pblk, struct pblk_c_ctx *c_ctx) c_ctx 195 drivers/lightnvm/pblk-write.c r_ctx->sentry = c_ctx->sentry; c_ctx 196 drivers/lightnvm/pblk-write.c r_ctx->nr_valid = c_ctx->nr_valid; c_ctx 197 drivers/lightnvm/pblk-write.c r_ctx->nr_padded = c_ctx->nr_padded; c_ctx 204 drivers/lightnvm/pblk-write.c atomic_long_add(c_ctx->nr_valid, &pblk->recov_writes); c_ctx 214 drivers/lightnvm/pblk-write.c struct pblk_c_ctx *c_ctx = nvm_rq_to_pdu(rqd); c_ctx 220 drivers/lightnvm/pblk-write.c pblk_queue_resubmit(pblk, c_ctx); c_ctx 222 drivers/lightnvm/pblk-write.c pblk_up_rq(pblk, c_ctx->lun_bitmap); c_ctx 223 drivers/lightnvm/pblk-write.c if (c_ctx->nr_padded) c_ctx 224 drivers/lightnvm/pblk-write.c pblk_bio_free_pages(pblk, rqd->bio, c_ctx->nr_valid, c_ctx 225 drivers/lightnvm/pblk-write.c c_ctx->nr_padded); c_ctx 255 drivers/lightnvm/pblk-write.c struct pblk_c_ctx *c_ctx = nvm_rq_to_pdu(rqd); c_ctx 268 drivers/lightnvm/pblk-write.c pblk_complete_write(pblk, rqd, c_ctx); c_ctx 320 drivers/lightnvm/pblk-write.c struct pblk_c_ctx *c_ctx = nvm_rq_to_pdu(rqd); c_ctx 321 drivers/lightnvm/pblk-write.c unsigned int valid = c_ctx->nr_valid; c_ctx 322 drivers/lightnvm/pblk-write.c unsigned int padded = c_ctx->nr_padded; c_ctx 330 drivers/lightnvm/pblk-write.c c_ctx->lun_bitmap = lun_bitmap; c_ctx 339 drivers/lightnvm/pblk-write.c ret = pblk_map_rq(pblk, rqd, c_ctx->sentry, lun_bitmap, c_ctx 342 drivers/lightnvm/pblk-write.c ret = pblk_map_erase_rq(pblk, rqd, c_ctx->sentry, lun_bitmap, c_ctx 547 drivers/lightnvm/pblk-write.c struct pblk_c_ctx *c_ctx = nvm_rq_to_pdu(rqd); c_ctx 550 drivers/lightnvm/pblk-write.c if (c_ctx->nr_padded) c_ctx 551 drivers/lightnvm/pblk-write.c pblk_bio_free_pages(pblk, bio, c_ctx->nr_valid, c_ctx 552 drivers/lightnvm/pblk-write.c c_ctx->nr_padded); c_ctx 770 drivers/lightnvm/pblk.h struct pblk_c_ctx *c_ctx); c_ctx 929 drivers/lightnvm/pblk.h static inline struct nvm_rq *nvm_rq_from_c_ctx(void *c_ctx) c_ctx 931 drivers/lightnvm/pblk.h return c_ctx - sizeof(struct nvm_rq);