cxt 20 arch/arm64/kvm/hyp/tlb.c struct tlb_inv_context *cxt) cxt 24 arch/arm64/kvm/hyp/tlb.c local_irq_save(cxt->flags); cxt 36 arch/arm64/kvm/hyp/tlb.c val = cxt->tcr = read_sysreg_el1(SYS_TCR); cxt 39 arch/arm64/kvm/hyp/tlb.c val = cxt->sctlr = read_sysreg_el1(SYS_SCTLR); cxt 64 arch/arm64/kvm/hyp/tlb.c struct tlb_inv_context *cxt) cxt 71 arch/arm64/kvm/hyp/tlb.c struct tlb_inv_context *cxt) cxt 74 arch/arm64/kvm/hyp/tlb.c __tlb_switch_to_guest_vhe(kvm, cxt); cxt 76 arch/arm64/kvm/hyp/tlb.c __tlb_switch_to_guest_nvhe(kvm, cxt); cxt 80 arch/arm64/kvm/hyp/tlb.c struct tlb_inv_context *cxt) cxt 92 arch/arm64/kvm/hyp/tlb.c write_sysreg_el1(cxt->tcr, SYS_TCR); cxt 93 arch/arm64/kvm/hyp/tlb.c write_sysreg_el1(cxt->sctlr, SYS_SCTLR); cxt 96 arch/arm64/kvm/hyp/tlb.c local_irq_restore(cxt->flags); cxt 100 arch/arm64/kvm/hyp/tlb.c struct tlb_inv_context *cxt) cxt 106 arch/arm64/kvm/hyp/tlb.c struct tlb_inv_context *cxt) cxt 109 arch/arm64/kvm/hyp/tlb.c __tlb_switch_to_host_vhe(kvm, cxt); cxt 111 arch/arm64/kvm/hyp/tlb.c __tlb_switch_to_host_nvhe(kvm, cxt); cxt 116 arch/arm64/kvm/hyp/tlb.c struct tlb_inv_context cxt; cxt 122 arch/arm64/kvm/hyp/tlb.c __tlb_switch_to_guest(kvm, &cxt); cxt 165 arch/arm64/kvm/hyp/tlb.c __tlb_switch_to_host(kvm, &cxt); cxt 170 arch/arm64/kvm/hyp/tlb.c struct tlb_inv_context cxt; cxt 176 arch/arm64/kvm/hyp/tlb.c __tlb_switch_to_guest(kvm, &cxt); cxt 182 arch/arm64/kvm/hyp/tlb.c __tlb_switch_to_host(kvm, &cxt); cxt 188 arch/arm64/kvm/hyp/tlb.c struct tlb_inv_context cxt; cxt 191 arch/arm64/kvm/hyp/tlb.c __tlb_switch_to_guest(kvm, &cxt); cxt 197 arch/arm64/kvm/hyp/tlb.c __tlb_switch_to_host(kvm, &cxt); cxt 2415 drivers/gpu/drm/i915/i915_drv.h int intel_engine_cmd_parser(struct i915_gem_context *cxt, cxt 67 drivers/infiniband/hw/qedr/qedr_roce_cm.c static void qedr_ll2_complete_tx_packet(void *cxt, u8 connection_handle, cxt 73 drivers/infiniband/hw/qedr/qedr_roce_cm.c struct qedr_dev *dev = (struct qedr_dev *)cxt; cxt 96 drivers/infiniband/hw/qedr/qedr_roce_cm.c static void qedr_ll2_complete_rx_packet(void *cxt, cxt 99 drivers/infiniband/hw/qedr/qedr_roce_cm.c struct qedr_dev *dev = (struct qedr_dev *)cxt; cxt 125 drivers/infiniband/hw/qedr/qedr_roce_cm.c static void qedr_ll2_release_rx_packet(void *cxt, u8 connection_handle, cxt 58 drivers/mtd/mtdoops.c static void mark_page_used(struct mtdoops_context *cxt, int page) cxt 60 drivers/mtd/mtdoops.c set_bit(page, cxt->oops_page_used); cxt 63 drivers/mtd/mtdoops.c static void mark_page_unused(struct mtdoops_context *cxt, int page) cxt 65 drivers/mtd/mtdoops.c clear_bit(page, cxt->oops_page_used); cxt 68 drivers/mtd/mtdoops.c static int page_is_used(struct mtdoops_context *cxt, int page) cxt 70 drivers/mtd/mtdoops.c return test_bit(page, cxt->oops_page_used); cxt 73 drivers/mtd/mtdoops.c static int mtdoops_erase_block(struct mtdoops_context *cxt, int offset) cxt 75 drivers/mtd/mtdoops.c struct mtd_info *mtd = cxt->mtd; cxt 96 drivers/mtd/mtdoops.c mark_page_unused(cxt, page); cxt 101 drivers/mtd/mtdoops.c static void mtdoops_inc_counter(struct mtdoops_context *cxt) cxt 103 drivers/mtd/mtdoops.c cxt->nextpage++; cxt 104 drivers/mtd/mtdoops.c if (cxt->nextpage >= cxt->oops_pages) cxt 105 drivers/mtd/mtdoops.c cxt->nextpage = 0; cxt 106 drivers/mtd/mtdoops.c cxt->nextcount++; cxt 107 drivers/mtd/mtdoops.c if (cxt->nextcount == 0xffffffff) cxt 108 drivers/mtd/mtdoops.c cxt->nextcount = 0; cxt 110 drivers/mtd/mtdoops.c if (page_is_used(cxt, cxt->nextpage)) { cxt 111 drivers/mtd/mtdoops.c schedule_work(&cxt->work_erase); cxt 116 drivers/mtd/mtdoops.c cxt->nextpage, cxt->nextcount); cxt 122 drivers/mtd/mtdoops.c struct mtdoops_context *cxt = cxt 124 drivers/mtd/mtdoops.c struct mtd_info *mtd = cxt->mtd; cxt 131 drivers/mtd/mtdoops.c mod = (cxt->nextpage * record_size) % mtd->erasesize; cxt 133 drivers/mtd/mtdoops.c cxt->nextpage = cxt->nextpage + ((mtd->erasesize - mod) / record_size); cxt 134 drivers/mtd/mtdoops.c if (cxt->nextpage >= cxt->oops_pages) cxt 135 drivers/mtd/mtdoops.c cxt->nextpage = 0; cxt 138 drivers/mtd/mtdoops.c while ((ret = mtd_block_isbad(mtd, cxt->nextpage * record_size)) > 0) { cxt 141 drivers/mtd/mtdoops.c cxt->nextpage * record_size); cxt 143 drivers/mtd/mtdoops.c cxt->nextpage = cxt->nextpage + (mtd->erasesize / record_size); cxt 144 drivers/mtd/mtdoops.c if (cxt->nextpage >= cxt->oops_pages) cxt 145 drivers/mtd/mtdoops.c cxt->nextpage = 0; cxt 146 drivers/mtd/mtdoops.c if (i == cxt->oops_pages / (mtd->erasesize / record_size)) { cxt 158 drivers/mtd/mtdoops.c ret = mtdoops_erase_block(cxt, cxt->nextpage * record_size); cxt 162 drivers/mtd/mtdoops.c cxt->nextpage, cxt->nextcount); cxt 167 drivers/mtd/mtdoops.c ret = mtd_block_markbad(mtd, cxt->nextpage * record_size); cxt 176 drivers/mtd/mtdoops.c static void mtdoops_write(struct mtdoops_context *cxt, int panic) cxt 178 drivers/mtd/mtdoops.c struct mtd_info *mtd = cxt->mtd; cxt 184 drivers/mtd/mtdoops.c hdr = cxt->oops_buf; cxt 185 drivers/mtd/mtdoops.c hdr[0] = cxt->nextcount; cxt 189 drivers/mtd/mtdoops.c ret = mtd_panic_write(mtd, cxt->nextpage * record_size, cxt 190 drivers/mtd/mtdoops.c record_size, &retlen, cxt->oops_buf); cxt 196 drivers/mtd/mtdoops.c ret = mtd_write(mtd, cxt->nextpage * record_size, cxt 197 drivers/mtd/mtdoops.c record_size, &retlen, cxt->oops_buf); cxt 201 drivers/mtd/mtdoops.c cxt->nextpage * record_size, retlen, record_size, ret); cxt 202 drivers/mtd/mtdoops.c mark_page_used(cxt, cxt->nextpage); cxt 203 drivers/mtd/mtdoops.c memset(cxt->oops_buf, 0xff, record_size); cxt 205 drivers/mtd/mtdoops.c mtdoops_inc_counter(cxt); cxt 210 drivers/mtd/mtdoops.c struct mtdoops_context *cxt = cxt 213 drivers/mtd/mtdoops.c mtdoops_write(cxt, 0); cxt 216 drivers/mtd/mtdoops.c static void find_next_position(struct mtdoops_context *cxt) cxt 218 drivers/mtd/mtdoops.c struct mtd_info *mtd = cxt->mtd; cxt 223 drivers/mtd/mtdoops.c for (page = 0; page < cxt->oops_pages; page++) { cxt 227 drivers/mtd/mtdoops.c mark_page_used(cxt, page); cxt 239 drivers/mtd/mtdoops.c mark_page_unused(cxt, page); cxt 258 drivers/mtd/mtdoops.c cxt->nextpage = cxt->oops_pages - 1; cxt 259 drivers/mtd/mtdoops.c cxt->nextcount = 0; cxt 262 drivers/mtd/mtdoops.c cxt->nextpage = maxpos; cxt 263 drivers/mtd/mtdoops.c cxt->nextcount = maxcount; cxt 266 drivers/mtd/mtdoops.c mtdoops_inc_counter(cxt); cxt 272 drivers/mtd/mtdoops.c struct mtdoops_context *cxt = container_of(dumper, cxt 279 drivers/mtd/mtdoops.c kmsg_dump_get_buffer(dumper, true, cxt->oops_buf + MTDOOPS_HEADER_SIZE, cxt 284 drivers/mtd/mtdoops.c mtdoops_write(cxt, 1); cxt 287 drivers/mtd/mtdoops.c schedule_work(&cxt->work_write); cxt 292 drivers/mtd/mtdoops.c struct mtdoops_context *cxt = &oops_cxt; cxt 297 drivers/mtd/mtdoops.c cxt->mtd_index = mtd->index; cxt 299 drivers/mtd/mtdoops.c if (mtd->index != cxt->mtd_index || cxt->mtd_index < 0) cxt 319 drivers/mtd/mtdoops.c cxt->oops_page_used = cxt 323 drivers/mtd/mtdoops.c if (!cxt->oops_page_used) { cxt 328 drivers/mtd/mtdoops.c cxt->dump.max_reason = KMSG_DUMP_OOPS; cxt 329 drivers/mtd/mtdoops.c cxt->dump.dump = mtdoops_do_dump; cxt 330 drivers/mtd/mtdoops.c err = kmsg_dump_register(&cxt->dump); cxt 333 drivers/mtd/mtdoops.c vfree(cxt->oops_page_used); cxt 334 drivers/mtd/mtdoops.c cxt->oops_page_used = NULL; cxt 338 drivers/mtd/mtdoops.c cxt->mtd = mtd; cxt 339 drivers/mtd/mtdoops.c cxt->oops_pages = (int)mtd->size / record_size; cxt 340 drivers/mtd/mtdoops.c find_next_position(cxt); cxt 346 drivers/mtd/mtdoops.c struct mtdoops_context *cxt = &oops_cxt; cxt 348 drivers/mtd/mtdoops.c if (mtd->index != cxt->mtd_index || cxt->mtd_index < 0) cxt 351 drivers/mtd/mtdoops.c if (kmsg_dump_unregister(&cxt->dump) < 0) cxt 354 drivers/mtd/mtdoops.c cxt->mtd = NULL; cxt 355 drivers/mtd/mtdoops.c flush_work(&cxt->work_erase); cxt 356 drivers/mtd/mtdoops.c flush_work(&cxt->work_write); cxt 367 drivers/mtd/mtdoops.c struct mtdoops_context *cxt = &oops_cxt; cxt 385 drivers/mtd/mtdoops.c cxt->mtd_index = -1; cxt 388 drivers/mtd/mtdoops.c cxt->mtd_index = mtd_index; cxt 390 drivers/mtd/mtdoops.c cxt->oops_buf = vmalloc(record_size); cxt 391 drivers/mtd/mtdoops.c if (!cxt->oops_buf) { cxt 395 drivers/mtd/mtdoops.c memset(cxt->oops_buf, 0xff, record_size); cxt 397 drivers/mtd/mtdoops.c INIT_WORK(&cxt->work_erase, mtdoops_workfunc_erase); cxt 398 drivers/mtd/mtdoops.c INIT_WORK(&cxt->work_write, mtdoops_workfunc_write); cxt 406 drivers/mtd/mtdoops.c struct mtdoops_context *cxt = &oops_cxt; cxt 409 drivers/mtd/mtdoops.c vfree(cxt->oops_buf); cxt 410 drivers/mtd/mtdoops.c vfree(cxt->oops_page_used); cxt 5057 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c void bnx2x_set_ctx_validation(struct bnx2x *bp, struct eth_context *cxt, cxt 5060 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c if (!cxt) { cxt 5061 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c BNX2X_ERR("bad context pointer %p\n", cxt); cxt 5066 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c cxt->ustorm_ag_context.cdu_usage = cxt 5070 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c cxt->xstorm_ag_context.cdu_reserved = cxt 1272 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.h void bnx2x_set_ctx_validation(struct bnx2x *bp, struct eth_context *cxt, cxt 189 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c init_p->cxts[0] = q->cxt; cxt 307 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c if (bnx2x_vfq(vf, qid, cxt)) { cxt 308 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c bnx2x_vfq(vf, qid, cxt)->ustorm_ag_context.cdu_usage = 0; cxt 309 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c bnx2x_vfq(vf, qid, cxt)->xstorm_ag_context.cdu_reserved = 0; cxt 1345 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c struct hw_dma *cxt = &bp->vfdb->context[i]; cxt 1346 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c BNX2X_PCI_FREE(cxt->addr, cxt->mapping, cxt->size); cxt 1375 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c struct hw_dma *cxt = BP_VF_CXT_PAGE(bp, i); cxt 1376 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c cxt->size = min_t(size_t, tot_size, CDU_ILT_PAGE_SZ); cxt 1378 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c if (cxt->size) { cxt 1379 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c cxt->addr = BNX2X_PCI_ALLOC(&cxt->mapping, cxt->size); cxt 1380 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c if (!cxt->addr) cxt 1383 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c cxt->addr = NULL; cxt 1384 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c cxt->mapping = 0; cxt 1386 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c tot_size -= cxt->size; cxt 2092 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c q->cxt = &((base_cxt + i)->eth); cxt 2096 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c vf->abs_vfid, i, q->index, q->cid, q->cxt); cxt 73 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.h struct eth_context *cxt; cxt 2300 drivers/net/ethernet/qlogic/qed/qed_iwarp.c qed_iwarp_ll2_comp_mpa_pkt(void *cxt, struct qed_ll2_comp_rx_data *data) cxt 2304 drivers/net/ethernet/qlogic/qed/qed_iwarp.c struct qed_hwfn *p_hwfn = cxt; cxt 2341 drivers/net/ethernet/qlogic/qed/qed_iwarp.c qed_iwarp_ll2_comp_syn_pkt(void *cxt, struct qed_ll2_comp_rx_data *data) cxt 2347 drivers/net/ethernet/qlogic/qed/qed_iwarp.c struct qed_hwfn *p_hwfn = cxt; cxt 2453 drivers/net/ethernet/qlogic/qed/qed_iwarp.c static void qed_iwarp_ll2_rel_rx_pkt(void *cxt, u8 connection_handle, cxt 2458 drivers/net/ethernet/qlogic/qed/qed_iwarp.c struct qed_hwfn *p_hwfn = cxt; cxt 2465 drivers/net/ethernet/qlogic/qed/qed_iwarp.c static void qed_iwarp_ll2_comp_tx_pkt(void *cxt, u8 connection_handle, cxt 2471 drivers/net/ethernet/qlogic/qed/qed_iwarp.c struct qed_hwfn *p_hwfn = cxt; cxt 2491 drivers/net/ethernet/qlogic/qed/qed_iwarp.c static void qed_iwarp_ll2_rel_tx_pkt(void *cxt, u8 connection_handle, cxt 2496 drivers/net/ethernet/qlogic/qed/qed_iwarp.c struct qed_hwfn *p_hwfn = cxt; cxt 2520 drivers/net/ethernet/qlogic/qed/qed_iwarp.c qed_iwarp_ll2_slowpath(void *cxt, cxt 2525 drivers/net/ethernet/qlogic/qed/qed_iwarp.c struct qed_hwfn *p_hwfn = cxt; cxt 91 drivers/net/ethernet/qlogic/qed/qed_ll2.c static void qed_ll2b_complete_tx_packet(void *cxt, cxt 98 drivers/net/ethernet/qlogic/qed/qed_ll2.c struct qed_hwfn *p_hwfn = cxt; cxt 161 drivers/net/ethernet/qlogic/qed/qed_ll2.c static void qed_ll2b_complete_rx_packet(void *cxt, cxt 164 drivers/net/ethernet/qlogic/qed/qed_ll2.c struct qed_hwfn *p_hwfn = cxt; cxt 1335 drivers/net/ethernet/qlogic/qed/qed_ll2.c int qed_ll2_acquire_connection(void *cxt, struct qed_ll2_acquire_data *data) cxt 1337 drivers/net/ethernet/qlogic/qed/qed_ll2.c struct qed_hwfn *p_hwfn = cxt; cxt 1478 drivers/net/ethernet/qlogic/qed/qed_ll2.c int qed_ll2_establish_connection(void *cxt, u8 connection_handle) cxt 1480 drivers/net/ethernet/qlogic/qed/qed_ll2.c struct qed_hwfn *p_hwfn = cxt; cxt 1627 drivers/net/ethernet/qlogic/qed/qed_ll2.c int qed_ll2_post_rx_buffer(void *cxt, cxt 1632 drivers/net/ethernet/qlogic/qed/qed_ll2.c struct qed_hwfn *p_hwfn = cxt; cxt 1844 drivers/net/ethernet/qlogic/qed/qed_ll2.c int qed_ll2_prepare_tx_packet(void *cxt, cxt 1849 drivers/net/ethernet/qlogic/qed/qed_ll2.c struct qed_hwfn *p_hwfn = cxt; cxt 1896 drivers/net/ethernet/qlogic/qed/qed_ll2.c int qed_ll2_set_fragment_of_tx_packet(void *cxt, cxt 1901 drivers/net/ethernet/qlogic/qed/qed_ll2.c struct qed_hwfn *p_hwfn = cxt; cxt 1936 drivers/net/ethernet/qlogic/qed/qed_ll2.c int qed_ll2_terminate_connection(void *cxt, u8 connection_handle) cxt 1938 drivers/net/ethernet/qlogic/qed/qed_ll2.c struct qed_hwfn *p_hwfn = cxt; cxt 2013 drivers/net/ethernet/qlogic/qed/qed_ll2.c void qed_ll2_release_connection(void *cxt, u8 connection_handle) cxt 2015 drivers/net/ethernet/qlogic/qed/qed_ll2.c struct qed_hwfn *p_hwfn = cxt; cxt 2159 drivers/net/ethernet/qlogic/qed/qed_ll2.c static int __qed_ll2_get_stats(void *cxt, u8 connection_handle, cxt 2162 drivers/net/ethernet/qlogic/qed/qed_ll2.c struct qed_hwfn *p_hwfn = cxt; cxt 2193 drivers/net/ethernet/qlogic/qed/qed_ll2.c int qed_ll2_get_stats(void *cxt, cxt 2197 drivers/net/ethernet/qlogic/qed/qed_ll2.c return __qed_ll2_get_stats(cxt, connection_handle, p_stats); cxt 2200 drivers/net/ethernet/qlogic/qed/qed_ll2.c static void qed_ll2b_release_rx_packet(void *cxt, cxt 2206 drivers/net/ethernet/qlogic/qed/qed_ll2.c struct qed_hwfn *p_hwfn = cxt; cxt 141 drivers/net/ethernet/qlogic/qed/qed_ll2.h int qed_ll2_acquire_connection(void *cxt, struct qed_ll2_acquire_data *data); cxt 154 drivers/net/ethernet/qlogic/qed/qed_ll2.h int qed_ll2_establish_connection(void *cxt, u8 connection_handle); cxt 168 drivers/net/ethernet/qlogic/qed/qed_ll2.h int qed_ll2_post_rx_buffer(void *cxt, cxt 184 drivers/net/ethernet/qlogic/qed/qed_ll2.h int qed_ll2_prepare_tx_packet(void *cxt, cxt 197 drivers/net/ethernet/qlogic/qed/qed_ll2.h void qed_ll2_release_connection(void *cxt, u8 connection_handle); cxt 213 drivers/net/ethernet/qlogic/qed/qed_ll2.h int qed_ll2_set_fragment_of_tx_packet(void *cxt, cxt 228 drivers/net/ethernet/qlogic/qed/qed_ll2.h int qed_ll2_terminate_connection(void *cxt, u8 connection_handle); cxt 241 drivers/net/ethernet/qlogic/qed/qed_ll2.h int qed_ll2_get_stats(void *cxt, cxt 490 drivers/scsi/qedi/qedi_fw_api.c struct e4_iscsi_task_context *cxt; cxt 501 drivers/scsi/qedi/qedi_fw_api.c cxt = task_params->context; cxt 505 drivers/scsi/qedi/qedi_fw_api.c set_local_completion_context(cxt); cxt 509 drivers/scsi/qedi/qedi_fw_api.c cxt->ystorm_st_context.pdu_hdr.r2t.desired_data_trns_len = val; cxt 510 drivers/scsi/qedi/qedi_fw_api.c cxt->mstorm_st_context.expected_itt = cxt 514 drivers/scsi/qedi/qedi_fw_api.c cxt->ystorm_st_context.pdu_hdr.cmd.expected_transfer_length = cxt 516 drivers/scsi/qedi/qedi_fw_api.c init_initiator_rw_cdb_ystorm_context(&cxt->ystorm_st_context, cxt 519 drivers/scsi/qedi/qedi_fw_api.c cxt->mstorm_st_context.sense_db.lo = val; cxt 522 drivers/scsi/qedi/qedi_fw_api.c cxt->mstorm_st_context.sense_db.hi = val; cxt 526 drivers/scsi/qedi/qedi_fw_api.c init_dif_context_flags(&cxt->ystorm_st_context.state.dif_flags, cxt 528 drivers/scsi/qedi/qedi_fw_api.c init_dif_context_flags(&cxt->ustorm_st_context.dif_flags, cxt 530 drivers/scsi/qedi/qedi_fw_api.c init_scsi_sgl_context(&cxt->ystorm_st_context.state.sgl_params, cxt 531 drivers/scsi/qedi/qedi_fw_api.c &cxt->ystorm_st_context.state.data_desc, cxt 542 drivers/scsi/qedi/qedi_fw_api.c SET_FIELD(cxt->ystorm_st_context.state.flags, cxt 546 drivers/scsi/qedi/qedi_fw_api.c init_dif_context_flags(&cxt->mstorm_st_context.dif_flags, cxt 548 drivers/scsi/qedi/qedi_fw_api.c init_scsi_sgl_context(&cxt->mstorm_st_context.sgl_params, cxt 549 drivers/scsi/qedi/qedi_fw_api.c &cxt->mstorm_st_context.data_desc, cxt 556 drivers/scsi/qedi/qedi_fw_api.c cxt->mstorm_st_context.rem_task_size = cpu_to_le32(task_size); cxt 618 drivers/scsi/qedi/qedi_fw_api.c struct e4_iscsi_task_context *cxt; cxt 620 drivers/scsi/qedi/qedi_fw_api.c cxt = task_params->context; cxt 626 drivers/scsi/qedi/qedi_fw_api.c init_ustorm_task_contexts(&cxt->ustorm_st_context, cxt 627 drivers/scsi/qedi/qedi_fw_api.c &cxt->ustorm_ag_context, cxt 635 drivers/scsi/qedi/qedi_fw_api.c init_scsi_sgl_context(&cxt->ystorm_st_context.state.sgl_params, cxt 636 drivers/scsi/qedi/qedi_fw_api.c &cxt->ystorm_st_context.state.data_desc, cxt 640 drivers/scsi/qedi/qedi_fw_api.c init_scsi_sgl_context(&cxt->mstorm_st_context.sgl_params, cxt 641 drivers/scsi/qedi/qedi_fw_api.c &cxt->mstorm_st_context.data_desc, cxt 644 drivers/scsi/qedi/qedi_fw_api.c cxt->mstorm_st_context.rem_task_size = cxt 660 drivers/scsi/qedi/qedi_fw_api.c struct e4_iscsi_task_context *cxt; cxt 662 drivers/scsi/qedi/qedi_fw_api.c cxt = task_params->context; cxt 672 drivers/scsi/qedi/qedi_fw_api.c init_scsi_sgl_context(&cxt->ystorm_st_context.state.sgl_params, cxt 673 drivers/scsi/qedi/qedi_fw_api.c &cxt->ystorm_st_context.state.data_desc, cxt 677 drivers/scsi/qedi/qedi_fw_api.c init_scsi_sgl_context(&cxt->mstorm_st_context.sgl_params, cxt 678 drivers/scsi/qedi/qedi_fw_api.c &cxt->mstorm_st_context.data_desc, cxt 681 drivers/scsi/qedi/qedi_fw_api.c init_ustorm_task_contexts(&cxt->ustorm_st_context, cxt 682 drivers/scsi/qedi/qedi_fw_api.c &cxt->ustorm_ag_context, cxt 689 drivers/scsi/qedi/qedi_fw_api.c cxt->mstorm_st_context.rem_task_size = cxt 706 drivers/scsi/qedi/qedi_fw_api.c struct e4_iscsi_task_context *cxt; cxt 708 drivers/scsi/qedi/qedi_fw_api.c cxt = task_params->context; cxt 715 drivers/scsi/qedi/qedi_fw_api.c init_scsi_sgl_context(&cxt->ystorm_st_context.state.sgl_params, cxt 716 drivers/scsi/qedi/qedi_fw_api.c &cxt->ystorm_st_context.state.data_desc, cxt 720 drivers/scsi/qedi/qedi_fw_api.c init_scsi_sgl_context(&cxt->mstorm_st_context.sgl_params, cxt 721 drivers/scsi/qedi/qedi_fw_api.c &cxt->mstorm_st_context.data_desc, cxt 724 drivers/scsi/qedi/qedi_fw_api.c init_ustorm_task_contexts(&cxt->ustorm_st_context, cxt 725 drivers/scsi/qedi/qedi_fw_api.c &cxt->ustorm_ag_context, cxt 732 drivers/scsi/qedi/qedi_fw_api.c cxt->mstorm_st_context.rem_task_size = cxt 761 drivers/scsi/qedi/qedi_fw_api.c struct e4_iscsi_task_context *cxt; cxt 763 drivers/scsi/qedi/qedi_fw_api.c cxt = task_params->context; cxt 770 drivers/scsi/qedi/qedi_fw_api.c init_scsi_sgl_context(&cxt->ystorm_st_context.state.sgl_params, cxt 771 drivers/scsi/qedi/qedi_fw_api.c &cxt->ystorm_st_context.state.data_desc, cxt 775 drivers/scsi/qedi/qedi_fw_api.c init_scsi_sgl_context(&cxt->mstorm_st_context.sgl_params, cxt 776 drivers/scsi/qedi/qedi_fw_api.c &cxt->mstorm_st_context.data_desc, cxt 779 drivers/scsi/qedi/qedi_fw_api.c cxt->mstorm_st_context.rem_task_size = cxt 783 drivers/scsi/qedi/qedi_fw_api.c init_ustorm_task_contexts(&cxt->ustorm_st_context, cxt 784 drivers/scsi/qedi/qedi_fw_api.c &cxt->ustorm_ag_context, cxt 102 fs/pstore/ram.c struct ramoops_context *cxt = psi->data; cxt 104 fs/pstore/ram.c cxt->dump_read_cnt = 0; cxt 105 fs/pstore/ram.c cxt->console_read_cnt = 0; cxt 106 fs/pstore/ram.c cxt->ftrace_read_cnt = 0; cxt 107 fs/pstore/ram.c cxt->pmsg_read_cnt = 0; cxt 226 fs/pstore/ram.c struct ramoops_context *cxt = record->psi->data; cxt 241 fs/pstore/ram.c while (cxt->dump_read_cnt < cxt->max_dump_cnt && !prz) { cxt 242 fs/pstore/ram.c prz = ramoops_get_next_prz(cxt->dprzs, cxt->dump_read_cnt++, cxt 257 fs/pstore/ram.c if (!prz_ok(prz) && !cxt->console_read_cnt++) cxt 258 fs/pstore/ram.c prz = ramoops_get_next_prz(&cxt->cprz, 0 /* single */, record); cxt 260 fs/pstore/ram.c if (!prz_ok(prz) && !cxt->pmsg_read_cnt++) cxt 261 fs/pstore/ram.c prz = ramoops_get_next_prz(&cxt->mprz, 0 /* single */, record); cxt 265 fs/pstore/ram.c if (!(cxt->flags & RAMOOPS_FLAG_FTRACE_PER_CPU) && cxt 266 fs/pstore/ram.c !cxt->ftrace_read_cnt++) { cxt 267 fs/pstore/ram.c prz = ramoops_get_next_prz(cxt->fprzs, 0 /* single */, cxt 283 fs/pstore/ram.c while (cxt->ftrace_read_cnt < cxt->max_ftrace_cnt) { cxt 284 fs/pstore/ram.c prz_next = ramoops_get_next_prz(cxt->fprzs, cxt 285 fs/pstore/ram.c cxt->ftrace_read_cnt++, record); cxt 351 fs/pstore/ram.c struct ramoops_context *cxt = record->psi->data; cxt 356 fs/pstore/ram.c if (!cxt->cprz) cxt 358 fs/pstore/ram.c persistent_ram_write(cxt->cprz, record->buf, record->size); cxt 363 fs/pstore/ram.c if (!cxt->fprzs) cxt 368 fs/pstore/ram.c if (cxt->flags & RAMOOPS_FLAG_FTRACE_PER_CPU) cxt 373 fs/pstore/ram.c persistent_ram_write(cxt->fprzs[zonenum], record->buf, cxt 393 fs/pstore/ram.c if (record->reason == KMSG_DUMP_OOPS && !cxt->dump_oops) cxt 405 fs/pstore/ram.c if (!cxt->dprzs) cxt 408 fs/pstore/ram.c prz = cxt->dprzs[cxt->dump_write_cnt]; cxt 431 fs/pstore/ram.c cxt->dump_write_cnt = (cxt->dump_write_cnt + 1) % cxt->max_dump_cnt; cxt 440 fs/pstore/ram.c struct ramoops_context *cxt = record->psi->data; cxt 442 fs/pstore/ram.c if (!cxt->mprz) cxt 444 fs/pstore/ram.c return persistent_ram_write_user(cxt->mprz, buf, record->size); cxt 452 fs/pstore/ram.c struct ramoops_context *cxt = record->psi->data; cxt 457 fs/pstore/ram.c if (record->id >= cxt->max_dump_cnt) cxt 459 fs/pstore/ram.c prz = cxt->dprzs[record->id]; cxt 462 fs/pstore/ram.c prz = cxt->cprz; cxt 465 fs/pstore/ram.c if (record->id >= cxt->max_ftrace_cnt) cxt 467 fs/pstore/ram.c prz = cxt->fprzs[record->id]; cxt 470 fs/pstore/ram.c prz = cxt->mprz; cxt 494 fs/pstore/ram.c static void ramoops_free_przs(struct ramoops_context *cxt) cxt 499 fs/pstore/ram.c if (cxt->dprzs) { cxt 500 fs/pstore/ram.c for (i = 0; i < cxt->max_dump_cnt; i++) cxt 501 fs/pstore/ram.c persistent_ram_free(cxt->dprzs[i]); cxt 503 fs/pstore/ram.c kfree(cxt->dprzs); cxt 504 fs/pstore/ram.c cxt->max_dump_cnt = 0; cxt 508 fs/pstore/ram.c if (cxt->fprzs) { cxt 509 fs/pstore/ram.c for (i = 0; i < cxt->max_ftrace_cnt; i++) cxt 510 fs/pstore/ram.c persistent_ram_free(cxt->fprzs[i]); cxt 511 fs/pstore/ram.c kfree(cxt->fprzs); cxt 512 fs/pstore/ram.c cxt->max_ftrace_cnt = 0; cxt 517 fs/pstore/ram.c struct device *dev, struct ramoops_context *cxt, cxt 557 fs/pstore/ram.c if (*paddr + mem_sz - cxt->phys_addr > cxt->size) { cxt 561 fs/pstore/ram.c cxt->size, (unsigned long long)cxt->phys_addr); cxt 584 fs/pstore/ram.c &cxt->ecc_info, cxt 585 fs/pstore/ram.c cxt->memtype, flags, label); cxt 613 fs/pstore/ram.c struct device *dev, struct ramoops_context *cxt, cxt 622 fs/pstore/ram.c if (*paddr + sz - cxt->phys_addr > cxt->size) { cxt 625 fs/pstore/ram.c cxt->size, (unsigned long long)cxt->phys_addr); cxt 630 fs/pstore/ram.c *prz = persistent_ram_new(*paddr, sz, sig, &cxt->ecc_info, cxt 631 fs/pstore/ram.c cxt->memtype, PRZ_FLAG_ZAP_OLD, label); cxt 736 fs/pstore/ram.c struct ramoops_context *cxt = &oops_cxt; cxt 745 fs/pstore/ram.c if (cxt->max_dump_cnt) { cxt 781 fs/pstore/ram.c cxt->size = pdata->mem_size; cxt 782 fs/pstore/ram.c cxt->phys_addr = pdata->mem_address; cxt 783 fs/pstore/ram.c cxt->memtype = pdata->mem_type; cxt 784 fs/pstore/ram.c cxt->record_size = pdata->record_size; cxt 785 fs/pstore/ram.c cxt->console_size = pdata->console_size; cxt 786 fs/pstore/ram.c cxt->ftrace_size = pdata->ftrace_size; cxt 787 fs/pstore/ram.c cxt->pmsg_size = pdata->pmsg_size; cxt 788 fs/pstore/ram.c cxt->dump_oops = pdata->dump_oops; cxt 789 fs/pstore/ram.c cxt->flags = pdata->flags; cxt 790 fs/pstore/ram.c cxt->ecc_info = pdata->ecc_info; cxt 792 fs/pstore/ram.c paddr = cxt->phys_addr; cxt 794 fs/pstore/ram.c dump_mem_sz = cxt->size - cxt->console_size - cxt->ftrace_size cxt 795 fs/pstore/ram.c - cxt->pmsg_size; cxt 796 fs/pstore/ram.c err = ramoops_init_przs("dmesg", dev, cxt, &cxt->dprzs, &paddr, cxt 797 fs/pstore/ram.c dump_mem_sz, cxt->record_size, cxt 798 fs/pstore/ram.c &cxt->max_dump_cnt, 0, 0); cxt 802 fs/pstore/ram.c err = ramoops_init_prz("console", dev, cxt, &cxt->cprz, &paddr, cxt 803 fs/pstore/ram.c cxt->console_size, 0); cxt 807 fs/pstore/ram.c cxt->max_ftrace_cnt = (cxt->flags & RAMOOPS_FLAG_FTRACE_PER_CPU) cxt 810 fs/pstore/ram.c err = ramoops_init_przs("ftrace", dev, cxt, &cxt->fprzs, &paddr, cxt 811 fs/pstore/ram.c cxt->ftrace_size, -1, cxt 812 fs/pstore/ram.c &cxt->max_ftrace_cnt, LINUX_VERSION_CODE, cxt 813 fs/pstore/ram.c (cxt->flags & RAMOOPS_FLAG_FTRACE_PER_CPU) cxt 818 fs/pstore/ram.c err = ramoops_init_prz("pmsg", dev, cxt, &cxt->mprz, &paddr, cxt 819 fs/pstore/ram.c cxt->pmsg_size, 0); cxt 823 fs/pstore/ram.c cxt->pstore.data = cxt; cxt 830 fs/pstore/ram.c cxt->pstore.flags = 0; cxt 831 fs/pstore/ram.c if (cxt->max_dump_cnt) cxt 832 fs/pstore/ram.c cxt->pstore.flags |= PSTORE_FLAGS_DMESG; cxt 833 fs/pstore/ram.c if (cxt->console_size) cxt 834 fs/pstore/ram.c cxt->pstore.flags |= PSTORE_FLAGS_CONSOLE; cxt 835 fs/pstore/ram.c if (cxt->max_ftrace_cnt) cxt 836 fs/pstore/ram.c cxt->pstore.flags |= PSTORE_FLAGS_FTRACE; cxt 837 fs/pstore/ram.c if (cxt->pmsg_size) cxt 838 fs/pstore/ram.c cxt->pstore.flags |= PSTORE_FLAGS_PMSG; cxt 845 fs/pstore/ram.c if (cxt->pstore.flags & PSTORE_FLAGS_DMESG) { cxt 846 fs/pstore/ram.c cxt->pstore.bufsize = cxt->dprzs[0]->buffer_size; cxt 847 fs/pstore/ram.c cxt->pstore.buf = kzalloc(cxt->pstore.bufsize, GFP_KERNEL); cxt 848 fs/pstore/ram.c if (!cxt->pstore.buf) { cxt 855 fs/pstore/ram.c err = pstore_register(&cxt->pstore); cxt 874 fs/pstore/ram.c cxt->size, (unsigned long long)cxt->phys_addr, cxt 875 fs/pstore/ram.c cxt->ecc_info.ecc_size); cxt 880 fs/pstore/ram.c kfree(cxt->pstore.buf); cxt 882 fs/pstore/ram.c cxt->pstore.bufsize = 0; cxt 883 fs/pstore/ram.c persistent_ram_free(cxt->mprz); cxt 886 fs/pstore/ram.c persistent_ram_free(cxt->cprz); cxt 888 fs/pstore/ram.c ramoops_free_przs(cxt); cxt 895 fs/pstore/ram.c struct ramoops_context *cxt = &oops_cxt; cxt 897 fs/pstore/ram.c pstore_unregister(&cxt->pstore); cxt 899 fs/pstore/ram.c kfree(cxt->pstore.buf); cxt 900 fs/pstore/ram.c cxt->pstore.bufsize = 0; cxt 902 fs/pstore/ram.c persistent_ram_free(cxt->mprz); cxt 903 fs/pstore/ram.c persistent_ram_free(cxt->cprz); cxt 904 fs/pstore/ram.c ramoops_free_przs(cxt); cxt 129 include/linux/qed/qed_ll2_if.h void (*qed_ll2_complete_rx_packet_cb)(void *cxt, cxt 133 include/linux/qed/qed_ll2_if.h void (*qed_ll2_release_rx_packet_cb)(void *cxt, cxt 140 include/linux/qed/qed_ll2_if.h void (*qed_ll2_complete_tx_packet_cb)(void *cxt, cxt 148 include/linux/qed/qed_ll2_if.h void (*qed_ll2_release_tx_packet_cb)(void *cxt, cxt 155 include/linux/qed/qed_ll2_if.h void (*qed_ll2_slowpath_cb)(void *cxt, u8 connection_handle, cxt 99 kernel/locking/locktorture.c static struct lock_torture_cxt cxt = { 0, 0, false, cxt 117 kernel/locking/locktorture.c (cxt.nrealwriters_stress * 2000 * longdelay_ms))) cxt 119 kernel/locking/locktorture.c if (!(torture_random(trsp) % (cxt.nrealwriters_stress * 20000))) cxt 161 kernel/locking/locktorture.c (cxt.nrealwriters_stress * 2000 * longdelay_ms))) cxt 164 kernel/locking/locktorture.c (cxt.nrealwriters_stress * 2 * shortdelay_us))) cxt 166 kernel/locking/locktorture.c if (!(torture_random(trsp) % (cxt.nrealwriters_stress * 20000))) cxt 192 kernel/locking/locktorture.c cxt.cur_ops->flags = flags; cxt 199 kernel/locking/locktorture.c spin_unlock_irqrestore(&torture_spinlock, cxt.cur_ops->flags); cxt 230 kernel/locking/locktorture.c (cxt.nrealwriters_stress * 2000 * longdelay_ms))) cxt 256 kernel/locking/locktorture.c (cxt.nrealreaders_stress * 2000 * longdelay_ms))) cxt 283 kernel/locking/locktorture.c cxt.cur_ops->flags = flags; cxt 290 kernel/locking/locktorture.c write_unlock_irqrestore(&torture_rwlock, cxt.cur_ops->flags); cxt 298 kernel/locking/locktorture.c cxt.cur_ops->flags = flags; cxt 305 kernel/locking/locktorture.c read_unlock_irqrestore(&torture_rwlock, cxt.cur_ops->flags); cxt 333 kernel/locking/locktorture.c (cxt.nrealwriters_stress * 2000 * longdelay_ms))) cxt 337 kernel/locking/locktorture.c if (!(torture_random(trsp) % (cxt.nrealwriters_stress * 20000))) cxt 451 kernel/locking/locktorture.c (cxt.nrealwriters_stress * factor))) { cxt 465 kernel/locking/locktorture.c (cxt.nrealwriters_stress * factor * 2))) { cxt 485 kernel/locking/locktorture.c (cxt.nrealwriters_stress * 2000 * longdelay_ms))) cxt 488 kernel/locking/locktorture.c (cxt.nrealwriters_stress * 2 * shortdelay_us))) cxt 490 kernel/locking/locktorture.c if (!(torture_random(trsp) % (cxt.nrealwriters_stress * 20000))) cxt 524 kernel/locking/locktorture.c (cxt.nrealwriters_stress * 2000 * longdelay_ms))) cxt 528 kernel/locking/locktorture.c if (!(torture_random(trsp) % (cxt.nrealwriters_stress * 20000))) cxt 549 kernel/locking/locktorture.c (cxt.nrealreaders_stress * 2000 * longdelay_ms))) cxt 553 kernel/locking/locktorture.c if (!(torture_random(trsp) % (cxt.nrealreaders_stress * 20000))) cxt 631 kernel/locking/locktorture.c cxt.cur_ops->task_boost(&rand); cxt 632 kernel/locking/locktorture.c cxt.cur_ops->writelock(); cxt 640 kernel/locking/locktorture.c cxt.cur_ops->write_delay(&rand); cxt 642 kernel/locking/locktorture.c cxt.cur_ops->writeunlock(); cxt 647 kernel/locking/locktorture.c cxt.cur_ops->task_boost(NULL); /* reset prio */ cxt 668 kernel/locking/locktorture.c cxt.cur_ops->readlock(); cxt 674 kernel/locking/locktorture.c cxt.cur_ops->read_delay(&rand); cxt 676 kernel/locking/locktorture.c cxt.cur_ops->readunlock(); cxt 695 kernel/locking/locktorture.c n_stress = write ? cxt.nrealwriters_stress : cxt.nrealreaders_stress; cxt 711 kernel/locking/locktorture.c atomic_inc(&cxt.n_lock_torture_errors); cxt 724 kernel/locking/locktorture.c int size = cxt.nrealwriters_stress * 200 + 8192; cxt 727 kernel/locking/locktorture.c if (cxt.cur_ops->readlock) cxt 728 kernel/locking/locktorture.c size += cxt.nrealreaders_stress * 200 + 8192; cxt 737 kernel/locking/locktorture.c __torture_print_stats(buf, cxt.lwsa, true); cxt 741 kernel/locking/locktorture.c if (cxt.cur_ops->readlock) { cxt 749 kernel/locking/locktorture.c __torture_print_stats(buf, cxt.lrsa, false); cxt 780 kernel/locking/locktorture.c torture_type, tag, cxt.debug_lock ? " [debug]": "", cxt 781 kernel/locking/locktorture.c cxt.nrealwriters_stress, cxt.nrealreaders_stress, stat_interval, cxt 799 kernel/locking/locktorture.c if (!cxt.lwsa && !cxt.lrsa) cxt 803 kernel/locking/locktorture.c for (i = 0; i < cxt.nrealwriters_stress; i++) cxt 811 kernel/locking/locktorture.c for (i = 0; i < cxt.nrealreaders_stress; i++) cxt 821 kernel/locking/locktorture.c if (atomic_read(&cxt.n_lock_torture_errors)) cxt 822 kernel/locking/locktorture.c lock_torture_print_module_parms(cxt.cur_ops, cxt 825 kernel/locking/locktorture.c lock_torture_print_module_parms(cxt.cur_ops, cxt 828 kernel/locking/locktorture.c lock_torture_print_module_parms(cxt.cur_ops, cxt 831 kernel/locking/locktorture.c kfree(cxt.lwsa); cxt 832 kernel/locking/locktorture.c cxt.lwsa = NULL; cxt 833 kernel/locking/locktorture.c kfree(cxt.lrsa); cxt 834 kernel/locking/locktorture.c cxt.lrsa = NULL; cxt 862 kernel/locking/locktorture.c cxt.cur_ops = torture_ops[i]; cxt 863 kernel/locking/locktorture.c if (strcmp(torture_type, cxt.cur_ops->name) == 0) cxt 883 kernel/locking/locktorture.c if (cxt.cur_ops->init) cxt 884 kernel/locking/locktorture.c cxt.cur_ops->init(); cxt 887 kernel/locking/locktorture.c cxt.nrealwriters_stress = nwriters_stress; cxt 889 kernel/locking/locktorture.c cxt.nrealwriters_stress = 2 * num_online_cpus(); cxt 893 kernel/locking/locktorture.c cxt.debug_lock = true; cxt 897 kernel/locking/locktorture.c cxt.debug_lock = true; cxt 902 kernel/locking/locktorture.c cxt.debug_lock = true; cxt 908 kernel/locking/locktorture.c cxt.lwsa = kmalloc_array(cxt.nrealwriters_stress, cxt 909 kernel/locking/locktorture.c sizeof(*cxt.lwsa), cxt 911 kernel/locking/locktorture.c if (cxt.lwsa == NULL) { cxt 917 kernel/locking/locktorture.c for (i = 0; i < cxt.nrealwriters_stress; i++) { cxt 918 kernel/locking/locktorture.c cxt.lwsa[i].n_lock_fail = 0; cxt 919 kernel/locking/locktorture.c cxt.lwsa[i].n_lock_acquired = 0; cxt 923 kernel/locking/locktorture.c if (cxt.cur_ops->readlock) { cxt 925 kernel/locking/locktorture.c cxt.nrealreaders_stress = nreaders_stress; cxt 933 kernel/locking/locktorture.c cxt.nrealwriters_stress = num_online_cpus(); cxt 934 kernel/locking/locktorture.c cxt.nrealreaders_stress = cxt.nrealwriters_stress; cxt 939 kernel/locking/locktorture.c cxt.lrsa = kmalloc_array(cxt.nrealreaders_stress, cxt 940 kernel/locking/locktorture.c sizeof(*cxt.lrsa), cxt 942 kernel/locking/locktorture.c if (cxt.lrsa == NULL) { cxt 945 kernel/locking/locktorture.c kfree(cxt.lwsa); cxt 946 kernel/locking/locktorture.c cxt.lwsa = NULL; cxt 950 kernel/locking/locktorture.c for (i = 0; i < cxt.nrealreaders_stress; i++) { cxt 951 kernel/locking/locktorture.c cxt.lrsa[i].n_lock_fail = 0; cxt 952 kernel/locking/locktorture.c cxt.lrsa[i].n_lock_acquired = 0; cxt 957 kernel/locking/locktorture.c lock_torture_print_module_parms(cxt.cur_ops, "Start of test"); cxt 984 kernel/locking/locktorture.c writer_tasks = kcalloc(cxt.nrealwriters_stress, cxt 994 kernel/locking/locktorture.c if (cxt.cur_ops->readlock) { cxt 995 kernel/locking/locktorture.c reader_tasks = kcalloc(cxt.nrealreaders_stress, cxt 1015 kernel/locking/locktorture.c for (i = 0, j = 0; i < cxt.nrealwriters_stress || cxt 1016 kernel/locking/locktorture.c j < cxt.nrealreaders_stress; i++, j++) { cxt 1017 kernel/locking/locktorture.c if (i >= cxt.nrealwriters_stress) cxt 1021 kernel/locking/locktorture.c firsterr = torture_create_kthread(lock_torture_writer, &cxt.lwsa[i], cxt 1027 kernel/locking/locktorture.c if (cxt.cur_ops->readlock == NULL || (j >= cxt.nrealreaders_stress)) cxt 1030 kernel/locking/locktorture.c firsterr = torture_create_kthread(lock_torture_reader, &cxt.lrsa[j], cxt 32 tools/testing/selftests/gpio/gpio-mockup-chardev.c struct libmnt_context *cxt; cxt 38 tools/testing/selftests/gpio/gpio-mockup-chardev.c cxt = mnt_new_context(); cxt 39 tools/testing/selftests/gpio/gpio-mockup-chardev.c if (!cxt) cxt 46 tools/testing/selftests/gpio/gpio-mockup-chardev.c if (mnt_context_get_mtab(cxt, &tb)) cxt 64 tools/testing/selftests/gpio/gpio-mockup-chardev.c mnt_free_context(cxt);