cxt                20 arch/arm64/kvm/hyp/tlb.c 						 struct tlb_inv_context *cxt)
cxt                24 arch/arm64/kvm/hyp/tlb.c 	local_irq_save(cxt->flags);
cxt                36 arch/arm64/kvm/hyp/tlb.c 		val = cxt->tcr = read_sysreg_el1(SYS_TCR);
cxt                39 arch/arm64/kvm/hyp/tlb.c 		val = cxt->sctlr = read_sysreg_el1(SYS_SCTLR);
cxt                64 arch/arm64/kvm/hyp/tlb.c 						  struct tlb_inv_context *cxt)
cxt                71 arch/arm64/kvm/hyp/tlb.c 					     struct tlb_inv_context *cxt)
cxt                74 arch/arm64/kvm/hyp/tlb.c 		__tlb_switch_to_guest_vhe(kvm, cxt);
cxt                76 arch/arm64/kvm/hyp/tlb.c 		__tlb_switch_to_guest_nvhe(kvm, cxt);
cxt                80 arch/arm64/kvm/hyp/tlb.c 						struct tlb_inv_context *cxt)
cxt                92 arch/arm64/kvm/hyp/tlb.c 		write_sysreg_el1(cxt->tcr, SYS_TCR);
cxt                93 arch/arm64/kvm/hyp/tlb.c 		write_sysreg_el1(cxt->sctlr, SYS_SCTLR);
cxt                96 arch/arm64/kvm/hyp/tlb.c 	local_irq_restore(cxt->flags);
cxt               100 arch/arm64/kvm/hyp/tlb.c 						 struct tlb_inv_context *cxt)
cxt               106 arch/arm64/kvm/hyp/tlb.c 					    struct tlb_inv_context *cxt)
cxt               109 arch/arm64/kvm/hyp/tlb.c 		__tlb_switch_to_host_vhe(kvm, cxt);
cxt               111 arch/arm64/kvm/hyp/tlb.c 		__tlb_switch_to_host_nvhe(kvm, cxt);
cxt               116 arch/arm64/kvm/hyp/tlb.c 	struct tlb_inv_context cxt;
cxt               122 arch/arm64/kvm/hyp/tlb.c 	__tlb_switch_to_guest(kvm, &cxt);
cxt               165 arch/arm64/kvm/hyp/tlb.c 	__tlb_switch_to_host(kvm, &cxt);
cxt               170 arch/arm64/kvm/hyp/tlb.c 	struct tlb_inv_context cxt;
cxt               176 arch/arm64/kvm/hyp/tlb.c 	__tlb_switch_to_guest(kvm, &cxt);
cxt               182 arch/arm64/kvm/hyp/tlb.c 	__tlb_switch_to_host(kvm, &cxt);
cxt               188 arch/arm64/kvm/hyp/tlb.c 	struct tlb_inv_context cxt;
cxt               191 arch/arm64/kvm/hyp/tlb.c 	__tlb_switch_to_guest(kvm, &cxt);
cxt               197 arch/arm64/kvm/hyp/tlb.c 	__tlb_switch_to_host(kvm, &cxt);
cxt              2415 drivers/gpu/drm/i915/i915_drv.h int intel_engine_cmd_parser(struct i915_gem_context *cxt,
cxt                67 drivers/infiniband/hw/qedr/qedr_roce_cm.c static void qedr_ll2_complete_tx_packet(void *cxt, u8 connection_handle,
cxt                73 drivers/infiniband/hw/qedr/qedr_roce_cm.c 	struct qedr_dev *dev = (struct qedr_dev *)cxt;
cxt                96 drivers/infiniband/hw/qedr/qedr_roce_cm.c static void qedr_ll2_complete_rx_packet(void *cxt,
cxt                99 drivers/infiniband/hw/qedr/qedr_roce_cm.c 	struct qedr_dev *dev = (struct qedr_dev *)cxt;
cxt               125 drivers/infiniband/hw/qedr/qedr_roce_cm.c static void qedr_ll2_release_rx_packet(void *cxt, u8 connection_handle,
cxt                58 drivers/mtd/mtdoops.c static void mark_page_used(struct mtdoops_context *cxt, int page)
cxt                60 drivers/mtd/mtdoops.c 	set_bit(page, cxt->oops_page_used);
cxt                63 drivers/mtd/mtdoops.c static void mark_page_unused(struct mtdoops_context *cxt, int page)
cxt                65 drivers/mtd/mtdoops.c 	clear_bit(page, cxt->oops_page_used);
cxt                68 drivers/mtd/mtdoops.c static int page_is_used(struct mtdoops_context *cxt, int page)
cxt                70 drivers/mtd/mtdoops.c 	return test_bit(page, cxt->oops_page_used);
cxt                73 drivers/mtd/mtdoops.c static int mtdoops_erase_block(struct mtdoops_context *cxt, int offset)
cxt                75 drivers/mtd/mtdoops.c 	struct mtd_info *mtd = cxt->mtd;
cxt                96 drivers/mtd/mtdoops.c 		mark_page_unused(cxt, page);
cxt               101 drivers/mtd/mtdoops.c static void mtdoops_inc_counter(struct mtdoops_context *cxt)
cxt               103 drivers/mtd/mtdoops.c 	cxt->nextpage++;
cxt               104 drivers/mtd/mtdoops.c 	if (cxt->nextpage >= cxt->oops_pages)
cxt               105 drivers/mtd/mtdoops.c 		cxt->nextpage = 0;
cxt               106 drivers/mtd/mtdoops.c 	cxt->nextcount++;
cxt               107 drivers/mtd/mtdoops.c 	if (cxt->nextcount == 0xffffffff)
cxt               108 drivers/mtd/mtdoops.c 		cxt->nextcount = 0;
cxt               110 drivers/mtd/mtdoops.c 	if (page_is_used(cxt, cxt->nextpage)) {
cxt               111 drivers/mtd/mtdoops.c 		schedule_work(&cxt->work_erase);
cxt               116 drivers/mtd/mtdoops.c 	       cxt->nextpage, cxt->nextcount);
cxt               122 drivers/mtd/mtdoops.c 	struct mtdoops_context *cxt =
cxt               124 drivers/mtd/mtdoops.c 	struct mtd_info *mtd = cxt->mtd;
cxt               131 drivers/mtd/mtdoops.c 	mod = (cxt->nextpage * record_size) % mtd->erasesize;
cxt               133 drivers/mtd/mtdoops.c 		cxt->nextpage = cxt->nextpage + ((mtd->erasesize - mod) / record_size);
cxt               134 drivers/mtd/mtdoops.c 		if (cxt->nextpage >= cxt->oops_pages)
cxt               135 drivers/mtd/mtdoops.c 			cxt->nextpage = 0;
cxt               138 drivers/mtd/mtdoops.c 	while ((ret = mtd_block_isbad(mtd, cxt->nextpage * record_size)) > 0) {
cxt               141 drivers/mtd/mtdoops.c 		       cxt->nextpage * record_size);
cxt               143 drivers/mtd/mtdoops.c 		cxt->nextpage = cxt->nextpage + (mtd->erasesize / record_size);
cxt               144 drivers/mtd/mtdoops.c 		if (cxt->nextpage >= cxt->oops_pages)
cxt               145 drivers/mtd/mtdoops.c 			cxt->nextpage = 0;
cxt               146 drivers/mtd/mtdoops.c 		if (i == cxt->oops_pages / (mtd->erasesize / record_size)) {
cxt               158 drivers/mtd/mtdoops.c 		ret = mtdoops_erase_block(cxt, cxt->nextpage * record_size);
cxt               162 drivers/mtd/mtdoops.c 		       cxt->nextpage, cxt->nextcount);
cxt               167 drivers/mtd/mtdoops.c 		ret = mtd_block_markbad(mtd, cxt->nextpage * record_size);
cxt               176 drivers/mtd/mtdoops.c static void mtdoops_write(struct mtdoops_context *cxt, int panic)
cxt               178 drivers/mtd/mtdoops.c 	struct mtd_info *mtd = cxt->mtd;
cxt               184 drivers/mtd/mtdoops.c 	hdr = cxt->oops_buf;
cxt               185 drivers/mtd/mtdoops.c 	hdr[0] = cxt->nextcount;
cxt               189 drivers/mtd/mtdoops.c 		ret = mtd_panic_write(mtd, cxt->nextpage * record_size,
cxt               190 drivers/mtd/mtdoops.c 				      record_size, &retlen, cxt->oops_buf);
cxt               196 drivers/mtd/mtdoops.c 		ret = mtd_write(mtd, cxt->nextpage * record_size,
cxt               197 drivers/mtd/mtdoops.c 				record_size, &retlen, cxt->oops_buf);
cxt               201 drivers/mtd/mtdoops.c 		       cxt->nextpage * record_size, retlen, record_size, ret);
cxt               202 drivers/mtd/mtdoops.c 	mark_page_used(cxt, cxt->nextpage);
cxt               203 drivers/mtd/mtdoops.c 	memset(cxt->oops_buf, 0xff, record_size);
cxt               205 drivers/mtd/mtdoops.c 	mtdoops_inc_counter(cxt);
cxt               210 drivers/mtd/mtdoops.c 	struct mtdoops_context *cxt =
cxt               213 drivers/mtd/mtdoops.c 	mtdoops_write(cxt, 0);
cxt               216 drivers/mtd/mtdoops.c static void find_next_position(struct mtdoops_context *cxt)
cxt               218 drivers/mtd/mtdoops.c 	struct mtd_info *mtd = cxt->mtd;
cxt               223 drivers/mtd/mtdoops.c 	for (page = 0; page < cxt->oops_pages; page++) {
cxt               227 drivers/mtd/mtdoops.c 		mark_page_used(cxt, page);
cxt               239 drivers/mtd/mtdoops.c 			mark_page_unused(cxt, page);
cxt               258 drivers/mtd/mtdoops.c 		cxt->nextpage = cxt->oops_pages - 1;
cxt               259 drivers/mtd/mtdoops.c 		cxt->nextcount = 0;
cxt               262 drivers/mtd/mtdoops.c 		cxt->nextpage = maxpos;
cxt               263 drivers/mtd/mtdoops.c 		cxt->nextcount = maxcount;
cxt               266 drivers/mtd/mtdoops.c 	mtdoops_inc_counter(cxt);
cxt               272 drivers/mtd/mtdoops.c 	struct mtdoops_context *cxt = container_of(dumper,
cxt               279 drivers/mtd/mtdoops.c 	kmsg_dump_get_buffer(dumper, true, cxt->oops_buf + MTDOOPS_HEADER_SIZE,
cxt               284 drivers/mtd/mtdoops.c 		mtdoops_write(cxt, 1);
cxt               287 drivers/mtd/mtdoops.c 	schedule_work(&cxt->work_write);
cxt               292 drivers/mtd/mtdoops.c 	struct mtdoops_context *cxt = &oops_cxt;
cxt               297 drivers/mtd/mtdoops.c 		cxt->mtd_index = mtd->index;
cxt               299 drivers/mtd/mtdoops.c 	if (mtd->index != cxt->mtd_index || cxt->mtd_index < 0)
cxt               319 drivers/mtd/mtdoops.c 	cxt->oops_page_used =
cxt               323 drivers/mtd/mtdoops.c 	if (!cxt->oops_page_used) {
cxt               328 drivers/mtd/mtdoops.c 	cxt->dump.max_reason = KMSG_DUMP_OOPS;
cxt               329 drivers/mtd/mtdoops.c 	cxt->dump.dump = mtdoops_do_dump;
cxt               330 drivers/mtd/mtdoops.c 	err = kmsg_dump_register(&cxt->dump);
cxt               333 drivers/mtd/mtdoops.c 		vfree(cxt->oops_page_used);
cxt               334 drivers/mtd/mtdoops.c 		cxt->oops_page_used = NULL;
cxt               338 drivers/mtd/mtdoops.c 	cxt->mtd = mtd;
cxt               339 drivers/mtd/mtdoops.c 	cxt->oops_pages = (int)mtd->size / record_size;
cxt               340 drivers/mtd/mtdoops.c 	find_next_position(cxt);
cxt               346 drivers/mtd/mtdoops.c 	struct mtdoops_context *cxt = &oops_cxt;
cxt               348 drivers/mtd/mtdoops.c 	if (mtd->index != cxt->mtd_index || cxt->mtd_index < 0)
cxt               351 drivers/mtd/mtdoops.c 	if (kmsg_dump_unregister(&cxt->dump) < 0)
cxt               354 drivers/mtd/mtdoops.c 	cxt->mtd = NULL;
cxt               355 drivers/mtd/mtdoops.c 	flush_work(&cxt->work_erase);
cxt               356 drivers/mtd/mtdoops.c 	flush_work(&cxt->work_write);
cxt               367 drivers/mtd/mtdoops.c 	struct mtdoops_context *cxt = &oops_cxt;
cxt               385 drivers/mtd/mtdoops.c 	cxt->mtd_index = -1;
cxt               388 drivers/mtd/mtdoops.c 		cxt->mtd_index = mtd_index;
cxt               390 drivers/mtd/mtdoops.c 	cxt->oops_buf = vmalloc(record_size);
cxt               391 drivers/mtd/mtdoops.c 	if (!cxt->oops_buf) {
cxt               395 drivers/mtd/mtdoops.c 	memset(cxt->oops_buf, 0xff, record_size);
cxt               397 drivers/mtd/mtdoops.c 	INIT_WORK(&cxt->work_erase, mtdoops_workfunc_erase);
cxt               398 drivers/mtd/mtdoops.c 	INIT_WORK(&cxt->work_write, mtdoops_workfunc_write);
cxt               406 drivers/mtd/mtdoops.c 	struct mtdoops_context *cxt = &oops_cxt;
cxt               409 drivers/mtd/mtdoops.c 	vfree(cxt->oops_buf);
cxt               410 drivers/mtd/mtdoops.c 	vfree(cxt->oops_page_used);
cxt              5057 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c void bnx2x_set_ctx_validation(struct bnx2x *bp, struct eth_context *cxt,
cxt              5060 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c 	if (!cxt) {
cxt              5061 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c 		BNX2X_ERR("bad context pointer %p\n", cxt);
cxt              5066 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c 	cxt->ustorm_ag_context.cdu_usage =
cxt              5070 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c 	cxt->xstorm_ag_context.cdu_reserved =
cxt              1272 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.h void bnx2x_set_ctx_validation(struct bnx2x *bp, struct eth_context *cxt,
cxt               189 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 	init_p->cxts[0] = q->cxt;
cxt               307 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 	if (bnx2x_vfq(vf, qid, cxt)) {
cxt               308 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 		bnx2x_vfq(vf, qid, cxt)->ustorm_ag_context.cdu_usage = 0;
cxt               309 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 		bnx2x_vfq(vf, qid, cxt)->xstorm_ag_context.cdu_reserved = 0;
cxt              1345 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 		struct hw_dma *cxt = &bp->vfdb->context[i];
cxt              1346 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 		BNX2X_PCI_FREE(cxt->addr, cxt->mapping, cxt->size);
cxt              1375 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 		struct hw_dma *cxt = BP_VF_CXT_PAGE(bp, i);
cxt              1376 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 		cxt->size = min_t(size_t, tot_size, CDU_ILT_PAGE_SZ);
cxt              1378 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 		if (cxt->size) {
cxt              1379 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 			cxt->addr = BNX2X_PCI_ALLOC(&cxt->mapping, cxt->size);
cxt              1380 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 			if (!cxt->addr)
cxt              1383 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 			cxt->addr = NULL;
cxt              1384 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 			cxt->mapping = 0;
cxt              1386 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 		tot_size -= cxt->size;
cxt              2092 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 		q->cxt = &((base_cxt + i)->eth);
cxt              2096 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 		   vf->abs_vfid, i, q->index, q->cid, q->cxt);
cxt                73 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.h 	struct eth_context		*cxt;
cxt              2300 drivers/net/ethernet/qlogic/qed/qed_iwarp.c qed_iwarp_ll2_comp_mpa_pkt(void *cxt, struct qed_ll2_comp_rx_data *data)
cxt              2304 drivers/net/ethernet/qlogic/qed/qed_iwarp.c 	struct qed_hwfn *p_hwfn = cxt;
cxt              2341 drivers/net/ethernet/qlogic/qed/qed_iwarp.c qed_iwarp_ll2_comp_syn_pkt(void *cxt, struct qed_ll2_comp_rx_data *data)
cxt              2347 drivers/net/ethernet/qlogic/qed/qed_iwarp.c 	struct qed_hwfn *p_hwfn = cxt;
cxt              2453 drivers/net/ethernet/qlogic/qed/qed_iwarp.c static void qed_iwarp_ll2_rel_rx_pkt(void *cxt, u8 connection_handle,
cxt              2458 drivers/net/ethernet/qlogic/qed/qed_iwarp.c 	struct qed_hwfn *p_hwfn = cxt;
cxt              2465 drivers/net/ethernet/qlogic/qed/qed_iwarp.c static void qed_iwarp_ll2_comp_tx_pkt(void *cxt, u8 connection_handle,
cxt              2471 drivers/net/ethernet/qlogic/qed/qed_iwarp.c 	struct qed_hwfn *p_hwfn = cxt;
cxt              2491 drivers/net/ethernet/qlogic/qed/qed_iwarp.c static void qed_iwarp_ll2_rel_tx_pkt(void *cxt, u8 connection_handle,
cxt              2496 drivers/net/ethernet/qlogic/qed/qed_iwarp.c 	struct qed_hwfn *p_hwfn = cxt;
cxt              2520 drivers/net/ethernet/qlogic/qed/qed_iwarp.c qed_iwarp_ll2_slowpath(void *cxt,
cxt              2525 drivers/net/ethernet/qlogic/qed/qed_iwarp.c 	struct qed_hwfn *p_hwfn = cxt;
cxt                91 drivers/net/ethernet/qlogic/qed/qed_ll2.c static void qed_ll2b_complete_tx_packet(void *cxt,
cxt                98 drivers/net/ethernet/qlogic/qed/qed_ll2.c 	struct qed_hwfn *p_hwfn = cxt;
cxt               161 drivers/net/ethernet/qlogic/qed/qed_ll2.c static void qed_ll2b_complete_rx_packet(void *cxt,
cxt               164 drivers/net/ethernet/qlogic/qed/qed_ll2.c 	struct qed_hwfn *p_hwfn = cxt;
cxt              1335 drivers/net/ethernet/qlogic/qed/qed_ll2.c int qed_ll2_acquire_connection(void *cxt, struct qed_ll2_acquire_data *data)
cxt              1337 drivers/net/ethernet/qlogic/qed/qed_ll2.c 	struct qed_hwfn *p_hwfn = cxt;
cxt              1478 drivers/net/ethernet/qlogic/qed/qed_ll2.c int qed_ll2_establish_connection(void *cxt, u8 connection_handle)
cxt              1480 drivers/net/ethernet/qlogic/qed/qed_ll2.c 	struct qed_hwfn *p_hwfn = cxt;
cxt              1627 drivers/net/ethernet/qlogic/qed/qed_ll2.c int qed_ll2_post_rx_buffer(void *cxt,
cxt              1632 drivers/net/ethernet/qlogic/qed/qed_ll2.c 	struct qed_hwfn *p_hwfn = cxt;
cxt              1844 drivers/net/ethernet/qlogic/qed/qed_ll2.c int qed_ll2_prepare_tx_packet(void *cxt,
cxt              1849 drivers/net/ethernet/qlogic/qed/qed_ll2.c 	struct qed_hwfn *p_hwfn = cxt;
cxt              1896 drivers/net/ethernet/qlogic/qed/qed_ll2.c int qed_ll2_set_fragment_of_tx_packet(void *cxt,
cxt              1901 drivers/net/ethernet/qlogic/qed/qed_ll2.c 	struct qed_hwfn *p_hwfn = cxt;
cxt              1936 drivers/net/ethernet/qlogic/qed/qed_ll2.c int qed_ll2_terminate_connection(void *cxt, u8 connection_handle)
cxt              1938 drivers/net/ethernet/qlogic/qed/qed_ll2.c 	struct qed_hwfn *p_hwfn = cxt;
cxt              2013 drivers/net/ethernet/qlogic/qed/qed_ll2.c void qed_ll2_release_connection(void *cxt, u8 connection_handle)
cxt              2015 drivers/net/ethernet/qlogic/qed/qed_ll2.c 	struct qed_hwfn *p_hwfn = cxt;
cxt              2159 drivers/net/ethernet/qlogic/qed/qed_ll2.c static int __qed_ll2_get_stats(void *cxt, u8 connection_handle,
cxt              2162 drivers/net/ethernet/qlogic/qed/qed_ll2.c 	struct qed_hwfn *p_hwfn = cxt;
cxt              2193 drivers/net/ethernet/qlogic/qed/qed_ll2.c int qed_ll2_get_stats(void *cxt,
cxt              2197 drivers/net/ethernet/qlogic/qed/qed_ll2.c 	return __qed_ll2_get_stats(cxt, connection_handle, p_stats);
cxt              2200 drivers/net/ethernet/qlogic/qed/qed_ll2.c static void qed_ll2b_release_rx_packet(void *cxt,
cxt              2206 drivers/net/ethernet/qlogic/qed/qed_ll2.c 	struct qed_hwfn *p_hwfn = cxt;
cxt               141 drivers/net/ethernet/qlogic/qed/qed_ll2.h int qed_ll2_acquire_connection(void *cxt, struct qed_ll2_acquire_data *data);
cxt               154 drivers/net/ethernet/qlogic/qed/qed_ll2.h int qed_ll2_establish_connection(void *cxt, u8 connection_handle);
cxt               168 drivers/net/ethernet/qlogic/qed/qed_ll2.h int qed_ll2_post_rx_buffer(void *cxt,
cxt               184 drivers/net/ethernet/qlogic/qed/qed_ll2.h int qed_ll2_prepare_tx_packet(void *cxt,
cxt               197 drivers/net/ethernet/qlogic/qed/qed_ll2.h void qed_ll2_release_connection(void *cxt, u8 connection_handle);
cxt               213 drivers/net/ethernet/qlogic/qed/qed_ll2.h int qed_ll2_set_fragment_of_tx_packet(void *cxt,
cxt               228 drivers/net/ethernet/qlogic/qed/qed_ll2.h int qed_ll2_terminate_connection(void *cxt, u8 connection_handle);
cxt               241 drivers/net/ethernet/qlogic/qed/qed_ll2.h int qed_ll2_get_stats(void *cxt,
cxt               490 drivers/scsi/qedi/qedi_fw_api.c 	struct e4_iscsi_task_context *cxt;
cxt               501 drivers/scsi/qedi/qedi_fw_api.c 	cxt = task_params->context;
cxt               505 drivers/scsi/qedi/qedi_fw_api.c 		set_local_completion_context(cxt);
cxt               509 drivers/scsi/qedi/qedi_fw_api.c 		cxt->ystorm_st_context.pdu_hdr.r2t.desired_data_trns_len = val;
cxt               510 drivers/scsi/qedi/qedi_fw_api.c 		cxt->mstorm_st_context.expected_itt =
cxt               514 drivers/scsi/qedi/qedi_fw_api.c 		cxt->ystorm_st_context.pdu_hdr.cmd.expected_transfer_length =
cxt               516 drivers/scsi/qedi/qedi_fw_api.c 		init_initiator_rw_cdb_ystorm_context(&cxt->ystorm_st_context,
cxt               519 drivers/scsi/qedi/qedi_fw_api.c 		cxt->mstorm_st_context.sense_db.lo = val;
cxt               522 drivers/scsi/qedi/qedi_fw_api.c 		cxt->mstorm_st_context.sense_db.hi = val;
cxt               526 drivers/scsi/qedi/qedi_fw_api.c 		init_dif_context_flags(&cxt->ystorm_st_context.state.dif_flags,
cxt               528 drivers/scsi/qedi/qedi_fw_api.c 		init_dif_context_flags(&cxt->ustorm_st_context.dif_flags,
cxt               530 drivers/scsi/qedi/qedi_fw_api.c 		init_scsi_sgl_context(&cxt->ystorm_st_context.state.sgl_params,
cxt               531 drivers/scsi/qedi/qedi_fw_api.c 				      &cxt->ystorm_st_context.state.data_desc,
cxt               542 drivers/scsi/qedi/qedi_fw_api.c 			SET_FIELD(cxt->ystorm_st_context.state.flags,
cxt               546 drivers/scsi/qedi/qedi_fw_api.c 		init_dif_context_flags(&cxt->mstorm_st_context.dif_flags,
cxt               548 drivers/scsi/qedi/qedi_fw_api.c 		init_scsi_sgl_context(&cxt->mstorm_st_context.sgl_params,
cxt               549 drivers/scsi/qedi/qedi_fw_api.c 				      &cxt->mstorm_st_context.data_desc,
cxt               556 drivers/scsi/qedi/qedi_fw_api.c 		cxt->mstorm_st_context.rem_task_size = cpu_to_le32(task_size);
cxt               618 drivers/scsi/qedi/qedi_fw_api.c 	struct e4_iscsi_task_context *cxt;
cxt               620 drivers/scsi/qedi/qedi_fw_api.c 	cxt = task_params->context;
cxt               626 drivers/scsi/qedi/qedi_fw_api.c 	init_ustorm_task_contexts(&cxt->ustorm_st_context,
cxt               627 drivers/scsi/qedi/qedi_fw_api.c 				  &cxt->ustorm_ag_context,
cxt               635 drivers/scsi/qedi/qedi_fw_api.c 		init_scsi_sgl_context(&cxt->ystorm_st_context.state.sgl_params,
cxt               636 drivers/scsi/qedi/qedi_fw_api.c 				      &cxt->ystorm_st_context.state.data_desc,
cxt               640 drivers/scsi/qedi/qedi_fw_api.c 		init_scsi_sgl_context(&cxt->mstorm_st_context.sgl_params,
cxt               641 drivers/scsi/qedi/qedi_fw_api.c 				      &cxt->mstorm_st_context.data_desc,
cxt               644 drivers/scsi/qedi/qedi_fw_api.c 	cxt->mstorm_st_context.rem_task_size =
cxt               660 drivers/scsi/qedi/qedi_fw_api.c 	struct e4_iscsi_task_context *cxt;
cxt               662 drivers/scsi/qedi/qedi_fw_api.c 	cxt = task_params->context;
cxt               672 drivers/scsi/qedi/qedi_fw_api.c 		init_scsi_sgl_context(&cxt->ystorm_st_context.state.sgl_params,
cxt               673 drivers/scsi/qedi/qedi_fw_api.c 				      &cxt->ystorm_st_context.state.data_desc,
cxt               677 drivers/scsi/qedi/qedi_fw_api.c 		init_scsi_sgl_context(&cxt->mstorm_st_context.sgl_params,
cxt               678 drivers/scsi/qedi/qedi_fw_api.c 				      &cxt->mstorm_st_context.data_desc,
cxt               681 drivers/scsi/qedi/qedi_fw_api.c 	init_ustorm_task_contexts(&cxt->ustorm_st_context,
cxt               682 drivers/scsi/qedi/qedi_fw_api.c 				  &cxt->ustorm_ag_context,
cxt               689 drivers/scsi/qedi/qedi_fw_api.c 	cxt->mstorm_st_context.rem_task_size =
cxt               706 drivers/scsi/qedi/qedi_fw_api.c 	struct e4_iscsi_task_context *cxt;
cxt               708 drivers/scsi/qedi/qedi_fw_api.c 	cxt = task_params->context;
cxt               715 drivers/scsi/qedi/qedi_fw_api.c 		init_scsi_sgl_context(&cxt->ystorm_st_context.state.sgl_params,
cxt               716 drivers/scsi/qedi/qedi_fw_api.c 				      &cxt->ystorm_st_context.state.data_desc,
cxt               720 drivers/scsi/qedi/qedi_fw_api.c 		init_scsi_sgl_context(&cxt->mstorm_st_context.sgl_params,
cxt               721 drivers/scsi/qedi/qedi_fw_api.c 				      &cxt->mstorm_st_context.data_desc,
cxt               724 drivers/scsi/qedi/qedi_fw_api.c 	init_ustorm_task_contexts(&cxt->ustorm_st_context,
cxt               725 drivers/scsi/qedi/qedi_fw_api.c 				  &cxt->ustorm_ag_context,
cxt               732 drivers/scsi/qedi/qedi_fw_api.c 	cxt->mstorm_st_context.rem_task_size =
cxt               761 drivers/scsi/qedi/qedi_fw_api.c 	struct e4_iscsi_task_context *cxt;
cxt               763 drivers/scsi/qedi/qedi_fw_api.c 	cxt = task_params->context;
cxt               770 drivers/scsi/qedi/qedi_fw_api.c 		init_scsi_sgl_context(&cxt->ystorm_st_context.state.sgl_params,
cxt               771 drivers/scsi/qedi/qedi_fw_api.c 				      &cxt->ystorm_st_context.state.data_desc,
cxt               775 drivers/scsi/qedi/qedi_fw_api.c 		init_scsi_sgl_context(&cxt->mstorm_st_context.sgl_params,
cxt               776 drivers/scsi/qedi/qedi_fw_api.c 				      &cxt->mstorm_st_context.data_desc,
cxt               779 drivers/scsi/qedi/qedi_fw_api.c 	cxt->mstorm_st_context.rem_task_size =
cxt               783 drivers/scsi/qedi/qedi_fw_api.c 	init_ustorm_task_contexts(&cxt->ustorm_st_context,
cxt               784 drivers/scsi/qedi/qedi_fw_api.c 				  &cxt->ustorm_ag_context,
cxt               102 fs/pstore/ram.c 	struct ramoops_context *cxt = psi->data;
cxt               104 fs/pstore/ram.c 	cxt->dump_read_cnt = 0;
cxt               105 fs/pstore/ram.c 	cxt->console_read_cnt = 0;
cxt               106 fs/pstore/ram.c 	cxt->ftrace_read_cnt = 0;
cxt               107 fs/pstore/ram.c 	cxt->pmsg_read_cnt = 0;
cxt               226 fs/pstore/ram.c 	struct ramoops_context *cxt = record->psi->data;
cxt               241 fs/pstore/ram.c 	while (cxt->dump_read_cnt < cxt->max_dump_cnt && !prz) {
cxt               242 fs/pstore/ram.c 		prz = ramoops_get_next_prz(cxt->dprzs, cxt->dump_read_cnt++,
cxt               257 fs/pstore/ram.c 	if (!prz_ok(prz) && !cxt->console_read_cnt++)
cxt               258 fs/pstore/ram.c 		prz = ramoops_get_next_prz(&cxt->cprz, 0 /* single */, record);
cxt               260 fs/pstore/ram.c 	if (!prz_ok(prz) && !cxt->pmsg_read_cnt++)
cxt               261 fs/pstore/ram.c 		prz = ramoops_get_next_prz(&cxt->mprz, 0 /* single */, record);
cxt               265 fs/pstore/ram.c 		if (!(cxt->flags & RAMOOPS_FLAG_FTRACE_PER_CPU) &&
cxt               266 fs/pstore/ram.c 		    !cxt->ftrace_read_cnt++) {
cxt               267 fs/pstore/ram.c 			prz = ramoops_get_next_prz(cxt->fprzs, 0 /* single */,
cxt               283 fs/pstore/ram.c 			while (cxt->ftrace_read_cnt < cxt->max_ftrace_cnt) {
cxt               284 fs/pstore/ram.c 				prz_next = ramoops_get_next_prz(cxt->fprzs,
cxt               285 fs/pstore/ram.c 						cxt->ftrace_read_cnt++, record);
cxt               351 fs/pstore/ram.c 	struct ramoops_context *cxt = record->psi->data;
cxt               356 fs/pstore/ram.c 		if (!cxt->cprz)
cxt               358 fs/pstore/ram.c 		persistent_ram_write(cxt->cprz, record->buf, record->size);
cxt               363 fs/pstore/ram.c 		if (!cxt->fprzs)
cxt               368 fs/pstore/ram.c 		if (cxt->flags & RAMOOPS_FLAG_FTRACE_PER_CPU)
cxt               373 fs/pstore/ram.c 		persistent_ram_write(cxt->fprzs[zonenum], record->buf,
cxt               393 fs/pstore/ram.c 	if (record->reason == KMSG_DUMP_OOPS && !cxt->dump_oops)
cxt               405 fs/pstore/ram.c 	if (!cxt->dprzs)
cxt               408 fs/pstore/ram.c 	prz = cxt->dprzs[cxt->dump_write_cnt];
cxt               431 fs/pstore/ram.c 	cxt->dump_write_cnt = (cxt->dump_write_cnt + 1) % cxt->max_dump_cnt;
cxt               440 fs/pstore/ram.c 		struct ramoops_context *cxt = record->psi->data;
cxt               442 fs/pstore/ram.c 		if (!cxt->mprz)
cxt               444 fs/pstore/ram.c 		return persistent_ram_write_user(cxt->mprz, buf, record->size);
cxt               452 fs/pstore/ram.c 	struct ramoops_context *cxt = record->psi->data;
cxt               457 fs/pstore/ram.c 		if (record->id >= cxt->max_dump_cnt)
cxt               459 fs/pstore/ram.c 		prz = cxt->dprzs[record->id];
cxt               462 fs/pstore/ram.c 		prz = cxt->cprz;
cxt               465 fs/pstore/ram.c 		if (record->id >= cxt->max_ftrace_cnt)
cxt               467 fs/pstore/ram.c 		prz = cxt->fprzs[record->id];
cxt               470 fs/pstore/ram.c 		prz = cxt->mprz;
cxt               494 fs/pstore/ram.c static void ramoops_free_przs(struct ramoops_context *cxt)
cxt               499 fs/pstore/ram.c 	if (cxt->dprzs) {
cxt               500 fs/pstore/ram.c 		for (i = 0; i < cxt->max_dump_cnt; i++)
cxt               501 fs/pstore/ram.c 			persistent_ram_free(cxt->dprzs[i]);
cxt               503 fs/pstore/ram.c 		kfree(cxt->dprzs);
cxt               504 fs/pstore/ram.c 		cxt->max_dump_cnt = 0;
cxt               508 fs/pstore/ram.c 	if (cxt->fprzs) {
cxt               509 fs/pstore/ram.c 		for (i = 0; i < cxt->max_ftrace_cnt; i++)
cxt               510 fs/pstore/ram.c 			persistent_ram_free(cxt->fprzs[i]);
cxt               511 fs/pstore/ram.c 		kfree(cxt->fprzs);
cxt               512 fs/pstore/ram.c 		cxt->max_ftrace_cnt = 0;
cxt               517 fs/pstore/ram.c 			     struct device *dev, struct ramoops_context *cxt,
cxt               557 fs/pstore/ram.c 	if (*paddr + mem_sz - cxt->phys_addr > cxt->size) {
cxt               561 fs/pstore/ram.c 			cxt->size, (unsigned long long)cxt->phys_addr);
cxt               584 fs/pstore/ram.c 					       &cxt->ecc_info,
cxt               585 fs/pstore/ram.c 					       cxt->memtype, flags, label);
cxt               613 fs/pstore/ram.c 			    struct device *dev, struct ramoops_context *cxt,
cxt               622 fs/pstore/ram.c 	if (*paddr + sz - cxt->phys_addr > cxt->size) {
cxt               625 fs/pstore/ram.c 			cxt->size, (unsigned long long)cxt->phys_addr);
cxt               630 fs/pstore/ram.c 	*prz = persistent_ram_new(*paddr, sz, sig, &cxt->ecc_info,
cxt               631 fs/pstore/ram.c 				  cxt->memtype, PRZ_FLAG_ZAP_OLD, label);
cxt               736 fs/pstore/ram.c 	struct ramoops_context *cxt = &oops_cxt;
cxt               745 fs/pstore/ram.c 	if (cxt->max_dump_cnt) {
cxt               781 fs/pstore/ram.c 	cxt->size = pdata->mem_size;
cxt               782 fs/pstore/ram.c 	cxt->phys_addr = pdata->mem_address;
cxt               783 fs/pstore/ram.c 	cxt->memtype = pdata->mem_type;
cxt               784 fs/pstore/ram.c 	cxt->record_size = pdata->record_size;
cxt               785 fs/pstore/ram.c 	cxt->console_size = pdata->console_size;
cxt               786 fs/pstore/ram.c 	cxt->ftrace_size = pdata->ftrace_size;
cxt               787 fs/pstore/ram.c 	cxt->pmsg_size = pdata->pmsg_size;
cxt               788 fs/pstore/ram.c 	cxt->dump_oops = pdata->dump_oops;
cxt               789 fs/pstore/ram.c 	cxt->flags = pdata->flags;
cxt               790 fs/pstore/ram.c 	cxt->ecc_info = pdata->ecc_info;
cxt               792 fs/pstore/ram.c 	paddr = cxt->phys_addr;
cxt               794 fs/pstore/ram.c 	dump_mem_sz = cxt->size - cxt->console_size - cxt->ftrace_size
cxt               795 fs/pstore/ram.c 			- cxt->pmsg_size;
cxt               796 fs/pstore/ram.c 	err = ramoops_init_przs("dmesg", dev, cxt, &cxt->dprzs, &paddr,
cxt               797 fs/pstore/ram.c 				dump_mem_sz, cxt->record_size,
cxt               798 fs/pstore/ram.c 				&cxt->max_dump_cnt, 0, 0);
cxt               802 fs/pstore/ram.c 	err = ramoops_init_prz("console", dev, cxt, &cxt->cprz, &paddr,
cxt               803 fs/pstore/ram.c 			       cxt->console_size, 0);
cxt               807 fs/pstore/ram.c 	cxt->max_ftrace_cnt = (cxt->flags & RAMOOPS_FLAG_FTRACE_PER_CPU)
cxt               810 fs/pstore/ram.c 	err = ramoops_init_przs("ftrace", dev, cxt, &cxt->fprzs, &paddr,
cxt               811 fs/pstore/ram.c 				cxt->ftrace_size, -1,
cxt               812 fs/pstore/ram.c 				&cxt->max_ftrace_cnt, LINUX_VERSION_CODE,
cxt               813 fs/pstore/ram.c 				(cxt->flags & RAMOOPS_FLAG_FTRACE_PER_CPU)
cxt               818 fs/pstore/ram.c 	err = ramoops_init_prz("pmsg", dev, cxt, &cxt->mprz, &paddr,
cxt               819 fs/pstore/ram.c 				cxt->pmsg_size, 0);
cxt               823 fs/pstore/ram.c 	cxt->pstore.data = cxt;
cxt               830 fs/pstore/ram.c 	cxt->pstore.flags = 0;
cxt               831 fs/pstore/ram.c 	if (cxt->max_dump_cnt)
cxt               832 fs/pstore/ram.c 		cxt->pstore.flags |= PSTORE_FLAGS_DMESG;
cxt               833 fs/pstore/ram.c 	if (cxt->console_size)
cxt               834 fs/pstore/ram.c 		cxt->pstore.flags |= PSTORE_FLAGS_CONSOLE;
cxt               835 fs/pstore/ram.c 	if (cxt->max_ftrace_cnt)
cxt               836 fs/pstore/ram.c 		cxt->pstore.flags |= PSTORE_FLAGS_FTRACE;
cxt               837 fs/pstore/ram.c 	if (cxt->pmsg_size)
cxt               838 fs/pstore/ram.c 		cxt->pstore.flags |= PSTORE_FLAGS_PMSG;
cxt               845 fs/pstore/ram.c 	if (cxt->pstore.flags & PSTORE_FLAGS_DMESG) {
cxt               846 fs/pstore/ram.c 		cxt->pstore.bufsize = cxt->dprzs[0]->buffer_size;
cxt               847 fs/pstore/ram.c 		cxt->pstore.buf = kzalloc(cxt->pstore.bufsize, GFP_KERNEL);
cxt               848 fs/pstore/ram.c 		if (!cxt->pstore.buf) {
cxt               855 fs/pstore/ram.c 	err = pstore_register(&cxt->pstore);
cxt               874 fs/pstore/ram.c 		cxt->size, (unsigned long long)cxt->phys_addr,
cxt               875 fs/pstore/ram.c 		cxt->ecc_info.ecc_size);
cxt               880 fs/pstore/ram.c 	kfree(cxt->pstore.buf);
cxt               882 fs/pstore/ram.c 	cxt->pstore.bufsize = 0;
cxt               883 fs/pstore/ram.c 	persistent_ram_free(cxt->mprz);
cxt               886 fs/pstore/ram.c 	persistent_ram_free(cxt->cprz);
cxt               888 fs/pstore/ram.c 	ramoops_free_przs(cxt);
cxt               895 fs/pstore/ram.c 	struct ramoops_context *cxt = &oops_cxt;
cxt               897 fs/pstore/ram.c 	pstore_unregister(&cxt->pstore);
cxt               899 fs/pstore/ram.c 	kfree(cxt->pstore.buf);
cxt               900 fs/pstore/ram.c 	cxt->pstore.bufsize = 0;
cxt               902 fs/pstore/ram.c 	persistent_ram_free(cxt->mprz);
cxt               903 fs/pstore/ram.c 	persistent_ram_free(cxt->cprz);
cxt               904 fs/pstore/ram.c 	ramoops_free_przs(cxt);
cxt               129 include/linux/qed/qed_ll2_if.h void (*qed_ll2_complete_rx_packet_cb)(void *cxt,
cxt               133 include/linux/qed/qed_ll2_if.h void (*qed_ll2_release_rx_packet_cb)(void *cxt,
cxt               140 include/linux/qed/qed_ll2_if.h void (*qed_ll2_complete_tx_packet_cb)(void *cxt,
cxt               148 include/linux/qed/qed_ll2_if.h void (*qed_ll2_release_tx_packet_cb)(void *cxt,
cxt               155 include/linux/qed/qed_ll2_if.h void (*qed_ll2_slowpath_cb)(void *cxt, u8 connection_handle,
cxt                99 kernel/locking/locktorture.c static struct lock_torture_cxt cxt = { 0, 0, false,
cxt               117 kernel/locking/locktorture.c 	      (cxt.nrealwriters_stress * 2000 * longdelay_ms)))
cxt               119 kernel/locking/locktorture.c 	if (!(torture_random(trsp) % (cxt.nrealwriters_stress * 20000)))
cxt               161 kernel/locking/locktorture.c 	      (cxt.nrealwriters_stress * 2000 * longdelay_ms)))
cxt               164 kernel/locking/locktorture.c 	      (cxt.nrealwriters_stress * 2 * shortdelay_us)))
cxt               166 kernel/locking/locktorture.c 	if (!(torture_random(trsp) % (cxt.nrealwriters_stress * 20000)))
cxt               192 kernel/locking/locktorture.c 	cxt.cur_ops->flags = flags;
cxt               199 kernel/locking/locktorture.c 	spin_unlock_irqrestore(&torture_spinlock, cxt.cur_ops->flags);
cxt               230 kernel/locking/locktorture.c 	      (cxt.nrealwriters_stress * 2000 * longdelay_ms)))
cxt               256 kernel/locking/locktorture.c 	      (cxt.nrealreaders_stress * 2000 * longdelay_ms)))
cxt               283 kernel/locking/locktorture.c 	cxt.cur_ops->flags = flags;
cxt               290 kernel/locking/locktorture.c 	write_unlock_irqrestore(&torture_rwlock, cxt.cur_ops->flags);
cxt               298 kernel/locking/locktorture.c 	cxt.cur_ops->flags = flags;
cxt               305 kernel/locking/locktorture.c 	read_unlock_irqrestore(&torture_rwlock, cxt.cur_ops->flags);
cxt               333 kernel/locking/locktorture.c 	      (cxt.nrealwriters_stress * 2000 * longdelay_ms)))
cxt               337 kernel/locking/locktorture.c 	if (!(torture_random(trsp) % (cxt.nrealwriters_stress * 20000)))
cxt               451 kernel/locking/locktorture.c 			      (cxt.nrealwriters_stress * factor))) {
cxt               465 kernel/locking/locktorture.c 			       (cxt.nrealwriters_stress * factor * 2))) {
cxt               485 kernel/locking/locktorture.c 	      (cxt.nrealwriters_stress * 2000 * longdelay_ms)))
cxt               488 kernel/locking/locktorture.c 	      (cxt.nrealwriters_stress * 2 * shortdelay_us)))
cxt               490 kernel/locking/locktorture.c 	if (!(torture_random(trsp) % (cxt.nrealwriters_stress * 20000)))
cxt               524 kernel/locking/locktorture.c 	      (cxt.nrealwriters_stress * 2000 * longdelay_ms)))
cxt               528 kernel/locking/locktorture.c 	if (!(torture_random(trsp) % (cxt.nrealwriters_stress * 20000)))
cxt               549 kernel/locking/locktorture.c 	      (cxt.nrealreaders_stress * 2000 * longdelay_ms)))
cxt               553 kernel/locking/locktorture.c 	if (!(torture_random(trsp) % (cxt.nrealreaders_stress * 20000)))
cxt               631 kernel/locking/locktorture.c 		cxt.cur_ops->task_boost(&rand);
cxt               632 kernel/locking/locktorture.c 		cxt.cur_ops->writelock();
cxt               640 kernel/locking/locktorture.c 		cxt.cur_ops->write_delay(&rand);
cxt               642 kernel/locking/locktorture.c 		cxt.cur_ops->writeunlock();
cxt               647 kernel/locking/locktorture.c 	cxt.cur_ops->task_boost(NULL); /* reset prio */
cxt               668 kernel/locking/locktorture.c 		cxt.cur_ops->readlock();
cxt               674 kernel/locking/locktorture.c 		cxt.cur_ops->read_delay(&rand);
cxt               676 kernel/locking/locktorture.c 		cxt.cur_ops->readunlock();
cxt               695 kernel/locking/locktorture.c 	n_stress = write ? cxt.nrealwriters_stress : cxt.nrealreaders_stress;
cxt               711 kernel/locking/locktorture.c 		atomic_inc(&cxt.n_lock_torture_errors);
cxt               724 kernel/locking/locktorture.c 	int size = cxt.nrealwriters_stress * 200 + 8192;
cxt               727 kernel/locking/locktorture.c 	if (cxt.cur_ops->readlock)
cxt               728 kernel/locking/locktorture.c 		size += cxt.nrealreaders_stress * 200 + 8192;
cxt               737 kernel/locking/locktorture.c 	__torture_print_stats(buf, cxt.lwsa, true);
cxt               741 kernel/locking/locktorture.c 	if (cxt.cur_ops->readlock) {
cxt               749 kernel/locking/locktorture.c 		__torture_print_stats(buf, cxt.lrsa, false);
cxt               780 kernel/locking/locktorture.c 		 torture_type, tag, cxt.debug_lock ? " [debug]": "",
cxt               781 kernel/locking/locktorture.c 		 cxt.nrealwriters_stress, cxt.nrealreaders_stress, stat_interval,
cxt               799 kernel/locking/locktorture.c 	if (!cxt.lwsa && !cxt.lrsa)
cxt               803 kernel/locking/locktorture.c 		for (i = 0; i < cxt.nrealwriters_stress; i++)
cxt               811 kernel/locking/locktorture.c 		for (i = 0; i < cxt.nrealreaders_stress; i++)
cxt               821 kernel/locking/locktorture.c 	if (atomic_read(&cxt.n_lock_torture_errors))
cxt               822 kernel/locking/locktorture.c 		lock_torture_print_module_parms(cxt.cur_ops,
cxt               825 kernel/locking/locktorture.c 		lock_torture_print_module_parms(cxt.cur_ops,
cxt               828 kernel/locking/locktorture.c 		lock_torture_print_module_parms(cxt.cur_ops,
cxt               831 kernel/locking/locktorture.c 	kfree(cxt.lwsa);
cxt               832 kernel/locking/locktorture.c 	cxt.lwsa = NULL;
cxt               833 kernel/locking/locktorture.c 	kfree(cxt.lrsa);
cxt               834 kernel/locking/locktorture.c 	cxt.lrsa = NULL;
cxt               862 kernel/locking/locktorture.c 		cxt.cur_ops = torture_ops[i];
cxt               863 kernel/locking/locktorture.c 		if (strcmp(torture_type, cxt.cur_ops->name) == 0)
cxt               883 kernel/locking/locktorture.c 	if (cxt.cur_ops->init)
cxt               884 kernel/locking/locktorture.c 		cxt.cur_ops->init();
cxt               887 kernel/locking/locktorture.c 		cxt.nrealwriters_stress = nwriters_stress;
cxt               889 kernel/locking/locktorture.c 		cxt.nrealwriters_stress = 2 * num_online_cpus();
cxt               893 kernel/locking/locktorture.c 		cxt.debug_lock = true;
cxt               897 kernel/locking/locktorture.c 		cxt.debug_lock = true;
cxt               902 kernel/locking/locktorture.c 		cxt.debug_lock = true;
cxt               908 kernel/locking/locktorture.c 		cxt.lwsa = kmalloc_array(cxt.nrealwriters_stress,
cxt               909 kernel/locking/locktorture.c 					 sizeof(*cxt.lwsa),
cxt               911 kernel/locking/locktorture.c 		if (cxt.lwsa == NULL) {
cxt               917 kernel/locking/locktorture.c 		for (i = 0; i < cxt.nrealwriters_stress; i++) {
cxt               918 kernel/locking/locktorture.c 			cxt.lwsa[i].n_lock_fail = 0;
cxt               919 kernel/locking/locktorture.c 			cxt.lwsa[i].n_lock_acquired = 0;
cxt               923 kernel/locking/locktorture.c 	if (cxt.cur_ops->readlock) {
cxt               925 kernel/locking/locktorture.c 			cxt.nrealreaders_stress = nreaders_stress;
cxt               933 kernel/locking/locktorture.c 				cxt.nrealwriters_stress = num_online_cpus();
cxt               934 kernel/locking/locktorture.c 			cxt.nrealreaders_stress = cxt.nrealwriters_stress;
cxt               939 kernel/locking/locktorture.c 			cxt.lrsa = kmalloc_array(cxt.nrealreaders_stress,
cxt               940 kernel/locking/locktorture.c 						 sizeof(*cxt.lrsa),
cxt               942 kernel/locking/locktorture.c 			if (cxt.lrsa == NULL) {
cxt               945 kernel/locking/locktorture.c 				kfree(cxt.lwsa);
cxt               946 kernel/locking/locktorture.c 				cxt.lwsa = NULL;
cxt               950 kernel/locking/locktorture.c 			for (i = 0; i < cxt.nrealreaders_stress; i++) {
cxt               951 kernel/locking/locktorture.c 				cxt.lrsa[i].n_lock_fail = 0;
cxt               952 kernel/locking/locktorture.c 				cxt.lrsa[i].n_lock_acquired = 0;
cxt               957 kernel/locking/locktorture.c 	lock_torture_print_module_parms(cxt.cur_ops, "Start of test");
cxt               984 kernel/locking/locktorture.c 		writer_tasks = kcalloc(cxt.nrealwriters_stress,
cxt               994 kernel/locking/locktorture.c 	if (cxt.cur_ops->readlock) {
cxt               995 kernel/locking/locktorture.c 		reader_tasks = kcalloc(cxt.nrealreaders_stress,
cxt              1015 kernel/locking/locktorture.c 	for (i = 0, j = 0; i < cxt.nrealwriters_stress ||
cxt              1016 kernel/locking/locktorture.c 		    j < cxt.nrealreaders_stress; i++, j++) {
cxt              1017 kernel/locking/locktorture.c 		if (i >= cxt.nrealwriters_stress)
cxt              1021 kernel/locking/locktorture.c 		firsterr = torture_create_kthread(lock_torture_writer, &cxt.lwsa[i],
cxt              1027 kernel/locking/locktorture.c 		if (cxt.cur_ops->readlock == NULL || (j >= cxt.nrealreaders_stress))
cxt              1030 kernel/locking/locktorture.c 		firsterr = torture_create_kthread(lock_torture_reader, &cxt.lrsa[j],
cxt                32 tools/testing/selftests/gpio/gpio-mockup-chardev.c 	struct libmnt_context *cxt;
cxt                38 tools/testing/selftests/gpio/gpio-mockup-chardev.c 	cxt = mnt_new_context();
cxt                39 tools/testing/selftests/gpio/gpio-mockup-chardev.c 	if (!cxt)
cxt                46 tools/testing/selftests/gpio/gpio-mockup-chardev.c 	if (mnt_context_get_mtab(cxt, &tb))
cxt                64 tools/testing/selftests/gpio/gpio-mockup-chardev.c 	mnt_free_context(cxt);