ctxs              241 arch/x86/include/asm/tlbflush.h 	struct tlb_context ctxs[TLB_NR_DYN_ASIDS];
ctxs               68 arch/x86/mm/tlb.c 		this_cpu_write(cpu_tlbstate.ctxs[asid].ctx_id, 0);
ctxs               91 arch/x86/mm/tlb.c 		if (this_cpu_read(cpu_tlbstate.ctxs[asid].ctx_id) !=
ctxs               96 arch/x86/mm/tlb.c 		*need_flush = (this_cpu_read(cpu_tlbstate.ctxs[asid].tlb_gen) <
ctxs              334 arch/x86/mm/tlb.c 		VM_WARN_ON(this_cpu_read(cpu_tlbstate.ctxs[prev_asid].ctx_id) !=
ctxs              362 arch/x86/mm/tlb.c 		if (this_cpu_read(cpu_tlbstate.ctxs[prev_asid].tlb_gen) ==
ctxs              415 arch/x86/mm/tlb.c 		this_cpu_write(cpu_tlbstate.ctxs[new_asid].ctx_id, next->context.ctx_id);
ctxs              416 arch/x86/mm/tlb.c 		this_cpu_write(cpu_tlbstate.ctxs[new_asid].tlb_gen, next_tlb_gen);
ctxs              507 arch/x86/mm/tlb.c 	this_cpu_write(cpu_tlbstate.ctxs[0].ctx_id, mm->context.ctx_id);
ctxs              508 arch/x86/mm/tlb.c 	this_cpu_write(cpu_tlbstate.ctxs[0].tlb_gen, tlb_gen);
ctxs              511 arch/x86/mm/tlb.c 		this_cpu_write(cpu_tlbstate.ctxs[i].ctx_id, 0);
ctxs              536 arch/x86/mm/tlb.c 	u64 local_tlb_gen = this_cpu_read(cpu_tlbstate.ctxs[loaded_mm_asid].tlb_gen);
ctxs              544 arch/x86/mm/tlb.c 	VM_WARN_ON(this_cpu_read(cpu_tlbstate.ctxs[loaded_mm_asid].ctx_id) !=
ctxs              635 arch/x86/mm/tlb.c 	this_cpu_write(cpu_tlbstate.ctxs[loaded_mm_asid].tlb_gen, mm_tlb_gen);
ctxs              126 block/blk-mq-sched.c 	return hctx->ctxs[idx];
ctxs               20 block/blk-mq-sysfs.c 	struct blk_mq_ctxs *ctxs = container_of(kobj, struct blk_mq_ctxs, kobj);
ctxs               22 block/blk-mq-sysfs.c 	free_percpu(ctxs->queue_ctx);
ctxs               23 block/blk-mq-sysfs.c 	kfree(ctxs);
ctxs               31 block/blk-mq-sysfs.c 	kobject_put(&ctx->ctxs->kobj);
ctxs               46 block/blk-mq-sysfs.c 	kfree(hctx->ctxs);
ctxs              983 block/blk-mq.c 	struct blk_mq_ctx *ctx = hctx->ctxs[bitnr];
ctxs             1018 block/blk-mq.c 	struct blk_mq_ctx *ctx = hctx->ctxs[bitnr];
ctxs             2381 block/blk-mq.c 	hctx->ctxs = kmalloc_array_node(nr_cpu_ids, sizeof(void *),
ctxs             2383 block/blk-mq.c 	if (!hctx->ctxs)
ctxs             2409 block/blk-mq.c 	kfree(hctx->ctxs);
ctxs             2530 block/blk-mq.c 			hctx->ctxs[hctx->nr_ctx++] = ctx;
ctxs             2650 block/blk-mq.c 	struct blk_mq_ctxs *ctxs;
ctxs             2653 block/blk-mq.c 	ctxs = kzalloc(sizeof(*ctxs), GFP_KERNEL);
ctxs             2654 block/blk-mq.c 	if (!ctxs)
ctxs             2657 block/blk-mq.c 	ctxs->queue_ctx = alloc_percpu(struct blk_mq_ctx);
ctxs             2658 block/blk-mq.c 	if (!ctxs->queue_ctx)
ctxs             2662 block/blk-mq.c 		struct blk_mq_ctx *ctx = per_cpu_ptr(ctxs->queue_ctx, cpu);
ctxs             2663 block/blk-mq.c 		ctx->ctxs = ctxs;
ctxs             2666 block/blk-mq.c 	q->mq_kobj = &ctxs->kobj;
ctxs             2667 block/blk-mq.c 	q->queue_ctx = ctxs->queue_ctx;
ctxs             2671 block/blk-mq.c 	kfree(ctxs);
ctxs               36 block/blk-mq.h 	struct blk_mq_ctxs      *ctxs;
ctxs              101 drivers/gpu/drm/nouveau/nvkm/engine/fifo/dmanv04.c 			u32 cm = ((1ULL << c->bits) - 1) << c->ctxs;
ctxs              104 drivers/gpu/drm/nouveau/nvkm/engine/fifo/dmanv04.c 			nvkm_wo32(fctx, c->ctxp + data, cv | (rv << c->ctxs));
ctxs                9 drivers/gpu/drm/nouveau/nvkm/engine/fifo/nv04.h 	unsigned ctxs:5;
ctxs               51 drivers/iommu/qcom_iommu.c 	struct qcom_iommu_ctx	*ctxs[0];   /* indexed by asid-1 */
ctxs               89 drivers/iommu/qcom_iommu.c 	return qcom_iommu->ctxs[asid - 1];
ctxs              752 drivers/iommu/qcom_iommu.c 	qcom_iommu->ctxs[ctx->asid - 1] = ctx;
ctxs              764 drivers/iommu/qcom_iommu.c 	qcom_iommu->ctxs[ctx->asid - 1] = NULL;
ctxs              809 drivers/iommu/qcom_iommu.c 	qcom_iommu = devm_kzalloc(dev, struct_size(qcom_iommu, ctxs, max_asid),
ctxs               67 drivers/video/fbdev/omap2/omapfb/vrfb.c static struct vrfb_ctx *ctxs;
ctxs               88 drivers/video/fbdev/omap2/omapfb/vrfb.c 	omap2_sms_write_rot_control(ctxs[ctx].control, ctx);
ctxs               89 drivers/video/fbdev/omap2/omapfb/vrfb.c 	omap2_sms_write_rot_size(ctxs[ctx].size, ctx);
ctxs               90 drivers/video/fbdev/omap2/omapfb/vrfb.c 	omap2_sms_write_rot_physical_ba(ctxs[ctx].physical_ba, ctx);
ctxs              211 drivers/video/fbdev/omap2/omapfb/vrfb.c 	ctxs[ctx].physical_ba = paddr;
ctxs              212 drivers/video/fbdev/omap2/omapfb/vrfb.c 	ctxs[ctx].size = size;
ctxs              213 drivers/video/fbdev/omap2/omapfb/vrfb.c 	ctxs[ctx].control = control;
ctxs              308 drivers/video/fbdev/omap2/omapfb/vrfb.c 		paddr = ctxs[ctx].base + SMS_ROT_VIRT_BASE(rot);
ctxs              350 drivers/video/fbdev/omap2/omapfb/vrfb.c 	ctxs = devm_kcalloc(&pdev->dev,
ctxs              354 drivers/video/fbdev/omap2/omapfb/vrfb.c 	if (!ctxs)
ctxs              365 drivers/video/fbdev/omap2/omapfb/vrfb.c 		ctxs[i].base = mem->start;
ctxs               42 include/linux/blk-mq.h 	struct blk_mq_ctx	**ctxs;
ctxs              365 include/linux/blk-mq.h 	     ({ ctx = (hctx)->ctxs[(i)]; 1; }); (i)++)
ctxs             5243 kernel/workqueue.c 	LIST_HEAD(ctxs);
ctxs             5263 kernel/workqueue.c 		list_add_tail(&ctx->list, &ctxs);
ctxs             5266 kernel/workqueue.c 	list_for_each_entry_safe(ctx, n, &ctxs, list) {