ctx1 607 arch/sparc/mm/srmmu.c int cctx, ctx1; ctx1 610 arch/sparc/mm/srmmu.c if ((ctx1 = vma->vm_mm->context) != -1) { ctx1 613 arch/sparc/mm/srmmu.c if (cctx != ctx1) { ctx1 614 arch/sparc/mm/srmmu.c printk("flush ctx %02x curr %02x\n", ctx1, cctx); ctx1 615 arch/sparc/mm/srmmu.c srmmu_set_context(ctx1); ctx1 194 drivers/gpu/drm/i915/gvt/execlist.c struct execlist_ctx_descriptor_format *ctx1 = &running->ctx[1]; ctx1 210 drivers/gpu/drm/i915/gvt/execlist.c if (valid_context(ctx1) && same_context(ctx0, ctx)) { ctx1 213 drivers/gpu/drm/i915/gvt/execlist.c execlist->running_context = ctx1; ctx1 229 drivers/gpu/drm/i915/gvt/execlist.c } else if ((!valid_context(ctx1) && same_context(ctx0, ctx)) ctx1 230 drivers/gpu/drm/i915/gvt/execlist.c || (valid_context(ctx1) && same_context(ctx1, ctx))) { ctx1 287 drivers/gpu/drm/i915/gvt/execlist.c struct execlist_ctx_descriptor_format *ctx0, *ctx1; ctx1 333 drivers/gpu/drm/i915/gvt/execlist.c ctx1 = &running->ctx[1]; ctx1 336 drivers/gpu/drm/i915/gvt/execlist.c running->index, ctx0->context_id, ctx1->context_id); ctx1 346 drivers/gpu/drm/i915/gvt/execlist.c if ((valid_context(ctx1) && same_context(ctx1, &slot->ctx[0]) && ctx1 349 drivers/gpu/drm/i915/gvt/execlist.c (!valid_context(ctx1) && ctx1 463 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv04.c u32 tmp, ctx1; ctx1 465 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv04.c ctx1 = nvkm_rd32(device, 0x700000 + inst); ctx1 466 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv04.c class = ctx1 & 0xff; ctx1 467 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv04.c op = (ctx1 >> 15) & 7; ctx1 495 drivers/infiniband/hw/cxgb3/cxio_hal.c u64 sge_cmd, ctx0, ctx1; ctx1 536 drivers/infiniband/hw/cxgb3/cxio_hal.c ctx1 = (u32) base_addr; ctx1 538 drivers/infiniband/hw/cxgb3/cxio_hal.c ctx1 |= ((u64) (V_EC_BASE_HI((u32) base_addr & 0xf) | V_EC_RESPQ(0) | ctx1 547 drivers/infiniband/hw/cxgb3/cxio_hal.c wqe->ctx1 = cpu_to_be64(ctx1); ctx1 291 drivers/infiniband/hw/cxgb3/cxio_wr.h __be64 ctx1; /* 5 */ ctx1 685 drivers/scsi/qla2xxx/qla_os.c struct ct6_dsd *ctx1 = sp->u.scmd.ct6_ctx; ctx1 687 drivers/scsi/qla2xxx/qla_os.c dma_pool_free(ha->fcp_cmnd_dma_pool, ctx1->fcp_cmnd, ctx1 688 drivers/scsi/qla2xxx/qla_os.c ctx1->fcp_cmnd_dma); ctx1 689 drivers/scsi/qla2xxx/qla_os.c list_splice(&ctx1->dsd_list, &ha->gbl_dsd_list); ctx1 690 drivers/scsi/qla2xxx/qla_os.c ha->gbl_dsd_inuse -= ctx1->dsd_use_cnt; ctx1 691 drivers/scsi/qla2xxx/qla_os.c ha->gbl_dsd_avail += ctx1->dsd_use_cnt; ctx1 692 drivers/scsi/qla2xxx/qla_os.c mempool_free(ctx1, ha->ctx_mempool); ctx1 768 drivers/scsi/qla2xxx/qla_os.c struct ct6_dsd *ctx1 = sp->u.scmd.ct6_ctx; ctx1 770 drivers/scsi/qla2xxx/qla_os.c dma_pool_free(ha->fcp_cmnd_dma_pool, ctx1->fcp_cmnd, ctx1 771 drivers/scsi/qla2xxx/qla_os.c ctx1->fcp_cmnd_dma); ctx1 772 drivers/scsi/qla2xxx/qla_os.c list_splice(&ctx1->dsd_list, &ha->gbl_dsd_list); ctx1 773 drivers/scsi/qla2xxx/qla_os.c ha->gbl_dsd_inuse -= ctx1->dsd_use_cnt; ctx1 774 drivers/scsi/qla2xxx/qla_os.c ha->gbl_dsd_avail += ctx1->dsd_use_cnt; ctx1 775 drivers/scsi/qla2xxx/qla_os.c mempool_free(ctx1, ha->ctx_mempool); ctx1 264 fs/nfs/internal.h static inline bool nfs_match_open_context(const struct nfs_open_context *ctx1, ctx1 267 fs/nfs/internal.h return cred_fscmp(ctx1->cred, ctx2->cred) == 0 && ctx1->state == ctx2->state; ctx1 3077 kernel/events/core.c static int context_equiv(struct perf_event_context *ctx1, ctx1 3080 kernel/events/core.c lockdep_assert_held(&ctx1->lock); ctx1 3084 kernel/events/core.c if (ctx1->pin_count || ctx2->pin_count) ctx1 3088 kernel/events/core.c if (ctx1 == ctx2->parent_ctx && ctx1->generation == ctx2->parent_gen) ctx1 3092 kernel/events/core.c if (ctx1->parent_ctx == ctx2 && ctx1->parent_gen == ctx2->generation) ctx1 3099 kernel/events/core.c if (ctx1->parent_ctx && ctx1->parent_ctx == ctx2->parent_ctx && ctx1 3100 kernel/events/core.c ctx1->parent_gen == ctx2->parent_gen)