ctx_id 18 arch/x86/include/asm/mmu.h u64 ctx_id; ctx_id 61 arch/x86/include/asm/mmu.h .ctx_id = 1, \ ctx_id 191 arch/x86/include/asm/mmu_context.h mm->context.ctx_id = atomic64_inc_return(&last_mm_ctx_id); ctx_id 152 arch/x86/include/asm/tlbflush.h u64 ctx_id; ctx_id 68 arch/x86/mm/tlb.c this_cpu_write(cpu_tlbstate.ctxs[asid].ctx_id, 0); ctx_id 91 arch/x86/mm/tlb.c if (this_cpu_read(cpu_tlbstate.ctxs[asid].ctx_id) != ctx_id 92 arch/x86/mm/tlb.c next->context.ctx_id) ctx_id 334 arch/x86/mm/tlb.c VM_WARN_ON(this_cpu_read(cpu_tlbstate.ctxs[prev_asid].ctx_id) != ctx_id 335 arch/x86/mm/tlb.c next->context.ctx_id); ctx_id 415 arch/x86/mm/tlb.c this_cpu_write(cpu_tlbstate.ctxs[new_asid].ctx_id, next->context.ctx_id); ctx_id 507 arch/x86/mm/tlb.c this_cpu_write(cpu_tlbstate.ctxs[0].ctx_id, mm->context.ctx_id); ctx_id 511 arch/x86/mm/tlb.c this_cpu_write(cpu_tlbstate.ctxs[i].ctx_id, 0); ctx_id 544 arch/x86/mm/tlb.c VM_WARN_ON(this_cpu_read(cpu_tlbstate.ctxs[loaded_mm_asid].ctx_id) != ctx_id 545 arch/x86/mm/tlb.c loaded_mm->context.ctx_id); ctx_id 85 drivers/crypto/picoxcell_crypto.c unsigned ctx_id; ctx_id 568 drivers/crypto/picoxcell_crypto.c req->ctx_id = spacc_load_ctx(&ctx->generic, ctx->cipher_key, ctx_id 593 drivers/crypto/picoxcell_crypto.c ctrl = spacc_alg->ctrl_default | (req->ctx_id << SPA_CTRL_CTX_IDX) | ctx_id 886 drivers/crypto/picoxcell_crypto.c req->ctx_id = spacc_load_ctx(&ctx->generic, ctx->key, ctx_id 899 drivers/crypto/picoxcell_crypto.c ctrl = spacc_alg->ctrl_default | (req->ctx_id << SPA_CTRL_CTX_IDX) | ctx_id 123 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c p->ctx = amdgpu_ctx_get(fpriv, cs->in.ctx_id); ctx_id 1031 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c ctx = amdgpu_ctx_get(fpriv, deps[i].ctx_id); ctx_id 1425 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c ctx = amdgpu_ctx_get(filp->driver_priv, wait->in.ctx_id); ctx_id 1473 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c ctx = amdgpu_ctx_get(filp->driver_priv, user->ctx_id); ctx_id 410 drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c id = args->in.ctx_id; ctx_id 421 drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c args->out.alloc.ctx_id = id; ctx_id 83 drivers/gpu/drm/amd/amdgpu/amdgpu_sched.c unsigned ctx_id, ctx_id 100 drivers/gpu/drm/amd/amdgpu/amdgpu_sched.c ctx = amdgpu_ctx_get(fpriv, ctx_id); ctx_id 135 drivers/gpu/drm/amd/amdgpu/amdgpu_sched.c args->in.ctx_id, ctx_id 174 drivers/gpu/drm/amd/powerplay/inc/power_state.h unsigned long ctx_id; ctx_id 191 drivers/gpu/drm/drm_context.c map = idr_find(&dev->ctx_idr, request->ctx_id); ctx_id 252 drivers/gpu/drm/drm_context.c if (IS_ERR(idr_replace(&dev->ctx_idr, map, request->ctx_id))) ctx_id 502 drivers/gpu/drm/drm_ioc32.c unsigned int ctx_id; /**< Context requesting private mapping */ ctx_id 516 drivers/gpu/drm/drm_ioc32.c request.ctx_id = req32.ctx_id; ctx_id 533 drivers/gpu/drm/drm_ioc32.c req.ctx_id = req32.ctx_id; ctx_id 1845 drivers/gpu/drm/i915/gem/i915_gem_context.c if (local.param.ctx_id) ctx_id 2123 drivers/gpu/drm/i915/gem/i915_gem_context.c args->ctx_id = ret; ctx_id 2124 drivers/gpu/drm/i915/gem/i915_gem_context.c DRM_DEBUG("HW context %d created\n", args->ctx_id); ctx_id 2143 drivers/gpu/drm/i915/gem/i915_gem_context.c if (!args->ctx_id) ctx_id 2149 drivers/gpu/drm/i915/gem/i915_gem_context.c ctx = idr_remove(&file_priv->context_idr, args->ctx_id); ctx_id 2221 drivers/gpu/drm/i915/gem/i915_gem_context.c ctx = i915_gem_context_lookup(file_priv, args->ctx_id); ctx_id 2291 drivers/gpu/drm/i915/gem/i915_gem_context.c ctx = i915_gem_context_lookup(file_priv, args->ctx_id); ctx_id 2314 drivers/gpu/drm/i915/gem/i915_gem_context.c ctx = __i915_gem_context_lookup_rcu(file->driver_priv, args->ctx_id); ctx_id 714 drivers/gpu/drm/i915/i915_perf.c u32 ctx_id; ctx_id 748 drivers/gpu/drm/i915/i915_perf.c ctx_id = report32[2] & stream->specific_ctx_id_mask; ctx_id 759 drivers/gpu/drm/i915/i915_perf.c ctx_id = report32[2] = INVALID_CTX_ID; ctx_id 793 drivers/gpu/drm/i915/i915_perf.c stream->specific_ctx_id == ctx_id || ctx_id 802 drivers/gpu/drm/i915/i915_perf.c stream->specific_ctx_id != ctx_id) { ctx_id 811 drivers/gpu/drm/i915/i915_perf.c stream->oa_buffer.last_ctx_id = ctx_id; ctx_id 2025 drivers/gpu/drm/i915/i915_perf.c u32 ctx_id = stream->specific_ctx_id; ctx_id 2042 drivers/gpu/drm/i915/i915_perf.c (ctx_id & GEN7_OACONTROL_CTX_MASK) | ctx_id 214 drivers/gpu/drm/virtio/virtgpu_drv.h uint32_t ctx_id; ctx_id 297 drivers/gpu/drm/virtio/virtgpu_drv.h uint32_t ctx_id, ctx_id 300 drivers/gpu/drm/virtio/virtgpu_drv.h uint32_t ctx_id, ctx_id 304 drivers/gpu/drm/virtio/virtgpu_drv.h uint32_t ctx_id, struct virtio_gpu_fence *fence); ctx_id 306 drivers/gpu/drm/virtio/virtgpu_drv.h uint32_t resource_id, uint32_t ctx_id, ctx_id 312 drivers/gpu/drm/virtio/virtgpu_drv.h uint32_t ctx_id, ctx_id 149 drivers/gpu/drm/virtio/virtgpu_gem.c virtio_gpu_cmd_context_attach_resource(vgdev, vfpriv->ctx_id, ctx_id 170 drivers/gpu/drm/virtio/virtgpu_gem.c virtio_gpu_cmd_context_detach_resource(vgdev, vfpriv->ctx_id, ctx_id 225 drivers/gpu/drm/virtio/virtgpu_ioctl.c vfpriv->ctx_id, out_fence); ctx_id 397 drivers/gpu/drm/virtio/virtgpu_ioctl.c vfpriv->ctx_id, offset, args->level, ctx_id 451 drivers/gpu/drm/virtio/virtgpu_ioctl.c vfpriv ? vfpriv->ctx_id : 0, offset, ctx_id 67 drivers/gpu/drm/virtio/virtgpu_kms.c uint32_t ctx_id) ctx_id 69 drivers/gpu/drm/virtio/virtgpu_kms.c virtio_gpu_cmd_context_destroy(vgdev, ctx_id); ctx_id 70 drivers/gpu/drm/virtio/virtgpu_kms.c ida_free(&vgdev->ctx_id_ida, ctx_id - 1); ctx_id 275 drivers/gpu/drm/virtio/virtgpu_kms.c vfpriv->ctx_id = id; ctx_id 290 drivers/gpu/drm/virtio/virtgpu_kms.c virtio_gpu_context_destroy(vgdev, vfpriv->ctx_id); ctx_id 21 drivers/gpu/drm/virtio/virtgpu_trace.h __field(u32, ctx_id) ctx_id 30 drivers/gpu/drm/virtio/virtgpu_trace.h __entry->ctx_id = le32_to_cpu(hdr->ctx_id); ctx_id 35 drivers/gpu/drm/virtio/virtgpu_trace.h __entry->ctx_id) ctx_id 806 drivers/gpu/drm/virtio/virtgpu_vq.c cmd_p->hdr.ctx_id = cpu_to_le32(id); ctx_id 823 drivers/gpu/drm/virtio/virtgpu_vq.c cmd_p->hdr.ctx_id = cpu_to_le32(id); ctx_id 828 drivers/gpu/drm/virtio/virtgpu_vq.c uint32_t ctx_id, ctx_id 838 drivers/gpu/drm/virtio/virtgpu_vq.c cmd_p->hdr.ctx_id = cpu_to_le32(ctx_id); ctx_id 845 drivers/gpu/drm/virtio/virtgpu_vq.c uint32_t ctx_id, ctx_id 855 drivers/gpu/drm/virtio/virtgpu_vq.c cmd_p->hdr.ctx_id = cpu_to_le32(ctx_id); ctx_id 892 drivers/gpu/drm/virtio/virtgpu_vq.c uint32_t ctx_id, ctx_id 910 drivers/gpu/drm/virtio/virtgpu_vq.c cmd_p->hdr.ctx_id = cpu_to_le32(ctx_id); ctx_id 920 drivers/gpu/drm/virtio/virtgpu_vq.c uint32_t resource_id, uint32_t ctx_id, ctx_id 932 drivers/gpu/drm/virtio/virtgpu_vq.c cmd_p->hdr.ctx_id = cpu_to_le32(ctx_id); ctx_id 943 drivers/gpu/drm/virtio/virtgpu_vq.c uint32_t ctx_id, struct virtio_gpu_fence *fence) ctx_id 955 drivers/gpu/drm/virtio/virtgpu_vq.c cmd_p->hdr.ctx_id = cpu_to_le32(ctx_id); ctx_id 1000 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c int ctx_id, ctx_id 1012 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c ctx_id != cur->cb_header->dxContext))) ctx_id 1027 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c if (ctx_id != SVGA3D_INVALID_ID) { ctx_id 1029 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c cur->cb_header->dxContext = ctx_id; ctx_id 1074 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c int ctx_id, bool interruptible, ctx_id 1078 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c return vmw_cmdbuf_reserve_cur(man, size, ctx_id, interruptible); ctx_id 1083 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c if (ctx_id != SVGA3D_INVALID_ID) { ctx_id 1085 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c header->cb_header->dxContext = ctx_id; ctx_id 883 drivers/gpu/drm/vmwgfx/vmwgfx_drv.h vmw_fifo_reserve_dx(struct vmw_private *dev_priv, uint32_t bytes, int ctx_id); ctx_id 1326 drivers/gpu/drm/vmwgfx/vmwgfx_drv.h int ctx_id, bool interruptible, ctx_id 387 drivers/gpu/drm/vmwgfx/vmwgfx_fifo.c int ctx_id) ctx_id 393 drivers/gpu/drm/vmwgfx/vmwgfx_fifo.c ctx_id, false, NULL); ctx_id 394 drivers/gpu/drm/vmwgfx/vmwgfx_fifo.c else if (ctx_id == SVGA3D_INVALID_ID) ctx_id 261 drivers/infiniband/hw/efa/efa_com.c u16 ctx_id; ctx_id 264 drivers/infiniband/hw/efa/efa_com.c ctx_id = aq->comp_ctx_pool[aq->comp_ctx_pool_next]; ctx_id 268 drivers/infiniband/hw/efa/efa_com.c return ctx_id; ctx_id 272 drivers/infiniband/hw/efa/efa_com.c u16 ctx_id) ctx_id 276 drivers/infiniband/hw/efa/efa_com.c aq->comp_ctx_pool[aq->comp_ctx_pool_next] = ctx_id; ctx_id 285 drivers/infiniband/hw/efa/efa_com.c u16 ctx_id = cmd_id & (aq->depth - 1); ctx_id 289 drivers/infiniband/hw/efa/efa_com.c efa_com_dealloc_ctx_id(aq, ctx_id); ctx_id 295 drivers/infiniband/hw/efa/efa_com.c u16 ctx_id = cmd_id & (aq->depth - 1); ctx_id 297 drivers/infiniband/hw/efa/efa_com.c if (aq->comp_ctx[ctx_id].occupied && capture) { ctx_id 306 drivers/infiniband/hw/efa/efa_com.c aq->comp_ctx[ctx_id].occupied = 1; ctx_id 311 drivers/infiniband/hw/efa/efa_com.c return &aq->comp_ctx[ctx_id]; ctx_id 324 drivers/infiniband/hw/efa/efa_com.c u16 ctx_id; ctx_id 330 drivers/infiniband/hw/efa/efa_com.c ctx_id = efa_com_alloc_ctx_id(aq); ctx_id 333 drivers/infiniband/hw/efa/efa_com.c cmd_id = ctx_id & queue_size_mask; ctx_id 343 drivers/infiniband/hw/efa/efa_com.c efa_com_dealloc_ctx_id(aq, ctx_id); ctx_id 14399 drivers/infiniband/hw/hfi1/chip.c u8 ctx_id = 0; ctx_id 14421 drivers/infiniband/hw/hfi1/chip.c reg |= (u64)dd->vnic.ctxt[ctx_id++]->ctxt << (j * 8); ctx_id 14423 drivers/infiniband/hw/hfi1/chip.c ctx_id %= dd->vnic.num_ctxt; ctx_id 120 drivers/media/platform/sti/hva/hva-hw.c u8 ctx_id = 0; ctx_id 130 drivers/media/platform/sti/hva/hva-hw.c ctx_id = (hva->sts_reg & 0xFF00) >> 8; ctx_id 131 drivers/media/platform/sti/hva/hva-hw.c if (ctx_id >= HVA_MAX_INSTANCES) { ctx_id 133 drivers/media/platform/sti/hva/hva-hw.c ctx->name, __func__, ctx_id); ctx_id 138 drivers/media/platform/sti/hva/hva-hw.c ctx = hva->instances[ctx_id]; ctx_id 224 drivers/media/platform/sti/hva/hva-hw.c u8 ctx_id = 0; ctx_id 234 drivers/media/platform/sti/hva/hva-hw.c ctx_id = (hva->sts_reg & 0xFF00) >> 8; ctx_id 235 drivers/media/platform/sti/hva/hva-hw.c if (ctx_id >= HVA_MAX_INSTANCES) { ctx_id 237 drivers/media/platform/sti/hva/hva-hw.c ctx_id); ctx_id 241 drivers/media/platform/sti/hva/hva-hw.c ctx = hva->instances[ctx_id]; ctx_id 47 drivers/misc/habanalabs/command_buffer.c int ctx_id) ctx_id 60 drivers/misc/habanalabs/command_buffer.c if (ctx_id == HL_KERNEL_ASID_ID) ctx_id 68 drivers/misc/habanalabs/command_buffer.c if (ctx_id == HL_KERNEL_ASID_ID) ctx_id 90 drivers/misc/habanalabs/command_buffer.c u32 cb_size, u64 *handle, int ctx_id) ctx_id 101 drivers/misc/habanalabs/command_buffer.c (ctx_id != HL_KERNEL_ASID_ID))) { ctx_id 120 drivers/misc/habanalabs/command_buffer.c if (ctx_id == HL_KERNEL_ASID_ID && ctx_id 137 drivers/misc/habanalabs/command_buffer.c cb = hl_cb_alloc(hdev, cb_size, ctx_id); ctx_id 145 drivers/misc/habanalabs/command_buffer.c cb->ctx_id = ctx_id; ctx_id 385 drivers/misc/habanalabs/command_buffer.c id, cb->ctx_id); ctx_id 85 drivers/misc/habanalabs/command_submission.c parser.ctx_id = job->cs->ctx->asid; ctx_id 114 drivers/misc/habanalabs/debugfs.c cb->id, cb->ctx_id, cb->size, ctx_id 280 drivers/misc/habanalabs/habanalabs.h u32 ctx_id; ctx_id 796 drivers/misc/habanalabs/habanalabs.h u32 ctx_id; ctx_id 1491 drivers/misc/habanalabs/habanalabs.h u64 *handle, int ctx_id); ctx_id 386 drivers/net/ethernet/netronome/nfp/flower/main.h nfp_flower_get_fl_payload_from_ctx(struct nfp_app *app, u32 ctx_id); ctx_id 118 drivers/net/ethernet/netronome/nfp/flower/metadata.c u32 ctx_id; ctx_id 126 drivers/net/ethernet/netronome/nfp/flower/metadata.c ctx_id = be32_to_cpu(stats->stats_con_id); ctx_id 127 drivers/net/ethernet/netronome/nfp/flower/metadata.c priv->stats[ctx_id].pkts += be32_to_cpu(stats->pkt_count); ctx_id 128 drivers/net/ethernet/netronome/nfp/flower/metadata.c priv->stats[ctx_id].bytes += be64_to_cpu(stats->byte_count); ctx_id 129 drivers/net/ethernet/netronome/nfp/flower/metadata.c priv->stats[ctx_id].used = jiffies; ctx_id 433 drivers/net/ethernet/netronome/nfp/flower/metadata.c nfp_flower_get_fl_payload_from_ctx(struct nfp_app *app, u32 ctx_id) ctx_id 438 drivers/net/ethernet/netronome/nfp/flower/metadata.c ctx_entry = rhashtable_lookup_fast(&priv->stats_ctx_table, &ctx_id, ctx_id 1348 drivers/net/ethernet/netronome/nfp/flower/offload.c u32 ctx_id; ctx_id 1350 drivers/net/ethernet/netronome/nfp/flower/offload.c ctx_id = be32_to_cpu(merge_flow->meta.host_ctx_id); ctx_id 1351 drivers/net/ethernet/netronome/nfp/flower/offload.c pkts = priv->stats[ctx_id].pkts; ctx_id 1355 drivers/net/ethernet/netronome/nfp/flower/offload.c bytes = priv->stats[ctx_id].bytes; ctx_id 1356 drivers/net/ethernet/netronome/nfp/flower/offload.c used = priv->stats[ctx_id].used; ctx_id 1359 drivers/net/ethernet/netronome/nfp/flower/offload.c priv->stats[ctx_id].pkts = 0; ctx_id 1360 drivers/net/ethernet/netronome/nfp/flower/offload.c priv->stats[ctx_id].bytes = 0; ctx_id 1368 drivers/net/ethernet/netronome/nfp/flower/offload.c ctx_id = be32_to_cpu(sub_flow->meta.host_ctx_id); ctx_id 1369 drivers/net/ethernet/netronome/nfp/flower/offload.c priv->stats[ctx_id].pkts += pkts; ctx_id 1370 drivers/net/ethernet/netronome/nfp/flower/offload.c priv->stats[ctx_id].bytes += bytes; ctx_id 1371 drivers/net/ethernet/netronome/nfp/flower/offload.c priv->stats[ctx_id].used = max_t(u64, used, ctx_id 1372 drivers/net/ethernet/netronome/nfp/flower/offload.c priv->stats[ctx_id].used); ctx_id 1405 drivers/net/ethernet/netronome/nfp/flower/offload.c u32 ctx_id; ctx_id 1414 drivers/net/ethernet/netronome/nfp/flower/offload.c ctx_id = be32_to_cpu(nfp_flow->meta.host_ctx_id); ctx_id 1421 drivers/net/ethernet/netronome/nfp/flower/offload.c flow_stats_update(&flow->stats, priv->stats[ctx_id].bytes, ctx_id 1422 drivers/net/ethernet/netronome/nfp/flower/offload.c priv->stats[ctx_id].pkts, priv->stats[ctx_id].used); ctx_id 1424 drivers/net/ethernet/netronome/nfp/flower/offload.c priv->stats[ctx_id].pkts = 0; ctx_id 1425 drivers/net/ethernet/netronome/nfp/flower/offload.c priv->stats[ctx_id].bytes = 0; ctx_id 268 drivers/net/ethernet/qlogic/netxen/netxen_nic.h __le32 ctx_id; ctx_id 757 drivers/net/ethernet/qlogic/netxen/netxen_nic_ctx.c recv_ctx->hwctx->ctx_id = cpu_to_le32(port); ctx_id 620 drivers/net/ethernet/qlogic/qlcnic/qlcnic.h u16 ctx_id; ctx_id 1259 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c recv_ctx->context_id = mbx_out->ctx_id; ctx_id 1307 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c cmd.req.arg[1] = tx_ring->ctx_id | temp; ctx_id 1379 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c tx->ctx_id = mbx_out->ctx_id; ctx_id 1387 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c tx->ctx_id, mbx_out->state); ctx_id 2228 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c temp = adapter->tx_ring->ctx_id; ctx_id 3533 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c cmd.req.arg[1] = BIT_1 | (adapter->tx_ring->ctx_id << 16); ctx_id 173 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.h u16 ctx_id; ctx_id 179 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.h u16 ctx_id; ctx_id 195 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.h u16 ctx_id; ctx_id 197 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.h u16 ctx_id; ctx_id 239 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.h u16 ctx_id; ctx_id 245 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.h u16 ctx_id; ctx_id 497 drivers/net/ethernet/qlogic/qlcnic/qlcnic_ctx.c tx_ring->ctx_id = le16_to_cpu(prsp->context_id); ctx_id 507 drivers/net/ethernet/qlogic/qlcnic/qlcnic_ctx.c tx_ring->ctx_id, tx_ring->state); ctx_id 534 drivers/net/ethernet/qlogic/qlcnic/qlcnic_ctx.c cmd.req.arg[1] = tx_ring->ctx_id; ctx_id 3044 drivers/net/ethernet/qlogic/qlcnic/qlcnic_main.c ring, tx_ring->ctx_id); ctx_id 902 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c vf->rx_ctx_id = mbx_out->ctx_id; ctx_id 961 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c vf->tx_ctx_id = mbx_out->ctx_id; ctx_id 1198 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c u16 ctx_id, pkts, time; ctx_id 1203 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c ctx_id = cmd->req.arg[1] >> 16; ctx_id 1209 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c if (ctx_id != vf->rx_ctx_id || pkts > coal->rx_packets || ctx_id 1214 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c if (ctx_id != vf->tx_ctx_id || pkts > coal->tx_packets || ctx_id 1231 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c ctx_id, pkts, time, type); ctx_id 499 drivers/scsi/cxlflash/main.c cmd->rcb.ctx_id = hwq->ctx_hndl; ctx_id 627 drivers/scsi/cxlflash/main.c cmd->rcb.ctx_id = hwq->ctx_hndl; ctx_id 2339 drivers/scsi/cxlflash/main.c cmd->rcb.ctx_id = hwq->ctx_hndl; ctx_id 27 drivers/scsi/cxlflash/sislite.h u16 ctx_id; /* ctx_hndl_t */ ctx_id 304 drivers/scsi/cxlflash/sislite.h #define SISL_RHT_CNT_ID(cnt, ctx_id) (((cnt) << 48) | ((ctx_id) << 32)) ctx_id 1048 fs/aio.c static struct kioctx *lookup_ioctx(unsigned long ctx_id) ctx_id 1050 fs/aio.c struct aio_ring __user *ring = (void __user *)ctx_id; ctx_id 1067 fs/aio.c if (ctx && ctx->user_id == ctx_id) { ctx_id 1915 fs/aio.c SYSCALL_DEFINE3(io_submit, aio_context_t, ctx_id, long, nr, ctx_id 1926 fs/aio.c ctx = lookup_ioctx(ctx_id); ctx_id 1957 fs/aio.c COMPAT_SYSCALL_DEFINE3(io_submit, compat_aio_context_t, ctx_id, ctx_id 1968 fs/aio.c ctx = lookup_ioctx(ctx_id); ctx_id 2009 fs/aio.c SYSCALL_DEFINE3(io_cancel, aio_context_t, ctx_id, struct iocb __user *, iocb, ctx_id 2023 fs/aio.c ctx = lookup_ioctx(ctx_id); ctx_id 2052 fs/aio.c static long do_io_getevents(aio_context_t ctx_id, ctx_id 2059 fs/aio.c struct kioctx *ioctx = lookup_ioctx(ctx_id); ctx_id 2085 fs/aio.c SYSCALL_DEFINE5(io_getevents, aio_context_t, ctx_id, ctx_id 2097 fs/aio.c ret = do_io_getevents(ctx_id, min_nr, nr, events, timeout ? &ts : NULL); ctx_id 2111 fs/aio.c aio_context_t, ctx_id, ctx_id 2133 fs/aio.c ret = do_io_getevents(ctx_id, min_nr, nr, events, timeout ? &ts : NULL); ctx_id 2146 fs/aio.c aio_context_t, ctx_id, ctx_id 2169 fs/aio.c ret = do_io_getevents(ctx_id, min_nr, nr, events, timeout ? &ts : NULL); ctx_id 2183 fs/aio.c SYSCALL_DEFINE5(io_getevents_time32, __u32, ctx_id, ctx_id 2195 fs/aio.c ret = do_io_getevents(ctx_id, min_nr, nr, events, timeout ? &t : NULL); ctx_id 2213 fs/aio.c compat_aio_context_t, ctx_id, ctx_id 2235 fs/aio.c ret = do_io_getevents(ctx_id, min_nr, nr, events, timeout ? &t : NULL); ctx_id 2248 fs/aio.c compat_aio_context_t, ctx_id, ctx_id 2270 fs/aio.c ret = do_io_getevents(ctx_id, min_nr, nr, events, timeout ? &t : NULL); ctx_id 518 include/linux/compat.h asmlinkage long compat_sys_io_submit(compat_aio_context_t ctx_id, int nr, ctx_id 520 include/linux/compat.h asmlinkage long compat_sys_io_pgetevents(compat_aio_context_t ctx_id, ctx_id 526 include/linux/compat.h asmlinkage long compat_sys_io_pgetevents_time64(compat_aio_context_t ctx_id, ctx_id 52 include/linux/sunrpc/gss_api.h struct gss_ctx **ctx_id, ctx_id 56 include/linux/sunrpc/gss_api.h struct gss_ctx *ctx_id, ctx_id 60 include/linux/sunrpc/gss_api.h struct gss_ctx *ctx_id, ctx_id 64 include/linux/sunrpc/gss_api.h struct gss_ctx *ctx_id, ctx_id 69 include/linux/sunrpc/gss_api.h struct gss_ctx *ctx_id, ctx_id 74 include/linux/sunrpc/gss_api.h struct gss_ctx **ctx_id); ctx_id 113 include/linux/sunrpc/gss_api.h struct gss_ctx *ctx_id, ctx_id 117 include/linux/sunrpc/gss_api.h struct gss_ctx *ctx_id, ctx_id 121 include/linux/sunrpc/gss_api.h struct gss_ctx *ctx_id, ctx_id 125 include/linux/sunrpc/gss_api.h struct gss_ctx *ctx_id, ctx_id 130 include/linux/sunrpc/gss_api.h struct gss_ctx *ctx_id, ctx_id 254 include/linux/sunrpc/gss_krb5.h gss_wrap_kerberos(struct gss_ctx *ctx_id, int offset, ctx_id 258 include/linux/sunrpc/gss_krb5.h gss_unwrap_kerberos(struct gss_ctx *ctx_id, int offset, int len, ctx_id 293 include/linux/syscalls.h asmlinkage long sys_io_cancel(aio_context_t ctx_id, struct iocb __user *iocb, ctx_id 295 include/linux/syscalls.h asmlinkage long sys_io_getevents(aio_context_t ctx_id, ctx_id 300 include/linux/syscalls.h asmlinkage long sys_io_getevents_time32(__u32 ctx_id, ctx_id 305 include/linux/syscalls.h asmlinkage long sys_io_pgetevents(aio_context_t ctx_id, ctx_id 311 include/linux/syscalls.h asmlinkage long sys_io_pgetevents_time32(aio_context_t ctx_id, ctx_id 238 include/uapi/drm/amdgpu_drm.h __u32 ctx_id; ctx_id 245 include/uapi/drm/amdgpu_drm.h __u32 ctx_id; ctx_id 294 include/uapi/drm/amdgpu_drm.h __u32 ctx_id; ctx_id 422 include/uapi/drm/amdgpu_drm.h __u32 ctx_id; ctx_id 436 include/uapi/drm/amdgpu_drm.h __u32 ctx_id; ctx_id 551 include/uapi/drm/amdgpu_drm.h __u32 ctx_id; ctx_id 609 include/uapi/drm/amdgpu_drm.h __u32 ctx_id; ctx_id 213 include/uapi/drm/drm.h unsigned int ctx_id; /**< Context requesting private mapping */ ctx_id 1481 include/uapi/drm/i915_drm.h __u32 ctx_id; /* output: id of new context*/ ctx_id 1486 include/uapi/drm/i915_drm.h __u32 ctx_id; /* output: id of new context*/ ctx_id 1496 include/uapi/drm/i915_drm.h __u32 ctx_id; ctx_id 1742 include/uapi/drm/i915_drm.h __u32 ctx_id; ctx_id 1796 include/uapi/drm/i915_drm.h __u32 ctx_id; ctx_id 106 include/uapi/linux/virtio_gpu.h __le32 ctx_id; ctx_id 169 include/uapi/misc/habanalabs.h __u32 ctx_id; ctx_id 194 include/uapi/misc/habanalabs.h __u32 ctx_id; ctx_id 254 include/uapi/misc/habanalabs.h __u32 ctx_id; ctx_id 279 include/uapi/misc/habanalabs.h __u32 ctx_id; ctx_id 374 include/uapi/misc/habanalabs.h __u32 ctx_id; ctx_id 505 include/uapi/misc/habanalabs.h __u32 ctx_id; ctx_id 661 net/sunrpc/auth_gss/gss_krb5_mech.c struct gss_ctx *ctx_id, ctx_id 679 net/sunrpc/auth_gss/gss_krb5_mech.c ctx_id->internal_ctx_id = ctx; ctx_id 376 net/sunrpc/auth_gss/gss_mech_switch.c struct gss_ctx **ctx_id, ctx_id 380 net/sunrpc/auth_gss/gss_mech_switch.c if (!(*ctx_id = kzalloc(sizeof(**ctx_id), gfp_mask))) ctx_id 382 net/sunrpc/auth_gss/gss_mech_switch.c (*ctx_id)->mech_type = gss_mech_get(mech); ctx_id 385 net/sunrpc/auth_gss/gss_mech_switch.c *ctx_id, endtime, gfp_mask); ctx_id 429 net/sunrpc/auth_gss/gss_mech_switch.c gss_wrap(struct gss_ctx *ctx_id, ctx_id 434 net/sunrpc/auth_gss/gss_mech_switch.c return ctx_id->mech_type->gm_ops ctx_id 435 net/sunrpc/auth_gss/gss_mech_switch.c ->gss_wrap(ctx_id, offset, buf, inpages); ctx_id 439 net/sunrpc/auth_gss/gss_mech_switch.c gss_unwrap(struct gss_ctx *ctx_id, ctx_id 444 net/sunrpc/auth_gss/gss_mech_switch.c return ctx_id->mech_type->gm_ops ctx_id 445 net/sunrpc/auth_gss/gss_mech_switch.c ->gss_unwrap(ctx_id, offset, len, buf); ctx_id 684 net/sunrpc/auth_gss/svcauth_gss.c struct gss_ctx *ctx_id = rsci->mechctx; ctx_id 707 net/sunrpc/auth_gss/svcauth_gss.c if (gss_verify_mic(ctx_id, &rpchdr, &checksum) != GSS_S_COMPLETE) { ctx_id 741 net/sunrpc/auth_gss/svcauth_gss.c gss_write_verf(struct svc_rqst *rqstp, struct gss_ctx *ctx_id, u32 seq) ctx_id 762 net/sunrpc/auth_gss/svcauth_gss.c maj_stat = gss_get_mic(ctx_id, &verf_data, &mic); ctx_id 213 tools/include/uapi/drm/drm.h unsigned int ctx_id; /**< Context requesting private mapping */ ctx_id 1481 tools/include/uapi/drm/i915_drm.h __u32 ctx_id; /* output: id of new context*/ ctx_id 1486 tools/include/uapi/drm/i915_drm.h __u32 ctx_id; /* output: id of new context*/ ctx_id 1496 tools/include/uapi/drm/i915_drm.h __u32 ctx_id; ctx_id 1742 tools/include/uapi/drm/i915_drm.h __u32 ctx_id; ctx_id 1796 tools/include/uapi/drm/i915_drm.h __u32 ctx_id;