/linux-4.1.27/include/linux/ |
H A D | aio.h | 17 extern long do_io_submit(aio_context_t ctx_id, long nr, 22 static inline long do_io_submit(aio_context_t ctx_id, long nr, do_io_submit() argument
|
H A D | compat.h | 507 asmlinkage long compat_sys_io_getevents(compat_aio_context_t ctx_id, 512 asmlinkage long compat_sys_io_submit(compat_aio_context_t ctx_id, int nr,
|
H A D | syscalls.h | 501 asmlinkage long sys_io_getevents(aio_context_t ctx_id, 508 asmlinkage long sys_io_cancel(aio_context_t ctx_id, struct iocb __user *iocb,
|
/linux-4.1.27/include/linux/sunrpc/ |
H A D | gss_api.h | 50 struct gss_ctx **ctx_id, 54 struct gss_ctx *ctx_id, 58 struct gss_ctx *ctx_id, 62 struct gss_ctx *ctx_id, 67 struct gss_ctx *ctx_id, 71 struct gss_ctx **ctx_id); 108 struct gss_ctx *ctx_id, 112 struct gss_ctx *ctx_id, 116 struct gss_ctx *ctx_id, 120 struct gss_ctx *ctx_id, 125 struct gss_ctx *ctx_id,
|
H A D | gss_krb5.h | 256 gss_wrap_kerberos(struct gss_ctx *ctx_id, int offset, 260 gss_unwrap_kerberos(struct gss_ctx *ctx_id, int offset,
|
/linux-4.1.27/net/sunrpc/auth_gss/ |
H A D | gss_mech_switch.c | 389 struct gss_ctx **ctx_id, gss_import_sec_context() 393 if (!(*ctx_id = kzalloc(sizeof(**ctx_id), gfp_mask))) gss_import_sec_context() 395 (*ctx_id)->mech_type = gss_mech_get(mech); gss_import_sec_context() 398 *ctx_id, endtime, gfp_mask); gss_import_sec_context() 442 gss_wrap(struct gss_ctx *ctx_id, gss_wrap() argument 447 return ctx_id->mech_type->gm_ops gss_wrap() 448 ->gss_wrap(ctx_id, offset, buf, inpages); gss_wrap() 452 gss_unwrap(struct gss_ctx *ctx_id, gss_unwrap() argument 456 return ctx_id->mech_type->gm_ops gss_unwrap() 457 ->gss_unwrap(ctx_id, offset, buf); gss_unwrap() 387 gss_import_sec_context(const void *input_token, size_t bufsize, struct gss_api_mech *mech, struct gss_ctx **ctx_id, time_t *endtime, gfp_t gfp_mask) gss_import_sec_context() argument
|
H A D | gss_krb5_mech.c | 681 struct gss_ctx *ctx_id, gss_import_sec_context_kerberos() 699 ctx_id->internal_ctx_id = ctx; gss_import_sec_context_kerberos() 680 gss_import_sec_context_kerberos(const void *p, size_t len, struct gss_ctx *ctx_id, time_t *endtime, gfp_t gfp_mask) gss_import_sec_context_kerberos() argument
|
H A D | svcauth_gss.c | 662 struct gss_ctx *ctx_id = rsci->mechctx; gss_verify_header() local 685 if (gss_verify_mic(ctx_id, &rpchdr, &checksum) != GSS_S_COMPLETE) { gss_verify_header() 719 gss_write_verf(struct svc_rqst *rqstp, struct gss_ctx *ctx_id, u32 seq) gss_write_verf() argument 736 maj_stat = gss_get_mic(ctx_id, &verf_data, &mic); gss_write_verf()
|
/linux-4.1.27/drivers/net/ethernet/qlogic/qlcnic/ |
H A D | qlcnic_83xx_hw.h | 172 u16 ctx_id; member in struct:qlcnic_rcv_mbx_out 178 u16 ctx_id; 194 u16 ctx_id; member in struct:qlcnic_add_rings_mbx_out 196 u16 ctx_id; 231 * @ctx_id: transmit context id 238 u16 ctx_id; member in struct:qlcnic_tx_mbx_out 244 u16 ctx_id; member in struct:qlcnic_tx_mbx_out
|
H A D | qlcnic_sriov_pf.c | 901 vf->rx_ctx_id = mbx_out->ctx_id; qlcnic_sriov_pf_create_rx_ctx_cmd() 960 vf->tx_ctx_id = mbx_out->ctx_id; qlcnic_sriov_pf_create_tx_ctx_cmd() 1200 u16 ctx_id, pkts, time; qlcnic_sriov_validate_cfg_intrcoal() local 1205 ctx_id = cmd->req.arg[1] >> 16; qlcnic_sriov_validate_cfg_intrcoal() 1211 if (ctx_id != vf->rx_ctx_id || pkts > coal->rx_packets || qlcnic_sriov_validate_cfg_intrcoal() 1216 if (ctx_id != vf->tx_ctx_id || pkts > coal->tx_packets || qlcnic_sriov_validate_cfg_intrcoal() 1232 netdev_err(adapter->netdev, "Received: ctx_id 0x%x packets 0x%x time_us 0x%x type 0x%x\n", qlcnic_sriov_validate_cfg_intrcoal() 1233 ctx_id, pkts, time, type); qlcnic_sriov_validate_cfg_intrcoal()
|
H A D | qlcnic_ctx.c | 499 tx_ring->ctx_id = le16_to_cpu(prsp->context_id); qlcnic_82xx_fw_cmd_create_tx_ctx() 509 tx_ring->ctx_id, tx_ring->state); qlcnic_82xx_fw_cmd_create_tx_ctx() 536 cmd.req.arg[1] = tx_ring->ctx_id; qlcnic_82xx_fw_cmd_del_tx_ctx()
|
H A D | qlcnic_83xx_hw.c | 1256 recv_ctx->context_id = mbx_out->ctx_id; qlcnic_83xx_create_rx_ctx() 1304 cmd.req.arg[1] = tx_ring->ctx_id | temp; qlcnic_83xx_del_tx_ctx() 1376 tx->ctx_id = mbx_out->ctx_id; qlcnic_83xx_create_tx_ctx() 1384 tx->ctx_id, mbx_out->state); qlcnic_83xx_create_tx_ctx() 2225 temp = adapter->tx_ring->ctx_id; qlcnic_83xx_set_tx_intr_coal() 3486 cmd.req.arg[1] = BIT_1 | (adapter->tx_ring->ctx_id << 16); qlcnic_83xx_get_stats()
|
H A D | qlcnic.h | 620 u16 ctx_id; member in struct:qlcnic_host_tx_ring
|
H A D | qlcnic_main.c | 3033 ring, tx_ring->ctx_id); qlcnic_dump_rings()
|
/linux-4.1.27/fs/ |
H A D | aio.c | 1020 static struct kioctx *lookup_ioctx(unsigned long ctx_id) lookup_ioctx() argument 1022 struct aio_ring __user *ring = (void __user *)ctx_id; lookup_ioctx() 1038 if (ctx && ctx->user_id == ctx_id) { lookup_ioctx() 1563 long do_io_submit(aio_context_t ctx_id, long nr, do_io_submit() argument 1580 ctx = lookup_ioctx(ctx_id); do_io_submit() 1619 * specified by ctx_id is invalid, if nr is < 0, if the iocb at 1628 SYSCALL_DEFINE3(io_submit, aio_context_t, ctx_id, long, nr, SYSCALL_DEFINE3() 1631 return do_io_submit(ctx_id, nr, iocbpp, 0); SYSCALL_DEFINE3() 1661 * May fail with -EINVAL if aio_context specified by ctx_id is 1665 SYSCALL_DEFINE3(io_cancel, aio_context_t, ctx_id, struct iocb __user *, iocb, SYSCALL_DEFINE3() 1677 ctx = lookup_ioctx(ctx_id); SYSCALL_DEFINE3() 1707 * the completion queue for the aio_context specified by ctx_id. If 1709 * -EINVAL if ctx_id is invalid, if min_nr is out of range, if nr is 1717 SYSCALL_DEFINE5(io_getevents, aio_context_t, ctx_id, SYSCALL_DEFINE5() 1723 struct kioctx *ioctx = lookup_ioctx(ctx_id); SYSCALL_DEFINE5()
|
H A D | compat.c | 523 COMPAT_SYSCALL_DEFINE5(io_getevents, compat_aio_context_t, ctx_id, COMPAT_SYSCALL_DEFINE5() 540 return sys_io_getevents(ctx_id, min_nr, nr, events, ut); COMPAT_SYSCALL_DEFINE5() 636 COMPAT_SYSCALL_DEFINE3(io_submit, compat_aio_context_t, ctx_id, COMPAT_SYSCALL_DEFINE3() 651 ret = do_io_submit(ctx_id, nr, iocb64, 1); COMPAT_SYSCALL_DEFINE3()
|
/linux-4.1.27/drivers/gpu/drm/ |
H A D | drm_ioc32.c | 579 unsigned int ctx_id; /**< Context requesting private mapping */ member in struct:drm_ctx_priv_map32 596 if (__put_user(req32.ctx_id, &request->ctx_id) compat_drm_setsareactx() 610 unsigned int ctx_id; compat_drm_getsareactx() local 614 || __get_user(ctx_id, &argp->ctx_id)) compat_drm_getsareactx() 620 if (__put_user(ctx_id, &request->ctx_id)) compat_drm_getsareactx()
|
H A D | drm_context.c | 166 map = idr_find(&dev->ctx_idr, request->ctx_id); drm_legacy_getsareactx() 223 if (IS_ERR(idr_replace(&dev->ctx_idr, map, request->ctx_id))) drm_legacy_setsareactx()
|
/linux-4.1.27/drivers/gpu/drm/i915/ |
H A D | i915_gem_context.c | 866 args->ctx_id = ctx->user_handle; i915_gem_context_create_ioctl() 867 DRM_DEBUG_DRIVER("HW context %d created\n", args->ctx_id); i915_gem_context_create_ioctl() 880 if (args->ctx_id == DEFAULT_CONTEXT_HANDLE) i915_gem_context_destroy_ioctl() 887 ctx = i915_gem_context_get(file_priv, args->ctx_id); i915_gem_context_destroy_ioctl() 897 DRM_DEBUG_DRIVER("HW context %d destroyed\n", args->ctx_id); i915_gem_context_destroy_ioctl() 913 ctx = i915_gem_context_get(file_priv, args->ctx_id); i915_gem_context_getparam_ioctl() 945 ctx = i915_gem_context_get(file_priv, args->ctx_id); i915_gem_context_setparam_ioctl()
|
H A D | i915_gem_execbuffer.c | 986 struct intel_engine_cs *ring, const u32 ctx_id) i915_gem_validate_context() 991 if (ring->id != RCS && ctx_id != DEFAULT_CONTEXT_HANDLE) i915_gem_validate_context() 994 ctx = i915_gem_context_get(file->driver_priv, ctx_id); i915_gem_validate_context() 1000 DRM_DEBUG("Context %u tried to submit while banned\n", ctx_id); i915_gem_validate_context() 1007 DRM_DEBUG("Could not create LRC %u: %d\n", ctx_id, ret); i915_gem_validate_context() 1414 const u32 ctx_id = i915_execbuffer2_get_context_id(*args); i915_gem_do_execbuffer() local 1494 ctx = i915_gem_validate_context(dev, file, ring, ctx_id); i915_gem_do_execbuffer() 985 i915_gem_validate_context(struct drm_device *dev, struct drm_file *file, struct intel_engine_cs *ring, const u32 ctx_id) i915_gem_validate_context() argument
|
H A D | intel_uncore.c | 1288 if (args->ctx_id == DEFAULT_CONTEXT_HANDLE && !capable(CAP_SYS_ADMIN)) i915_get_reset_stats_ioctl() 1295 ctx = i915_gem_context_get(file->driver_priv, args->ctx_id); i915_get_reset_stats_ioctl()
|
H A D | i915_debugfs.c | 1977 u32 ctx_id; i915_execlists() local 2001 ctx_id = I915_READ(RING_EXECLIST_STATUS(ring) + 4); for_each_ring() 2003 status, ctx_id); for_each_ring() 2017 ctx_id = I915_READ(RING_CONTEXT_STATUS_BUF(ring) + 8*i + 4); for_each_ring() 2020 i, status, ctx_id); for_each_ring()
|
/linux-4.1.27/arch/um/os-Linux/ |
H A D | aio.c | 46 static long io_getevents(aio_context_t ctx_id, long min_nr, long nr, io_getevents() argument 49 return syscall(__NR_io_getevents, ctx_id, min_nr, nr, events, timeout); io_getevents()
|
/linux-4.1.27/include/uapi/drm/ |
H A D | i915_drm.h | 1055 __u32 ctx_id; member in struct:drm_i915_gem_context_create 1060 __u32 ctx_id; member in struct:drm_i915_gem_context_destroy 1078 __u32 ctx_id; member in struct:drm_i915_reset_stats 1108 __u32 ctx_id; member in struct:drm_i915_gem_context_param
|
H A D | drm.h | 201 unsigned int ctx_id; /**< Context requesting private mapping */ member in struct:drm_ctx_priv_map
|
/linux-4.1.27/drivers/crypto/ |
H A D | picoxcell_crypto.c | 97 unsigned ctx_id; member in struct:spacc_req 607 req->ctx_id = spacc_load_ctx(&ctx->generic, ctx->cipher_key, spacc_aead_submit() 642 ctrl = spacc_alg->ctrl_default | (req->ctx_id << SPA_CTRL_CTX_IDX) | spacc_aead_submit() 952 req->ctx_id = spacc_load_ctx(&ctx->generic, ctx->key, spacc_ablk_submit() 965 ctrl = spacc_alg->ctrl_default | (req->ctx_id << SPA_CTRL_CTX_IDX) | spacc_ablk_submit()
|
/linux-4.1.27/arch/s390/kernel/ |
H A D | compat_wrapper.c | 172 COMPAT_SYSCALL_WRAP3(io_cancel, aio_context_t, ctx_id, struct iocb __user *, iocb, struct io_event __user *, result);
|
/linux-4.1.27/drivers/net/ethernet/qlogic/netxen/ |
H A D | netxen_nic_ctx.c | 781 recv_ctx->hwctx->ctx_id = cpu_to_le32(port); netxen_alloc_hw_resources()
|
H A D | netxen_nic.h | 240 * Bit 18-27 : ctx_id => Context id 284 __le32 ctx_id; member in struct:netxen_ring_ctx
|