ticket 45 arch/ia64/include/asm/spinlock.h int *p = (int *)&lock->lock, ticket, serve; ticket 47 arch/ia64/include/asm/spinlock.h ticket = ia64_fetchadd(1, p, acq); ticket 49 arch/ia64/include/asm/spinlock.h if (!(((ticket >> TICKET_SHIFT) ^ ticket) & TICKET_MASK)) ticket 57 arch/ia64/include/asm/spinlock.h if (!(((serve >> TICKET_SHIFT) ^ ticket) & TICKET_MASK)) ticket 113 arch/ia64/mm/tlb.c unsigned long ticket; ticket 119 arch/ia64/mm/tlb.c ss->ticket = 0; ticket 125 arch/ia64/mm/tlb.c unsigned long t = ia64_fetchadd(1, &ss->ticket, acq), serve; ticket 192 arch/mips/cavium-octeon/executive/cvmx-cmd-queue.c ticket[__cvmx_cmd_queue_get_index(queue_id)] = 0; ticket 154 arch/mips/include/asm/octeon/cvmx-cmd-queue.h uint64_t ticket[(CVMX_CMD_QUEUE_END >> 16) * 256]; ticket 278 arch/mips/include/asm/octeon/cvmx-cmd-queue.h [ticket_ptr] "=" GCC_OFF_SMALL_ASM()(__cvmx_cmd_queue_state_ptr->ticket[__cvmx_cmd_queue_get_index(queue_id)]), ticket 279 arch/mips/include/asm/octeon/cvmx-cmd-queue.h [now_serving] "=m"(qptr->now_serving), [ticket] "=r"(tmp), ticket 470 drivers/gpu/drm/amd/amdgpu/amdgpu.h struct ww_acquire_ctx ticket; ticket 543 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c struct ww_acquire_ctx ticket; /* Reservation ticket */ ticket 588 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c ret = ttm_eu_reserve_buffers(&ctx->ticket, &ctx->list, ticket 661 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c ret = ttm_eu_reserve_buffers(&ctx->ticket, &ctx->list, ticket 695 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c ttm_eu_backoff_reservation(&ctx->ticket, &ctx->list); ticket 1765 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c struct ww_acquire_ctx ticket; ticket 1801 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c ret = ttm_eu_reserve_buffers(&ticket, &resv_list, false, &duplicates, ticket 1860 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c ttm_eu_backoff_reservation(&ticket, &resv_list); ticket 1999 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c ret = ttm_eu_reserve_buffers(&ctx.ticket, &ctx.list, ticket 2091 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c ttm_eu_backoff_reservation(&ctx.ticket, &ctx.list); ticket 648 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c r = ttm_eu_reserve_buffers(&p->ticket, &p->validated, true, ticket 717 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c ttm_eu_backoff_reservation(&p->ticket, &p->validated); ticket 754 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c ttm_eu_backoff_reservation(&parser->ticket, ticket 1327 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c amdgpu_vm_bo_trace_cs(&fpriv->vm, &p->ticket); ticket 1336 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c ttm_eu_fence_buffer_objects(&p->ticket, &p->validated, p->fence); ticket 1731 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c if (dma_resv_locking_ctx((*bo)->tbo.base.resv) != &parser->ticket) ticket 69 drivers/gpu/drm/amd/amdgpu/amdgpu_csa.c struct ww_acquire_ctx ticket; ticket 83 drivers/gpu/drm/amd/amdgpu/amdgpu_csa.c r = ttm_eu_reserve_buffers(&ticket, &list, true, NULL, false); ticket 91 drivers/gpu/drm/amd/amdgpu/amdgpu_csa.c ttm_eu_backoff_reservation(&ticket, &list); ticket 103 drivers/gpu/drm/amd/amdgpu/amdgpu_csa.c ttm_eu_backoff_reservation(&ticket, &list); ticket 107 drivers/gpu/drm/amd/amdgpu/amdgpu_csa.c ttm_eu_backoff_reservation(&ticket, &list); ticket 165 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c struct ww_acquire_ctx ticket; ticket 178 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c r = ttm_eu_reserve_buffers(&ticket, &list, false, &duplicates, false); ticket 203 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c ttm_eu_backoff_reservation(&ticket, &list); ticket 554 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c struct ww_acquire_ctx ticket; ticket 616 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c r = ttm_eu_reserve_buffers(&ticket, &list, true, &duplicates, false); ticket 662 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c ttm_eu_backoff_reservation(&ticket, &list); ticket 2410 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c void amdgpu_vm_bo_trace_cs(struct amdgpu_vm *vm, struct ww_acquire_ctx *ticket) ticket 2424 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c ticket) ticket 394 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h void amdgpu_vm_bo_trace_cs(struct amdgpu_vm *vm, struct ww_acquire_ctx *ticket); ticket 4496 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c struct ww_acquire_ctx ticket; ticket 4520 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c r = ttm_eu_reserve_buffers(&ticket, &list, false, NULL, true); ticket 4535 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c ttm_eu_backoff_reservation(&ticket, &list); ticket 4542 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c ttm_eu_backoff_reservation(&ticket, &list); ticket 4549 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c ttm_eu_backoff_reservation(&ticket, &list); ticket 122 drivers/gpu/drm/etnaviv/etnaviv_gem_submit.c struct ww_acquire_ctx *ticket) ticket 137 drivers/gpu/drm/etnaviv/etnaviv_gem_submit.c ticket); ticket 147 drivers/gpu/drm/etnaviv/etnaviv_gem_submit.c ww_acquire_done(ticket); ticket 165 drivers/gpu/drm/etnaviv/etnaviv_gem_submit.c ticket); ticket 438 drivers/gpu/drm/etnaviv/etnaviv_gem_submit.c struct ww_acquire_ctx ticket; ticket 523 drivers/gpu/drm/etnaviv/etnaviv_gem_submit.c ww_acquire_init(&ticket, &reservation_ww_class); ticket 576 drivers/gpu/drm/etnaviv/etnaviv_gem_submit.c ret = submit_lock_objects(submit, &ticket); ticket 612 drivers/gpu/drm/etnaviv/etnaviv_gem_submit.c ww_acquire_fini(&ticket); ticket 136 drivers/gpu/drm/msm/msm_gem.h struct ww_acquire_ctx ticket; ticket 57 drivers/gpu/drm/msm/msm_gem_submit.c ww_acquire_init(&submit->ticket, &reservation_ww_class); ticket 185 drivers/gpu/drm/msm/msm_gem_submit.c &submit->ticket); ticket 192 drivers/gpu/drm/msm/msm_gem_submit.c ww_acquire_done(&submit->ticket); ticket 207 drivers/gpu/drm/msm/msm_gem_submit.c &submit->ticket); ticket 394 drivers/gpu/drm/msm/msm_gem_submit.c ww_acquire_fini(&submit->ticket); ticket 326 drivers/gpu/drm/nouveau/nouveau_gem.c struct ww_acquire_ctx ticket; ticket 371 drivers/gpu/drm/nouveau/nouveau_gem.c ww_acquire_fini(&op->ticket); ticket 387 drivers/gpu/drm/nouveau/nouveau_gem.c ww_acquire_init(&op->ticket, &reservation_ww_class); ticket 420 drivers/gpu/drm/nouveau/nouveau_gem.c ret = ttm_bo_reserve(&nvbo->bo, true, false, &op->ticket); ticket 428 drivers/gpu/drm/nouveau/nouveau_gem.c &op->ticket); ticket 475 drivers/gpu/drm/nouveau/nouveau_gem.c ww_acquire_done(&op->ticket); ticket 156 drivers/gpu/drm/qxl/qxl_drv.h struct ww_acquire_ctx ticket; ticket 262 drivers/gpu/drm/qxl/qxl_release.c ret = ttm_eu_reserve_buffers(&release->ticket, &release->bos, ticket 272 drivers/gpu/drm/qxl/qxl_release.c ttm_eu_backoff_reservation(&release->ticket, &release->bos); ticket 286 drivers/gpu/drm/qxl/qxl_release.c ttm_eu_backoff_reservation(&release->ticket, &release->bos); ticket 466 drivers/gpu/drm/qxl/qxl_release.c ww_acquire_fini(&release->ticket); ticket 1086 drivers/gpu/drm/radeon/radeon.h struct ww_acquire_ctx ticket; ticket 201 drivers/gpu/drm/radeon/radeon_cs.c r = radeon_bo_list_validate(p->rdev, &p->ticket, &p->validated, p->ring); ticket 432 drivers/gpu/drm/radeon/radeon_cs.c ttm_eu_fence_buffer_objects(&parser->ticket, ticket 436 drivers/gpu/drm/radeon/radeon_cs.c ttm_eu_backoff_reservation(&parser->ticket, ticket 554 drivers/gpu/drm/radeon/radeon_gem.c struct ww_acquire_ctx ticket; ticket 569 drivers/gpu/drm/radeon/radeon_gem.c r = ttm_eu_reserve_buffers(&ticket, &list, true, NULL, true); ticket 593 drivers/gpu/drm/radeon/radeon_gem.c ttm_eu_backoff_reservation(&ticket, &list); ticket 534 drivers/gpu/drm/radeon/radeon_object.c struct ww_acquire_ctx *ticket, ticket 545 drivers/gpu/drm/radeon/radeon_object.c r = ttm_eu_reserve_buffers(ticket, head, true, &duplicates, true); ticket 589 drivers/gpu/drm/radeon/radeon_object.c ttm_eu_backoff_reservation(ticket, head); ticket 144 drivers/gpu/drm/radeon/radeon_object.h struct ww_acquire_ctx *ticket, ticket 816 drivers/gpu/drm/ttm/ttm_bo.c struct ww_acquire_ctx *ticket) ticket 820 drivers/gpu/drm/ttm/ttm_bo.c if (!busy_bo || !ticket) ticket 825 drivers/gpu/drm/ttm/ttm_bo.c ticket); ticket 827 drivers/gpu/drm/ttm/ttm_bo.c r = dma_resv_lock(busy_bo->base.resv, ticket); ticket 844 drivers/gpu/drm/ttm/ttm_bo.c struct ww_acquire_ctx *ticket) ticket 860 drivers/gpu/drm/ttm/ttm_bo.c if (busy && !busy_bo && ticket != ticket 886 drivers/gpu/drm/ttm/ttm_bo.c ret = ttm_mem_evict_wait_busy(busy_bo, ctx, ticket); ticket 971 drivers/gpu/drm/ttm/ttm_bo.c struct ww_acquire_ctx *ticket; ticket 974 drivers/gpu/drm/ttm/ttm_bo.c ticket = dma_resv_locking_ctx(bo->base.resv); ticket 982 drivers/gpu/drm/ttm/ttm_bo.c ticket); ticket 56 drivers/gpu/drm/ttm/ttm_execbuf_util.c void ttm_eu_backoff_reservation(struct ww_acquire_ctx *ticket, ticket 78 drivers/gpu/drm/ttm/ttm_execbuf_util.c if (ticket) ticket 79 drivers/gpu/drm/ttm/ttm_execbuf_util.c ww_acquire_fini(ticket); ticket 95 drivers/gpu/drm/ttm/ttm_execbuf_util.c int ttm_eu_reserve_buffers(struct ww_acquire_ctx *ticket, ticket 109 drivers/gpu/drm/ttm/ttm_execbuf_util.c if (ticket) ticket 110 drivers/gpu/drm/ttm/ttm_execbuf_util.c ww_acquire_init(ticket, &reservation_ww_class); ticket 115 drivers/gpu/drm/ttm/ttm_execbuf_util.c ret = __ttm_bo_reserve(bo, intr, (ticket == NULL), ticket); ticket 148 drivers/gpu/drm/ttm/ttm_execbuf_util.c ticket); ticket 150 drivers/gpu/drm/ttm/ttm_execbuf_util.c dma_resv_lock_slow(bo->base.resv, ticket); ticket 162 drivers/gpu/drm/ttm/ttm_execbuf_util.c if (ticket) { ticket 163 drivers/gpu/drm/ttm/ttm_execbuf_util.c ww_acquire_done(ticket); ticket 164 drivers/gpu/drm/ttm/ttm_execbuf_util.c ww_acquire_fini(ticket); ticket 185 drivers/gpu/drm/ttm/ttm_execbuf_util.c void ttm_eu_fence_buffer_objects(struct ww_acquire_ctx *ticket, ticket 214 drivers/gpu/drm/ttm/ttm_execbuf_util.c if (ticket) ticket 215 drivers/gpu/drm/ttm/ttm_execbuf_util.c ww_acquire_fini(ticket); ticket 220 drivers/gpu/drm/virtio/virtgpu_drv.h int virtio_gpu_object_list_validate(struct ww_acquire_ctx *ticket, ticket 59 drivers/gpu/drm/virtio/virtgpu_ioctl.c int virtio_gpu_object_list_validate(struct ww_acquire_ctx *ticket, ticket 68 drivers/gpu/drm/virtio/virtgpu_ioctl.c ret = ttm_eu_reserve_buffers(ticket, head, true, NULL, true); ticket 77 drivers/gpu/drm/virtio/virtgpu_ioctl.c ttm_eu_backoff_reservation(ticket, head); ticket 119 drivers/gpu/drm/virtio/virtgpu_ioctl.c struct ww_acquire_ctx ticket; ticket 196 drivers/gpu/drm/virtio/virtgpu_ioctl.c ret = virtio_gpu_object_list_validate(&ticket, &validate_list); ticket 227 drivers/gpu/drm/virtio/virtgpu_ioctl.c ttm_eu_fence_buffer_objects(&ticket, &validate_list, &out_fence->f); ticket 237 drivers/gpu/drm/virtio/virtgpu_ioctl.c ttm_eu_backoff_reservation(&ticket, &validate_list); ticket 151 drivers/gpu/drm/virtio/virtgpu_object.c struct ww_acquire_ctx ticket; ticket 163 drivers/gpu/drm/virtio/virtgpu_object.c ret = virtio_gpu_object_list_validate(&ticket, &validate_list); ticket 169 drivers/gpu/drm/virtio/virtgpu_object.c ttm_eu_fence_buffer_objects(&ticket, &validate_list, ticket 174 drivers/gpu/drm/virtio/virtgpu_object.c ttm_eu_backoff_reservation(&ticket, &validate_list); ticket 474 drivers/gpu/drm/vmwgfx/vmwgfx_resource.c vmw_resource_check_buffer(struct ww_acquire_ctx *ticket, ticket 495 drivers/gpu/drm/vmwgfx/vmwgfx_resource.c ret = ttm_eu_reserve_buffers(ticket, &val_list, interruptible, NULL, ticket 514 drivers/gpu/drm/vmwgfx/vmwgfx_resource.c ttm_eu_backoff_reservation(ticket, &val_list); ticket 566 drivers/gpu/drm/vmwgfx/vmwgfx_resource.c vmw_resource_backoff_reservation(struct ww_acquire_ctx *ticket, ticket 576 drivers/gpu/drm/vmwgfx/vmwgfx_resource.c ttm_eu_backoff_reservation(ticket, &val_list); ticket 589 drivers/gpu/drm/vmwgfx/vmwgfx_resource.c static int vmw_resource_do_evict(struct ww_acquire_ctx *ticket, ticket 600 drivers/gpu/drm/vmwgfx/vmwgfx_resource.c ret = vmw_resource_check_buffer(ticket, res, interruptible, &val_buf); ticket 615 drivers/gpu/drm/vmwgfx/vmwgfx_resource.c vmw_resource_backoff_reservation(ticket, &val_buf); ticket 853 drivers/gpu/drm/vmwgfx/vmwgfx_resource.c struct ww_acquire_ctx ticket; ticket 868 drivers/gpu/drm/vmwgfx/vmwgfx_resource.c ret = vmw_resource_do_evict(&ticket, evict_res, false); ticket 81 drivers/gpu/drm/vmwgfx/vmwgfx_validation.h struct ww_acquire_ctx ticket; ticket 172 drivers/gpu/drm/vmwgfx/vmwgfx_validation.h return ttm_eu_reserve_buffers(&ctx->ticket, &ctx->bo_list, intr, ticket 187 drivers/gpu/drm/vmwgfx/vmwgfx_validation.h ttm_eu_backoff_reservation(&ctx->ticket, &ctx->bo_list); ticket 202 drivers/gpu/drm/vmwgfx/vmwgfx_validation.h ttm_eu_fence_buffer_objects(&ctx->ticket, &ctx->bo_list, ticket 224 fs/btrfs/space-info.c struct reserve_ticket *ticket; ticket 227 fs/btrfs/space-info.c ticket = list_first_entry(head, struct reserve_ticket, list); ticket 230 fs/btrfs/space-info.c if ((used + ticket->bytes <= space_info->total_bytes) || ticket 231 fs/btrfs/space-info.c can_overcommit(fs_info, space_info, ticket->bytes, flush, ticket 235 fs/btrfs/space-info.c ticket->bytes); ticket 236 fs/btrfs/space-info.c list_del_init(&ticket->list); ticket 237 fs/btrfs/space-info.c ticket->bytes = 0; ticket 239 fs/btrfs/space-info.c wake_up(&ticket->wait); ticket 462 fs/btrfs/space-info.c struct reserve_ticket *ticket = NULL; ticket 482 fs/btrfs/space-info.c ticket = list_first_entry(&space_info->priority_tickets, ticket 485 fs/btrfs/space-info.c ticket = list_first_entry(&space_info->tickets, ticket 487 fs/btrfs/space-info.c bytes_needed = (ticket) ? ticket->bytes : 0; ticket 634 fs/btrfs/space-info.c struct reserve_ticket *ticket; ticket 639 fs/btrfs/space-info.c list_for_each_entry(ticket, &space_info->tickets, list) ticket 640 fs/btrfs/space-info.c to_reclaim += ticket->bytes; ticket 641 fs/btrfs/space-info.c list_for_each_entry(ticket, &space_info->priority_tickets, list) ticket 642 fs/btrfs/space-info.c to_reclaim += ticket->bytes; ticket 704 fs/btrfs/space-info.c struct reserve_ticket *ticket; ticket 715 fs/btrfs/space-info.c ticket = list_first_entry(&space_info->tickets, ticket 729 fs/btrfs/space-info.c first_ticket_bytes = ticket->bytes; ticket 730 fs/btrfs/space-info.c else if (first_ticket_bytes > ticket->bytes) ticket 735 fs/btrfs/space-info.c ticket->bytes); ticket 737 fs/btrfs/space-info.c list_del_init(&ticket->list); ticket 738 fs/btrfs/space-info.c ticket->error = -ENOSPC; ticket 739 fs/btrfs/space-info.c wake_up(&ticket->wait); ticket 855 fs/btrfs/space-info.c struct reserve_ticket *ticket, ticket 876 fs/btrfs/space-info.c if (ticket->bytes == 0) { ticket 886 fs/btrfs/space-info.c struct reserve_ticket *ticket) ticket 893 fs/btrfs/space-info.c while (ticket->bytes > 0 && ticket->error == 0) { ticket 894 fs/btrfs/space-info.c ret = prepare_to_wait_event(&ticket->wait, &wait, TASK_KILLABLE); ticket 904 fs/btrfs/space-info.c list_del_init(&ticket->list); ticket 905 fs/btrfs/space-info.c ticket->error = -EINTR; ticket 912 fs/btrfs/space-info.c finish_wait(&ticket->wait, &wait); ticket 930 fs/btrfs/space-info.c struct reserve_ticket *ticket, ticket 937 fs/btrfs/space-info.c wait_reserve_ticket(fs_info, space_info, ticket); ticket 940 fs/btrfs/space-info.c priority_reclaim_metadata_space(fs_info, space_info, ticket, ticket 945 fs/btrfs/space-info.c priority_reclaim_metadata_space(fs_info, space_info, ticket, ticket 955 fs/btrfs/space-info.c ret = ticket->error; ticket 956 fs/btrfs/space-info.c if (ticket->bytes || ticket->error) { ticket 962 fs/btrfs/space-info.c list_del_init(&ticket->list); ticket 967 fs/btrfs/space-info.c ASSERT(list_empty(&ticket->list)); ticket 974 fs/btrfs/space-info.c ASSERT(!(ticket->bytes == 0 && ticket->error)); ticket 998 fs/btrfs/space-info.c struct reserve_ticket ticket; ticket 1033 fs/btrfs/space-info.c ticket.bytes = orig_bytes; ticket 1034 fs/btrfs/space-info.c ticket.error = 0; ticket 1035 fs/btrfs/space-info.c init_waitqueue_head(&ticket.wait); ticket 1037 fs/btrfs/space-info.c list_add_tail(&ticket.list, &space_info->tickets); ticket 1048 fs/btrfs/space-info.c list_add_tail(&ticket.list, ticket 1072 fs/btrfs/space-info.c return handle_reserve_ticket(fs_info, space_info, &ticket, flush); ticket 30 fs/xfs/xfs_log.c struct xlog_ticket *ticket, ticket 57 fs/xfs/xfs_log.c struct xlog_ticket *ticket, ticket 81 fs/xfs/xfs_log.c struct xlog_ticket *ticket); ticket 85 fs/xfs/xfs_log.c struct xlog_ticket *ticket); ticket 513 fs/xfs/xfs_log.c struct xlog_ticket *ticket, ticket 525 fs/xfs/xfs_log.c (((ticket->t_flags & XLOG_TIC_INITED) == 0) && ticket 526 fs/xfs/xfs_log.c (xlog_commit_record(log, ticket, iclog, &lsn)))) { ticket 533 fs/xfs/xfs_log.c trace_xfs_log_done_nonperm(log, ticket); ticket 539 fs/xfs/xfs_log.c xlog_ungrant_log_space(log, ticket); ticket 541 fs/xfs/xfs_log.c trace_xfs_log_done_perm(log, ticket); ticket 543 fs/xfs/xfs_log.c xlog_regrant_reserve_log_space(log, ticket); ticket 548 fs/xfs/xfs_log.c ticket->t_flags |= XLOG_TIC_INITED; ticket 551 fs/xfs/xfs_log.c xfs_log_ticket_put(ticket); ticket 1515 fs/xfs/xfs_log.c struct xlog_ticket *ticket, ticket 1532 fs/xfs/xfs_log.c error = xlog_write(log, &vec, ticket, commitlsnp, iclog, ticket 2000 fs/xfs/xfs_log.c struct xlog_ticket *ticket) ticket 2003 fs/xfs/xfs_log.c uint ophdr_spc = ticket->t_res_num_ophdrs * (uint)sizeof(xlog_op_header_t); ticket 2040 fs/xfs/xfs_log.c ticket->t_unit_res); ticket 2042 fs/xfs/xfs_log.c ticket->t_curr_res); ticket 2044 fs/xfs/xfs_log.c ticket->t_res_arr_sum, ticket->t_res_o_flow); ticket 2046 fs/xfs/xfs_log.c ticket->t_res_num_ophdrs, ophdr_spc); ticket 2048 fs/xfs/xfs_log.c ticket->t_res_arr_sum + ticket->t_res_o_flow + ophdr_spc); ticket 2050 fs/xfs/xfs_log.c ticket->t_res_num); ticket 2052 fs/xfs/xfs_log.c for (i = 0; i < ticket->t_res_num; i++) { ticket 2053 fs/xfs/xfs_log.c uint r_type = ticket->t_res_arr[i].r_type; ticket 2057 fs/xfs/xfs_log.c ticket->t_res_arr[i].r_len); ticket 2117 fs/xfs/xfs_log.c struct xlog_ticket *ticket, ticket 2126 fs/xfs/xfs_log.c if (ticket->t_flags & XLOG_TIC_INITED) ticket 2140 fs/xfs/xfs_log.c xlog_tic_add_region(ticket, vecp->i_len, vecp->i_type); ticket 2144 fs/xfs/xfs_log.c ticket->t_res_num_ophdrs += headers; ticket 2157 fs/xfs/xfs_log.c struct xlog_ticket *ticket) ticket 2159 fs/xfs/xfs_log.c if (!(ticket->t_flags & XLOG_TIC_INITED)) ticket 2162 fs/xfs/xfs_log.c ophdr->oh_tid = cpu_to_be32(ticket->t_tid); ticket 2163 fs/xfs/xfs_log.c ophdr->oh_clientid = ticket->t_clientid; ticket 2168 fs/xfs/xfs_log.c ticket->t_flags &= ~XLOG_TIC_INITED; ticket 2177 fs/xfs/xfs_log.c struct xlog_ticket *ticket, ticket 2180 fs/xfs/xfs_log.c ophdr->oh_tid = cpu_to_be32(ticket->t_tid); ticket 2181 fs/xfs/xfs_log.c ophdr->oh_clientid = ticket->t_clientid; ticket 2200 fs/xfs/xfs_log.c ophdr->oh_clientid, ticket); ticket 2215 fs/xfs/xfs_log.c struct xlog_ticket *ticket, ticket 2250 fs/xfs/xfs_log.c ticket->t_curr_res -= sizeof(struct xlog_op_header); ticket 2251 fs/xfs/xfs_log.c ticket->t_res_num_ophdrs++; ticket 2345 fs/xfs/xfs_log.c struct xlog_ticket *ticket, ticket 2364 fs/xfs/xfs_log.c len = xlog_write_calc_vec_length(ticket, log_vector); ticket 2371 fs/xfs/xfs_log.c if (ticket->t_flags & XLOG_TIC_INITED) ticket 2372 fs/xfs/xfs_log.c ticket->t_curr_res -= sizeof(xlog_op_header_t); ticket 2379 fs/xfs/xfs_log.c ticket->t_curr_res -= sizeof(xlog_op_header_t); ticket 2381 fs/xfs/xfs_log.c if (ticket->t_curr_res < 0) { ticket 2384 fs/xfs/xfs_log.c xlog_print_tic_res(log->l_mp, ticket); ticket 2395 fs/xfs/xfs_log.c error = xlog_state_get_iclog_space(log, len, &iclog, ticket, ticket 2430 fs/xfs/xfs_log.c start_rec_copy = xlog_write_start_rec(ptr, ticket); ticket 2437 fs/xfs/xfs_log.c ophdr = xlog_write_setup_ophdr(log, ptr, ticket, flags); ticket 2444 fs/xfs/xfs_log.c len += xlog_write_setup_copy(ticket, ophdr, ticket 2981 fs/xfs/xfs_log.c struct xlog_ticket *ticket, ticket 3017 fs/xfs/xfs_log.c ticket->t_curr_res -= log->l_iclog_hsize; ticket 3018 fs/xfs/xfs_log.c xlog_tic_add_region(ticket, ticket 3090 fs/xfs/xfs_log.c struct xlog_ticket *ticket) ticket 3092 fs/xfs/xfs_log.c trace_xfs_log_regrant_reserve_enter(log, ticket); ticket 3094 fs/xfs/xfs_log.c if (ticket->t_cnt > 0) ticket 3095 fs/xfs/xfs_log.c ticket->t_cnt--; ticket 3098 fs/xfs/xfs_log.c ticket->t_curr_res); ticket 3100 fs/xfs/xfs_log.c ticket->t_curr_res); ticket 3101 fs/xfs/xfs_log.c ticket->t_curr_res = ticket->t_unit_res; ticket 3102 fs/xfs/xfs_log.c xlog_tic_reset_res(ticket); ticket 3104 fs/xfs/xfs_log.c trace_xfs_log_regrant_reserve_sub(log, ticket); ticket 3107 fs/xfs/xfs_log.c if (ticket->t_cnt > 0) ticket 3111 fs/xfs/xfs_log.c ticket->t_unit_res); ticket 3113 fs/xfs/xfs_log.c trace_xfs_log_regrant_reserve_exit(log, ticket); ticket 3115 fs/xfs/xfs_log.c ticket->t_curr_res = ticket->t_unit_res; ticket 3116 fs/xfs/xfs_log.c xlog_tic_reset_res(ticket); ticket 3137 fs/xfs/xfs_log.c struct xlog_ticket *ticket) ticket 3141 fs/xfs/xfs_log.c if (ticket->t_cnt > 0) ticket 3142 fs/xfs/xfs_log.c ticket->t_cnt--; ticket 3144 fs/xfs/xfs_log.c trace_xfs_log_ungrant_enter(log, ticket); ticket 3145 fs/xfs/xfs_log.c trace_xfs_log_ungrant_sub(log, ticket); ticket 3151 fs/xfs/xfs_log.c bytes = ticket->t_curr_res; ticket 3152 fs/xfs/xfs_log.c if (ticket->t_cnt > 0) { ticket 3153 fs/xfs/xfs_log.c ASSERT(ticket->t_flags & XLOG_TIC_PERM_RESERV); ticket 3154 fs/xfs/xfs_log.c bytes += ticket->t_unit_res*ticket->t_cnt; ticket 3160 fs/xfs/xfs_log.c trace_xfs_log_ungrant_exit(log, ticket); ticket 3540 fs/xfs/xfs_log.c xlog_ticket_t *ticket) ticket 3542 fs/xfs/xfs_log.c ASSERT(atomic_read(&ticket->t_ref) > 0); ticket 3543 fs/xfs/xfs_log.c if (atomic_dec_and_test(&ticket->t_ref)) ticket 3544 fs/xfs/xfs_log.c kmem_zone_free(xfs_log_ticket_zone, ticket); ticket 3549 fs/xfs/xfs_log.c xlog_ticket_t *ticket) ticket 3551 fs/xfs/xfs_log.c ASSERT(atomic_read(&ticket->t_ref) > 0); ticket 3552 fs/xfs/xfs_log.c atomic_inc(&ticket->t_ref); ticket 3553 fs/xfs/xfs_log.c return ticket; ticket 109 fs/xfs/xfs_log.h struct xlog_ticket *ticket, ticket 129 fs/xfs/xfs_log.h struct xlog_ticket **ticket, ticket 136 fs/xfs/xfs_log.h struct xlog_ticket *xfs_log_ticket_get(struct xlog_ticket *ticket); ticket 137 fs/xfs/xfs_log.h void xfs_log_ticket_put(struct xlog_ticket *ticket); ticket 65 fs/xfs/xfs_log_cil.c log->l_cilp->xc_ctx->ticket = xlog_cil_ticket_alloc(log); ticket 425 fs/xfs/xfs_log_cil.c if (ctx->ticket->t_curr_res == 0) { ticket 426 fs/xfs/xfs_log_cil.c ctx_res = ctx->ticket->t_unit_res; ticket 427 fs/xfs/xfs_log_cil.c ctx->ticket->t_curr_res = ctx_res; ticket 438 fs/xfs/xfs_log_cil.c ctx->ticket->t_unit_res += split_res; ticket 439 fs/xfs/xfs_log_cil.c ctx->ticket->t_curr_res += split_res; ticket 664 fs/xfs/xfs_log_cil.c new_ctx->ticket = xlog_cil_ticket_alloc(log); ticket 792 fs/xfs/xfs_log_cil.c tic = ctx->ticket; ticket 874 fs/xfs/xfs_log_cil.c xfs_log_ticket_put(new_ctx->ticket); ticket 1217 fs/xfs/xfs_log_cil.c if (log->l_cilp->xc_ctx->ticket) ticket 1218 fs/xfs/xfs_log_cil.c xfs_log_ticket_put(log->l_cilp->xc_ctx->ticket); ticket 243 fs/xfs/xfs_log_priv.h struct xlog_ticket *ticket; /* chkpt ticket */ ticket 446 fs/xfs/xfs_log_priv.h void xlog_print_tic_res(struct xfs_mount *mp, struct xlog_ticket *ticket); ticket 658 include/drm/ttm/ttm_bo_driver.h struct ww_acquire_ctx *ticket) ticket 664 include/drm/ttm/ttm_bo_driver.h if (WARN_ON(ticket)) ticket 672 include/drm/ttm/ttm_bo_driver.h ret = dma_resv_lock_interruptible(bo->base.resv, ticket); ticket 674 include/drm/ttm/ttm_bo_driver.h ret = dma_resv_lock(bo->base.resv, ticket); ticket 726 include/drm/ttm/ttm_bo_driver.h struct ww_acquire_ctx *ticket) ticket 732 include/drm/ttm/ttm_bo_driver.h ret = __ttm_bo_reserve(bo, interruptible, no_wait, ticket); ticket 751 include/drm/ttm/ttm_bo_driver.h struct ww_acquire_ctx *ticket) ticket 759 include/drm/ttm/ttm_bo_driver.h ticket); ticket 761 include/drm/ttm/ttm_bo_driver.h dma_resv_lock_slow(bo->base.resv, ticket); ticket 62 include/drm/ttm/ttm_execbuf_util.h extern void ttm_eu_backoff_reservation(struct ww_acquire_ctx *ticket, ticket 100 include/drm/ttm/ttm_execbuf_util.h extern int ttm_eu_reserve_buffers(struct ww_acquire_ctx *ticket, ticket 117 include/drm/ttm/ttm_execbuf_util.h extern void ttm_eu_fence_buffer_objects(struct ww_acquire_ctx *ticket, ticket 31 include/keys/rxrpc-type.h u8 ticket[0]; /* the encrypted ticket */ ticket 76 include/keys/rxrpc-type.h u8 *ticket; /* krb5 ticket */ ticket 103 include/keys/rxrpc-type.h u8 ticket[0]; ticket 132 net/rxrpc/key.c memcpy(&token->kad->ticket, &xdr[8], tktlen); ticket 146 net/rxrpc/key.c token->kad->ticket[0], token->kad->ticket[1], ticket 147 net/rxrpc/key.c token->kad->ticket[2], token->kad->ticket[3], ticket 148 net/rxrpc/key.c token->kad->ticket[4], token->kad->ticket[5], ticket 149 net/rxrpc/key.c token->kad->ticket[6], token->kad->ticket[7]); ticket 207 net/rxrpc/key.c kfree(rxk5->ticket); ticket 506 net/rxrpc/key.c ret = rxrpc_krb5_decode_ticket(&rxk5->ticket, &rxk5->ticket_len, ticket 748 net/rxrpc/key.c v1->ticket[0], v1->ticket[1], ticket 749 net/rxrpc/key.c v1->ticket[2], v1->ticket[3], ticket 750 net/rxrpc/key.c v1->ticket[4], v1->ticket[5], ticket 751 net/rxrpc/key.c v1->ticket[6], v1->ticket[7]); ticket 773 net/rxrpc/key.c memcpy(&token->kad->ticket, v1->ticket, v1->ticket_length); ticket 1173 net/rxrpc/key.c ENCODE_DATA(token->kad->ticket_len, token->kad->ticket); ticket 1207 net/rxrpc/key.c ENCODE_DATA(token->k5->ticket_len, token->k5->ticket); ticket 741 net/rxrpc/rxkad.c iov[2].iov_base = (void *)s2->ticket; ticket 896 net/rxrpc/rxkad.c void *ticket, size_t ticket_len, ticket 931 net/rxrpc/rxkad.c ASSERTCMP((unsigned long) ticket & 7UL, ==, 0); ticket 941 net/rxrpc/rxkad.c sg_init_one(&sg[0], ticket, ticket_len); ticket 947 net/rxrpc/rxkad.c p = ticket; ticket 1095 net/rxrpc/rxkad.c void *ticket; ticket 1138 net/rxrpc/rxkad.c ticket = kmalloc(ticket_len, GFP_NOFS); ticket 1139 net/rxrpc/rxkad.c if (!ticket) ticket 1145 net/rxrpc/rxkad.c ticket, ticket_len) < 0) ticket 1148 net/rxrpc/rxkad.c ret = rxkad_decrypt_ticket(conn, skb, ticket, ticket_len, &session_key, ticket 1216 net/rxrpc/rxkad.c kfree(ticket); ticket 1224 net/rxrpc/rxkad.c kfree(ticket); ticket 1232 net/rxrpc/rxkad.c kfree(ticket);