waiters 653 drivers/gpu/drm/amd/amdkfd/kfd_events.c static void free_waiters(uint32_t num_events, struct kfd_event_waiter *waiters) waiters 658 drivers/gpu/drm/amd/amdkfd/kfd_events.c if (waiters[i].event) waiters 659 drivers/gpu/drm/amd/amdkfd/kfd_events.c remove_wait_queue(&waiters[i].event->wq, waiters 660 drivers/gpu/drm/amd/amdkfd/kfd_events.c &waiters[i].wait); waiters 662 drivers/gpu/drm/amd/amdkfd/kfd_events.c kfree(waiters); waiters 594 drivers/gpu/drm/radeon/radeon.h signed waiters; waiters 50 drivers/gpu/drm/radeon/radeon_semaphore.c (*semaphore)->waiters = 0; waiters 66 drivers/gpu/drm/radeon/radeon_semaphore.c --semaphore->waiters; waiters 83 drivers/gpu/drm/radeon/radeon_semaphore.c ++semaphore->waiters; waiters 99 drivers/gpu/drm/radeon/radeon_semaphore.c if ((*semaphore)->waiters > 0) { waiters 176 drivers/gpu/drm/radeon/radeon_trace.h __field(signed, waiters) waiters 182 drivers/gpu/drm/radeon/radeon_trace.h __entry->waiters = sem->waiters; waiters 187 drivers/gpu/drm/radeon/radeon_trace.h __entry->waiters, __entry->gpu_addr) waiters 102 drivers/greybus/greybus_trace.h __field(int, waiters) waiters 112 drivers/greybus/greybus_trace.h __entry->waiters = atomic_read(&operation->waiters); waiters 118 drivers/greybus/greybus_trace.h __entry->active, __entry->waiters, __entry->errno) waiters 90 drivers/greybus/operation.c if (atomic_read(&operation->waiters)) waiters 554 drivers/greybus/operation.c atomic_set(&operation->waiters, 0); waiters 1082 drivers/greybus/operation.c atomic_inc(&operation->waiters); waiters 1085 drivers/greybus/operation.c atomic_dec(&operation->waiters); waiters 1109 drivers/greybus/operation.c atomic_inc(&operation->waiters); waiters 1112 drivers/greybus/operation.c atomic_dec(&operation->waiters); waiters 46 drivers/md/persistent-data/dm-block-manager.c struct list_head waiters; waiters 153 drivers/md/persistent-data/dm-block-manager.c list_for_each_entry_safe(w, tmp, &lock->waiters, list) { waiters 179 drivers/md/persistent-data/dm-block-manager.c INIT_LIST_HEAD(&lock->waiters); waiters 188 drivers/md/persistent-data/dm-block-manager.c list_empty(&lock->waiters); waiters 214 drivers/md/persistent-data/dm-block-manager.c list_add_tail(&w.list, &lock->waiters); waiters 249 drivers/md/persistent-data/dm-block-manager.c if (!list_empty(&lock->waiters)) waiters 266 drivers/md/persistent-data/dm-block-manager.c if (lock->count == 0 && list_empty(&lock->waiters)) { waiters 281 drivers/md/persistent-data/dm-block-manager.c list_add(&w.list, &lock->waiters); waiters 295 drivers/md/persistent-data/dm-block-manager.c if (!list_empty(&lock->waiters)) waiters 42 drivers/tee/optee/call.c list_add_tail(&w->list_node, &cq->waiters); waiters 57 drivers/tee/optee/call.c list_add_tail(&w->list_node, &cq->waiters); waiters 66 drivers/tee/optee/call.c list_for_each_entry(w, &cq->waiters, list_node) { waiters 635 drivers/tee/optee/core.c INIT_LIST_HEAD(&optee->call_queue.waiters); waiters 34 drivers/tee/optee/optee_private.h struct list_head waiters; waiters 176 fs/xfs/xfs_log.c INIT_LIST_HEAD(&head->waiters); waiters 187 fs/xfs/xfs_log.c list_for_each_entry(tic, &head->waiters, t_queue) waiters 219 fs/xfs/xfs_log.c list_for_each_entry(tic, &head->waiters, t_queue) { waiters 266 fs/xfs/xfs_log.c list_add_tail(&tic->t_queue, &head->waiters); waiters 331 fs/xfs/xfs_log.c if (!list_empty_careful(&head->waiters)) { waiters 1057 fs/xfs/xfs_log.c if (!list_empty_careful(&log->l_write_head.waiters)) { waiters 1066 fs/xfs/xfs_log.c if (!list_empty_careful(&log->l_reserve_head.waiters)) { waiters 340 fs/xfs/xfs_log_priv.h struct list_head waiters; waiters 957 fs/xfs/xfs_trace.h __entry->reserveq = list_empty(&log->l_reserve_head.waiters); waiters 958 fs/xfs/xfs_trace.h __entry->writeq = list_empty(&log->l_write_head.waiters); waiters 108 include/linux/greybus/operation.h atomic_t waiters; waiters 313 include/linux/page-flags.h PAGEFLAG(Waiters, waiters, PF_ONLY_HEAD) __CLEARPAGEFLAG(Waiters, waiters, PF_ONLY_HEAD) waiters 32 include/linux/rtmutex.h struct rb_root_cached waiters; waiters 87 include/linux/rtmutex.h , .waiters = RB_ROOT_CACHED \ waiters 254 kernel/futex.c atomic_t waiters; waiters 351 kernel/futex.c atomic_inc(&hb->waiters); waiters 366 kernel/futex.c atomic_dec(&hb->waiters); waiters 373 kernel/futex.c return atomic_read(&hb->waiters); waiters 4192 kernel/futex.c atomic_set(&futex_queues[i].waiters, 0); waiters 275 kernel/locking/rtmutex.c struct rb_node **link = &lock->waiters.rb_root.rb_node; waiters 292 kernel/locking/rtmutex.c rb_insert_color_cached(&waiter->tree_entry, &lock->waiters, leftmost); waiters 301 kernel/locking/rtmutex.c rb_erase_cached(&waiter->tree_entry, &lock->waiters); waiters 1680 kernel/locking/rtmutex.c lock->waiters = RB_ROOT_CACHED; waiters 49 kernel/locking/rtmutex_common.h return !RB_EMPTY_ROOT(&lock->waiters.rb_root); waiters 55 kernel/locking/rtmutex_common.h struct rb_node *leftmost = rb_first_cached(&lock->waiters); waiters 405 kernel/trace/ring_buffer.c wait_queue_head_t waiters; waiters 559 kernel/trace/ring_buffer.c wake_up_all(&rbwork->waiters); waiters 604 kernel/trace/ring_buffer.c prepare_to_wait(&work->waiters, &wait, TASK_INTERRUPTIBLE); waiters 668 kernel/trace/ring_buffer.c finish_wait(&work->waiters, &wait); waiters 703 kernel/trace/ring_buffer.c poll_wait(filp, &work->waiters, poll_table); waiters 1306 kernel/trace/ring_buffer.c init_waitqueue_head(&cpu_buffer->irq_work.waiters); waiters 1401 kernel/trace/ring_buffer.c init_waitqueue_head(&buffer->irq_work.waiters);