steal 210 arch/s390/kernel/vtime.c u64 steal, avg_steal; steal 215 arch/s390/kernel/vtime.c steal = S390_lowcore.steal_timer; steal 217 arch/s390/kernel/vtime.c if ((s64) steal > 0) { steal 219 arch/s390/kernel/vtime.c account_steal_time(steal); steal 220 arch/s390/kernel/vtime.c avg_steal += steal; steal 55 arch/x86/include/uapi/asm/kvm_para.h __u64 steal; steal 381 arch/x86/kernel/kvm.c u64 steal; steal 389 arch/x86/kernel/kvm.c steal = src->steal; steal 393 arch/x86/kernel/kvm.c return steal; steal 2681 arch/x86/kvm/x86.c st->steal += current->sched_info.run_delay - steal 610 drivers/gpu/drm/radeon/radeon_object.c int steal; steal 624 drivers/gpu/drm/radeon/radeon_object.c steal = -1; steal 633 drivers/gpu/drm/radeon/radeon_object.c steal = i; steal 638 drivers/gpu/drm/radeon/radeon_object.c if (steal == -1) steal 641 drivers/gpu/drm/radeon/radeon_object.c reg = &rdev->surface_regs[steal]; steal 644 drivers/gpu/drm/radeon/radeon_object.c DRM_DEBUG("stealing surface reg %d from %p\n", steal, old_object); steal 647 drivers/gpu/drm/radeon/radeon_object.c i = steal; steal 110 drivers/media/pci/ivtv/ivtv-queue.c int ivtv_queue_move(struct ivtv_stream *s, struct ivtv_queue *from, struct ivtv_queue *steal, steal 126 drivers/media/pci/ivtv/ivtv-queue.c bytes_steal = (from_free && steal) ? steal->length : 0; steal 132 drivers/media/pci/ivtv/ivtv-queue.c while (steal && bytes_available < needed_bytes) { steal 133 drivers/media/pci/ivtv/ivtv-queue.c struct ivtv_buffer *buf = list_entry(steal->list.prev, struct ivtv_buffer, list); steal 141 drivers/media/pci/ivtv/ivtv-queue.c list_move_tail(steal->list.prev, &from->list); steal 143 drivers/media/pci/ivtv/ivtv-queue.c steal->buffers--; steal 144 drivers/media/pci/ivtv/ivtv-queue.c steal->length -= s->buf_size; steal 145 drivers/media/pci/ivtv/ivtv-queue.c steal->bytesused -= buf->bytesused - buf->readpos; steal 150 drivers/media/pci/ivtv/ivtv-queue.c if (list_empty(&steal->list)) steal 152 drivers/media/pci/ivtv/ivtv-queue.c buf = list_entry(steal->list.prev, struct ivtv_buffer, list); steal 62 drivers/media/pci/ivtv/ivtv-queue.h int ivtv_queue_move(struct ivtv_stream *s, struct ivtv_queue *from, struct ivtv_queue *steal, steal 233 fs/pipe.c .steal = anon_pipe_buf_steal, steal 240 fs/pipe.c .steal = anon_pipe_buf_steal, steal 247 fs/pipe.c .steal = anon_pipe_buf_steal, steal 110 fs/proc/stat.c u64 user, nice, system, idle, iowait, irq, softirq, steal; steal 118 fs/proc/stat.c irq = softirq = steal = 0; steal 132 fs/proc/stat.c steal += kcs->cpustat[CPUTIME_STEAL]; steal 154 fs/proc/stat.c seq_put_decimal_ull(p, " ", nsec_to_clock_t(steal)); steal 170 fs/proc/stat.c steal = kcs->cpustat[CPUTIME_STEAL]; steal 181 fs/proc/stat.c seq_put_decimal_ull(p, " ", nsec_to_clock_t(steal)); steal 144 fs/splice.c .steal = page_cache_pipe_buf_steal, steal 161 fs/splice.c .steal = user_page_pipe_buf_steal, steal 330 fs/splice.c .steal = generic_pipe_buf_steal, steal 344 fs/splice.c .steal = generic_pipe_buf_nosteal, steal 99 include/linux/pipe_fs_i.h int (*steal)(struct pipe_inode_info *, struct pipe_buffer *); steal 153 include/linux/pipe_fs_i.h return buf->ops->steal(pipe, buf); steal 437 kernel/bpf/bpf_lru_list.c int steal, first_steal; steal 468 kernel/bpf/bpf_lru_list.c steal = first_steal; steal 470 kernel/bpf/bpf_lru_list.c steal_loc_l = per_cpu_ptr(clru->local_list, steal); steal 480 kernel/bpf/bpf_lru_list.c steal = get_next_cpu(steal); steal 481 kernel/bpf/bpf_lru_list.c } while (!node && steal != first_steal); steal 483 kernel/bpf/bpf_lru_list.c loc_l->next_steal = steal; steal 1187 kernel/relay.c .steal = generic_pipe_buf_steal, steal 151 kernel/sched/core.c s64 __maybe_unused steal = 0, irq_delta = 0; steal 179 kernel/sched/core.c steal = paravirt_steal_clock(cpu_of(rq)); steal 180 kernel/sched/core.c steal -= rq->prev_steal_time_rq; steal 182 kernel/sched/core.c if (unlikely(steal > delta)) steal 183 kernel/sched/core.c steal = delta; steal 185 kernel/sched/core.c rq->prev_steal_time_rq += steal; steal 186 kernel/sched/core.c delta -= steal; steal 193 kernel/sched/core.c if ((irq_delta + steal) && sched_feat(NONTASK_CAPACITY)) steal 194 kernel/sched/core.c update_irq_load_avg(rq, irq_delta + steal); steal 238 kernel/sched/cputime.c u64 steal; steal 240 kernel/sched/cputime.c steal = paravirt_steal_clock(smp_processor_id()); steal 241 kernel/sched/cputime.c steal -= this_rq()->prev_steal_time; steal 242 kernel/sched/cputime.c steal = min(steal, maxtime); steal 243 kernel/sched/cputime.c account_steal_time(steal); steal 244 kernel/sched/cputime.c this_rq()->prev_steal_time += steal; steal 246 kernel/sched/cputime.c return steal; steal 477 kernel/sched/cputime.c u64 cputime, steal; steal 489 kernel/sched/cputime.c steal = steal_account_process_time(ULONG_MAX); steal 491 kernel/sched/cputime.c if (steal >= cputime) steal 494 kernel/sched/cputime.c cputime -= steal; steal 510 kernel/sched/cputime.c u64 cputime, steal; steal 518 kernel/sched/cputime.c steal = steal_account_process_time(ULONG_MAX); steal 520 kernel/sched/cputime.c if (steal >= cputime) steal 523 kernel/sched/cputime.c cputime -= steal; steal 6112 kernel/trace/trace.c .steal = generic_pipe_buf_steal, steal 7388 kernel/trace/trace.c .steal = generic_pipe_buf_nosteal, steal 141 net/smc/smc_rx.c .steal = smc_rx_pipe_buf_nosteal,