vtime 104 arch/s390/kernel/vtime.c static inline u64 scale_vtime(u64 vtime) vtime 110 arch/s390/kernel/vtime.c return vtime * mult / div; vtime 111 arch/s390/kernel/vtime.c return vtime; vtime 57 block/bfq-iosched.h u64 vtime; vtime 117 block/bfq-wf2q.c !bfq_gt(new_entity->start, st->vtime) vtime 690 block/bfq-wf2q.c !bfq_gt(last_idle->finish, st->vtime)) { vtime 695 block/bfq-wf2q.c st->vtime = last_idle->finish; vtime 698 block/bfq-wf2q.c if (first_idle && !bfq_gt(first_idle->finish, st->vtime)) vtime 819 block/bfq-wf2q.c entity->start = new_st->vtime; vtime 852 block/bfq-wf2q.c st->vtime += bfq_delta(served, st->wsum); vtime 950 block/bfq-wf2q.c if (backshifted && bfq_gt(st->vtime, entity->finish)) { vtime 951 block/bfq-wf2q.c unsigned long delta = st->vtime - entity->finish; vtime 983 block/bfq-wf2q.c if (non_blocking_wait_rq && bfq_gt(st->vtime, entity->finish)) { vtime 987 block/bfq-wf2q.c min_vstart = st->vtime; vtime 1208 block/bfq-wf2q.c if (!ins_into_idle_tree || !bfq_gt(entity->finish, st->vtime)) vtime 1332 block/bfq-wf2q.c if (bfq_gt(root_entity->min_start, st->vtime)) vtime 1335 block/bfq-wf2q.c return st->vtime; vtime 1340 block/bfq-wf2q.c if (new_value > st->vtime) { vtime 1341 block/bfq-wf2q.c st->vtime = new_value; vtime 1359 block/bfq-wf2q.c u64 vtime) vtime 1367 block/bfq-wf2q.c if (!bfq_gt(entry->start, vtime)) vtime 1373 block/bfq-wf2q.c if (!bfq_gt(entry->min_start, vtime)) { vtime 470 block/blk-iocost.c atomic64_t vtime; vtime 675 block/blk-iocost.c atomic64_add(cost, &iocg->vtime); vtime 1031 block/blk-iocost.c u64 vtime, vmargin, vmin; vtime 1076 block/blk-iocost.c vtime = atomic64_read(&iocg->vtime); vtime 1081 block/blk-iocost.c time_before64(vtime, vmin)) { vtime 1082 block/blk-iocost.c atomic64_add(vmin - vtime, &iocg->vtime); vtime 1083 block/blk-iocost.c atomic64_add(vmin - vtime, &iocg->done_vtime); vtime 1084 block/blk-iocost.c vtime = vmin; vtime 1098 block/blk-iocost.c last_period, cur_period, vtime); vtime 1100 block/blk-iocost.c iocg->last_vtime = vtime; vtime 1155 block/blk-iocost.c vbudget = now->vnow - atomic64_read(&iocg->vtime); vtime 1164 block/blk-iocost.c atomic64_add(delta, &iocg->vtime); vtime 1216 block/blk-iocost.c u64 vtime = atomic64_read(&iocg->vtime); vtime 1226 block/blk-iocost.c vtime += abs_cost_to_cost(iocg->abs_vdebt, hw_inuse); vtime 1233 block/blk-iocost.c if (!iocg->abs_vdebt || time_before_eq64(vtime, now->vnow)) { vtime 1238 block/blk-iocost.c time_before_eq64(vtime, now->vnow + vmargin)) vtime 1249 block/blk-iocost.c expires = now->now_ns + DIV64_U64_ROUND_UP(vtime - now->vnow, vtime 1327 block/blk-iocost.c if (atomic64_read(&iocg->done_vtime) != atomic64_read(&iocg->vtime)) vtime 1403 block/blk-iocost.c u64 vdone, vtime, vusage, vmargin, vmin; vtime 1411 block/blk-iocost.c vtime = atomic64_read(&iocg->vtime); vtime 1422 block/blk-iocost.c time_after64(vtime, vdone) && vtime 1423 block/blk-iocost.c time_after64(vtime, now.vnow - vtime 1430 block/blk-iocost.c else if (time_before64(iocg->last_vtime, vtime)) vtime 1431 block/blk-iocost.c vusage = vtime - iocg->last_vtime; vtime 1441 block/blk-iocost.c vusage = max(vusage, vtime - vdone); vtime 1460 block/blk-iocost.c time_before64(vtime, vmin)) { vtime 1461 block/blk-iocost.c u64 delta = vmin - vtime; vtime 1464 block/blk-iocost.c atomic64_add(delta, &iocg->vtime); vtime 1692 block/blk-iocost.c u64 abs_cost, cost, vtime; vtime 1709 block/blk-iocost.c vtime = atomic64_read(&iocg->vtime); vtime 1713 block/blk-iocost.c time_after_eq64(vtime + ioc->inuse_margin_vtime, now.vnow)) { vtime 1730 block/blk-iocost.c time_before_eq64(vtime + cost, now.vnow)) { vtime 1843 block/blk-iocost.c time_before_eq64(atomic64_read(&iocg->vtime) + cost, now.vnow)) { vtime 2028 block/blk-iocost.c atomic64_set(&iocg->vtime, now.vnow); vtime 846 include/linux/sched.h struct vtime vtime; vtime 17 include/trace/events/iocost.h u64 last_period, u64 cur_period, u64 vtime), vtime 19 include/trace/events/iocost.h TP_ARGS(iocg, path, now, last_period, cur_period, vtime), vtime 30 include/trace/events/iocost.h __field(u64, vtime) vtime 46 include/trace/events/iocost.h __entry->vtime = vtime; vtime 59 include/trace/events/iocost.h __entry->last_vtime, __entry->vtime, vtime 152 init/init_task.c .vtime.seqcount = SEQCNT_ZERO(init_task.vtime_seqcount), vtime 153 init/init_task.c .vtime.starttime = 0, vtime 154 init/init_task.c .vtime.state = VTIME_SYS, vtime 1911 kernel/fork.c seqcount_init(&p->vtime.seqcount); vtime 1912 kernel/fork.c p->vtime.starttime = 0; vtime 1913 kernel/fork.c p->vtime.state = VTIME_INACTIVE; vtime 684 kernel/sched/cputime.c static u64 vtime_delta(struct vtime *vtime) vtime 689 kernel/sched/cputime.c if (clock < vtime->starttime) vtime 692 kernel/sched/cputime.c return clock - vtime->starttime; vtime 695 kernel/sched/cputime.c static u64 get_vtime_delta(struct vtime *vtime) vtime 697 kernel/sched/cputime.c u64 delta = vtime_delta(vtime); vtime 708 kernel/sched/cputime.c WARN_ON_ONCE(vtime->state == VTIME_INACTIVE); vtime 709 kernel/sched/cputime.c vtime->starttime += delta; vtime 715 kernel/sched/cputime.c struct vtime *vtime) vtime 717 kernel/sched/cputime.c vtime->stime += get_vtime_delta(vtime); vtime 718 kernel/sched/cputime.c if (vtime->stime >= TICK_NSEC) { vtime 719 kernel/sched/cputime.c account_system_time(tsk, irq_count(), vtime->stime); vtime 720 kernel/sched/cputime.c vtime->stime = 0; vtime 725 kernel/sched/cputime.c struct vtime *vtime) vtime 727 kernel/sched/cputime.c vtime->gtime += get_vtime_delta(vtime); vtime 728 kernel/sched/cputime.c if (vtime->gtime >= TICK_NSEC) { vtime 729 kernel/sched/cputime.c account_guest_time(tsk, vtime->gtime); vtime 730 kernel/sched/cputime.c vtime->gtime = 0; vtime 736 kernel/sched/cputime.c struct vtime *vtime = &tsk->vtime; vtime 738 kernel/sched/cputime.c if (!vtime_delta(vtime)) vtime 741 kernel/sched/cputime.c write_seqcount_begin(&vtime->seqcount); vtime 744 kernel/sched/cputime.c vtime_account_guest(tsk, vtime); vtime 746 kernel/sched/cputime.c __vtime_account_system(tsk, vtime); vtime 747 kernel/sched/cputime.c write_seqcount_end(&vtime->seqcount); vtime 752 kernel/sched/cputime.c struct vtime *vtime = &tsk->vtime; vtime 754 kernel/sched/cputime.c write_seqcount_begin(&vtime->seqcount); vtime 755 kernel/sched/cputime.c __vtime_account_system(tsk, vtime); vtime 756 kernel/sched/cputime.c vtime->state = VTIME_USER; vtime 757 kernel/sched/cputime.c write_seqcount_end(&vtime->seqcount); vtime 762 kernel/sched/cputime.c struct vtime *vtime = &tsk->vtime; vtime 764 kernel/sched/cputime.c write_seqcount_begin(&vtime->seqcount); vtime 765 kernel/sched/cputime.c vtime->utime += get_vtime_delta(vtime); vtime 766 kernel/sched/cputime.c if (vtime->utime >= TICK_NSEC) { vtime 767 kernel/sched/cputime.c account_user_time(tsk, vtime->utime); vtime 768 kernel/sched/cputime.c vtime->utime = 0; vtime 770 kernel/sched/cputime.c vtime->state = VTIME_SYS; vtime 771 kernel/sched/cputime.c write_seqcount_end(&vtime->seqcount); vtime 776 kernel/sched/cputime.c struct vtime *vtime = &tsk->vtime; vtime 784 kernel/sched/cputime.c write_seqcount_begin(&vtime->seqcount); vtime 785 kernel/sched/cputime.c __vtime_account_system(tsk, vtime); vtime 787 kernel/sched/cputime.c write_seqcount_end(&vtime->seqcount); vtime 793 kernel/sched/cputime.c struct vtime *vtime = &tsk->vtime; vtime 795 kernel/sched/cputime.c write_seqcount_begin(&vtime->seqcount); vtime 796 kernel/sched/cputime.c vtime_account_guest(tsk, vtime); vtime 798 kernel/sched/cputime.c write_seqcount_end(&vtime->seqcount); vtime 804 kernel/sched/cputime.c account_idle_time(get_vtime_delta(&tsk->vtime)); vtime 809 kernel/sched/cputime.c struct vtime *vtime = &prev->vtime; vtime 811 kernel/sched/cputime.c write_seqcount_begin(&vtime->seqcount); vtime 812 kernel/sched/cputime.c vtime->state = VTIME_INACTIVE; vtime 813 kernel/sched/cputime.c write_seqcount_end(&vtime->seqcount); vtime 815 kernel/sched/cputime.c vtime = ¤t->vtime; vtime 817 kernel/sched/cputime.c write_seqcount_begin(&vtime->seqcount); vtime 818 kernel/sched/cputime.c vtime->state = VTIME_SYS; vtime 819 kernel/sched/cputime.c vtime->starttime = sched_clock(); vtime 820 kernel/sched/cputime.c write_seqcount_end(&vtime->seqcount); vtime 825 kernel/sched/cputime.c struct vtime *vtime = &t->vtime; vtime 829 kernel/sched/cputime.c write_seqcount_begin(&vtime->seqcount); vtime 830 kernel/sched/cputime.c vtime->state = VTIME_SYS; vtime 831 kernel/sched/cputime.c vtime->starttime = sched_clock(); vtime 832 kernel/sched/cputime.c write_seqcount_end(&vtime->seqcount); vtime 838 kernel/sched/cputime.c struct vtime *vtime = &t->vtime; vtime 846 kernel/sched/cputime.c seq = read_seqcount_begin(&vtime->seqcount); vtime 849 kernel/sched/cputime.c if (vtime->state == VTIME_SYS && t->flags & PF_VCPU) vtime 850 kernel/sched/cputime.c gtime += vtime->gtime + vtime_delta(vtime); vtime 852 kernel/sched/cputime.c } while (read_seqcount_retry(&vtime->seqcount, seq)); vtime 864 kernel/sched/cputime.c struct vtime *vtime = &t->vtime; vtime 875 kernel/sched/cputime.c seq = read_seqcount_begin(&vtime->seqcount); vtime 881 kernel/sched/cputime.c if (vtime->state == VTIME_INACTIVE || is_idle_task(t)) vtime 884 kernel/sched/cputime.c delta = vtime_delta(vtime); vtime 890 kernel/sched/cputime.c if (vtime->state == VTIME_USER || t->flags & PF_VCPU) vtime 891 kernel/sched/cputime.c *utime += vtime->utime + delta; vtime 892 kernel/sched/cputime.c else if (vtime->state == VTIME_SYS) vtime 893 kernel/sched/cputime.c *stime += vtime->stime + delta; vtime 894 kernel/sched/cputime.c } while (read_seqcount_retry(&vtime->seqcount, seq)); vtime 63 tools/testing/selftests/x86/test_vsyscall.c const time_func_t vtime = (time_func_t)VSYS(0xffffffffff600400); vtime 276 tools/testing/selftests/x86/test_vsyscall.c t_vsys = vtime(&t2_vsys); vtime 531 tools/testing/selftests/x86/test_vsyscall.c vtime(&tmp);