vtime             104 arch/s390/kernel/vtime.c static inline u64 scale_vtime(u64 vtime)
vtime             110 arch/s390/kernel/vtime.c 		return vtime * mult / div;
vtime             111 arch/s390/kernel/vtime.c 	return vtime;
vtime              57 block/bfq-iosched.h 	u64 vtime;
vtime             117 block/bfq-wf2q.c 				 !bfq_gt(new_entity->start, st->vtime)
vtime             690 block/bfq-wf2q.c 	    !bfq_gt(last_idle->finish, st->vtime)) {
vtime             695 block/bfq-wf2q.c 		st->vtime = last_idle->finish;
vtime             698 block/bfq-wf2q.c 	if (first_idle && !bfq_gt(first_idle->finish, st->vtime))
vtime             819 block/bfq-wf2q.c 			entity->start = new_st->vtime;
vtime             852 block/bfq-wf2q.c 		st->vtime += bfq_delta(served, st->wsum);
vtime             950 block/bfq-wf2q.c 	if (backshifted && bfq_gt(st->vtime, entity->finish)) {
vtime             951 block/bfq-wf2q.c 		unsigned long delta = st->vtime - entity->finish;
vtime             983 block/bfq-wf2q.c 	if (non_blocking_wait_rq && bfq_gt(st->vtime, entity->finish)) {
vtime             987 block/bfq-wf2q.c 		min_vstart = st->vtime;
vtime            1208 block/bfq-wf2q.c 	if (!ins_into_idle_tree || !bfq_gt(entity->finish, st->vtime))
vtime            1332 block/bfq-wf2q.c 	if (bfq_gt(root_entity->min_start, st->vtime))
vtime            1335 block/bfq-wf2q.c 	return st->vtime;
vtime            1340 block/bfq-wf2q.c 	if (new_value > st->vtime) {
vtime            1341 block/bfq-wf2q.c 		st->vtime = new_value;
vtime            1359 block/bfq-wf2q.c 						  u64 vtime)
vtime            1367 block/bfq-wf2q.c 		if (!bfq_gt(entry->start, vtime))
vtime            1373 block/bfq-wf2q.c 			if (!bfq_gt(entry->min_start, vtime)) {
vtime             470 block/blk-iocost.c 	atomic64_t			vtime;
vtime             675 block/blk-iocost.c 	atomic64_add(cost, &iocg->vtime);
vtime            1031 block/blk-iocost.c 	u64 vtime, vmargin, vmin;
vtime            1076 block/blk-iocost.c 	vtime = atomic64_read(&iocg->vtime);
vtime            1081 block/blk-iocost.c 	    time_before64(vtime, vmin)) {
vtime            1082 block/blk-iocost.c 		atomic64_add(vmin - vtime, &iocg->vtime);
vtime            1083 block/blk-iocost.c 		atomic64_add(vmin - vtime, &iocg->done_vtime);
vtime            1084 block/blk-iocost.c 		vtime = vmin;
vtime            1098 block/blk-iocost.c 			last_period, cur_period, vtime);
vtime            1100 block/blk-iocost.c 	iocg->last_vtime = vtime;
vtime            1155 block/blk-iocost.c 	vbudget = now->vnow - atomic64_read(&iocg->vtime);
vtime            1164 block/blk-iocost.c 		atomic64_add(delta, &iocg->vtime);
vtime            1216 block/blk-iocost.c 	u64 vtime = atomic64_read(&iocg->vtime);
vtime            1226 block/blk-iocost.c 	vtime += abs_cost_to_cost(iocg->abs_vdebt, hw_inuse);
vtime            1233 block/blk-iocost.c 	if (!iocg->abs_vdebt || time_before_eq64(vtime, now->vnow)) {
vtime            1238 block/blk-iocost.c 	    time_before_eq64(vtime, now->vnow + vmargin))
vtime            1249 block/blk-iocost.c 	expires = now->now_ns + DIV64_U64_ROUND_UP(vtime - now->vnow,
vtime            1327 block/blk-iocost.c 	if (atomic64_read(&iocg->done_vtime) != atomic64_read(&iocg->vtime))
vtime            1403 block/blk-iocost.c 		u64 vdone, vtime, vusage, vmargin, vmin;
vtime            1411 block/blk-iocost.c 		vtime = atomic64_read(&iocg->vtime);
vtime            1422 block/blk-iocost.c 		    time_after64(vtime, vdone) &&
vtime            1423 block/blk-iocost.c 		    time_after64(vtime, now.vnow -
vtime            1430 block/blk-iocost.c 		else if (time_before64(iocg->last_vtime, vtime))
vtime            1431 block/blk-iocost.c 			vusage = vtime - iocg->last_vtime;
vtime            1441 block/blk-iocost.c 		vusage = max(vusage, vtime - vdone);
vtime            1460 block/blk-iocost.c 		    time_before64(vtime, vmin)) {
vtime            1461 block/blk-iocost.c 			u64 delta = vmin - vtime;
vtime            1464 block/blk-iocost.c 			atomic64_add(delta, &iocg->vtime);
vtime            1692 block/blk-iocost.c 	u64 abs_cost, cost, vtime;
vtime            1709 block/blk-iocost.c 	vtime = atomic64_read(&iocg->vtime);
vtime            1713 block/blk-iocost.c 	    time_after_eq64(vtime + ioc->inuse_margin_vtime, now.vnow)) {
vtime            1730 block/blk-iocost.c 	    time_before_eq64(vtime + cost, now.vnow)) {
vtime            1843 block/blk-iocost.c 	    time_before_eq64(atomic64_read(&iocg->vtime) + cost, now.vnow)) {
vtime            2028 block/blk-iocost.c 	atomic64_set(&iocg->vtime, now.vnow);
vtime             846 include/linux/sched.h 	struct vtime			vtime;
vtime              17 include/trace/events/iocost.h 		u64 last_period, u64 cur_period, u64 vtime),
vtime              19 include/trace/events/iocost.h 	TP_ARGS(iocg, path, now, last_period, cur_period, vtime),
vtime              30 include/trace/events/iocost.h 		__field(u64, vtime)
vtime              46 include/trace/events/iocost.h 		__entry->vtime = vtime;
vtime              59 include/trace/events/iocost.h 		__entry->last_vtime, __entry->vtime,
vtime             152 init/init_task.c 	.vtime.seqcount	= SEQCNT_ZERO(init_task.vtime_seqcount),
vtime             153 init/init_task.c 	.vtime.starttime = 0,
vtime             154 init/init_task.c 	.vtime.state	= VTIME_SYS,
vtime            1911 kernel/fork.c  	seqcount_init(&p->vtime.seqcount);
vtime            1912 kernel/fork.c  	p->vtime.starttime = 0;
vtime            1913 kernel/fork.c  	p->vtime.state = VTIME_INACTIVE;
vtime             684 kernel/sched/cputime.c static u64 vtime_delta(struct vtime *vtime)
vtime             689 kernel/sched/cputime.c 	if (clock < vtime->starttime)
vtime             692 kernel/sched/cputime.c 	return clock - vtime->starttime;
vtime             695 kernel/sched/cputime.c static u64 get_vtime_delta(struct vtime *vtime)
vtime             697 kernel/sched/cputime.c 	u64 delta = vtime_delta(vtime);
vtime             708 kernel/sched/cputime.c 	WARN_ON_ONCE(vtime->state == VTIME_INACTIVE);
vtime             709 kernel/sched/cputime.c 	vtime->starttime += delta;
vtime             715 kernel/sched/cputime.c 				   struct vtime *vtime)
vtime             717 kernel/sched/cputime.c 	vtime->stime += get_vtime_delta(vtime);
vtime             718 kernel/sched/cputime.c 	if (vtime->stime >= TICK_NSEC) {
vtime             719 kernel/sched/cputime.c 		account_system_time(tsk, irq_count(), vtime->stime);
vtime             720 kernel/sched/cputime.c 		vtime->stime = 0;
vtime             725 kernel/sched/cputime.c 				struct vtime *vtime)
vtime             727 kernel/sched/cputime.c 	vtime->gtime += get_vtime_delta(vtime);
vtime             728 kernel/sched/cputime.c 	if (vtime->gtime >= TICK_NSEC) {
vtime             729 kernel/sched/cputime.c 		account_guest_time(tsk, vtime->gtime);
vtime             730 kernel/sched/cputime.c 		vtime->gtime = 0;
vtime             736 kernel/sched/cputime.c 	struct vtime *vtime = &tsk->vtime;
vtime             738 kernel/sched/cputime.c 	if (!vtime_delta(vtime))
vtime             741 kernel/sched/cputime.c 	write_seqcount_begin(&vtime->seqcount);
vtime             744 kernel/sched/cputime.c 		vtime_account_guest(tsk, vtime);
vtime             746 kernel/sched/cputime.c 		__vtime_account_system(tsk, vtime);
vtime             747 kernel/sched/cputime.c 	write_seqcount_end(&vtime->seqcount);
vtime             752 kernel/sched/cputime.c 	struct vtime *vtime = &tsk->vtime;
vtime             754 kernel/sched/cputime.c 	write_seqcount_begin(&vtime->seqcount);
vtime             755 kernel/sched/cputime.c 	__vtime_account_system(tsk, vtime);
vtime             756 kernel/sched/cputime.c 	vtime->state = VTIME_USER;
vtime             757 kernel/sched/cputime.c 	write_seqcount_end(&vtime->seqcount);
vtime             762 kernel/sched/cputime.c 	struct vtime *vtime = &tsk->vtime;
vtime             764 kernel/sched/cputime.c 	write_seqcount_begin(&vtime->seqcount);
vtime             765 kernel/sched/cputime.c 	vtime->utime += get_vtime_delta(vtime);
vtime             766 kernel/sched/cputime.c 	if (vtime->utime >= TICK_NSEC) {
vtime             767 kernel/sched/cputime.c 		account_user_time(tsk, vtime->utime);
vtime             768 kernel/sched/cputime.c 		vtime->utime = 0;
vtime             770 kernel/sched/cputime.c 	vtime->state = VTIME_SYS;
vtime             771 kernel/sched/cputime.c 	write_seqcount_end(&vtime->seqcount);
vtime             776 kernel/sched/cputime.c 	struct vtime *vtime = &tsk->vtime;
vtime             784 kernel/sched/cputime.c 	write_seqcount_begin(&vtime->seqcount);
vtime             785 kernel/sched/cputime.c 	__vtime_account_system(tsk, vtime);
vtime             787 kernel/sched/cputime.c 	write_seqcount_end(&vtime->seqcount);
vtime             793 kernel/sched/cputime.c 	struct vtime *vtime = &tsk->vtime;
vtime             795 kernel/sched/cputime.c 	write_seqcount_begin(&vtime->seqcount);
vtime             796 kernel/sched/cputime.c 	vtime_account_guest(tsk, vtime);
vtime             798 kernel/sched/cputime.c 	write_seqcount_end(&vtime->seqcount);
vtime             804 kernel/sched/cputime.c 	account_idle_time(get_vtime_delta(&tsk->vtime));
vtime             809 kernel/sched/cputime.c 	struct vtime *vtime = &prev->vtime;
vtime             811 kernel/sched/cputime.c 	write_seqcount_begin(&vtime->seqcount);
vtime             812 kernel/sched/cputime.c 	vtime->state = VTIME_INACTIVE;
vtime             813 kernel/sched/cputime.c 	write_seqcount_end(&vtime->seqcount);
vtime             815 kernel/sched/cputime.c 	vtime = &current->vtime;
vtime             817 kernel/sched/cputime.c 	write_seqcount_begin(&vtime->seqcount);
vtime             818 kernel/sched/cputime.c 	vtime->state = VTIME_SYS;
vtime             819 kernel/sched/cputime.c 	vtime->starttime = sched_clock();
vtime             820 kernel/sched/cputime.c 	write_seqcount_end(&vtime->seqcount);
vtime             825 kernel/sched/cputime.c 	struct vtime *vtime = &t->vtime;
vtime             829 kernel/sched/cputime.c 	write_seqcount_begin(&vtime->seqcount);
vtime             830 kernel/sched/cputime.c 	vtime->state = VTIME_SYS;
vtime             831 kernel/sched/cputime.c 	vtime->starttime = sched_clock();
vtime             832 kernel/sched/cputime.c 	write_seqcount_end(&vtime->seqcount);
vtime             838 kernel/sched/cputime.c 	struct vtime *vtime = &t->vtime;
vtime             846 kernel/sched/cputime.c 		seq = read_seqcount_begin(&vtime->seqcount);
vtime             849 kernel/sched/cputime.c 		if (vtime->state == VTIME_SYS && t->flags & PF_VCPU)
vtime             850 kernel/sched/cputime.c 			gtime += vtime->gtime + vtime_delta(vtime);
vtime             852 kernel/sched/cputime.c 	} while (read_seqcount_retry(&vtime->seqcount, seq));
vtime             864 kernel/sched/cputime.c 	struct vtime *vtime = &t->vtime;
vtime             875 kernel/sched/cputime.c 		seq = read_seqcount_begin(&vtime->seqcount);
vtime             881 kernel/sched/cputime.c 		if (vtime->state == VTIME_INACTIVE || is_idle_task(t))
vtime             884 kernel/sched/cputime.c 		delta = vtime_delta(vtime);
vtime             890 kernel/sched/cputime.c 		if (vtime->state == VTIME_USER || t->flags & PF_VCPU)
vtime             891 kernel/sched/cputime.c 			*utime += vtime->utime + delta;
vtime             892 kernel/sched/cputime.c 		else if (vtime->state == VTIME_SYS)
vtime             893 kernel/sched/cputime.c 			*stime += vtime->stime + delta;
vtime             894 kernel/sched/cputime.c 	} while (read_seqcount_retry(&vtime->seqcount, seq));
vtime              63 tools/testing/selftests/x86/test_vsyscall.c const time_func_t vtime = (time_func_t)VSYS(0xffffffffff600400);
vtime             276 tools/testing/selftests/x86/test_vsyscall.c 		t_vsys = vtime(&t2_vsys);
vtime             531 tools/testing/selftests/x86/test_vsyscall.c 	vtime(&tmp);