cl_cumul          135 net/sched/sch_hfsc.c 	u64	cl_cumul;		/* cumulative work in bytes done by
cl_cumul          615 net/sched/sch_hfsc.c 	rtsc_min(&cl->cl_deadline, &cl->cl_rsc, cur_time, cl->cl_cumul);
cl_cumul          629 net/sched/sch_hfsc.c 	cl->cl_e = rtsc_y2x(&cl->cl_eligible, cl->cl_cumul);
cl_cumul          630 net/sched/sch_hfsc.c 	cl->cl_d = rtsc_y2x(&cl->cl_deadline, cl->cl_cumul + next_len);
cl_cumul          638 net/sched/sch_hfsc.c 	cl->cl_e = rtsc_y2x(&cl->cl_eligible, cl->cl_cumul);
cl_cumul          639 net/sched/sch_hfsc.c 	cl->cl_d = rtsc_y2x(&cl->cl_deadline, cl->cl_cumul + next_len);
cl_cumul          647 net/sched/sch_hfsc.c 	cl->cl_d = rtsc_y2x(&cl->cl_deadline, cl->cl_cumul + next_len);
cl_cumul          880 net/sched/sch_hfsc.c 	rtsc_init(&cl->cl_deadline, &cl->cl_rsc, cur_time, cl->cl_cumul);
cl_cumul         1328 net/sched/sch_hfsc.c 	xstats.rtwork  = cl->cl_cumul;
cl_cumul         1448 net/sched/sch_hfsc.c 	cl->cl_cumul        = 0;
cl_cumul         1623 net/sched/sch_hfsc.c 		cl->cl_cumul += qdisc_pkt_len(skb);