hlevel 616 arch/x86/kvm/paging_tmpl.h int write_fault, int hlevel, hlevel 685 arch/x86/kvm/paging_tmpl.h disallowed_hugepage_adjust(it, gfn, &pfn, &hlevel); hlevel 688 arch/x86/kvm/paging_tmpl.h if (it.level == hlevel) hlevel 39 drivers/gpio/gpio-mt7621.c u32 hlevel; hlevel 124 drivers/gpio/gpio-mt7621.c mtk_gpio_w32(rg, GPIO_REG_HLVL, high | (BIT(pin) & rg->hlevel)); hlevel 160 drivers/gpio/gpio-mt7621.c rg->hlevel | rg->llevel) & mask) hlevel 168 drivers/gpio/gpio-mt7621.c rg->hlevel &= ~mask; hlevel 183 drivers/gpio/gpio-mt7621.c rg->hlevel |= mask; hlevel 176 net/sched/sch_htb.c struct htb_level hlevel[TC_HTB_MAXDEPTH]; hlevel 303 net/sched/sch_htb.c struct rb_node **p = &q->hlevel[cl->level].wait_pq.rb_node, *parent = NULL; hlevel 323 net/sched/sch_htb.c rb_insert_color(&cl->pq_node, &q->hlevel[cl->level].wait_pq); hlevel 350 net/sched/sch_htb.c htb_add_to_id_tree(&q->hlevel[cl->level].hprio[prio].row, cl, prio); hlevel 376 net/sched/sch_htb.c struct htb_level *hlevel = &q->hlevel[cl->level]; hlevel 380 net/sched/sch_htb.c struct htb_prio *hprio = &hlevel->hprio[prio]; hlevel 680 net/sched/sch_htb.c htb_safe_rb_erase(&cl->pq_node, &q->hlevel[cl->level].wait_pq); hlevel 708 net/sched/sch_htb.c struct rb_root *wait_pq = &q->hlevel[level].wait_pq; hlevel 828 net/sched/sch_htb.c struct htb_level *hlevel = &q->hlevel[level]; hlevel 829 net/sched/sch_htb.c struct htb_prio *hprio = &hlevel->hprio[prio]; hlevel 866 net/sched/sch_htb.c &q->hlevel[0].hprio[prio].ptr); hlevel 877 net/sched/sch_htb.c &q->hlevel[0].hprio[prio].ptr); hlevel 971 net/sched/sch_htb.c memset(q->hlevel, 0, sizeof(q->hlevel)); hlevel 1196 net/sched/sch_htb.c &q->hlevel[parent->level].wait_pq); hlevel 1285 net/sched/sch_htb.c &q->hlevel[cl->level].wait_pq); hlevel 1411 net/sched/sch_htb.c htb_safe_rb_erase(&parent->pq_node, &q->hlevel[0].wait_pq);