prio 1389 arch/alpha/kernel/osf_sys.c int prio = sys_getpriority(which, who); prio 1390 arch/alpha/kernel/osf_sys.c if (prio >= 0) { prio 1395 arch/alpha/kernel/osf_sys.c prio = 20 - prio; prio 1397 arch/alpha/kernel/osf_sys.c return prio; prio 17 arch/arc/kernel/intc-arcv2.c unsigned int pad:3, firq:1, prio:4, exts:8, irqs:8, ver:8; prio 19 arch/arc/kernel/intc-arcv2.c unsigned int ver:8, irqs:8, exts:8, prio:4, firq:1, pad:3; prio 67 arch/arc/kernel/intc-arcv2.c irq_prio = irq_bcr.prio; /* Encoded as N-1 for N levels */ prio 3 arch/arm/mach-pxa/devices.h .prio = PXAD_PRIO_##_prio, .drcmr = _requestor }) prio 46 arch/c6x/include/asm/irq.h extern asmlinkage void c6x_do_IRQ(unsigned int prio, struct pt_regs *regs); prio 34 arch/c6x/kernel/irq.c unsigned int prio = data->hwirq; prio 37 arch/c6x/kernel/irq.c and_creg(IER, ~(1 << prio)); prio 43 arch/c6x/kernel/irq.c unsigned int prio = data->hwirq; prio 46 arch/c6x/kernel/irq.c or_creg(IER, 1 << prio); prio 58 arch/c6x/kernel/irq.c asmlinkage void c6x_do_IRQ(unsigned int prio, struct pt_regs *regs) prio 64 arch/c6x/kernel/irq.c generic_handle_irq(prio_to_virq[prio]); prio 73 arch/ia64/kernel/sys_ia64.c long prio; prio 75 arch/ia64/kernel/sys_ia64.c prio = sys_getpriority(which, who); prio 76 arch/ia64/kernel/sys_ia64.c if (prio >= 0) { prio 78 arch/ia64/kernel/sys_ia64.c prio = 20 - prio; prio 80 arch/ia64/kernel/sys_ia64.c return prio; prio 72 arch/mips/alchemy/common/irq.c int prio; /* irq priority, 0 highest, 3 lowest */ prio 910 arch/mips/alchemy/common/irq.c if (map->prio == 0) prio 949 arch/mips/alchemy/common/irq.c au1300_set_irq_priority(i, dints->prio); prio 315 arch/mips/include/asm/octeon/cvmx-sriox-defs.h uint64_t prio:4; prio 319 arch/mips/include/asm/octeon/cvmx-sriox-defs.h uint64_t prio:4; prio 729 arch/mips/include/asm/octeon/cvmx-sriox-defs.h uint64_t prio:2; prio 751 arch/mips/include/asm/octeon/cvmx-sriox-defs.h uint64_t prio:2; prio 760 arch/mips/include/asm/octeon/cvmx-sriox-defs.h uint64_t prio:2; prio 770 arch/mips/include/asm/octeon/cvmx-sriox-defs.h uint64_t prio:2; prio 173 arch/powerpc/include/asm/cell-regs.h u8 prio; prio 183 arch/powerpc/include/asm/cell-regs.h u64 prio; prio 472 arch/powerpc/include/asm/mpic.h extern void mpic_cpu_set_priority(int prio); prio 253 arch/powerpc/include/asm/opal.h int64_t opal_xive_set_irq_config(uint32_t girq, uint64_t vp, uint8_t prio, prio 255 arch/powerpc/include/asm/opal.h int64_t opal_xive_get_queue_info(uint64_t vp, uint32_t prio, prio 261 arch/powerpc/include/asm/opal.h int64_t opal_xive_set_queue_info(uint64_t vp, uint32_t prio, prio 280 arch/powerpc/include/asm/opal.h int64_t opal_xive_get_queue_state(uint64_t vp, uint32_t prio, prio 283 arch/powerpc/include/asm/opal.h int64_t opal_xive_set_queue_state(uint64_t vp, uint32_t prio, prio 57 arch/powerpc/include/asm/xics.h void (*set_priority)(unsigned char prio); prio 113 arch/powerpc/include/asm/xive.h extern int xive_native_configure_irq(u32 hw_irq, u32 target, u8 prio, u32 sw_irq); prio 115 arch/powerpc/include/asm/xive.h extern int xive_native_configure_queue(u32 vp_id, struct xive_q *q, u8 prio, prio 117 arch/powerpc/include/asm/xive.h extern void xive_native_disable_queue(u32 vp_id, struct xive_q *q, u8 prio); prio 127 arch/powerpc/include/asm/xive.h extern int xive_native_get_queue_info(u32 vp_id, uint32_t prio, prio 134 arch/powerpc/include/asm/xive.h extern int xive_native_get_queue_state(u32 vp_id, uint32_t prio, u32 *qtoggle, prio 136 arch/powerpc/include/asm/xive.h extern int xive_native_set_queue_state(u32 vp_id, uint32_t prio, u32 qtoggle, prio 146 arch/powerpc/kvm/book3s.c unsigned int prio; prio 149 arch/powerpc/kvm/book3s.c case 0x100: prio = BOOK3S_IRQPRIO_SYSTEM_RESET; break; prio 150 arch/powerpc/kvm/book3s.c case 0x200: prio = BOOK3S_IRQPRIO_MACHINE_CHECK; break; prio 151 arch/powerpc/kvm/book3s.c case 0x300: prio = BOOK3S_IRQPRIO_DATA_STORAGE; break; prio 152 arch/powerpc/kvm/book3s.c case 0x380: prio = BOOK3S_IRQPRIO_DATA_SEGMENT; break; prio 153 arch/powerpc/kvm/book3s.c case 0x400: prio = BOOK3S_IRQPRIO_INST_STORAGE; break; prio 154 arch/powerpc/kvm/book3s.c case 0x480: prio = BOOK3S_IRQPRIO_INST_SEGMENT; break; prio 155 arch/powerpc/kvm/book3s.c case 0x500: prio = BOOK3S_IRQPRIO_EXTERNAL; break; prio 156 arch/powerpc/kvm/book3s.c case 0x600: prio = BOOK3S_IRQPRIO_ALIGNMENT; break; prio 157 arch/powerpc/kvm/book3s.c case 0x700: prio = BOOK3S_IRQPRIO_PROGRAM; break; prio 158 arch/powerpc/kvm/book3s.c case 0x800: prio = BOOK3S_IRQPRIO_FP_UNAVAIL; break; prio 159 arch/powerpc/kvm/book3s.c case 0x900: prio = BOOK3S_IRQPRIO_DECREMENTER; break; prio 160 arch/powerpc/kvm/book3s.c case 0xc00: prio = BOOK3S_IRQPRIO_SYSCALL; break; prio 161 arch/powerpc/kvm/book3s.c case 0xd00: prio = BOOK3S_IRQPRIO_DEBUG; break; prio 162 arch/powerpc/kvm/book3s.c case 0xf20: prio = BOOK3S_IRQPRIO_ALTIVEC; break; prio 163 arch/powerpc/kvm/book3s.c case 0xf40: prio = BOOK3S_IRQPRIO_VSX; break; prio 164 arch/powerpc/kvm/book3s.c case 0xf60: prio = BOOK3S_IRQPRIO_FAC_UNAVAIL; break; prio 165 arch/powerpc/kvm/book3s.c default: prio = BOOK3S_IRQPRIO_MAX; break; prio 168 arch/powerpc/kvm/book3s.c return prio; prio 1192 arch/powerpc/kvm/book3s_xics.c u64 val, prio; prio 1205 arch/powerpc/kvm/book3s_xics.c prio = irqp->priority; prio 1206 arch/powerpc/kvm/book3s_xics.c if (prio == MASKED) { prio 1208 arch/powerpc/kvm/book3s_xics.c prio = irqp->saved_priority; prio 1210 arch/powerpc/kvm/book3s_xics.c val |= prio << KVM_XICS_PRIORITY_SHIFT; prio 1242 arch/powerpc/kvm/book3s_xics.c u8 prio; prio 1260 arch/powerpc/kvm/book3s_xics.c prio = val >> KVM_XICS_PRIORITY_SHIFT; prio 1261 arch/powerpc/kvm/book3s_xics.c if (prio != MASKED && prio 1268 arch/powerpc/kvm/book3s_xics.c irqp->saved_priority = prio; prio 1270 arch/powerpc/kvm/book3s_xics.c prio = MASKED; prio 1271 arch/powerpc/kvm/book3s_xics.c irqp->priority = prio; prio 175 arch/powerpc/kvm/book3s_xive.c int kvmppc_xive_attach_escalation(struct kvm_vcpu *vcpu, u8 prio, prio 179 arch/powerpc/kvm/book3s_xive.c struct xive_q *q = &xc->queues[prio]; prio 184 arch/powerpc/kvm/book3s_xive.c if (xc->esc_virq[prio]) prio 188 arch/powerpc/kvm/book3s_xive.c xc->esc_virq[prio] = irq_create_mapping(NULL, q->esc_irq); prio 189 arch/powerpc/kvm/book3s_xive.c if (!xc->esc_virq[prio]) { prio 191 arch/powerpc/kvm/book3s_xive.c prio, xc->server_num); prio 200 arch/powerpc/kvm/book3s_xive.c vcpu->kvm->arch.lpid, xc->server_num, prio); prio 203 arch/powerpc/kvm/book3s_xive.c prio, xc->server_num); prio 208 arch/powerpc/kvm/book3s_xive.c pr_devel("Escalation %s irq %d (prio %d)\n", name, xc->esc_virq[prio], prio); prio 210 arch/powerpc/kvm/book3s_xive.c rc = request_irq(xc->esc_virq[prio], xive_esc_irq, prio 214 arch/powerpc/kvm/book3s_xive.c prio, xc->server_num); prio 217 arch/powerpc/kvm/book3s_xive.c xc->esc_virq_names[prio] = name; prio 228 arch/powerpc/kvm/book3s_xive.c struct irq_data *d = irq_get_irq_data(xc->esc_virq[prio]); prio 239 arch/powerpc/kvm/book3s_xive.c irq_dispose_mapping(xc->esc_virq[prio]); prio 240 arch/powerpc/kvm/book3s_xive.c xc->esc_virq[prio] = 0; prio 245 arch/powerpc/kvm/book3s_xive.c static int xive_provision_queue(struct kvm_vcpu *vcpu, u8 prio) prio 249 arch/powerpc/kvm/book3s_xive.c struct xive_q *q = &xc->queues[prio]; prio 260 arch/powerpc/kvm/book3s_xive.c prio, xc->server_num); prio 272 arch/powerpc/kvm/book3s_xive.c rc = xive_native_configure_queue(xc->vp_id, q, prio, qpage, prio 276 arch/powerpc/kvm/book3s_xive.c prio, xc->server_num); prio 281 arch/powerpc/kvm/book3s_xive.c static int xive_check_provisioning(struct kvm *kvm, u8 prio) prio 290 arch/powerpc/kvm/book3s_xive.c if (xive->qmap & (1 << prio)) prio 293 arch/powerpc/kvm/book3s_xive.c pr_devel("Provisioning prio... %d\n", prio); prio 299 arch/powerpc/kvm/book3s_xive.c rc = xive_provision_queue(vcpu, prio); prio 301 arch/powerpc/kvm/book3s_xive.c kvmppc_xive_attach_escalation(vcpu, prio, prio 309 arch/powerpc/kvm/book3s_xive.c xive->qmap |= (1 << prio); prio 313 arch/powerpc/kvm/book3s_xive.c static void xive_inc_q_pending(struct kvm *kvm, u32 server, u8 prio) prio 329 arch/powerpc/kvm/book3s_xive.c q = &xc->queues[prio]; prio 333 arch/powerpc/kvm/book3s_xive.c static int xive_try_pick_queue(struct kvm_vcpu *vcpu, u8 prio) prio 344 arch/powerpc/kvm/book3s_xive.c q = &xc->queues[prio]; prio 353 arch/powerpc/kvm/book3s_xive.c int kvmppc_xive_select_target(struct kvm *kvm, u32 *server, u8 prio) prio 365 arch/powerpc/kvm/book3s_xive.c pr_devel("Finding irq target on 0x%x/%d...\n", *server, prio); prio 368 arch/powerpc/kvm/book3s_xive.c rc = xive_try_pick_queue(vcpu, prio); prio 378 arch/powerpc/kvm/book3s_xive.c rc = xive_try_pick_queue(vcpu, prio); prio 381 arch/powerpc/kvm/book3s_xive.c pr_devel(" found on 0x%x/%d\n", *server, prio); prio 474 arch/powerpc/kvm/book3s_xive.c u8 prio) prio 518 arch/powerpc/kvm/book3s_xive.c state->guest_priority = prio; prio 530 arch/powerpc/kvm/book3s_xive.c u32 server, u8 prio) prio 541 arch/powerpc/kvm/book3s_xive.c rc = kvmppc_xive_select_target(kvm, &server, prio); prio 562 arch/powerpc/kvm/book3s_xive.c state->act_priority = prio; prio 570 arch/powerpc/kvm/book3s_xive.c prio, state->number); prio 906 arch/powerpc/kvm/book3s_xive.c u8 prio; prio 943 arch/powerpc/kvm/book3s_xive.c prio = xive_lock_and_mask(xive, sb, state); prio 944 arch/powerpc/kvm/book3s_xive.c pr_devel(" old IPI prio %02x P:%d Q:%d\n", prio, prio 978 arch/powerpc/kvm/book3s_xive.c if (prio != MASKED && !state->old_p) prio 986 arch/powerpc/kvm/book3s_xive.c state->guest_priority = prio; prio 1001 arch/powerpc/kvm/book3s_xive.c u8 prio; prio 1019 arch/powerpc/kvm/book3s_xive.c prio = xive_lock_and_mask(xive, sb, state); prio 1020 arch/powerpc/kvm/book3s_xive.c pr_devel(" old IRQ prio %02x P:%d Q:%d\n", prio, prio 1060 arch/powerpc/kvm/book3s_xive.c if (prio == MASKED || state->old_p) prio 1067 arch/powerpc/kvm/book3s_xive.c state->guest_priority = prio; prio 1499 arch/powerpc/kvm/book3s_xive.c u64 val, prio; prio 1535 arch/powerpc/kvm/book3s_xive.c prio = state->saved_scan_prio; prio 1537 arch/powerpc/kvm/book3s_xive.c if (prio == MASKED) { prio 1539 arch/powerpc/kvm/book3s_xive.c prio = state->saved_priority; prio 1541 arch/powerpc/kvm/book3s_xive.c val |= prio << KVM_XICS_PRIORITY_SHIFT; prio 1559 arch/powerpc/kvm/book3s_xive.c if (state->in_queue || (prio == MASKED && state->saved_q)) prio 244 arch/powerpc/kvm/book3s_xive.h static inline u8 xive_prio_from_guest(u8 prio) prio 246 arch/powerpc/kvm/book3s_xive.h if (prio == 0xff || prio < 6) prio 247 arch/powerpc/kvm/book3s_xive.h return prio; prio 251 arch/powerpc/kvm/book3s_xive.h static inline u8 xive_prio_to_guest(u8 prio) prio 253 arch/powerpc/kvm/book3s_xive.h return prio; prio 293 arch/powerpc/kvm/book3s_xive.h int kvmppc_xive_select_target(struct kvm *kvm, u32 *server, u8 prio); prio 294 arch/powerpc/kvm/book3s_xive.h int kvmppc_xive_attach_escalation(struct kvm_vcpu *vcpu, u8 prio, prio 41 arch/powerpc/kvm/book3s_xive_native.c static void kvmppc_xive_native_cleanup_queue(struct kvm_vcpu *vcpu, int prio) prio 44 arch/powerpc/kvm/book3s_xive_native.c struct xive_q *q = &xc->queues[prio]; prio 46 arch/powerpc/kvm/book3s_xive_native.c xive_native_disable_queue(xc->vp_id, q, prio); prio 54 arch/powerpc/kvm/book3s_xive_native.c u8 prio, __be32 *qpage, prio 60 arch/powerpc/kvm/book3s_xive_native.c rc = xive_native_configure_queue(vp_id, q, prio, qpage, order, prio 809 arch/powerpc/kvm/book3s_xive_native.c unsigned int prio; prio 816 arch/powerpc/kvm/book3s_xive_native.c for (prio = 0; prio < KVMPPC_XIVE_Q_COUNT; prio++) { prio 819 arch/powerpc/kvm/book3s_xive_native.c if (prio == 7 && xive->single_escalation) prio 822 arch/powerpc/kvm/book3s_xive_native.c if (xc->esc_virq[prio]) { prio 823 arch/powerpc/kvm/book3s_xive_native.c free_irq(xc->esc_virq[prio], vcpu); prio 824 arch/powerpc/kvm/book3s_xive_native.c irq_dispose_mapping(xc->esc_virq[prio]); prio 825 arch/powerpc/kvm/book3s_xive_native.c kfree(xc->esc_virq_names[prio]); prio 826 arch/powerpc/kvm/book3s_xive_native.c xc->esc_virq[prio] = 0; prio 829 arch/powerpc/kvm/book3s_xive_native.c kvmppc_xive_native_cleanup_queue(vcpu, prio); prio 886 arch/powerpc/kvm/book3s_xive_native.c unsigned int prio; prio 892 arch/powerpc/kvm/book3s_xive_native.c for (prio = 0; prio < KVMPPC_XIVE_Q_COUNT; prio++) { prio 893 arch/powerpc/kvm/book3s_xive_native.c struct xive_q *q = &xc->queues[prio]; prio 116 arch/powerpc/kvm/book3s_xive_template.c u8 prio = 0xff; prio 128 arch/powerpc/kvm/book3s_xive_template.c prio = ffs(pending) - 1; prio 131 arch/powerpc/kvm/book3s_xive_template.c if (prio >= xc->cppr || prio > 7) { prio 133 arch/powerpc/kvm/book3s_xive_template.c prio = xc->mfrr; prio 140 arch/powerpc/kvm/book3s_xive_template.c q = &xc->queues[prio]; prio 173 arch/powerpc/kvm/book3s_xive_template.c if (hirq == XICS_IPI || (prio == 0 && !qpage)) { prio 194 arch/powerpc/kvm/book3s_xive_template.c pending &= ~(1 << prio); prio 216 arch/powerpc/kvm/book3s_xive_template.c if (prio >= xc->mfrr && xc->mfrr < xc->cppr) { prio 217 arch/powerpc/kvm/book3s_xive_template.c prio = xc->mfrr; prio 254 arch/powerpc/kvm/book3s_xive_template.c xc->cppr = prio; prio 357 arch/powerpc/kvm/book3s_xive_template.c u8 pending, prio; prio 368 arch/powerpc/kvm/book3s_xive_template.c prio = ffs(pending) - 1; prio 370 arch/powerpc/kvm/book3s_xive_template.c __x_writeb(prio, __x_tima + TM_SPC_SET_OS_PENDING); prio 376 arch/powerpc/kvm/book3s_xive_template.c unsigned int prio; prio 379 arch/powerpc/kvm/book3s_xive_template.c for (prio = xc->cppr; prio < KVMPPC_XIVE_Q_COUNT; prio++) { prio 380 arch/powerpc/kvm/book3s_xive_template.c struct xive_q *q = &xc->queues[prio]; prio 324 arch/powerpc/kvm/booke.c unsigned int prio = BOOKE_IRQPRIO_EXTERNAL; prio 327 arch/powerpc/kvm/booke.c prio = BOOKE_IRQPRIO_EXTERNAL_LEVEL; prio 329 arch/powerpc/kvm/booke.c kvmppc_booke_queue_irqprio(vcpu, prio); prio 35 arch/powerpc/kvm/e500_emulate.c int prio = -1; prio 39 arch/powerpc/kvm/e500_emulate.c prio = BOOKE_IRQPRIO_DBELL; prio 42 arch/powerpc/kvm/e500_emulate.c prio = BOOKE_IRQPRIO_DBELL_CRIT; prio 48 arch/powerpc/kvm/e500_emulate.c return prio; prio 54 arch/powerpc/kvm/e500_emulate.c int prio = dbell2prio(param); prio 56 arch/powerpc/kvm/e500_emulate.c if (prio < 0) prio 59 arch/powerpc/kvm/e500_emulate.c clear_bit(prio, &vcpu->arch.pending_exceptions); prio 66 arch/powerpc/kvm/e500_emulate.c int prio = dbell2prio(rb); prio 71 arch/powerpc/kvm/e500_emulate.c if (prio < 0) prio 77 arch/powerpc/kvm/e500_emulate.c set_bit(prio, &cvcpu->arch.pending_exceptions); prio 57 arch/powerpc/platforms/cell/interrupt.c return IIC_IRQ_TYPE_IPI | (bits.prio >> 4); prio 73 arch/powerpc/platforms/cell/interrupt.c out_be64(&iic->regs->prio, iic->eoi_stack[--iic->eoi_ptr]); prio 147 arch/powerpc/platforms/cell/interrupt.c iic->eoi_stack[++iic->eoi_ptr] = pending.prio; prio 154 arch/powerpc/platforms/cell/interrupt.c out_be64(&this_cpu_ptr(&cpu_iic)->regs->prio, 0xff); prio 291 arch/powerpc/platforms/cell/interrupt.c out_be64(&iic->regs->prio, 0); prio 2528 arch/powerpc/platforms/cell/spufs/file.c ctx->prio, prio 74 arch/powerpc/platforms/cell/spufs/sched.c #define SCALE_PRIO(x, prio) \ prio 75 arch/powerpc/platforms/cell/spufs/sched.c max(x * (MAX_PRIO - prio) / (MAX_USER_PRIO / 2), MIN_SPU_TIMESLICE) prio 87 arch/powerpc/platforms/cell/spufs/sched.c if (ctx->prio < NORMAL_PRIO) prio 88 arch/powerpc/platforms/cell/spufs/sched.c ctx->time_slice = SCALE_PRIO(DEF_SPU_TIMESLICE * 4, ctx->prio); prio 90 arch/powerpc/platforms/cell/spufs/sched.c ctx->time_slice = SCALE_PRIO(DEF_SPU_TIMESLICE, ctx->prio); prio 117 arch/powerpc/platforms/cell/spufs/sched.c if (rt_prio(current->prio)) prio 118 arch/powerpc/platforms/cell/spufs/sched.c ctx->prio = current->prio; prio 120 arch/powerpc/platforms/cell/spufs/sched.c ctx->prio = current->static_prio; prio 497 arch/powerpc/platforms/cell/spufs/sched.c list_add_tail(&ctx->rq, &spu_prio->runq[ctx->prio]); prio 498 arch/powerpc/platforms/cell/spufs/sched.c set_bit(ctx->prio, spu_prio->bitmap); prio 513 arch/powerpc/platforms/cell/spufs/sched.c int prio = ctx->prio; prio 520 arch/powerpc/platforms/cell/spufs/sched.c if (list_empty(&spu_prio->runq[prio])) prio 521 arch/powerpc/platforms/cell/spufs/sched.c clear_bit(prio, spu_prio->bitmap); prio 643 arch/powerpc/platforms/cell/spufs/sched.c if (tmp && tmp->prio > ctx->prio && prio 645 arch/powerpc/platforms/cell/spufs/sched.c (!victim || tmp->prio > victim->prio)) { prio 671 arch/powerpc/platforms/cell/spufs/sched.c if (!spu || victim->prio <= ctx->prio) { prio 796 arch/powerpc/platforms/cell/spufs/sched.c if (!spu && rt_prio(ctx->prio)) prio 825 arch/powerpc/platforms/cell/spufs/sched.c static struct spu_context *grab_runnable_context(int prio, int node) prio 831 arch/powerpc/platforms/cell/spufs/sched.c best = find_first_bit(spu_prio->bitmap, prio); prio 832 arch/powerpc/platforms/cell/spufs/sched.c while (best < prio) { prio 929 arch/powerpc/platforms/cell/spufs/sched.c new = grab_runnable_context(ctx->prio + 1, spu->node); prio 113 arch/powerpc/platforms/cell/spufs/spufs.h int prio; prio 74 arch/powerpc/sysdev/ehv_pic.c unsigned int config, prio, cpu_dest; prio 79 arch/powerpc/sysdev/ehv_pic.c ev_int_get_config(src, &config, &prio, &cpu_dest); prio 80 arch/powerpc/sysdev/ehv_pic.c ev_int_set_config(src, config, prio, cpuid); prio 114 arch/powerpc/sysdev/ehv_pic.c unsigned int vecpri, vold, vnew, prio, cpu_dest; prio 125 arch/powerpc/sysdev/ehv_pic.c ev_int_get_config(src, &vold, &prio, &cpu_dest); prio 136 arch/powerpc/sysdev/ehv_pic.c prio = 8; prio 138 arch/powerpc/sysdev/ehv_pic.c ev_int_set_config(src, vecpri, prio, cpu_dest); prio 35 arch/powerpc/sysdev/ipic.c .prio = IPIC_SIPRR_C, prio 42 arch/powerpc/sysdev/ipic.c .prio = IPIC_SIPRR_C, prio 49 arch/powerpc/sysdev/ipic.c .prio = IPIC_SIPRR_C, prio 56 arch/powerpc/sysdev/ipic.c .prio = IPIC_SIPRR_C, prio 63 arch/powerpc/sysdev/ipic.c .prio = IPIC_SIPRR_C, prio 70 arch/powerpc/sysdev/ipic.c .prio = IPIC_SIPRR_C, prio 77 arch/powerpc/sysdev/ipic.c .prio = IPIC_SIPRR_C, prio 84 arch/powerpc/sysdev/ipic.c .prio = IPIC_SIPRR_C, prio 91 arch/powerpc/sysdev/ipic.c .prio = IPIC_SIPRR_D, prio 98 arch/powerpc/sysdev/ipic.c .prio = IPIC_SIPRR_D, prio 105 arch/powerpc/sysdev/ipic.c .prio = IPIC_SIPRR_D, prio 112 arch/powerpc/sysdev/ipic.c .prio = IPIC_SIPRR_D, prio 119 arch/powerpc/sysdev/ipic.c .prio = IPIC_SIPRR_D, prio 126 arch/powerpc/sysdev/ipic.c .prio = IPIC_SIPRR_D, prio 133 arch/powerpc/sysdev/ipic.c .prio = IPIC_SIPRR_D, prio 140 arch/powerpc/sysdev/ipic.c .prio = IPIC_SIPRR_D, prio 148 arch/powerpc/sysdev/ipic.c .prio = IPIC_SMPRR_A, prio 156 arch/powerpc/sysdev/ipic.c .prio = IPIC_SMPRR_A, prio 164 arch/powerpc/sysdev/ipic.c .prio = IPIC_SMPRR_A, prio 172 arch/powerpc/sysdev/ipic.c .prio = IPIC_SMPRR_B, prio 180 arch/powerpc/sysdev/ipic.c .prio = IPIC_SMPRR_B, prio 188 arch/powerpc/sysdev/ipic.c .prio = IPIC_SMPRR_B, prio 196 arch/powerpc/sysdev/ipic.c .prio = IPIC_SMPRR_B, prio 203 arch/powerpc/sysdev/ipic.c .prio = IPIC_SIPRR_A, prio 210 arch/powerpc/sysdev/ipic.c .prio = IPIC_SIPRR_A, prio 217 arch/powerpc/sysdev/ipic.c .prio = IPIC_SIPRR_A, prio 224 arch/powerpc/sysdev/ipic.c .prio = IPIC_SIPRR_A, prio 231 arch/powerpc/sysdev/ipic.c .prio = IPIC_SIPRR_A, prio 238 arch/powerpc/sysdev/ipic.c .prio = IPIC_SIPRR_A, prio 245 arch/powerpc/sysdev/ipic.c .prio = IPIC_SIPRR_A, prio 252 arch/powerpc/sysdev/ipic.c .prio = IPIC_SIPRR_A, prio 259 arch/powerpc/sysdev/ipic.c .prio = IPIC_SIPRR_B, prio 266 arch/powerpc/sysdev/ipic.c .prio = IPIC_SIPRR_B, prio 273 arch/powerpc/sysdev/ipic.c .prio = IPIC_SIPRR_B, prio 280 arch/powerpc/sysdev/ipic.c .prio = IPIC_SIPRR_B, prio 287 arch/powerpc/sysdev/ipic.c .prio = IPIC_SIPRR_B, prio 294 arch/powerpc/sysdev/ipic.c .prio = IPIC_SIPRR_B, prio 301 arch/powerpc/sysdev/ipic.c .prio = IPIC_SIPRR_B, prio 308 arch/powerpc/sysdev/ipic.c .prio = IPIC_SIPRR_B, prio 316 arch/powerpc/sysdev/ipic.c .prio = IPIC_SMPRR_A, prio 323 arch/powerpc/sysdev/ipic.c .prio = IPIC_SMPRR_A, prio 330 arch/powerpc/sysdev/ipic.c .prio = IPIC_SMPRR_A, prio 337 arch/powerpc/sysdev/ipic.c .prio = IPIC_SMPRR_A, prio 344 arch/powerpc/sysdev/ipic.c .prio = IPIC_SMPRR_A, prio 351 arch/powerpc/sysdev/ipic.c .prio = IPIC_SMPRR_B, prio 358 arch/powerpc/sysdev/ipic.c .prio = IPIC_SMPRR_B, prio 365 arch/powerpc/sysdev/ipic.c .prio = IPIC_SMPRR_B, prio 372 arch/powerpc/sysdev/ipic.c .prio = IPIC_SMPRR_B, prio 379 arch/powerpc/sysdev/ipic.c .prio = 0, prio 385 arch/powerpc/sysdev/ipic.c .prio = 0, prio 391 arch/powerpc/sysdev/ipic.c .prio = 0, prio 397 arch/powerpc/sysdev/ipic.c .prio = 0, prio 403 arch/powerpc/sysdev/ipic.c .prio = 0, prio 409 arch/powerpc/sysdev/ipic.c .prio = 0, prio 415 arch/powerpc/sysdev/ipic.c .prio = 0, prio 421 arch/powerpc/sysdev/ipic.c .prio = 0, prio 427 arch/powerpc/sysdev/ipic.c .prio = 0, prio 433 arch/powerpc/sysdev/ipic.c .prio = 0, prio 439 arch/powerpc/sysdev/ipic.c .prio = 0, prio 445 arch/powerpc/sysdev/ipic.c .prio = 0, prio 451 arch/powerpc/sysdev/ipic.c .prio = 0, prio 457 arch/powerpc/sysdev/ipic.c .prio = 0, prio 463 arch/powerpc/sysdev/ipic.c .prio = 0, prio 469 arch/powerpc/sysdev/ipic.c .prio = 0, prio 475 arch/powerpc/sysdev/ipic.c .prio = 0, prio 481 arch/powerpc/sysdev/ipic.c .prio = 0, prio 487 arch/powerpc/sysdev/ipic.c .prio = 0, prio 493 arch/powerpc/sysdev/ipic.c .prio = 0, prio 499 arch/powerpc/sysdev/ipic.c .prio = 0, prio 49 arch/powerpc/sysdev/ipic.h u8 prio; /* priority register offset from base */ prio 1733 arch/powerpc/sysdev/mpic.c void mpic_cpu_set_priority(int prio) prio 1737 arch/powerpc/sysdev/mpic.c prio &= MPIC_CPU_TASKPRI_MASK; prio 1738 arch/powerpc/sysdev/mpic.c mpic_cpu_write(MPIC_INFO(CPU_CURRENT_TASK_PRI), prio); prio 132 arch/powerpc/sysdev/xive/common.c u8 prio = 0; prio 138 arch/powerpc/sysdev/xive/common.c prio = ffs(xc->pending_prio) - 1; prio 139 arch/powerpc/sysdev/xive/common.c DBG_VERBOSE("scan_irq: trying prio %d\n", prio); prio 142 arch/powerpc/sysdev/xive/common.c irq = xive_read_eq(&xc->queue[prio], just_peek); prio 160 arch/powerpc/sysdev/xive/common.c xc->pending_prio &= ~(1 << prio); prio 167 arch/powerpc/sysdev/xive/common.c q = &xc->queue[prio]; prio 179 arch/powerpc/sysdev/xive/common.c prio = 0xff; prio 182 arch/powerpc/sysdev/xive/common.c if (prio != xc->cppr) { prio 183 arch/powerpc/sysdev/xive/common.c DBG_VERBOSE("scan_irq: adjusting CPPR to %d\n", prio); prio 184 arch/powerpc/sysdev/xive/common.c xc->cppr = prio; prio 185 arch/powerpc/sysdev/xive/common.c out_8(xive_tima + xive_tima_offset + TM_CPPR, prio); prio 265 arch/powerpc/sysdev/xive/common.c u8 prio; prio 271 arch/powerpc/sysdev/xive/common.c rc = xive_ops->get_irq_config(hw_irq, &target, &prio, &lirq); prio 278 arch/powerpc/sysdev/xive/common.c hw_irq, target, prio, lirq); prio 100 arch/powerpc/sysdev/xive/native.c int xive_native_configure_irq(u32 hw_irq, u32 target, u8 prio, u32 sw_irq) prio 105 arch/powerpc/sysdev/xive/native.c rc = opal_xive_set_irq_config(hw_irq, target, prio, sw_irq); prio 114 arch/powerpc/sysdev/xive/native.c static int xive_native_get_irq_config(u32 hw_irq, u32 *target, u8 *prio, prio 121 arch/powerpc/sysdev/xive/native.c rc = opal_xive_get_irq_config(hw_irq, &vp, prio, &lirq); prio 130 arch/powerpc/sysdev/xive/native.c int xive_native_configure_queue(u32 vp_id, struct xive_q *q, u8 prio, prio 151 arch/powerpc/sysdev/xive/native.c rc = opal_xive_get_queue_info(vp_id, prio, NULL, NULL, prio 156 arch/powerpc/sysdev/xive/native.c pr_err("Error %lld getting queue info prio %d\n", rc, prio); prio 173 arch/powerpc/sysdev/xive/native.c rc = opal_xive_set_queue_info(vp_id, prio, qpage_phys, order, flags); prio 179 arch/powerpc/sysdev/xive/native.c pr_err("Error %lld setting queue for prio %d\n", rc, prio); prio 194 arch/powerpc/sysdev/xive/native.c static void __xive_native_disable_queue(u32 vp_id, struct xive_q *q, u8 prio) prio 200 arch/powerpc/sysdev/xive/native.c rc = opal_xive_set_queue_info(vp_id, prio, 0, 0, 0); prio 206 arch/powerpc/sysdev/xive/native.c pr_err("Error %lld disabling queue for prio %d\n", rc, prio); prio 209 arch/powerpc/sysdev/xive/native.c void xive_native_disable_queue(u32 vp_id, struct xive_q *q, u8 prio) prio 211 arch/powerpc/sysdev/xive/native.c __xive_native_disable_queue(vp_id, q, prio); prio 215 arch/powerpc/sysdev/xive/native.c static int xive_native_setup_queue(unsigned int cpu, struct xive_cpu *xc, u8 prio) prio 217 arch/powerpc/sysdev/xive/native.c struct xive_q *q = &xc->queue[prio]; prio 225 arch/powerpc/sysdev/xive/native.c q, prio, qpage, xive_queue_shift, false); prio 228 arch/powerpc/sysdev/xive/native.c static void xive_native_cleanup_queue(unsigned int cpu, struct xive_cpu *xc, u8 prio) prio 230 arch/powerpc/sysdev/xive/native.c struct xive_q *q = &xc->queue[prio]; prio 237 arch/powerpc/sysdev/xive/native.c __xive_native_disable_queue(get_hard_smp_processor_id(cpu), q, prio); prio 754 arch/powerpc/sysdev/xive/native.c int xive_native_get_queue_info(u32 vp_id, u32 prio, prio 768 arch/powerpc/sysdev/xive/native.c rc = opal_xive_get_queue_info(vp_id, prio, &qpage, &qsize, prio 772 arch/powerpc/sysdev/xive/native.c vp_id, prio, rc); prio 791 arch/powerpc/sysdev/xive/native.c int xive_native_get_queue_state(u32 vp_id, u32 prio, u32 *qtoggle, u32 *qindex) prio 797 arch/powerpc/sysdev/xive/native.c rc = opal_xive_get_queue_state(vp_id, prio, &opal_qtoggle, prio 801 arch/powerpc/sysdev/xive/native.c vp_id, prio, rc); prio 814 arch/powerpc/sysdev/xive/native.c int xive_native_set_queue_state(u32 vp_id, u32 prio, u32 qtoggle, u32 qindex) prio 818 arch/powerpc/sysdev/xive/native.c rc = opal_xive_set_queue_state(vp_id, prio, qtoggle, qindex); prio 821 arch/powerpc/sysdev/xive/native.c vp_id, prio, rc); prio 194 arch/powerpc/sysdev/xive/spapr.c unsigned long prio, prio 201 arch/powerpc/sysdev/xive/spapr.c flags, lisn, target, prio, sw_irq); prio 206 arch/powerpc/sysdev/xive/spapr.c target, prio, sw_irq); prio 211 arch/powerpc/sysdev/xive/spapr.c lisn, target, prio, rc); prio 221 arch/powerpc/sysdev/xive/spapr.c unsigned long *prio, prio 231 arch/powerpc/sysdev/xive/spapr.c target, prio, sw_irq); prio 241 arch/powerpc/sysdev/xive/spapr.c *prio = retbuf[1]; prio 431 arch/powerpc/sysdev/xive/spapr.c static int xive_spapr_configure_irq(u32 hw_irq, u32 target, u8 prio, u32 sw_irq) prio 436 arch/powerpc/sysdev/xive/spapr.c prio, sw_irq); prio 441 arch/powerpc/sysdev/xive/spapr.c static int xive_spapr_get_irq_config(u32 hw_irq, u32 *target, u8 *prio, prio 453 arch/powerpc/sysdev/xive/spapr.c *prio = h_prio; prio 460 arch/powerpc/sysdev/xive/spapr.c static int xive_spapr_configure_queue(u32 target, struct xive_q *q, u8 prio, prio 482 arch/powerpc/sysdev/xive/spapr.c rc = plpar_int_get_queue_info(0, target, prio, &esn_page, &esn_size); prio 485 arch/powerpc/sysdev/xive/spapr.c target, prio); prio 497 arch/powerpc/sysdev/xive/spapr.c rc = plpar_int_set_queue_config(flags, target, prio, qpage_phys, order); prio 500 arch/powerpc/sysdev/xive/spapr.c target, prio); prio 510 arch/powerpc/sysdev/xive/spapr.c u8 prio) prio 512 arch/powerpc/sysdev/xive/spapr.c struct xive_q *q = &xc->queue[prio]; prio 520 arch/powerpc/sysdev/xive/spapr.c q, prio, qpage, xive_queue_shift); prio 524 arch/powerpc/sysdev/xive/spapr.c u8 prio) prio 526 arch/powerpc/sysdev/xive/spapr.c struct xive_q *q = &xc->queue[prio]; prio 531 arch/powerpc/sysdev/xive/spapr.c rc = plpar_int_set_queue_config(0, hw_cpu, prio, 0, 0); prio 534 arch/powerpc/sysdev/xive/spapr.c hw_cpu, prio); prio 677 arch/powerpc/sysdev/xive/spapr.c int prio, found; prio 701 arch/powerpc/sysdev/xive/spapr.c for (prio = 0; prio < 8; prio++) { prio 709 arch/powerpc/sysdev/xive/spapr.c if (prio >= base && prio < base + range) prio 714 arch/powerpc/sysdev/xive/spapr.c found = prio; prio 42 arch/powerpc/sysdev/xive/xive-internal.h int (*configure_irq)(u32 hw_irq, u32 target, u8 prio, u32 sw_irq); prio 43 arch/powerpc/sysdev/xive/xive-internal.h int (*get_irq_config)(u32 hw_irq, u32 *target, u8 *prio, prio 45 arch/powerpc/sysdev/xive/xive-internal.h int (*setup_queue)(unsigned int cpu, struct xive_cpu *xc, u8 prio); prio 46 arch/powerpc/sysdev/xive/xive-internal.h void (*cleanup_queue)(unsigned int cpu, struct xive_cpu *xc, u8 prio); prio 269 arch/s390/include/asm/fcx.h u32 prio:8; prio 28 arch/unicore32/include/mach/dma.h puv3_dma_prio prio, prio 24 arch/unicore32/kernel/dma.c puv3_dma_prio prio; prio 32 arch/unicore32/kernel/dma.c int puv3_request_dma(char *name, puv3_dma_prio prio, prio 49 arch/unicore32/kernel/dma.c if ((dma_channels[i].prio == prio) && prio 56 arch/unicore32/kernel/dma.c } while (!found && prio--); prio 160 arch/unicore32/kernel/dma.c dma_channels[i].prio = min((i & 0x7) >> 1, DMA_PRIO_LOW); prio 173 arch/x86/include/asm/topology.h void sched_set_itmt_core_prio(int prio, int core_cpu); prio 184 arch/x86/include/asm/topology.h static inline void sched_set_itmt_core_prio(int prio, int core_cpu) prio 190 arch/x86/kernel/itmt.c void sched_set_itmt_core_prio(int prio, int core_cpu) prio 202 arch/x86/kernel/itmt.c smt_prio = prio * smp_num_siblings / i; prio 383 drivers/base/power/qos.c curr_value = req->data.pnode.prio; prio 833 drivers/base/power/qos.c dev->power.qos->latency_tolerance_req->data.pnode.prio; prio 209 drivers/bus/da8xx-mstpri.c const struct da8xx_mstpri_priority *prio; prio 230 drivers/bus/da8xx-mstpri.c prio = &prio_list->priorities[i]; prio 231 drivers/bus/da8xx-mstpri.c prio_descr = &da8xx_mstpri_priority_list[prio->which]; prio 240 drivers/bus/da8xx-mstpri.c reg |= prio->val << prio_descr->shift; prio 5028 drivers/crypto/caam/caamalg_qi2.c ppriv->prio = j; prio 88 drivers/crypto/caam/caamalg_qi2.h int prio; prio 1263 drivers/dma/imx-sdma.c int prio, ret; prio 1287 drivers/dma/imx-sdma.c prio = 3; prio 1290 drivers/dma/imx-sdma.c prio = 2; prio 1294 drivers/dma/imx-sdma.c prio = 1; prio 1309 drivers/dma/imx-sdma.c ret = sdma_set_channel_priority(sdmac, prio); prio 561 drivers/dma/ipu/ipu_idmac.c bool prio) prio 565 drivers/dma/ipu/ipu_idmac.c if (prio) prio 242 drivers/dma/mmp_pdma.c int prio, i; prio 256 drivers/dma/mmp_pdma.c for (prio = 0; prio <= ((pdev->dma_channels - 1) & 0xf) >> 2; prio++) { prio 258 drivers/dma/mmp_pdma.c if (prio != (i & 0xf) >> 2) prio 106 drivers/dma/pxa_dma.c enum pxad_chan_prio prio; /* Required priority of phy */ prio 371 drivers/dma/pxa_dma.c int prio, i; prio 385 drivers/dma/pxa_dma.c for (prio = pchan->prio; prio >= PXAD_PRIO_HIGHEST; prio--) { prio 387 drivers/dma/pxa_dma.c if (prio != (i & 0xf) >> 2) prio 717 drivers/dma/pxa_dma.c chan->prio = PXAD_PRIO_LOWEST; prio 1303 drivers/dma/pxa_dma.c to_pxad_chan(chan)->prio = dma_spec->args[1]; prio 1343 drivers/dma/pxa_dma.c c->prio = PXAD_PRIO_LOWEST; prio 1463 drivers/dma/pxa_dma.c c->prio = p->prio; prio 587 drivers/gpu/drm/i915/gem/i915_gem_context.c i915_gem_context_create_kernel(struct drm_i915_private *i915, int prio) prio 603 drivers/gpu/drm/i915/gem/i915_gem_context.c ctx->sched.priority = I915_USER_PRIORITY(prio); prio 162 drivers/gpu/drm/i915/gem/i915_gem_context.h i915_gem_context_create_kernel(struct drm_i915_private *i915, int prio); prio 203 drivers/gpu/drm/i915/gt/intel_lrc.c int prio; prio 276 drivers/gpu/drm/i915/gt/intel_lrc.c int prio = rq_prio(rq); prio 287 drivers/gpu/drm/i915/gt/intel_lrc.c prio = I915_PRIORITY_UNPREEMPTABLE; prio 296 drivers/gpu/drm/i915/gt/intel_lrc.c prio |= I915_PRIORITY_NOSEMAPHORE; prio 300 drivers/gpu/drm/i915/gt/intel_lrc.c return prio | __NO_PREEMPTION; prio 479 drivers/gpu/drm/i915/gt/intel_lrc.c int prio = I915_PRIORITY_INVALID; prio 503 drivers/gpu/drm/i915/gt/intel_lrc.c if (rq_prio(rq) != prio) { prio 504 drivers/gpu/drm/i915/gt/intel_lrc.c prio = rq_prio(rq); prio 505 drivers/gpu/drm/i915/gt/intel_lrc.c pl = i915_sched_lookup_priolist(engine, prio); prio 1656 drivers/gpu/drm/i915/gt/intel_lrc.c int prio) prio 1659 drivers/gpu/drm/i915/gt/intel_lrc.c list_add_tail(&node->link, i915_sched_lookup_priolist(engine, prio)); prio 3553 drivers/gpu/drm/i915/gt/intel_lrc.c const int prio = ve->base.execlists.queue_priority_hint; prio 3590 drivers/gpu/drm/i915/gt/intel_lrc.c if (prio == node->prio || (prio > node->prio && first)) prio 3604 drivers/gpu/drm/i915/gt/intel_lrc.c if (prio > other->prio) { prio 3619 drivers/gpu/drm/i915/gt/intel_lrc.c node->prio = prio; prio 3620 drivers/gpu/drm/i915/gt/intel_lrc.c if (first && prio > sibling->execlists.queue_priority_hint) { prio 3621 drivers/gpu/drm/i915/gt/intel_lrc.c sibling->execlists.queue_priority_hint = prio; prio 1435 drivers/gpu/drm/i915/gt/selftest_lrc.c struct i915_gem_context *ctx, int prio, prio 1452 drivers/gpu/drm/i915/gt/selftest_lrc.c ctx->sched.priority = prio; prio 72 drivers/gpu/drm/i915/i915_scheduler.c i915_sched_lookup_priolist(struct intel_engine_cs *engine, int prio) prio 84 drivers/gpu/drm/i915/i915_scheduler.c idx = I915_PRIORITY_COUNT - (prio & I915_PRIORITY_MASK) - 1; prio 85 drivers/gpu/drm/i915/i915_scheduler.c prio >>= I915_USER_PRIORITY_SHIFT; prio 87 drivers/gpu/drm/i915/i915_scheduler.c prio = I915_PRIORITY_NORMAL; prio 96 drivers/gpu/drm/i915/i915_scheduler.c if (prio > p->priority) { prio 98 drivers/gpu/drm/i915/i915_scheduler.c } else if (prio < p->priority) { prio 106 drivers/gpu/drm/i915/i915_scheduler.c if (prio == I915_PRIORITY_NORMAL) { prio 112 drivers/gpu/drm/i915/i915_scheduler.c prio = I915_PRIORITY_NORMAL; /* recurses just once */ prio 127 drivers/gpu/drm/i915/i915_scheduler.c p->priority = prio; prio 180 drivers/gpu/drm/i915/i915_scheduler.c static inline bool need_preempt(int prio, int active) prio 189 drivers/gpu/drm/i915/i915_scheduler.c return prio >= max(I915_PRIORITY_NORMAL, active); prio 194 drivers/gpu/drm/i915/i915_scheduler.c int prio) prio 202 drivers/gpu/drm/i915/i915_scheduler.c if (prio <= engine->execlists.queue_priority_hint) prio 222 drivers/gpu/drm/i915/i915_scheduler.c engine->execlists.queue_priority_hint = prio; prio 223 drivers/gpu/drm/i915/i915_scheduler.c if (need_preempt(prio, rq_prio(inflight))) prio 236 drivers/gpu/drm/i915/i915_scheduler.c const int prio = attr->priority; prio 242 drivers/gpu/drm/i915/i915_scheduler.c GEM_BUG_ON(prio == I915_PRIORITY_INVALID); prio 244 drivers/gpu/drm/i915/i915_scheduler.c if (prio <= READ_ONCE(node->attr.priority)) prio 290 drivers/gpu/drm/i915/i915_scheduler.c if (prio > READ_ONCE(p->signaler->attr.priority)) prio 325 drivers/gpu/drm/i915/i915_scheduler.c if (prio <= node->attr.priority || node_signaled(node)) prio 330 drivers/gpu/drm/i915/i915_scheduler.c node->attr.priority = prio; prio 349 drivers/gpu/drm/i915/i915_scheduler.c prio); prio 354 drivers/gpu/drm/i915/i915_scheduler.c kick_submission(engine, node_to_request(node), prio); prio 46 drivers/gpu/drm/i915/i915_scheduler.h i915_sched_lookup_priolist(struct intel_engine_cs *engine, int prio); prio 55 drivers/gpu/drm/i915/i915_scheduler.h static inline bool i915_scheduler_need_preempt(int prio, int active) prio 70 drivers/gpu/drm/i915/i915_scheduler.h return prio > max(I915_PRIORITY_NORMAL - 1, active); prio 747 drivers/gpu/drm/i915/i915_trace.h __field(u32, prio) prio 757 drivers/gpu/drm/i915/i915_trace.h __entry->prio = rq->sched.attr.priority; prio 764 drivers/gpu/drm/i915/i915_trace.h __entry->prio, __entry->port) prio 891 drivers/gpu/drm/msm/msm_drv.c ret = msm_wait_fence(gpu->rb[queue->prio]->fctx, args->fence, &timeout, prio 945 drivers/gpu/drm/msm/msm_drv.c return msm_submitqueue_create(dev, file->driver_priv, args->prio, prio 418 drivers/gpu/drm/msm/msm_drv.h u32 prio, u32 flags, u32 *id); prio 49 drivers/gpu/drm/msm/msm_gem_submit.c submit->ring = gpu->rb[queue->prio]; prio 438 drivers/gpu/drm/msm/msm_gem_submit.c ring = gpu->rb[queue->prio]; prio 174 drivers/gpu/drm/msm/msm_gpu.h u32 prio; prio 57 drivers/gpu/drm/msm/msm_submitqueue.c u32 prio, u32 flags, u32 *id) prio 74 drivers/gpu/drm/msm/msm_submitqueue.c if (prio >= priv->gpu->nr_rings) prio 77 drivers/gpu/drm/msm/msm_submitqueue.c queue->prio = prio; prio 320 drivers/gpu/drm/nouveau/nvkm/falcon/msgqueue.c nvkm_msgqueue_post(struct nvkm_msgqueue *priv, enum msgqueue_msg_priority prio, prio 332 drivers/gpu/drm/nouveau/nvkm/falcon/msgqueue.c queue = priv->func->cmd_queue(priv, prio); prio 339 drivers/gpu/drm/rcar-du/rcar_du_crtc.c unsigned int prio = 0; prio 359 drivers/gpu/drm/rcar-du/rcar_du_crtc.c prio += plane_format(plane)->planes * 4; prio 367 drivers/gpu/drm/rcar-du/rcar_du_crtc.c prio -= 4; prio 368 drivers/gpu/drm/rcar-du/rcar_du_crtc.c dspr |= (index + 1) << prio; prio 374 drivers/gpu/drm/rcar-du/rcar_du_crtc.c prio -= 4; prio 375 drivers/gpu/drm/rcar-du/rcar_du_crtc.c dspr |= (index + 1) << prio; prio 91 drivers/gpu/drm/vmwgfx/vmwgfx_context.c .prio = 3, prio 105 drivers/gpu/drm/vmwgfx/vmwgfx_context.c .prio = 3, prio 119 drivers/gpu/drm/vmwgfx/vmwgfx_cotable.c .prio = 3, prio 836 drivers/gpu/drm/vmwgfx/vmwgfx_drv.h static inline void vmw_bo_prio_add(struct vmw_buffer_object *vbo, int prio) prio 838 drivers/gpu/drm/vmwgfx/vmwgfx_drv.h if (vbo->res_prios[prio]++ == 0) prio 851 drivers/gpu/drm/vmwgfx/vmwgfx_drv.h static inline void vmw_bo_prio_del(struct vmw_buffer_object *vbo, int prio) prio 853 drivers/gpu/drm/vmwgfx/vmwgfx_drv.h if (--vbo->res_prios[prio] == 0) prio 46 drivers/gpu/drm/vmwgfx/vmwgfx_resource.c res->func->prio; prio 81 drivers/gpu/drm/vmwgfx/vmwgfx_resource_priv.h u32 prio; prio 98 drivers/gpu/drm/vmwgfx/vmwgfx_shader.c .prio = 3, prio 112 drivers/gpu/drm/vmwgfx/vmwgfx_shader.c .prio = 3, prio 115 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c .prio = 1, prio 129 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c .prio = 1, prio 2833 drivers/infiniband/core/cma.c int prio; prio 2836 drivers/infiniband/core/cma.c prio = rt_tos2priority(tos); prio 2839 drivers/infiniband/core/cma.c return netdev_get_prio_tc_map(dev, prio); prio 2843 drivers/infiniband/core/cma.c return (vlan_dev_get_egress_qos_mask(ndev, prio) & prio 1196 drivers/infiniband/hw/bnxt_re/main.c u16 prio; prio 1199 drivers/infiniband/hw/bnxt_re/main.c for (prio = 0, id = 0; prio < 8; prio++) { prio 1200 drivers/infiniband/hw/bnxt_re/main.c if (prio_map & (1 << prio)) { prio 1201 drivers/infiniband/hw/bnxt_re/main.c cosq[id] = cid_map[prio]; prio 1241 drivers/infiniband/hw/bnxt_re/main.c if ((prio_map == 0 && rdev->qplib_res.prio) || prio 1242 drivers/infiniband/hw/bnxt_re/main.c (prio_map != 0 && !rdev->qplib_res.prio)) { prio 1243 drivers/infiniband/hw/bnxt_re/main.c rdev->qplib_res.prio = prio_map ? true : false; prio 203 drivers/infiniband/hw/bnxt_re/qplib_res.h bool prio; prio 335 drivers/infiniband/hw/bnxt_re/qplib_sp.c if ((vlan_id != 0xFFFF) || res->prio) { prio 390 drivers/infiniband/hw/bnxt_re/qplib_sp.c if (res->prio) { prio 1563 drivers/infiniband/hw/mlx4/main.c ctrl->prio = cpu_to_be16(__mlx4_domain[domain] | prio 3170 drivers/infiniband/hw/mlx5/main.c struct mlx5_ib_flow_prio *prio, bool ft_added) prio 3172 drivers/infiniband/hw/mlx5/main.c prio->refcount -= !!ft_added; prio 3173 drivers/infiniband/hw/mlx5/main.c if (!prio->refcount) { prio 3174 drivers/infiniband/hw/mlx5/main.c mlx5_destroy_flow_table(prio->flow_table); prio 3175 drivers/infiniband/hw/mlx5/main.c prio->flow_table = NULL; prio 3202 drivers/infiniband/hw/mlx5/main.c put_flow_table(dev, iter->prio, true); prio 3208 drivers/infiniband/hw/mlx5/main.c put_flow_table(dev, handler->prio, true); prio 3238 drivers/infiniband/hw/mlx5/main.c struct mlx5_ib_flow_prio *prio, prio 3252 drivers/infiniband/hw/mlx5/main.c prio->flow_table = ft; prio 3253 drivers/infiniband/hw/mlx5/main.c prio->refcount = 0; prio 3254 drivers/infiniband/hw/mlx5/main.c return prio; prio 3263 drivers/infiniband/hw/mlx5/main.c struct mlx5_ib_flow_prio *prio; prio 3287 drivers/infiniband/hw/mlx5/main.c prio = &dev->flow_db->prios[priority]; prio 3300 drivers/infiniband/hw/mlx5/main.c prio = &dev->flow_db->egress_prios[priority]; prio 3315 drivers/infiniband/hw/mlx5/main.c prio = &dev->flow_db->prios[MLX5_IB_FLOW_LEFTOVERS_PRIO]; prio 3325 drivers/infiniband/hw/mlx5/main.c prio = &dev->flow_db->sniffer[ft_type]; prio 3336 drivers/infiniband/hw/mlx5/main.c ft = prio->flow_table; prio 3338 drivers/infiniband/hw/mlx5/main.c return _get_prio(ns, prio, priority, max_table_size, num_groups, prio 3341 drivers/infiniband/hw/mlx5/main.c return prio; prio 3634 drivers/infiniband/hw/mlx5/main.c handler->prio = ft_prio; prio 3934 drivers/infiniband/hw/mlx5/main.c struct mlx5_ib_flow_prio *prio = NULL; prio 3984 drivers/infiniband/hw/mlx5/main.c prio = &dev->flow_db->prios[priority]; prio 3986 drivers/infiniband/hw/mlx5/main.c prio = &dev->flow_db->egress_prios[priority]; prio 3988 drivers/infiniband/hw/mlx5/main.c prio = &dev->flow_db->fdb; prio 3990 drivers/infiniband/hw/mlx5/main.c prio = &dev->flow_db->rdma_rx[priority]; prio 3992 drivers/infiniband/hw/mlx5/main.c if (!prio) prio 3995 drivers/infiniband/hw/mlx5/main.c if (prio->flow_table) prio 3996 drivers/infiniband/hw/mlx5/main.c return prio; prio 3998 drivers/infiniband/hw/mlx5/main.c return _get_prio(ns, prio, priority, max_table_size, prio 4041 drivers/infiniband/hw/mlx5/main.c handler->prio = ft_prio; prio 179 drivers/infiniband/hw/mlx5/mlx5_ib.h struct mlx5_ib_flow_prio *prio; prio 2867 drivers/infiniband/hw/mlx5/qp.c MLX5_SET(modify_tis_in, in, bitmask.prio, 1); prio 2871 drivers/infiniband/hw/mlx5/qp.c MLX5_SET(tisc, tisc, prio, ((sl & 0x7) << 1)); prio 580 drivers/infiniband/hw/ocrdma/ocrdma.h static inline u8 ocrdma_get_pfc_prio(u8 *pfc, u8 prio) prio 582 drivers/infiniband/hw/ocrdma/ocrdma.h return *(pfc + prio); prio 585 drivers/infiniband/hw/ocrdma/ocrdma.h static inline u8 ocrdma_get_app_prio(u8 *app_prio, u8 prio) prio 587 drivers/infiniband/hw/ocrdma/ocrdma.h return *(app_prio + prio); prio 322 drivers/infiniband/sw/rxe/rxe_net.c __u8 proto, __u8 prio, __u8 ttl) prio 334 drivers/infiniband/sw/rxe/rxe_net.c ip6_flow_hdr(ip6h, prio, htonl(0)); prio 82 drivers/irqchip/irq-davinci-aintc.c unsigned int irq_off, reg_off, prio, shift; prio 127 drivers/irqchip/irq-davinci-aintc.c for (shift = 0, prio = 0; shift < 32; shift += 4, prios++) prio 128 drivers/irqchip/irq-davinci-aintc.c prio |= (*prios & 0x07) << shift; prio 129 drivers/irqchip/irq-davinci-aintc.c davinci_aintc_writel(prio, reg_off); prio 415 drivers/irqchip/irq-gic-v3.c static void gic_irq_set_prio(struct irq_data *d, u8 prio) prio 422 drivers/irqchip/irq-gic-v3.c writeb_relaxed(prio, base + offset + index); prio 109 drivers/md/bcache/alloc.c if (b->prio && prio 110 drivers/md/bcache/alloc.c b->prio != BTREE_PRIO && prio 112 drivers/md/bcache/alloc.c b->prio--; prio 113 drivers/md/bcache/alloc.c c->min_prio = min(c->min_prio, b->prio); prio 150 drivers/md/bcache/alloc.c b->prio = INITIAL_PRIO; prio 174 drivers/md/bcache/alloc.c (b->prio - ca->set->min_prio + min_prio) * GC_SECTORS_USED(b); \ prio 455 drivers/md/bcache/alloc.c b->prio = BTREE_PRIO; prio 459 drivers/md/bcache/alloc.c b->prio = INITIAL_PRIO; prio 199 drivers/md/bcache/bcache.h uint16_t prio; prio 417 drivers/md/bcache/bcache.h struct closure prio; prio 1302 drivers/md/bcache/btree.c b->prio = BTREE_PRIO; prio 1303 drivers/md/bcache/btree.c else if (!level && b->prio == BTREE_PRIO) prio 1304 drivers/md/bcache/btree.c b->prio = INITIAL_PRIO; prio 2371 drivers/md/bcache/btree.c BUG_ON(PTR_BUCKET(b->c, &b->key, i)->prio != BTREE_PRIO); prio 141 drivers/md/bcache/extents.c PTR_BUCKET(b->c, k, j)->prio); prio 185 drivers/md/bcache/extents.c g->prio != BTREE_PRIO || prio 201 drivers/md/bcache/extents.c g->prio, g->gen, g->last_gc, GC_MARK(g)); prio 520 drivers/md/bcache/extents.c if (g->prio == BTREE_PRIO) prio 533 drivers/md/bcache/extents.c g->prio, g->gen, g->last_gc, GC_MARK(g)); prio 556 drivers/md/bcache/request.c PTR_BUCKET(b->c, k, ptr)->prio = INITIAL_PRIO; prio 508 drivers/md/bcache/super.c closure_put(&ca->prio); prio 514 drivers/md/bcache/super.c struct closure *cl = &ca->prio; prio 528 drivers/md/bcache/super.c closure_bio_submit(ca->set, bio, &ca->prio); prio 573 drivers/md/bcache/super.c d->prio = cpu_to_le16(b->prio); prio 641 drivers/md/bcache/super.c b->prio = le16_to_cpu(d->prio); prio 1043 drivers/md/bcache/sysfs.c p[i] = ca->buckets[i].prio; prio 617 drivers/media/pci/ngene/ngene.h enum v4l2_priority prio; prio 712 drivers/media/pci/ngene/ngene.h struct v4l2_prio_state prio; prio 1727 drivers/media/usb/uvc/uvc_driver.c v4l2_prio_init(&chain->prio); prio 1990 drivers/media/usb/uvc/uvc_driver.c vdev->prio = &stream->chain->prio; prio 446 drivers/media/usb/uvc/uvcvideo.h struct v4l2_prio_state prio; /* V4L2 priority state */ prio 243 drivers/media/v4l2-core/v4l2-dev.c static inline bool prio_is_valid(enum v4l2_priority prio) prio 245 drivers/media/v4l2-core/v4l2-dev.c return prio == V4L2_PRIORITY_BACKGROUND || prio 246 drivers/media/v4l2-core/v4l2-dev.c prio == V4L2_PRIORITY_INTERACTIVE || prio 247 drivers/media/v4l2-core/v4l2-dev.c prio == V4L2_PRIORITY_RECORD; prio 905 drivers/media/v4l2-core/v4l2-dev.c if (vdev->prio == NULL) prio 906 drivers/media/v4l2-core/v4l2-dev.c vdev->prio = &vdev->v4l2_dev->prio; prio 24 drivers/media/v4l2-core/v4l2-device.c v4l2_prio_init(&v4l2_dev->prio); prio 35 drivers/media/v4l2-core/v4l2-fh.c fh->prio = V4L2_PRIORITY_UNSET; prio 48 drivers/media/v4l2-core/v4l2-fh.c v4l2_prio_open(fh->vdev->prio, &fh->prio); prio 76 drivers/media/v4l2-core/v4l2-fh.c v4l2_prio_close(fh->vdev->prio, fh->prio); prio 1103 drivers/media/v4l2-core/v4l2-ioctl.c *p = v4l2_prio_max(vfd->prio); prio 1118 drivers/media/v4l2-core/v4l2-ioctl.c return v4l2_prio_change(vfd->prio, &vfh->prio, *p); prio 2900 drivers/media/v4l2-core/v4l2-ioctl.c ret = v4l2_prio_check(vfd->prio, vfh->prio); prio 2918 drivers/media/v4l2-core/v4l2-ioctl.c vfh ? v4l2_prio_check(vfd->prio, vfh->prio) >= 0 : 0, prio 1014 drivers/net/caif/caif_hsi.c int prio; prio 1025 drivers/net/caif/caif_hsi.c prio = CFHSI_PRIO_BEBK; prio 1028 drivers/net/caif/caif_hsi.c prio = CFHSI_PRIO_VI; prio 1031 drivers/net/caif/caif_hsi.c prio = CFHSI_PRIO_VO; prio 1035 drivers/net/caif/caif_hsi.c prio = CFHSI_PRIO_CTL; prio 1045 drivers/net/caif/caif_hsi.c skb_queue_tail(&cfhsi->qhead[prio], skb); prio 290 drivers/net/can/at91_can.c unsigned int mb, enum at91_mb_mode mode, int prio) prio 292 drivers/net/can/at91_can.c at91_write(priv, AT91_MMR(mb), (mode << 24) | (prio << 16)); prio 454 drivers/net/can/at91_can.c unsigned int mb, prio; prio 461 drivers/net/can/at91_can.c prio = get_tx_next_prio(priv); prio 476 drivers/net/can/at91_can.c set_mb_mode_prio(priv, mb, AT91_MB_MODE_TX, prio); prio 465 drivers/net/ethernet/amd/xgbe/xgbe-dev.c unsigned int prio, tc; prio 467 drivers/net/ethernet/amd/xgbe/xgbe-dev.c for (prio = 0; prio < IEEE_8021QAZ_MAX_TCS; prio++) { prio 469 drivers/net/ethernet/amd/xgbe/xgbe-dev.c if (pdata->prio2q_map[prio] != queue) prio 473 drivers/net/ethernet/amd/xgbe/xgbe-dev.c tc = pdata->ets->prio_tc[prio]; prio 2604 drivers/net/ethernet/amd/xgbe/xgbe-dev.c unsigned int ppq, ppq_extra, prio; prio 2639 drivers/net/ethernet/amd/xgbe/xgbe-dev.c for (i = 0, prio = 0; i < prio_queues;) { prio 2643 drivers/net/ethernet/amd/xgbe/xgbe-dev.c "PRIO%u mapped to RXq%u\n", prio, i); prio 2644 drivers/net/ethernet/amd/xgbe/xgbe-dev.c mask |= (1 << prio); prio 2645 drivers/net/ethernet/amd/xgbe/xgbe-dev.c pdata->prio2q_map[prio++] = i; prio 2650 drivers/net/ethernet/amd/xgbe/xgbe-dev.c "PRIO%u mapped to RXq%u\n", prio, i); prio 2651 drivers/net/ethernet/amd/xgbe/xgbe-dev.c mask |= (1 << prio); prio 2652 drivers/net/ethernet/amd/xgbe/xgbe-dev.c pdata->prio2q_map[prio++] = i; prio 2683 drivers/net/ethernet/amd/xgbe/xgbe-dev.c unsigned int offset, queue, prio; prio 2706 drivers/net/ethernet/amd/xgbe/xgbe-dev.c for (prio = 0; prio < IEEE_8021QAZ_MAX_TCS; prio++) prio 2707 drivers/net/ethernet/amd/xgbe/xgbe-dev.c netdev_set_prio_tc_map(pdata->netdev, prio, prio 2708 drivers/net/ethernet/amd/xgbe/xgbe-dev.c pdata->ets->prio_tc[prio]); prio 2716 drivers/net/ethernet/amd/xgbe/xgbe-dev.c unsigned int i, prio; prio 2735 drivers/net/ethernet/amd/xgbe/xgbe-dev.c for (prio = 0; prio < IEEE_8021QAZ_MAX_TCS; prio++) { prio 2736 drivers/net/ethernet/amd/xgbe/xgbe-dev.c if (ets->prio_tc[prio] == i) prio 2737 drivers/net/ethernet/amd/xgbe/xgbe-dev.c mask |= (1 << prio); prio 4240 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c int cos, prio, count, offset; prio 4267 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c for (prio = 0; prio < BNX2X_MAX_PRIORITY; prio++) { prio 4268 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c int outer_prio = c2s_map[prio]; prio 4270 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c netdev_set_prio_tc_map(dev, prio, bp->prio_to_cos[outer_prio]); prio 728 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c u8 prio, cos; prio 730 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c for (prio = 0; prio < BNX2X_MAX_PRIORITY; prio++) { prio 732 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c & (1 << prio)) { prio 733 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c bp->prio_to_cos[prio] = cos; prio 735 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c "tx_mapping %d --> %d\n", prio, cos); prio 1979 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c static void bnx2x_dcbnl_set_pg_tccfg_tx(struct net_device *netdev, int prio, prio 1985 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c DP(BNX2X_MSG_DCB, "prio[%d] = %d\n", prio, pgid); prio 1986 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c if (!bnx2x_dcbnl_set_valid(bp) || prio >= DCBX_MAX_NUM_PRI_PG_ENTRIES) prio 2002 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c bp->dcbx_config_params.admin_configuration_ets_pg[prio] = pgid; prio 2019 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c static void bnx2x_dcbnl_set_pg_tccfg_rx(struct net_device *netdev, int prio, prio 2034 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c static void bnx2x_dcbnl_get_pg_tccfg_tx(struct net_device *netdev, int prio, prio 2039 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c DP(BNX2X_MSG_DCB, "prio = %d\n", prio); prio 2055 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c if (!bp->dcb_state || prio >= DCBX_MAX_NUM_PRI_PG_ENTRIES) prio 2058 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c *pgid = DCBX_PRI_PG_GET(bp->dcbx_local_feat.ets.pri_pg_tbl, prio); prio 2075 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c static void bnx2x_dcbnl_get_pg_tccfg_rx(struct net_device *netdev, int prio, prio 2094 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c static void bnx2x_dcbnl_set_pfc_cfg(struct net_device *netdev, int prio, prio 2098 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c DP(BNX2X_MSG_DCB, "prio[%d] = %d\n", prio, setting); prio 2100 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c if (!bnx2x_dcbnl_set_valid(bp) || prio >= MAX_PFC_PRIORITIES) prio 2104 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c bp->dcbx_config_params.admin_pfc_bitmap |= (1 << prio); prio 2107 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c bp->dcbx_config_params.admin_pfc_bitmap &= ~(1 << prio); prio 2111 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c static void bnx2x_dcbnl_get_pfc_cfg(struct net_device *netdev, int prio, prio 2115 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c DP(BNX2X_MSG_DCB, "prio = %d\n", prio); prio 2119 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c if (!bp->dcb_state || prio >= MAX_PFC_PRIORITIES) prio 2122 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c *setting = (bp->dcbx_local_feat.pfc.pri_en_bitmap >> prio) & 0x1; prio 28 drivers/net/ethernet/broadcom/bnxt/bnxt_tc.c #define VLAN_TCI(vid, prio) ((vid) | ((prio) << VLAN_PRIO_SHIFT)) prio 1177 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h uint32_t prio:1; /* filter has priority over active/server */ prio 1028 drivers/net/ethernet/chelsio/cxgb4/cxgb4_dcb.c int prio; prio 1036 drivers/net/ethernet/chelsio/cxgb4/cxgb4_dcb.c prio = __cxgb4_getapp(dev, app->selector - 1, app->protocol, 0); prio 1038 drivers/net/ethernet/chelsio/cxgb4/cxgb4_dcb.c if (prio < 0) prio 1039 drivers/net/ethernet/chelsio/cxgb4/cxgb4_dcb.c prio = dcb_ieee_getapp_mask(dev, app); prio 1041 drivers/net/ethernet/chelsio/cxgb4/cxgb4_dcb.c app->priority = ffs(prio) - 1; prio 2424 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c int prio, pgid; prio 2427 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c for (prio = 0; prio < 8; prio++) { prio 2428 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c pgid = (dcb->pgid >> 4 * (7 - prio)) & 0xf; prio 2456 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c int prio; prio 2459 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c for (prio = 0; prio < 8; prio++) prio 2460 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c seq_printf(seq, " %3d", dcb->priorate[prio]); prio 2465 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c int prio; prio 2468 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c for (prio = 0; prio < 8; prio++) { prio 2469 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c int pfcen = (dcb->pfcen >> 1 * (7 - prio)) prio 617 drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c FW_FILTER_WR_PRIO_V(f->fs.prio) | prio 408 drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_flower.c u8 prio = act->vlan.prio; prio 410 drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_flower.c u16 vlan_tci = (prio << VLAN_PRIO_SHIFT) | vid; prio 229 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h static inline void set_wr_txq(struct sk_buff *skb, int prio, int queue) prio 231 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h skb_set_queue_mapping(skb, (queue << 1) | prio); prio 760 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c u8 prio = 0; prio 820 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c prio = netdev_txq_to_tc(net_dev, queue_mapping); prio 824 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c prio = net_dev->num_tc - prio - 1; prio 840 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c err = priv->enqueue(priv, fq, &fd, prio); prio 2501 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c struct dpaa2_fd *fd, u8 prio) prio 2504 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c priv->tx_qdid, prio, prio 2510 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c struct dpaa2_fd *fd, u8 prio) prio 2513 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c fq->tx_fqid[prio], fd); prio 365 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.h struct dpaa2_fd *fd, u8 prio); prio 549 drivers/net/ethernet/freescale/enetc/enetc_hw.h int prio) prio 554 drivers/net/ethernet/freescale/enetc/enetc_hw.h val |= ENETC_TBMR_SET_PRIO(prio); prio 1101 drivers/net/ethernet/freescale/gianfar_ethtool.c u32 prio = 0, prio_mask = 0; prio 1118 drivers/net/ethernet/freescale/gianfar_ethtool.c prio = vlan_tci_prio(rule); prio 1169 drivers/net/ethernet/freescale/gianfar_ethtool.c gfar_set_attribute(prio, prio_mask, RQFCR_PID_PRI, tab); prio 152 drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c u8 prio, tc_map; prio 167 drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c prio = dcbxcfg->app[i].priority; prio 168 drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c tc_map = BIT(dcbxcfg->etscfg.prioritytable[prio]); prio 174 drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c sapp.priority = prio; prio 1620 drivers/net/ethernet/intel/igb/igb_main.c enum tx_queue_prio prio) prio 1629 drivers/net/ethernet/intel/igb/igb_main.c if (prio == TX_QUEUE_PRIO_HIGH) prio 168 drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c u8 prio, u8 bwg_id, u8 bw_pct, prio 173 drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c if (prio != DCB_ATTR_VALUE_UNDEFINED) prio 174 drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c adapter->temp_dcb_cfg.tc_config[tc].path[0].prio_type = prio; prio 194 drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c u8 prio, u8 bwg_id, u8 bw_pct, prio 199 drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c if (prio != DCB_ATTR_VALUE_UNDEFINED) prio 200 drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c adapter->temp_dcb_cfg.tc_config[tc].path[1].prio_type = prio; prio 220 drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c u8 *prio, u8 *bwg_id, u8 *bw_pct, prio 225 drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c *prio = adapter->dcb_cfg.tc_config[tc].path[0].prio_type; prio 240 drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c u8 *prio, u8 *bwg_id, u8 *bw_pct, prio 245 drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c *prio = adapter->dcb_cfg.tc_config[tc].path[1].prio_type; prio 9025 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c u8 prio; prio 9027 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c for (prio = 0; prio < MAX_USER_PRIORITY; prio++) { prio 9031 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c tc = ixgbe_dcb_get_tc_from_up(dcb_cfg, 0, prio); prio 9033 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c tc = ets->prio_tc[prio]; prio 9035 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c netdev_set_prio_tc_map(dev, prio, tc); prio 439 drivers/net/ethernet/marvell/mvpp2/mvpp2_cls.c static void mvpp2_cls_flow_pri_set(struct mvpp2_cls_flow_entry *fe, int prio) prio 442 drivers/net/ethernet/marvell/mvpp2/mvpp2_cls.c fe->data[1] |= MVPP2_CLS_FLOW_TBL1_PRIO(prio); prio 3318 drivers/net/ethernet/mellanox/mlx4/en_netdev.c u8 prio; prio 3320 drivers/net/ethernet/mellanox/mlx4/en_netdev.c for (prio = 0; prio < IEEE_8021QAZ_MAX_TCS; ++prio) { prio 3321 drivers/net/ethernet/mellanox/mlx4/en_netdev.c priv->ets.prio_tc[prio] = prio; prio 3322 drivers/net/ethernet/mellanox/mlx4/en_netdev.c priv->ets.tc_tsa[prio] = IEEE_8021QAZ_TSA_VENDOR; prio 836 drivers/net/ethernet/mellanox/mlx4/main.c ctrl->prio = cpu_to_be16(MLX4_DOMAIN_NIC); prio 782 drivers/net/ethernet/mellanox/mlx4/mcg.c hw->prio = cpu_to_be16(ctrl->priority); prio 1050 drivers/net/ethernet/mellanox/mlx4/mcg.c int port, int qpn, u16 prio, u64 *reg_id) prio 1068 drivers/net/ethernet/mellanox/mlx4/mcg.c rule.priority = prio; prio 279 drivers/net/ethernet/mellanox/mlx5/core/en/port.c int prio; prio 295 drivers/net/ethernet/mellanox/mlx5/core/en/port.c for (prio = 0; prio < 8; prio++) { prio 296 drivers/net/ethernet/mellanox/mlx5/core/en/port.c buffer[prio] = (u8)(prio_x_buff >> (4 * prio)) & 0xF; prio 297 drivers/net/ethernet/mellanox/mlx5/core/en/port.c mlx5_core_dbg(mdev, "prio %d, buffer %d\n", prio, buffer[prio]); prio 311 drivers/net/ethernet/mellanox/mlx5/core/en/port.c int prio; prio 334 drivers/net/ethernet/mellanox/mlx5/core/en/port.c for (prio = 0; prio < 8; prio++) prio 335 drivers/net/ethernet/mellanox/mlx5/core/en/port.c prio_x_buff |= (buffer[prio] << (4 * prio)); prio 200 drivers/net/ethernet/mellanox/mlx5/core/en/port_buffer.c int prio; prio 208 drivers/net/ethernet/mellanox/mlx5/core/en/port_buffer.c for (prio = 0; prio < MLX5E_MAX_PRIORITY; prio++) { prio 209 drivers/net/ethernet/mellanox/mlx5/core/en/port_buffer.c if (buffer[prio] != i) prio 213 drivers/net/ethernet/mellanox/mlx5/core/en/port_buffer.c lossy_count += !(pfc_en & (1 << prio)); prio 321 drivers/net/ethernet/mellanox/mlx5/core/en_arfs.c ft_attr.prio = MLX5E_NIC_PRIO; prio 57 drivers/net/ethernet/mellanox/mlx5/core/en_dcbnl.c static int mlx5e_set_dscp2prio(struct mlx5e_priv *priv, u8 dscp, u8 prio); prio 1148 drivers/net/ethernet/mellanox/mlx5/core/en_dcbnl.c static int mlx5e_set_dscp2prio(struct mlx5e_priv *priv, u8 dscp, u8 prio) prio 1152 drivers/net/ethernet/mellanox/mlx5/core/en_dcbnl.c err = mlx5_set_dscp2prio(priv->mdev, dscp, prio); prio 1156 drivers/net/ethernet/mellanox/mlx5/core/en_dcbnl.c priv->dcbx_dp.dscp2prio[dscp] = prio; prio 1124 drivers/net/ethernet/mellanox/mlx5/core/en_fs.c ft_attr->prio = MLX5E_NIC_PRIO; prio 1134 drivers/net/ethernet/mellanox/mlx5/core/en_fs.c ft_attr->prio = MLX5E_NIC_PRIO; prio 1363 drivers/net/ethernet/mellanox/mlx5/core/en_fs.c ft_attr.prio = MLX5E_NIC_PRIO; prio 1484 drivers/net/ethernet/mellanox/mlx5/core/en_fs.c ft_attr.prio = MLX5E_NIC_PRIO; prio 66 drivers/net/ethernet/mellanox/mlx5/core/en_fs_ethtool.c int prio; prio 74 drivers/net/ethernet/mellanox/mlx5/core/en_fs_ethtool.c prio = MLX5E_ETHTOOL_L3_L4_PRIO + (max_tuples - num_tuples); prio 75 drivers/net/ethernet/mellanox/mlx5/core/en_fs_ethtool.c eth_ft = &priv->fs.ethtool.l3_l4_ft[prio]; prio 80 drivers/net/ethernet/mellanox/mlx5/core/en_fs_ethtool.c prio = MLX5E_ETHTOOL_L3_L4_PRIO + (max_tuples - num_tuples); prio 81 drivers/net/ethernet/mellanox/mlx5/core/en_fs_ethtool.c eth_ft = &priv->fs.ethtool.l3_l4_ft[prio]; prio 85 drivers/net/ethernet/mellanox/mlx5/core/en_fs_ethtool.c prio = max_tuples - num_tuples; prio 86 drivers/net/ethernet/mellanox/mlx5/core/en_fs_ethtool.c eth_ft = &priv->fs.ethtool.l2_ft[prio]; prio 87 drivers/net/ethernet/mellanox/mlx5/core/en_fs_ethtool.c prio += MLX5E_ETHTOOL_L2_PRIO; prio 105 drivers/net/ethernet/mellanox/mlx5/core/en_fs_ethtool.c ft = mlx5_create_auto_grouped_flow_table(ns, prio, prio 3249 drivers/net/ethernet/mellanox/mlx5/core/en_main.c MLX5_SET(tisc, tisc, prio, tc << 1); prio 1025 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c int i, prio; prio 1030 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c for (prio = 0; prio < NUM_PPORT_PRIO; prio++) { prio 1033 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c pport_per_tc_prio_stats_desc[i].format, prio); prio 1036 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c pport_per_tc_congest_prio_stats_desc[i].format, prio); prio 1047 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c int i, prio; prio 1052 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c for (prio = 0; prio < NUM_PPORT_PRIO; prio++) { prio 1055 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c MLX5E_READ_CTR64_BE(&pport->per_tc_prio_counters[prio], prio 1059 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c MLX5E_READ_CTR64_BE(&pport->per_tc_congest_prio_counters[prio], prio 1073 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c int prio; prio 1080 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c for (prio = 0; prio < NUM_PPORT_PRIO; prio++) { prio 1081 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c out = pstats->per_tc_prio_counters[prio]; prio 1082 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c MLX5_SET(ppcnt_reg, in, prio_tc, prio); prio 1104 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c int prio; prio 1111 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c for (prio = 0; prio < NUM_PPORT_PRIO; prio++) { prio 1112 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c out = pstats->per_tc_congest_prio_counters[prio]; prio 1113 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c MLX5_SET(ppcnt_reg, in, prio_tc, prio); prio 1151 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c int i, prio; prio 1153 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c for (prio = 0; prio < NUM_PPORT_PRIO; prio++) { prio 1156 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c pport_per_prio_traffic_stats_desc[i].format, prio); prio 1166 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c int i, prio; prio 1168 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c for (prio = 0; prio < NUM_PPORT_PRIO; prio++) { prio 1171 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c MLX5E_READ_CTR64_BE(&priv->stats.pport.per_prio_counters[prio], prio 1240 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c int i, prio; prio 1243 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c for_each_set_bit(prio, &pfc_combined, NUM_PPORT_PRIO) { prio 1247 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c snprintf(pfc_string, sizeof(pfc_string), "prio%d", prio); prio 1272 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c int i, prio; prio 1275 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c for_each_set_bit(prio, &pfc_combined, NUM_PPORT_PRIO) { prio 1278 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c MLX5E_READ_CTR64_BE(&priv->stats.pport.per_prio_counters[prio], prio 1326 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c int prio; prio 1334 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c for (prio = 0; prio < NUM_PPORT_PRIO; prio++) { prio 1335 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c out = pstats->per_prio_counters[prio]; prio 1336 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c MLX5_SET(ppcnt_reg, in, prio_tc, prio); prio 195 drivers/net/ethernet/mellanox/mlx5/core/en_stats.h #define PPORT_PER_PRIO_GET(pstats, prio, c) \ prio 196 drivers/net/ethernet/mellanox/mlx5/core/en_stats.h MLX5_GET64(ppcnt_reg, pstats->per_prio_counters[prio], \ prio 182 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c u8 prio; prio 591 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c ft_attr->prio = MLX5E_TC_PRIO; prio 691 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c static inline u32 hash_hairpin_info(u16 peer_vhca_id, u8 prio) prio 693 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c return (peer_vhca_id << 16 | prio); prio 697 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c u16 peer_vhca_id, u8 prio) prio 700 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c u32 hash_key = hash_hairpin_info(peer_vhca_id, prio); prio 704 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c if (hpe->peer_vhca_id == peer_vhca_id && hpe->prio == prio) { prio 822 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c hpe->prio = match_prio; prio 1166 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c if (!mlx5_eswitch_prios_supported(esw) && attr->prio != 1) { prio 1176 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c if (attr->prio > max_prio) { prio 2775 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c if (act->vlan.prio != (match_prio_val & match_prio_mask)) { prio 2796 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c .vlan.prio = prio 3125 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c attr->vlan_prio[vlan_idx] = act->vlan.prio; prio 3139 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c act->vlan.prio)) prio 3164 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c .vlan.prio = 0, prio 3574 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c esw_attr->prio = f->common.prio; prio 4019 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c if (ma->common.prio != 1) { prio 342 drivers/net/ethernet/mellanox/mlx5/core/eswitch.c ft_attr.prio = LEGACY_FDB_PRIO; prio 398 drivers/net/ethernet/mellanox/mlx5/core/eswitch.h u16 prio; prio 51 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c #define fdb_prio_table(esw, chain, prio, level) \ prio 52 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c (esw)->fdb_table.offloads.fdb_prio[(chain)][(prio)][(level)] prio 66 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c esw_get_prio_table(struct mlx5_eswitch *esw, u32 chain, u16 prio, int level); prio 68 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c esw_put_prio_table(struct mlx5_eswitch *esw, u32 chain, u16 prio, int level); prio 161 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c flow_act.vlan[0].prio = attr->vlan_prio[0]; prio 165 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c flow_act.vlan[1].prio = attr->vlan_prio[1]; prio 218 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c fdb = esw_get_prio_table(esw, attr->chain, attr->prio, !!split); prio 237 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c esw_put_prio_table(esw, attr->chain, attr->prio, !!split); prio 257 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c fast_fdb = esw_get_prio_table(esw, attr->chain, attr->prio, 0); prio 263 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c fwd_fdb = esw_get_prio_table(esw, attr->chain, attr->prio, 1); prio 300 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c esw_put_prio_table(esw, attr->chain, attr->prio, 1); prio 302 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c esw_put_prio_table(esw, attr->chain, attr->prio, 0); prio 327 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c esw_put_prio_table(esw, attr->chain, attr->prio, 1); prio 328 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c esw_put_prio_table(esw, attr->chain, attr->prio, 0); prio 330 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c esw_put_prio_table(esw, attr->chain, attr->prio, !!split); prio 922 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c esw_get_prio_table(struct mlx5_eswitch *esw, u32 chain, u16 prio, int level) prio 935 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c fdb = fdb_prio_table(esw, chain, prio, level).fdb; prio 939 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c fdb_prio_table(esw, chain, prio, level--).num_rules++; prio 955 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c table_prio = (chain * FDB_MAX_PRIO) + prio - 1; prio 961 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c if (fdb_prio_table(esw, chain, prio, l).fdb) { prio 962 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c fdb_prio_table(esw, chain, prio, l).num_rules++; prio 972 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c fdb_prio_table(esw, chain, prio, l).fdb = fdb; prio 973 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c fdb_prio_table(esw, chain, prio, l).num_rules = 1; prio 982 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c esw_put_prio_table(esw, chain, prio, l); prio 988 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c esw_put_prio_table(struct mlx5_eswitch *esw, u32 chain, u16 prio, int level) prio 998 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c if (--(fdb_prio_table(esw, chain, prio, l).num_rules) > 0) prio 1001 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c put_sz_to_pool(esw, fdb_prio_table(esw, chain, prio, l).fdb->max_fte); prio 1002 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c mlx5_destroy_flow_table(fdb_prio_table(esw, chain, prio, l).fdb); prio 1003 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c fdb_prio_table(esw, chain, prio, l).fdb = NULL; prio 1103 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c ft_attr.prio = FDB_SLOW_PATH; prio 1778 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c flow_act.vlan[0].prio = 0; prio 54 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads_termtbl.c int prio, flags; prio 66 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads_termtbl.c prio = FDB_SLOW_PATH; prio 68 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads_termtbl.c tt->termtbl = mlx5_create_auto_grouped_flow_table(root_ns, prio, 1, 1, prio 471 drivers/net/ethernet/mellanox/mlx5/core/fs_cmd.c MLX5_SET(vlan, vlan, prio, fte->action.vlan[0].prio); prio 477 drivers/net/ethernet/mellanox/mlx5/core/fs_cmd.c MLX5_SET(vlan, vlan, prio, fte->action.vlan[1].prio); prio 133 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c int prio; prio 354 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c unsigned int prio) prio 359 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c if (iter_prio->prio == prio) prio 445 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c struct fs_prio *prio; prio 450 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c fs_get_obj(prio, ft->node.parent); prio 451 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c prio->num_ft--; prio 778 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static struct mlx5_flow_table *find_closest_ft(struct fs_prio *prio, bool reverse) prio 784 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c parent = prio->node.parent; prio 785 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c curr_node = &prio->node; prio 795 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static struct mlx5_flow_table *find_next_chained_ft(struct fs_prio *prio) prio 797 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c return find_closest_ft(prio, false); prio 801 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static struct mlx5_flow_table *find_prev_chained_ft(struct fs_prio *prio) prio 803 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c return find_closest_ft(prio, true); prio 807 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c struct fs_prio *prio, prio 810 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c struct mlx5_flow_root_namespace *root = find_root(&prio->node); prio 815 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c fs_for_each_ft(iter, prio) { prio 833 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c struct fs_prio *prio) prio 837 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c prev_ft = find_prev_chained_ft(prio); prio 848 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c *prio) prio 850 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c struct mlx5_flow_root_namespace *root = find_root(&prio->node); prio 965 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c struct fs_prio *prio) prio 972 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c if (list_empty(&prio->node.children)) { prio 973 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c err = connect_prev_fts(dev, ft, prio); prio 977 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c next_ft = find_next_chained_ft(prio); prio 985 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c err = update_root_ft_create(ft, prio); prio 990 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c struct fs_prio *prio) prio 992 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c struct list_head *prev = &prio->node.children; prio 995 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c fs_for_each_ft(iter, prio) { prio 1021 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c fs_prio = find_prio(ns, ft_attr->prio); prio 1080 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c int prio, int max_fte, prio 1087 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c ft_attr.prio = prio; prio 1094 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c int prio, u32 level) prio 1099 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c ft_attr.prio = prio; prio 1106 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c int prio, prio 1119 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c ft_attr.prio = prio; prio 1871 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c struct fs_prio *prio; prio 1873 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c fs_get_obj(prio, ft->node.parent); prio 1880 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c next_ft = find_next_chained_ft(prio); prio 1946 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c struct fs_prio *prio; prio 1948 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c fs_get_obj(prio, ft->node.parent); prio 1950 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c if (!list_is_last(&ft->node.list, &prio->node.children)) prio 1952 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c return find_next_chained_ft(prio); prio 2005 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c struct fs_prio *prio; prio 2012 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c fs_get_obj(prio, ft->node.parent); prio 2013 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c if (!(list_first_entry(&prio->node.children, prio 2018 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c next_ft = find_next_chained_ft(prio); prio 2023 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c err = connect_prev_fts(dev, next_ft, prio); prio 2074 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c int prio = 0; prio 2102 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c prio = RDMA_RX_BYPASS_PRIO; prio 2105 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c prio = RDMA_RX_KERNEL_PRIO; prio 2108 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c prio = type; prio 2114 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c fs_prio = find_prio(&root_ns->ns, prio); prio 2154 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c unsigned int prio, prio 2168 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c fs_prio->prio = prio; prio 2175 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c unsigned int prio, prio 2178 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c return _fs_create_prio(ns, prio, num_levels, FS_TYPE_PRIO_CHAINS); prio 2182 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c unsigned int prio, int num_levels) prio 2184 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c return _fs_create_prio(ns, prio, num_levels, FS_TYPE_PRIO); prio 2195 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static struct mlx5_flow_namespace *fs_create_namespace(struct fs_prio *prio, prio 2207 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c tree_add_node(&ns->node, &prio->node); prio 2208 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c list_add_tail(&ns->node.list, &prio->node.children); prio 2213 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static int create_leaf_prios(struct mlx5_flow_namespace *ns, int prio, prio 2220 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c fs_prio = fs_create_prio(ns, prio++, prio_metadata->num_levels); prio 2247 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c int prio) prio 2265 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c return create_leaf_prios(fs_ns, prio, init_node); prio 2266 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c fs_prio = fs_create_prio(fs_ns, prio, init_node->num_levels); prio 2279 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c prio = 0; prio 2282 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c base, init_node, prio); prio 2287 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c prio += init_node->children[i].num_leaf_prios; prio 2345 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static void set_prio_attrs_in_prio(struct fs_prio *prio, int acc_level); prio 2349 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c struct fs_prio *prio; prio 2351 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c fs_for_each_prio(prio, ns) { prio 2353 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c set_prio_attrs_in_prio(prio, acc_level); prio 2354 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c acc_level += prio->num_levels; prio 2359 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static void set_prio_attrs_in_prio(struct fs_prio *prio, int acc_level) prio 2364 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c prio->start_level = acc_level; prio 2365 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c fs_for_each_ns(ns, prio) prio 2368 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c if (!prio->num_levels) prio 2369 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c prio->num_levels = acc_level_ns - prio->start_level; prio 2370 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c WARN_ON(prio->num_levels < acc_level_ns - prio->start_level); prio 2376 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c struct fs_prio *prio; prio 2379 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c fs_for_each_prio(prio, ns) { prio 2380 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c set_prio_attrs_in_prio(prio, start_level); prio 2381 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c start_level += prio->num_levels; prio 2400 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c ft_attr.prio = ANCHOR_PRIO; prio 2510 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c struct fs_prio *prio; prio 2517 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c prio = fs_create_prio(&steering->sniffer_tx_root_ns->ns, 0, 1); prio 2518 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c return PTR_ERR_OR_ZERO(prio); prio 2523 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c struct fs_prio *prio; prio 2530 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c prio = fs_create_prio(&steering->sniffer_rx_root_ns->ns, 0, 1); prio 2531 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c return PTR_ERR_OR_ZERO(prio); prio 2563 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c int prio; prio 2598 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c for (prio = 0; prio < FDB_MAX_PRIO * (chain + 1); prio++) { prio 2599 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c min_prio = fs_create_prio(ns, prio, 2); prio 2628 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c struct fs_prio *prio; prio 2635 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c prio = fs_create_prio(&steering->esw_egress_root_ns[vport]->ns, 0, 1); prio 2636 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c return PTR_ERR_OR_ZERO(prio); prio 2641 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c struct fs_prio *prio; prio 2648 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c prio = fs_create_prio(&steering->esw_ingress_root_ns[vport]->ns, 0, 1); prio 2649 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c return PTR_ERR_OR_ZERO(prio); prio 214 drivers/net/ethernet/mellanox/mlx5/core/fs_core.h unsigned int prio; prio 283 drivers/net/ethernet/mellanox/mlx5/core/fs_core.h #define fs_for_each_ns_or_ft_reverse(pos, prio) \ prio 284 drivers/net/ethernet/mellanox/mlx5/core/fs_core.h list_for_each_entry_reverse(pos, &(prio)->node.children, list) prio 286 drivers/net/ethernet/mellanox/mlx5/core/fs_core.h #define fs_for_each_ns_or_ft(pos, prio) \ prio 287 drivers/net/ethernet/mellanox/mlx5/core/fs_core.h list_for_each_entry(pos, (&(prio)->node.children), list) prio 292 drivers/net/ethernet/mellanox/mlx5/core/fs_core.h #define fs_for_each_ns(pos, prio) \ prio 293 drivers/net/ethernet/mellanox/mlx5/core/fs_core.h fs_list_for_each_entry(pos, &(prio)->node.children) prio 295 drivers/net/ethernet/mellanox/mlx5/core/fs_core.h #define fs_for_each_ft(pos, prio) \ prio 296 drivers/net/ethernet/mellanox/mlx5/core/fs_core.h fs_list_for_each_entry(pos, &(prio)->node.children) prio 298 drivers/net/ethernet/mellanox/mlx5/core/fs_core.h #define fs_for_each_ft_safe(pos, tmp, prio) \ prio 299 drivers/net/ethernet/mellanox/mlx5/core/fs_core.h fs_list_for_each_entry_safe(pos, tmp, &(prio)->node.children) prio 584 drivers/net/ethernet/mellanox/mlx5/core/port.c MLX5_SET(qtct_reg, in, prio, i); prio 598 drivers/net/ethernet/mellanox/mlx5/core/port.c u8 prio, u8 *tc) prio 608 drivers/net/ethernet/mellanox/mlx5/core/port.c MLX5_SET(qtct_reg, in, prio, prio); prio 917 drivers/net/ethernet/mellanox/mlx5/core/port.c int mlx5_set_dscp2prio(struct mlx5_core_dev *mdev, u8 dscp, u8 prio) prio 942 drivers/net/ethernet/mellanox/mlx5/core/port.c MLX5_SET16(qpdpm_dscp_reg, qpdpm_dscp, prio, prio); prio 977 drivers/net/ethernet/mellanox/mlx5/core/port.c dscp2prio[i] = MLX5_GET16(qpdpm_dscp_reg, qpdpm_dscp, prio); prio 451 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c if (tmp_matcher->prio >= matcher->prio) { prio 651 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c matcher->prio = priority; prio 698 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_types.h u16 prio; prio 201 drivers/net/ethernet/mellanox/mlx5/core/steering/fs_dr.c u8 prio = vlan->prio; prio 205 drivers/net/ethernet/mellanox/mlx5/core/steering/fs_dr.c vlan_hdr = (u32)n_ethtype << 16 | (u32)(prio) << 12 | (u32)vid; prio 3742 drivers/net/ethernet/mellanox/mlxsw/reg.h mlxsw_reg_qpdsm_prio_pack(char *payload, unsigned short prio, u8 dscp) prio 3744 drivers/net/ethernet/mellanox/mlxsw/reg.h mlxsw_reg_qpdsm_prio_entry_color0_e_set(payload, prio, 1); prio 3745 drivers/net/ethernet/mellanox/mlxsw/reg.h mlxsw_reg_qpdsm_prio_entry_color0_dscp_set(payload, prio, dscp); prio 3746 drivers/net/ethernet/mellanox/mlxsw/reg.h mlxsw_reg_qpdsm_prio_entry_color1_e_set(payload, prio, 1); prio 3747 drivers/net/ethernet/mellanox/mlxsw/reg.h mlxsw_reg_qpdsm_prio_entry_color1_dscp_set(payload, prio, dscp); prio 3748 drivers/net/ethernet/mellanox/mlxsw/reg.h mlxsw_reg_qpdsm_prio_entry_color2_e_set(payload, prio, 1); prio 3749 drivers/net/ethernet/mellanox/mlxsw/reg.h mlxsw_reg_qpdsm_prio_entry_color2_dscp_set(payload, prio, dscp); prio 3797 drivers/net/ethernet/mellanox/mlxsw/reg.h mlxsw_reg_qpdpm_dscp_pack(char *payload, unsigned short dscp, u8 prio) prio 3800 drivers/net/ethernet/mellanox/mlxsw/reg.h mlxsw_reg_qpdpm_dscp_entry_prio_set(payload, dscp, prio); prio 5186 drivers/net/ethernet/mellanox/mlxsw/reg.h static inline void mlxsw_reg_pptb_prio_to_buff_pack(char *payload, u8 prio, prio 5189 drivers/net/ethernet/mellanox/mlxsw/reg.h mlxsw_reg_pptb_prio_to_buff_set(payload, prio, buff); prio 5190 drivers/net/ethernet/mellanox/mlxsw/reg.h mlxsw_reg_pptb_prio_to_buff_msb_set(payload, prio, buff); prio 6425 drivers/net/ethernet/mellanox/mlxsw/reg.h u8 prio) prio 6428 drivers/net/ethernet/mellanox/mlxsw/reg.h mlxsw_reg_rdpm_dscp_entry_prio_set(payload, index, prio); prio 10299 drivers/net/ethernet/mellanox/mlxsw/reg.h MLXSW_ITEM32(reg, sbmm, prio, 0x00, 8, 4); prio 10326 drivers/net/ethernet/mellanox/mlxsw/reg.h static inline void mlxsw_reg_sbmm_pack(char *payload, u8 prio, u32 min_buff, prio 10330 drivers/net/ethernet/mellanox/mlxsw/reg.h mlxsw_reg_sbmm_prio_set(payload, prio); prio 1067 drivers/net/ethernet/mellanox/mlxsw/spectrum.c int prio, char *ppcnt_pl) prio 1072 drivers/net/ethernet/mellanox/mlxsw/spectrum.c mlxsw_reg_ppcnt_pack(ppcnt_pl, mlxsw_sp_port->local_port, grp, prio); prio 2308 drivers/net/ethernet/mellanox/mlxsw/spectrum.c static void mlxsw_sp_port_get_prio_strings(u8 **p, int prio) prio 2314 drivers/net/ethernet/mellanox/mlxsw/spectrum.c mlxsw_sp_port_hw_prio_stats[i].str, prio); prio 2444 drivers/net/ethernet/mellanox/mlxsw/spectrum.c enum mlxsw_reg_ppcnt_grp grp, int prio, prio 2457 drivers/net/ethernet/mellanox/mlxsw/spectrum.c mlxsw_sp_port_get_stats_raw(dev, grp, prio, ppcnt_pl); prio 724 drivers/net/ethernet/mellanox/mlxsw/spectrum.h u32 action, u16 vid, u16 proto, u8 prio, prio 919 drivers/net/ethernet/mellanox/mlxsw/spectrum.h enum mlxsw_sp_mr_route_prio prio); prio 93 drivers/net/ethernet/mellanox/mlxsw/spectrum1_mr_tcam.c enum mlxsw_sp_mr_route_prio prio) prio 100 drivers/net/ethernet/mellanox/mlxsw/spectrum1_mr_tcam.c &tcam_region->parman_prios[prio], prio 105 drivers/net/ethernet/mellanox/mlxsw/spectrum1_mr_tcam.c route->parman_prio = &tcam_region->parman_prios[prio]; prio 126 drivers/net/ethernet/mellanox/mlxsw/spectrum1_mr_tcam.c enum mlxsw_sp_mr_route_prio prio) prio 133 drivers/net/ethernet/mellanox/mlxsw/spectrum1_mr_tcam.c key, prio); prio 210 drivers/net/ethernet/mellanox/mlxsw/spectrum2_mr_tcam.c enum mlxsw_sp_mr_route_prio prio) prio 229 drivers/net/ethernet/mellanox/mlxsw/spectrum2_mr_tcam.c mlxsw_sp2_mr_tcam_rule_parse(rule, key, prio); prio 587 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl.c u32 action, u16 vid, u16 proto, u8 prio, prio 608 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl.c vid, prio, ethertype, prio 230 drivers/net/ethernet/mellanox/mlxsw/spectrum_dcb.c int prio; prio 247 drivers/net/ethernet/mellanox/mlxsw/spectrum_dcb.c prio = fls(dcb_ieee_getapp_mask(dev, app)); prio 248 drivers/net/ethernet/mellanox/mlxsw/spectrum_dcb.c if (prio--) { prio 249 drivers/net/ethernet/mellanox/mlxsw/spectrum_dcb.c if (prio < app->priority) prio 251 drivers/net/ethernet/mellanox/mlxsw/spectrum_dcb.c app->priority, app->protocol, prio); prio 252 drivers/net/ethernet/mellanox/mlxsw/spectrum_dcb.c else if (prio > app->priority) prio 254 drivers/net/ethernet/mellanox/mlxsw/spectrum_dcb.c app->priority, app->protocol, prio); prio 532 drivers/net/ethernet/mellanox/mlxsw/spectrum_dcb.c u8 prio) prio 540 drivers/net/ethernet/mellanox/mlxsw/spectrum_dcb.c MLXSW_REG_PPCNT_PRIO_CNT, prio); prio 545 drivers/net/ethernet/mellanox/mlxsw/spectrum_dcb.c my_pfc->requests[prio] = mlxsw_reg_ppcnt_tx_pause_get(ppcnt_pl); prio 546 drivers/net/ethernet/mellanox/mlxsw/spectrum_dcb.c my_pfc->indications[prio] = mlxsw_reg_ppcnt_rx_pause_get(ppcnt_pl); prio 123 drivers/net/ethernet/mellanox/mlxsw/spectrum_flower.c u8 prio = act->vlan.prio; prio 128 drivers/net/ethernet/mellanox/mlxsw/spectrum_flower.c proto, prio, extack); prio 353 drivers/net/ethernet/mellanox/mlxsw/spectrum_flower.c mlxsw_sp_acl_rulei_priority(rulei, f->common.prio); prio 270 drivers/net/ethernet/mellanox/mlxsw/spectrum_mr.c route_params.prio = mlxsw_sp_mr_route_prio(mr_route); prio 889 drivers/net/ethernet/mellanox/mlxsw/spectrum_mr.c .prio = MLXSW_SP_MR_ROUTE_PRIO_CATCHALL, prio 38 drivers/net/ethernet/mellanox/mlxsw/spectrum_mr.h enum mlxsw_sp_mr_route_prio prio; prio 326 drivers/net/ethernet/mellanox/mlxsw/spectrum_mr_tcam.c route_params->prio); prio 407 drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c u32 prio; prio 4411 drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c fib4_entry->prio = fen_info->fi->fib_priority; prio 4691 drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c if (fib4_entry->prio >= new4_entry->prio || prio 4713 drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c fib4_entry->prio != new4_entry->prio) prio 2258 drivers/net/ethernet/micrel/ksz884x.c int prio; prio 2262 drivers/net/ethernet/micrel/ksz884x.c for (prio = 0; prio < PRIO_QUEUES; prio++) { prio 2263 drivers/net/ethernet/micrel/ksz884x.c sw->port_cfg[port].rx_rate[prio] = prio 2264 drivers/net/ethernet/micrel/ksz884x.c sw->port_cfg[port].tx_rate[prio] = 0; prio 2542 drivers/net/ethernet/micrel/ksz884x.c static void sw_cfg_port_based(struct ksz_hw *hw, int port, u8 prio) prio 2546 drivers/net/ethernet/micrel/ksz884x.c if (prio > PORT_BASED_PRIORITY_BASE) prio 2547 drivers/net/ethernet/micrel/ksz884x.c prio = PORT_BASED_PRIORITY_BASE; prio 2549 drivers/net/ethernet/micrel/ksz884x.c hw->ksz_switch->port_cfg[port].port_prio = prio; prio 2553 drivers/net/ethernet/micrel/ksz884x.c data |= prio << PORT_BASED_PRIORITY_SHIFT; prio 349 drivers/net/ethernet/mscc/ocelot_ace.c if (ace->prio != 0) prio 609 drivers/net/ethernet/mscc/ocelot_ace.c if (rule->prio < tmp->prio) prio 191 drivers/net/ethernet/mscc/ocelot_ace.h u16 prio; prio 165 drivers/net/ethernet/mscc/ocelot_flower.c ocelot_rule->prio = f->common.prio; prio 215 drivers/net/ethernet/mscc/ocelot_flower.c rule.prio = f->common.prio; prio 233 drivers/net/ethernet/mscc/ocelot_flower.c rule.prio = f->common.prio; prio 108 drivers/net/ethernet/netronome/nfp/abm/cls.c nfp_abm_find_band_for_prio(struct nfp_abm_link *alink, unsigned int prio) prio 113 drivers/net/ethernet/netronome/nfp/abm/cls.c if ((prio & iter->mask) == iter->val) prio 126 drivers/net/ethernet/netronome/nfp/flower/action.c FIELD_PREP(NFP_FL_PUSH_VLAN_PRIO, act->vlan.prio) | prio 96 drivers/net/ethernet/netronome/nfp/flower/qos_conf.c if (flow->common.prio != 1) { prio 64 drivers/net/ethernet/qlogic/qed/qed_dcbx.c #define QED_DCBX_PRIO2TC(prio_tc_tbl, prio) \ prio 65 drivers/net/ethernet/qlogic/qed/qed_dcbx.c ((u32)(prio_tc_tbl >> ((7 - prio) * 4)) & 0x7) prio 194 drivers/net/ethernet/qlogic/qed/qed_dcbx.c bool app_tlv, bool enable, u8 prio, u8 tc, prio 200 drivers/net/ethernet/qlogic/qed/qed_dcbx.c p_data->arr[type].priority = prio; prio 218 drivers/net/ethernet/qlogic/qed/qed_dcbx.c qed_wr(p_hwfn, p_ptt, DORQ_REG_PF_PCP_BB_K2, prio << 1); prio 226 drivers/net/ethernet/qlogic/qed/qed_dcbx.c bool app_tlv, bool enable, u8 prio, u8 tc, prio 242 drivers/net/ethernet/qlogic/qed/qed_dcbx.c prio, tc, type, personality); prio 532 drivers/net/ethernet/qlogic/qed/qed_dcbx.c entry->prio = ffs(pri_map) - 1; prio 556 drivers/net/ethernet/qlogic/qed/qed_dcbx.c p_params->pfc.prio[0] = !!(pfc_map & DCBX_PFC_PRI_EN_BITMAP_PRI_0); prio 557 drivers/net/ethernet/qlogic/qed/qed_dcbx.c p_params->pfc.prio[1] = !!(pfc_map & DCBX_PFC_PRI_EN_BITMAP_PRI_1); prio 558 drivers/net/ethernet/qlogic/qed/qed_dcbx.c p_params->pfc.prio[2] = !!(pfc_map & DCBX_PFC_PRI_EN_BITMAP_PRI_2); prio 559 drivers/net/ethernet/qlogic/qed/qed_dcbx.c p_params->pfc.prio[3] = !!(pfc_map & DCBX_PFC_PRI_EN_BITMAP_PRI_3); prio 560 drivers/net/ethernet/qlogic/qed/qed_dcbx.c p_params->pfc.prio[4] = !!(pfc_map & DCBX_PFC_PRI_EN_BITMAP_PRI_4); prio 561 drivers/net/ethernet/qlogic/qed/qed_dcbx.c p_params->pfc.prio[5] = !!(pfc_map & DCBX_PFC_PRI_EN_BITMAP_PRI_5); prio 562 drivers/net/ethernet/qlogic/qed/qed_dcbx.c p_params->pfc.prio[6] = !!(pfc_map & DCBX_PFC_PRI_EN_BITMAP_PRI_6); prio 563 drivers/net/ethernet/qlogic/qed/qed_dcbx.c p_params->pfc.prio[7] = !!(pfc_map & DCBX_PFC_PRI_EN_BITMAP_PRI_7); prio 1069 drivers/net/ethernet/qlogic/qed/qed_dcbx.c if (p_params->pfc.prio[i]) prio 1190 drivers/net/ethernet/qlogic/qed/qed_dcbx.c *entry |= ((u32)(p_params->app_entry[i].prio) << prio 1418 drivers/net/ethernet/qlogic/qed/qed_dcbx.c static void qed_dcbnl_getpgtccfgrx(struct qed_dev *cdev, int tc, u8 *prio, prio 1422 drivers/net/ethernet/qlogic/qed/qed_dcbx.c *prio = *bwg_id = *bw_pct = *up_map = 0; prio 1448 drivers/net/ethernet/qlogic/qed/qed_dcbx.c *setting = dcbx_info->operational.params.pfc.prio[priority]; prio 1473 drivers/net/ethernet/qlogic/qed/qed_dcbx.c dcbx_set.config.params.pfc.prio[priority] = !!setting; prio 1624 drivers/net/ethernet/qlogic/qed/qed_dcbx.c static void qed_dcbnl_setpgtccfgrx(struct qed_dev *cdev, int prio, prio 1758 drivers/net/ethernet/qlogic/qed/qed_dcbx.c u8 prio = 0; prio 1769 drivers/net/ethernet/qlogic/qed/qed_dcbx.c prio = entry->prio; prio 1782 drivers/net/ethernet/qlogic/qed/qed_dcbx.c return prio; prio 1820 drivers/net/ethernet/qlogic/qed/qed_dcbx.c dcbx_set.config.params.app_entry[i].prio = pri_map; prio 2005 drivers/net/ethernet/qlogic/qed/qed_dcbx.c table[i].priority = dcbx_info->remote.params.app_entry[i].prio; prio 2026 drivers/net/ethernet/qlogic/qed/qed_dcbx.c if (dcbx_info->remote.params.pfc.prio[i]) prio 2094 drivers/net/ethernet/qlogic/qed/qed_dcbx.c if (params->pfc.prio[i]) prio 2134 drivers/net/ethernet/qlogic/qed/qed_dcbx.c dcbx_set.config.params.pfc.prio[i] = !!(pfc->pfc_en & BIT(i)); prio 2283 drivers/net/ethernet/qlogic/qed/qed_dcbx.c u8 prio = 0; prio 2310 drivers/net/ethernet/qlogic/qed/qed_dcbx.c prio = entry->prio; prio 2322 drivers/net/ethernet/qlogic/qed/qed_dcbx.c app->priority = ffs(prio) - 1; prio 2389 drivers/net/ethernet/qlogic/qed/qed_dcbx.c dcbx_set.config.params.app_entry[i].prio = BIT(app->priority); prio 32 drivers/net/ethernet/qlogic/qede/qede_dcbnl.c static void qede_dcbnl_getpgtccfgtx(struct net_device *netdev, int prio, prio 38 drivers/net/ethernet/qlogic/qede/qede_dcbnl.c edev->ops->dcb->getpgtccfgtx(edev->cdev, prio, prio_type, prio 50 drivers/net/ethernet/qlogic/qede/qede_dcbnl.c static void qede_dcbnl_getpgtccfgrx(struct net_device *netdev, int prio, prio 56 drivers/net/ethernet/qlogic/qede/qede_dcbnl.c edev->ops->dcb->getpgtccfgrx(edev->cdev, prio, prio_type, pgid, bw_pct, prio 68 drivers/net/ethernet/qlogic/qede/qede_dcbnl.c static void qede_dcbnl_getpfccfg(struct net_device *netdev, int prio, prio 73 drivers/net/ethernet/qlogic/qede/qede_dcbnl.c edev->ops->dcb->getpfccfg(edev->cdev, prio, setting); prio 76 drivers/net/ethernet/qlogic/qede/qede_dcbnl.c static void qede_dcbnl_setpfccfg(struct net_device *netdev, int prio, prio 81 drivers/net/ethernet/qlogic/qede/qede_dcbnl.c edev->ops->dcb->setpfccfg(edev->cdev, prio, setting); prio 119 drivers/net/ethernet/qlogic/qede/qede_dcbnl.c static void qede_dcbnl_setpgtccfgtx(struct net_device *netdev, int prio, prio 124 drivers/net/ethernet/qlogic/qede/qede_dcbnl.c return edev->ops->dcb->setpgtccfgtx(edev->cdev, prio, pri_type, pgid, prio 128 drivers/net/ethernet/qlogic/qede/qede_dcbnl.c static void qede_dcbnl_setpgtccfgrx(struct net_device *netdev, int prio, prio 133 drivers/net/ethernet/qlogic/qede/qede_dcbnl.c return edev->ops->dcb->setpgtccfgrx(edev->cdev, prio, pri_type, pgid, prio 793 drivers/net/ethernet/qlogic/qlcnic/qlcnic_dcb.c qlcnic_dcb_get_pg_tc_cfg_tx(struct net_device *netdev, int tc, u8 *prio, prio 802 drivers/net/ethernet/qlogic/qlcnic/qlcnic_dcb.c *prio = *pgid = *bw_per = *up_tc_map = 0; prio 816 drivers/net/ethernet/qlogic/qlcnic/qlcnic_dcb.c *prio = tc_cfg->prio_type; prio 854 drivers/net/ethernet/qlogic/qlcnic/qlcnic_dcb.c static void qlcnic_dcb_get_pfc_cfg(struct net_device *netdev, int prio, prio 859 drivers/net/ethernet/qlogic/qlcnic/qlcnic_dcb.c u8 val = QLC_DCB_GET_MAP(prio); prio 875 drivers/net/ethernet/qlogic/qlcnic/qlcnic_dcb.c if ((val & tc_cfg->up_tc_map) && (tc_cfg->prio_cfg[prio].valid)) prio 876 drivers/net/ethernet/qlogic/qlcnic/qlcnic_dcb.c *setting = tc_cfg->prio_cfg[prio].pfc_type; prio 1003 drivers/net/ethernet/qlogic/qlcnic/qlcnic_dcb.c qlcnic_dcb_get_pg_tc_cfg_rx(struct net_device *netdev, int prio, u8 *prio_type, prio 1103 drivers/net/ethernet/qlogic/qlcnic/qlcnic_dcb.c u8 i, setting, prio; prio 1114 drivers/net/ethernet/qlogic/qlcnic/qlcnic_dcb.c prio = qlcnic_dcb_prio_count(tc->up_tc_map); prio 1117 drivers/net/ethernet/qlogic/qlcnic/qlcnic_dcb.c qlcnic_dcb_get_pfc_cfg(netdev, prio, &setting); prio 75 drivers/net/ethernet/stmicro/stmmac/dwmac4_core.c u32 prio, u32 queue) prio 88 drivers/net/ethernet/stmicro/stmmac/dwmac4_core.c value |= (prio << GMAC_RXQCTRL_PSRQX_SHIFT(queue)) & prio 94 drivers/net/ethernet/stmicro/stmmac/dwmac4_core.c u32 prio, u32 queue) prio 107 drivers/net/ethernet/stmicro/stmmac/dwmac4_core.c value |= (prio << GMAC_TXQCTRL_PSTQX_SHIFT(queue)) & prio 397 drivers/net/ethernet/stmicro/stmmac/dwmac5.c if (entry->prio < curr_prio) prio 400 drivers/net/ethernet/stmicro/stmmac/dwmac5.c if (entry->prio < min_prio) { prio 401 drivers/net/ethernet/stmicro/stmmac/dwmac5.c min_prio = entry->prio; prio 442 drivers/net/ethernet/stmicro/stmmac/dwmac5.c curr_prio = entry->prio; prio 95 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c static void dwxgmac2_rx_queue_prio(struct mac_device_info *hw, u32 prio, prio 107 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c value |= (prio << XGMAC_PSRQ_SHIFT(queue)) & XGMAC_PSRQ(queue); prio 112 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c static void dwxgmac2_tx_queue_prio(struct mac_device_info *hw, u32 prio, prio 124 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c value |= (prio << XGMAC_PSTC_SHIFT(queue)) & XGMAC_PSTC(queue); prio 962 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c if (entry->prio < curr_prio) prio 965 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c if (entry->prio < min_prio) { prio 966 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c min_prio = entry->prio; prio 1008 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c curr_prio = entry->prio; prio 289 drivers/net/ethernet/stmicro/stmmac/hwif.h void (*rx_queue_prio)(struct mac_device_info *hw, u32 prio, u32 queue); prio 291 drivers/net/ethernet/stmicro/stmmac/hwif.h void (*tx_queue_prio)(struct mac_device_info *hw, u32 prio, u32 queue); prio 102 drivers/net/ethernet/stmicro/stmmac/stmmac.h u32 prio; prio 2386 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c u32 prio; prio 2392 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c prio = priv->plat->rx_queues_cfg[queue].prio; prio 2393 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c stmmac_rx_queue_prio(priv, priv->hw, prio, queue); prio 2406 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c u32 prio; prio 2412 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c prio = priv->plat->tx_queues_cfg[queue].prio; prio 2413 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c stmmac_tx_queue_prio(priv, priv->hw, prio, queue); prio 192 drivers/net/ethernet/stmicro/stmmac/stmmac_platform.c &plat->rx_queues_cfg[queue].prio)) { prio 193 drivers/net/ethernet/stmicro/stmmac/stmmac_platform.c plat->rx_queues_cfg[queue].prio = 0; prio 272 drivers/net/ethernet/stmicro/stmmac/stmmac_platform.c &plat->tx_queues_cfg[queue].prio)) { prio 273 drivers/net/ethernet/stmicro/stmmac/stmmac_platform.c plat->tx_queues_cfg[queue].prio = 0; prio 20 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c entry->prio = ~0x0; prio 97 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c u32 prio = cls->common.prio << 16; prio 141 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c entry->prio = prio; prio 148 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c frag->prio = prio; prio 155 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c entry->prio = prio; prio 1506 drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c unsigned int ppq, ppq_extra, prio, prio_queues; prio 1556 drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c for (i = 0, prio = 0; i < prio_queues;) { prio 1560 drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c "PRIO%u mapped to RXq%u\n", prio, i); prio 1561 drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c mask |= (1 << prio); prio 1562 drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c prio++; prio 1567 drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c "PRIO%u mapped to RXq%u\n", prio, i); prio 1568 drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c mask |= (1 << prio); prio 1569 drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c prio++; prio 412 drivers/net/ethernet/ti/cpsw.c u16 vtag, vid, prio, pkt_type; prio 436 drivers/net/ethernet/ti/cpsw.c prio = (rx_vlan_encap_hdr >> prio 440 drivers/net/ethernet/ti/cpsw.c vtag = (prio << VLAN_PRIO_SHIFT) | vid; prio 832 drivers/net/geneve.c __u8 prio; prio 842 drivers/net/geneve.c prio = info->key.tos; prio 843 drivers/net/geneve.c if ((prio == 1) && !geneve->collect_md) { prio 844 drivers/net/geneve.c prio = ip_tunnel_get_dsfield(ip_hdr(skb), skb); prio 848 drivers/net/geneve.c fl6->flowlabel = ip6_make_flowinfo(RT_TOS(prio), prio 946 drivers/net/geneve.c __u8 prio, ttl; prio 958 drivers/net/geneve.c prio = ip_tunnel_ecn_encap(key->tos, ip_hdr(skb), skb); prio 961 drivers/net/geneve.c prio = ip_tunnel_ecn_encap(ip6_tclass(fl6.flowlabel), prio 974 drivers/net/geneve.c &fl6.saddr, &fl6.daddr, prio, ttl, prio 6715 drivers/net/wireless/ath/ath10k/mac.c u32 prio = 0, acc = 0; prio 6728 drivers/net/wireless/ath/ath10k/mac.c prio = 7; prio 6734 drivers/net/wireless/ath/ath10k/mac.c prio = 5; prio 6740 drivers/net/wireless/ath/ath10k/mac.c prio = 2; prio 6746 drivers/net/wireless/ath/ath10k/mac.c prio = 0; prio 6798 drivers/net/wireless/ath/ath10k/mac.c arg.user_priority = prio; prio 60 drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c u8 prio, u8 ifidx) prio 70 drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c fifo = brcmf_flowring_prio2fifo[prio]; prio 104 drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c u8 prio, u8 ifidx) prio 115 drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c fifo = brcmf_flowring_prio2fifo[prio]; prio 51 drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.h u8 prio, u8 ifidx); prio 53 drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.h u8 prio, u8 ifidx); prio 319 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c static uint prio2prec(u32 prio) prio 321 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c return (prio == PRIO_8021D_NONE || prio == PRIO_8021D_BE) ? prio 322 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c (prio^2) : prio; prio 86 drivers/net/wireless/broadcom/brcm80211/brcmfmac/tracepoint.h __field(u8, prio) prio 93 drivers/net/wireless/broadcom/brcm80211/brcmfmac/tracepoint.h __entry->prio = *((u8 *)data + 1); prio 99 drivers/net/wireless/broadcom/brcm80211/brcmfmac/tracepoint.h TP_printk("bcdc: prio=%d siglen=%d", __entry->prio, __entry->siglen) prio 1342 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c bool state, int prio) prio 99 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.h bool state, int prio); prio 3500 drivers/net/wireless/intel/iwlwifi/mvm/rs.c int prio = -1; prio 3506 drivers/net/wireless/intel/iwlwifi/mvm/rs.c prio = 3; prio 3509 drivers/net/wireless/intel/iwlwifi/mvm/rs.c prio = 2; prio 3512 drivers/net/wireless/intel/iwlwifi/mvm/rs.c prio = 1; prio 3516 drivers/net/wireless/intel/iwlwifi/mvm/rs.c prio = -1; prio 3519 drivers/net/wireless/intel/iwlwifi/mvm/rs.c return prio; prio 342 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c static void rtl8180_handle_tx(struct ieee80211_hw *dev, unsigned int prio) prio 345 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c struct rtl8180_tx_ring *ring = &priv->tx_ring[prio]; prio 372 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c ieee80211_wake_queue(dev, prio); prio 464 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c unsigned int idx, prio, hw_prio; prio 473 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c prio = skb_get_queue_mapping(skb); prio 474 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c ring = &priv->tx_ring[prio]; prio 575 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c ieee80211_stop_queue(dev, prio); prio 581 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c hw_prio = rtl8187se_queues_map[prio]; prio 585 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c hw_prio = rtl8180_queues_map[prio]; prio 1069 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c unsigned int prio, unsigned int entries) prio 1080 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c prio); prio 1084 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c priv->tx_ring[prio].desc = ring; prio 1085 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c priv->tx_ring[prio].dma = dma; prio 1086 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c priv->tx_ring[prio].idx = 0; prio 1087 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c priv->tx_ring[prio].entries = entries; prio 1088 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c skb_queue_head_init(&priv->tx_ring[prio].queue); prio 1097 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c static void rtl8180_free_tx_ring(struct ieee80211_hw *dev, unsigned int prio) prio 1100 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c struct rtl8180_tx_ring *ring = &priv->tx_ring[prio]; prio 526 drivers/net/wireless/realtek/rtlwifi/pci.c static void _rtl_pci_tx_isr(struct ieee80211_hw *hw, int prio) prio 531 drivers/net/wireless/realtek/rtlwifi/pci.c struct rtl8192_tx_ring *ring = &rtlpci->tx_ring[prio]; prio 545 drivers/net/wireless/realtek/rtlwifi/pci.c if (!rtlpriv->cfg->ops->is_tx_desc_closed(hw, prio, ring->idx)) prio 566 drivers/net/wireless/realtek/rtlwifi/pci.c if (prio == TXCMD_QUEUE) { prio 613 drivers/net/wireless/realtek/rtlwifi/pci.c prio, ring->idx, prio 1208 drivers/net/wireless/realtek/rtlwifi/pci.c unsigned int prio, unsigned int entries) prio 1227 drivers/net/wireless/realtek/rtlwifi/pci.c prio); prio 1231 drivers/net/wireless/realtek/rtlwifi/pci.c rtlpci->tx_ring[prio].buffer_desc = buffer_desc; prio 1232 drivers/net/wireless/realtek/rtlwifi/pci.c rtlpci->tx_ring[prio].buffer_desc_dma = buffer_desc_dma; prio 1234 drivers/net/wireless/realtek/rtlwifi/pci.c rtlpci->tx_ring[prio].cur_tx_rp = 0; prio 1235 drivers/net/wireless/realtek/rtlwifi/pci.c rtlpci->tx_ring[prio].cur_tx_wp = 0; prio 1243 drivers/net/wireless/realtek/rtlwifi/pci.c pr_err("Cannot allocate TX ring (prio = %d)\n", prio); prio 1247 drivers/net/wireless/realtek/rtlwifi/pci.c rtlpci->tx_ring[prio].desc = desc; prio 1248 drivers/net/wireless/realtek/rtlwifi/pci.c rtlpci->tx_ring[prio].dma = desc_dma; prio 1250 drivers/net/wireless/realtek/rtlwifi/pci.c rtlpci->tx_ring[prio].idx = 0; prio 1251 drivers/net/wireless/realtek/rtlwifi/pci.c rtlpci->tx_ring[prio].entries = entries; prio 1252 drivers/net/wireless/realtek/rtlwifi/pci.c skb_queue_head_init(&rtlpci->tx_ring[prio].queue); prio 1255 drivers/net/wireless/realtek/rtlwifi/pci.c prio, desc); prio 1334 drivers/net/wireless/realtek/rtlwifi/pci.c unsigned int prio) prio 1338 drivers/net/wireless/realtek/rtlwifi/pci.c struct rtl8192_tx_ring *ring = &rtlpci->tx_ring[prio]; prio 53 drivers/s390/cio/orb.h u32 prio:8; prio 981 drivers/scsi/csiostor/csio_wr.c csio_wr_issue(struct csio_hw *hw, int qidx, bool prio) prio 990 drivers/scsi/csiostor/csio_wr.c csio_wr_reg32(hw, DBPRIO_V(prio) | QID_V(q->un.eq.physeqid) | prio 1799 drivers/scsi/fcoe/fcoe.c int prio; prio 1816 drivers/scsi/fcoe/fcoe.c prio = ffs(entry->app.priority) - 1; prio 1818 drivers/scsi/fcoe/fcoe.c prio = entry->app.priority; prio 1820 drivers/scsi/fcoe/fcoe.c if (prio < 0) prio 1825 drivers/scsi/fcoe/fcoe.c ctlr->priority = prio; prio 1828 drivers/scsi/fcoe/fcoe.c fcoe->priority = prio; prio 324 drivers/scsi/qedf/qedf.h u8 prio; prio 336 drivers/scsi/qedf/qedf_debugfs.c seq_printf(s, "FIP 802.1Q Priority: %d\n", qedf->prio); prio 116 drivers/scsi/qedf/qedf_main.c vlan_id_tmp = vlan_id | (qedf->prio << VLAN_PRIO_SHIFT); prio 120 drivers/scsi/qedf/qedf_main.c vlan_id_tmp, qedf->prio); prio 598 drivers/scsi/qedf/qedf_main.c qedf->prio = qedf_default_prio; prio 603 drivers/scsi/qedf/qedf_main.c qedf->prio = QEDF_DEFAULT_PRIO; prio 605 drivers/scsi/qedf/qedf_main.c qedf->prio = tmp_prio; prio 1286 drivers/scsi/qedf/qedf_main.c qedf->prio << FCOE_CONN_OFFLOAD_RAMROD_DATA_PRIORITY_SHIFT; prio 3237 drivers/scsi/qedf/qedf_main.c qedf->prio = qedf_default_prio; prio 3239 drivers/scsi/qedf/qedf_main.c qedf->prio = QEDF_DEFAULT_PRIO; prio 502 drivers/scsi/virtio_scsi.c cmd->prio = 0; prio 129 drivers/sh/intc/chip.c int intc_set_priority(unsigned int irq, unsigned int prio) prio 135 drivers/sh/intc/chip.c if (!intc_get_prio_level(irq) || prio <= 1) prio 138 drivers/sh/intc/chip.c ihp = intc_find_irq(d->prio, d->nr_prio, irq); prio 140 drivers/sh/intc/chip.c if (prio >= (1 << _INTC_WIDTH(ihp->handle))) prio 143 drivers/sh/intc/chip.c intc_set_prio_level(irq, prio); prio 132 drivers/sh/intc/core.c hp = d->prio + d->nr_prio; prio 256 drivers/sh/intc/core.c d->prio = kcalloc(hw->nr_vectors, sizeof(*d->prio), prio 258 drivers/sh/intc/core.c if (!d->prio) prio 267 drivers/sh/intc/core.c sort(d->prio, hw->nr_prio_regs, sizeof(*d->prio), prio 387 drivers/sh/intc/core.c kfree(d->prio); prio 65 drivers/sh/intc/internals.h struct intc_handle_int *prio; prio 447 drivers/soc/fsl/dpio/dpio-service.c u32 qdid, u8 prio, u16 qdbin, prio 458 drivers/soc/fsl/dpio/dpio-service.c qbman_eq_desc_set_qd(&ed, qdid, qdbin, prio); prio 80 drivers/soc/ti/knav_dma.c u32 prio; prio 168 drivers/soc/ti/knav_dma.c writel_relaxed(cfg->u.tx.priority, &chan->reg_tx_sched->prio); prio 260 drivers/staging/rtl8192e/rtl8192e/rtl_core.c static short _rtl92e_check_nic_enough_desc(struct net_device *dev, int prio) prio 263 drivers/staging/rtl8192e/rtl8192e/rtl_core.c struct rtl8192_tx_ring *ring = &priv->tx_ring[prio]; prio 1577 drivers/staging/rtl8192e/rtl8192e/rtl_core.c static void _rtl92e_free_tx_ring(struct net_device *dev, unsigned int prio) prio 1580 drivers/staging/rtl8192e/rtl8192e/rtl_core.c struct rtl8192_tx_ring *ring = &priv->tx_ring[prio]; prio 1663 drivers/staging/rtl8192e/rtl8192e/rtl_core.c static void _rtl92e_tx_isr(struct net_device *dev, int prio) prio 1667 drivers/staging/rtl8192e/rtl8192e/rtl_core.c struct rtl8192_tx_ring *ring = &priv->tx_ring[prio]; prio 1673 drivers/staging/rtl8192e/rtl8192e/rtl_core.c if (prio != BEACON_QUEUE) { prio 1685 drivers/staging/rtl8192e/rtl8192e/rtl_core.c if (prio != BEACON_QUEUE) prio 1827 drivers/staging/rtl8192e/rtl8192e/rtl_core.c static int _rtl92e_alloc_tx_ring(struct net_device *dev, unsigned int prio, prio 1837 drivers/staging/rtl8192e/rtl8192e/rtl_core.c netdev_warn(dev, "Cannot allocate TX ring (prio = %d)\n", prio); prio 1841 drivers/staging/rtl8192e/rtl8192e/rtl_core.c priv->tx_ring[prio].desc = ring; prio 1842 drivers/staging/rtl8192e/rtl8192e/rtl_core.c priv->tx_ring[prio].dma = dma; prio 1843 drivers/staging/rtl8192e/rtl8192e/rtl_core.c priv->tx_ring[prio].idx = 0; prio 1844 drivers/staging/rtl8192e/rtl8192e/rtl_core.c priv->tx_ring[prio].entries = entries; prio 1845 drivers/staging/rtl8192e/rtl8192e/rtl_core.c skb_queue_head_init(&priv->tx_ring[prio].queue); prio 1314 drivers/tty/ipwireless/hardware.c static int set_control_line(struct ipw_hardware *hw, int prio, prio 1320 drivers/tty/ipwireless/hardware.c if (prio == PRIO_SETUP) prio 1329 drivers/tty/ipwireless/hardware.c send_packet(hw, prio, &packet->header); prio 386 fs/jffs2/nodelist.h uint32_t *len, int prio, uint32_t sumsize); prio 75 fs/jffs2/nodemgmt.c uint32_t *len, int prio, uint32_t sumsize) prio 93 fs/jffs2/nodemgmt.c if (prio != ALLOC_DELETION && !jffs2_rp_can_write(c)) { prio 117 fs/jffs2/nodemgmt.c if (prio == ALLOC_DELETION && c->nr_free_blocks + c->nr_erasing_blocks >= c->resv_blocks_deletion) { prio 142 fs/jffs2/nodemgmt.c if (prio == ALLOC_DELETION && c->nr_free_blocks + c->nr_erasing_blocks >= c->resv_blocks_deletion) { prio 54 include/linux/audit.h u64 prio; prio 27 include/linux/bio.h #define bio_set_prio(bio, prio) ((bio)->bi_ioprio = prio) prio 135 include/linux/buffer_head.h BUFFER_FNS(Prio, prio) prio 96 include/linux/compaction.h const struct alloc_context *ac, enum compact_priority prio, prio 23 include/linux/dma/pxa-dma.h enum pxad_chan_prio prio; prio 1278 include/linux/mlx4/device.h __be16 prio; prio 1460 include/linux/mlx4/device.h int port, int qpn, u16 prio, u64 *reg_id); prio 149 include/linux/mlx5/fs.h int prio, prio 156 include/linux/mlx5/fs.h int prio; prio 168 include/linux/mlx5/fs.h int prio, prio 173 include/linux/mlx5/fs.h int prio, u32 level); prio 189 include/linux/mlx5/fs.h u8 prio; prio 2783 include/linux/mlx5/mlx5_ifc.h u8 prio[0x3]; prio 2906 include/linux/mlx5/mlx5_ifc.h u8 prio[0x4]; prio 5875 include/linux/mlx5/mlx5_ifc.h u8 prio[0x1]; prio 8325 include/linux/mlx5/mlx5_ifc.h u8 prio[0x3]; prio 9615 include/linux/mlx5/mlx5_ifc.h u8 prio[0x04]; prio 9671 include/linux/mlx5/mlx5_ifc.h u8 prio[0x3]; prio 217 include/linux/mlx5/mlx5_ifc_fpga.h u8 prio[0x3]; prio 170 include/linux/mlx5/port.h u8 prio, u8 *tc); prio 199 include/linux/mlx5/port.h int mlx5_set_dscp2prio(struct mlx5_core_dev *mdev, u8 dscp, u8 prio); prio 2082 include/linux/netdevice.h int netdev_get_prio_tc_map(const struct net_device *dev, u32 prio) prio 2084 include/linux/netdevice.h return dev->prio_tc_map[prio & TC_BITMASK]; prio 2088 include/linux/netdevice.h int netdev_set_prio_tc_map(struct net_device *dev, u8 prio, u8 tc) prio 2093 include/linux/netdevice.h dev->prio_tc_map[prio & TC_BITMASK] = tc & TC_BITMASK; prio 68 include/linux/parman.h void parman_prio_init(struct parman *parman, struct parman_prio *prio, prio 70 include/linux/parman.h void parman_prio_fini(struct parman_prio *prio); prio 71 include/linux/parman.h int parman_item_add(struct parman *parman, struct parman_prio *prio, prio 73 include/linux/parman.h void parman_item_remove(struct parman *parman, struct parman_prio *prio, prio 84 include/linux/plist.h int prio; prio 112 include/linux/plist.h .prio = (__prio), \ prio 132 include/linux/plist.h static inline void plist_node_init(struct plist_node *node, int prio) prio 134 include/linux/plist.h node->prio = prio; prio 165 include/linux/pm_qos.h return dev->power.qos->resume_latency_req->data.pnode.prio; prio 243 include/linux/qed/qed_eth_if.h void (*getpgtccfgtx)(struct qed_dev *cdev, int prio, u8 *prio_type, prio 246 include/linux/qed/qed_eth_if.h void (*getpgtccfgrx)(struct qed_dev *cdev, int prio, u8 *prio_type, prio 249 include/linux/qed/qed_eth_if.h void (*getpfccfg)(struct qed_dev *cdev, int prio, u8 *setting); prio 250 include/linux/qed/qed_eth_if.h void (*setpfccfg)(struct qed_dev *cdev, int prio, u8 setting); prio 259 include/linux/qed/qed_eth_if.h void (*setpgtccfgtx)(struct qed_dev *cdev, int prio, prio 261 include/linux/qed/qed_eth_if.h void (*setpgtccfgrx)(struct qed_dev *cdev, int prio, prio 93 include/linux/qed/qed_if.h u8 prio[QED_MAX_PFC_PRIORITIES]; prio 108 include/linux/qed/qed_if.h u8 prio; prio 670 include/linux/sched.h int prio; prio 11 include/linux/sched/deadline.h static inline int dl_prio(int prio) prio 13 include/linux/sched/deadline.h if (unlikely(prio < MAX_DL_PRIO)) prio 20 include/linux/sched/deadline.h return dl_prio(p->prio); prio 34 include/linux/sched/prio.h #define PRIO_TO_NICE(prio) ((prio) - DEFAULT_PRIO) prio 56 include/linux/sched/prio.h static inline long rlimit_to_nice(long prio) prio 58 include/linux/sched/prio.h return (MAX_NICE - prio + 1); prio 9 include/linux/sched/rt.h static inline int rt_prio(int prio) prio 11 include/linux/sched/rt.h if (unlikely(prio < MAX_RT_PRIO)) prio 18 include/linux/sched/rt.h return rt_prio(p->prio); prio 137 include/linux/sh_intc.h int intc_set_priority(unsigned int irq, unsigned int prio); prio 3056 include/linux/skbuff.h bool skb_page_frag_refill(unsigned int sz, struct page_frag *pfrag, gfp_t prio); prio 115 include/linux/stmmac.h u32 prio; prio 127 include/linux/stmmac.h u32 prio; prio 237 include/linux/swap.h signed short prio; /* swap priority of this type */ prio 27 include/linux/tracepoint-defs.h int prio; prio 41 include/linux/tracepoint.h int prio); prio 256 include/linux/tracepoint.h int prio) \ prio 259 include/linux/tracepoint.h (void *)probe, data, prio); \ prio 35 include/linux/vmpressure.h extern void vmpressure_prio(gfp_t gfp, struct mem_cgroup *memcg, int prio); prio 50 include/linux/vmpressure.h int prio) {} prio 278 include/media/v4l2-dev.h struct v4l2_prio_state *prio; prio 56 include/media/v4l2-device.h struct v4l2_prio_state prio; prio 46 include/media/v4l2-fh.h enum v4l2_priority prio; prio 235 include/net/bluetooth/rfcomm.h struct rfcomm_dlc *rfcomm_dlc_alloc(gfp_t prio); prio 230 include/net/caif/cfpkt.h void cfpkt_set_prio(struct cfpkt *pkt, int prio); prio 169 include/net/flow_offload.h u8 prio; prio 356 include/net/flow_offload.h u32 prio; prio 547 include/net/pkt_cls.h u32 prio; prio 636 include/net/pkt_cls.h cls_common->prio = tp->prio >> 16; prio 754 include/net/pkt_cls.h u32 prio; prio 361 include/net/sch_generic.h u32 prio; prio 1399 include/net/sctp/structs.h __u16 prio; prio 151 include/net/udp_tunnel.h __u8 prio, __u8 ttl, __be32 label, prio 337 include/scsi/sas.h u8 prio:4; prio 568 include/scsi/sas.h u8 prio:4; prio 112 include/soc/fsl/dpaa2-io.h int dpaa2_io_service_enqueue_qd(struct dpaa2_io *d, u32 qdid, u8 prio, prio 175 include/trace/events/compaction.h int prio), prio 177 include/trace/events/compaction.h TP_ARGS(order, gfp_mask, prio), prio 182 include/trace/events/compaction.h __field(int, prio) prio 188 include/trace/events/compaction.h __entry->prio = prio; prio 194 include/trace/events/compaction.h __entry->prio) prio 66 include/trace/events/sched.h __field( int, prio ) prio 74 include/trace/events/sched.h __entry->prio = p->prio; /* XXX SCHED_DEADLINE */ prio 80 include/trace/events/sched.h __entry->comm, __entry->pid, __entry->prio, prio 159 include/trace/events/sched.h __entry->prev_prio = prev->prio; prio 163 include/trace/events/sched.h __entry->next_prio = next->prio; prio 198 include/trace/events/sched.h __field( int, prio ) prio 206 include/trace/events/sched.h __entry->prio = p->prio; /* XXX SCHED_DEADLINE */ prio 212 include/trace/events/sched.h __entry->comm, __entry->pid, __entry->prio, prio 225 include/trace/events/sched.h __field( int, prio ) prio 231 include/trace/events/sched.h __entry->prio = p->prio; /* XXX SCHED_DEADLINE */ prio 235 include/trace/events/sched.h __entry->comm, __entry->pid, __entry->prio) prio 271 include/trace/events/sched.h __field( int, prio ) prio 277 include/trace/events/sched.h __entry->prio = current->prio; /* XXX SCHED_DEADLINE */ prio 281 include/trace/events/sched.h __entry->comm, __entry->pid, __entry->prio) prio 459 include/trace/events/sched.h __entry->oldprio = tsk->prio; prio 461 include/trace/events/sched.h min(tsk->normal_prio, pi_task->prio) : prio 287 include/uapi/drm/msm_drm.h __u32 prio; /* in, Priority level */ prio 311 include/uapi/linux/bcache.h __u16 prio; prio 298 include/uapi/linux/if_link.h __u8 prio[2]; prio 15 include/uapi/linux/netfilter_bridge/ebt_vlan.h __u8 prio; /* VLAN User Priority {0-7} */ prio 341 include/uapi/linux/pkt_sched.h __u8 prio; /* prio of this VQ */ prio 397 include/uapi/linux/pkt_sched.h __u32 prio; prio 48 include/uapi/linux/virtio_scsi.h __u8 prio; /* SAM command priority field */ prio 58 include/uapi/linux/virtio_scsi.h __u8 prio; /* SAM command priority field */ prio 69 init/init_task.c .prio = MAX_PRIO - 20, prio 627 ipc/mqueue.c if (walk->task->prio <= current->prio) { prio 107 kernel/audit.h u64 prio; prio 830 kernel/auditfilter.c new->prio = old->prio; prio 982 kernel/auditfilter.c entry->rule.prio = ~0ULL; prio 985 kernel/auditfilter.c entry->rule.prio = ++prio_high; prio 987 kernel/auditfilter.c entry->rule.prio = --prio_low; prio 205 kernel/auditsc.c if (!ctx->prio) { prio 206 kernel/auditsc.c ctx->prio = 1; prio 717 kernel/auditsc.c if (rule->prio <= ctx->prio) prio 723 kernel/auditsc.c ctx->prio = rule->prio; prio 900 kernel/auditsc.c context->prio = state == AUDIT_RECORD_CONTEXT ? ~0ULL : 0; prio 1645 kernel/auditsc.c context->prio = 0; prio 1717 kernel/auditsc.c context->prio = context->state == AUDIT_RECORD_CONTEXT ? ~0ULL : 0; prio 2179 kernel/auditsc.c if (!ctx->prio) { prio 2180 kernel/auditsc.c ctx->prio = 1; prio 2348 kernel/futex.c int prio; prio 2358 kernel/futex.c prio = min(current->normal_prio, MAX_RT_PRIO); prio 2360 kernel/futex.c plist_node_init(&q->list, prio); prio 38 kernel/locking/rtmutex-debug.c printk("%16s:%5d [%p, %3d]", p->comm, task_pid_nr(p), p, p->prio); prio 232 kernel/locking/rtmutex.c &(struct rt_mutex_waiter){ .prio = (p)->prio, .deadline = (p)->dl.deadline } prio 238 kernel/locking/rtmutex.c if (left->prio < right->prio) prio 247 kernel/locking/rtmutex.c if (dl_prio(left->prio)) prio 257 kernel/locking/rtmutex.c if (left->prio != right->prio) prio 266 kernel/locking/rtmutex.c if (dl_prio(left->prio)) prio 682 kernel/locking/rtmutex.c waiter->prio = task->prio; prio 955 kernel/locking/rtmutex.c waiter->prio = task->prio; prio 37 kernel/locking/rtmutex_common.h int prio; prio 112 kernel/power/qos.c return plist_first(&c->list)->prio; prio 115 kernel/power/qos.c return plist_last(&c->list)->prio; prio 119 kernel/power/qos.c total_value += node->prio; prio 184 kernel/power/qos.c if ((req->node).prio != c->default_value) { prio 190 kernel/power/qos.c (req->node).prio, state); prio 355 kernel/power/qos.c if (new_value != req->node.prio) prio 454 kernel/power/qos.c if (new_value != req->node.prio) prio 798 kernel/power/qos.c if (req->pnode.prio == new_value) prio 2138 kernel/rcu/rcutorture.c int prio; prio 2143 kernel/rcu/rcutorture.c prio = rcu_get_gp_kthreads_prio(); prio 2144 kernel/rcu/rcutorture.c if (!prio) prio 2147 kernel/rcu/rcutorture.c if (prio < 2) { prio 749 kernel/sched/core.c int prio = p->static_prio - MAX_RT_PRIO; prio 767 kernel/sched/core.c reweight_task(p, prio); prio 769 kernel/sched/core.c load->weight = scale_load(sched_prio_to_weight[prio]); prio 770 kernel/sched/core.c load->inv_weight = sched_prio_to_wmult[prio]; prio 1350 kernel/sched/core.c int prio; prio 1353 kernel/sched/core.c prio = MAX_DL_PRIO-1; prio 1355 kernel/sched/core.c prio = MAX_RT_PRIO-1 - p->rt_priority; prio 1357 kernel/sched/core.c prio = __normal_prio(p); prio 1358 kernel/sched/core.c return prio; prio 1376 kernel/sched/core.c if (!rt_prio(p->prio)) prio 1378 kernel/sched/core.c return p->prio; prio 1408 kernel/sched/core.c } else if (oldprio != p->prio || dl_task(p)) prio 2850 kernel/sched/core.c p->prio = current->normal_prio; prio 2865 kernel/sched/core.c p->prio = p->normal_prio = __normal_prio(p); prio 2875 kernel/sched/core.c if (dl_prio(p->prio)) prio 2877 kernel/sched/core.c else if (rt_prio(p->prio)) prio 4344 kernel/sched/core.c static inline int __rt_effective_prio(struct task_struct *pi_task, int prio) prio 4347 kernel/sched/core.c prio = min(prio, pi_task->prio); prio 4349 kernel/sched/core.c return prio; prio 4352 kernel/sched/core.c static inline int rt_effective_prio(struct task_struct *p, int prio) prio 4356 kernel/sched/core.c return __rt_effective_prio(pi_task, prio); prio 4372 kernel/sched/core.c int prio, oldprio, queued, running, queue_flag = prio 4379 kernel/sched/core.c prio = __rt_effective_prio(pi_task, p->normal_prio); prio 4384 kernel/sched/core.c if (p->pi_top_task == pi_task && prio == p->prio && !dl_prio(prio)) prio 4404 kernel/sched/core.c if (prio == p->prio && !dl_prio(prio)) prio 4426 kernel/sched/core.c oldprio = p->prio; prio 4428 kernel/sched/core.c if (oldprio == prio) prio 4448 kernel/sched/core.c if (dl_prio(prio)) { prio 4456 kernel/sched/core.c } else if (rt_prio(prio)) { prio 4459 kernel/sched/core.c if (oldprio < prio) prio 4470 kernel/sched/core.c p->prio = prio; prio 4487 kernel/sched/core.c static inline int rt_effective_prio(struct task_struct *p, int prio) prio 4489 kernel/sched/core.c return prio; prio 4528 kernel/sched/core.c old_prio = p->prio; prio 4529 kernel/sched/core.c p->prio = effective_prio(p); prio 4530 kernel/sched/core.c delta = p->prio - old_prio; prio 4607 kernel/sched/core.c return p->prio - MAX_RT_PRIO; prio 4721 kernel/sched/core.c p->prio = normal_prio(p); prio 4723 kernel/sched/core.c p->prio = rt_effective_prio(p, p->prio); prio 4725 kernel/sched/core.c if (dl_prio(p->prio)) prio 4727 kernel/sched/core.c else if (rt_prio(p->prio)) prio 4952 kernel/sched/core.c oldprio = p->prio; prio 4984 kernel/sched/core.c if (oldprio < p->prio) prio 7771 kernel/sched/core.c int prio, delta; prio 7774 kernel/sched/core.c for (prio = 0; prio < ARRAY_SIZE(sched_prio_to_weight); prio++) { prio 7775 kernel/sched/core.c delta = abs(sched_prio_to_weight[prio] - weight); prio 7781 kernel/sched/core.c return PRIO_TO_NICE(prio - 1 + MAX_RT_PRIO); prio 28 kernel/sched/cpupri.c static int convert_prio(int prio) prio 32 kernel/sched/cpupri.c if (prio == CPUPRI_INVALID) prio 34 kernel/sched/cpupri.c else if (prio == MAX_PRIO) prio 36 kernel/sched/cpupri.c else if (prio >= MAX_RT_PRIO) prio 39 kernel/sched/cpupri.c cpupri = MAX_RT_PRIO - prio + 1; prio 63 kernel/sched/cpupri.c int task_pri = convert_prio(p->prio); prio 1377 kernel/sched/deadline.c int prio = dl_task_of(dl_se)->prio; prio 1380 kernel/sched/deadline.c WARN_ON(!dl_prio(prio)); prio 1391 kernel/sched/deadline.c int prio = dl_task_of(dl_se)->prio; prio 1393 kernel/sched/deadline.c WARN_ON(!dl_prio(prio)); prio 445 kernel/sched/debug.c p->prio); prio 956 kernel/sched/debug.c P(prio); prio 2908 kernel/sched/fair.c void reweight_task(struct task_struct *p, int prio) prio 2913 kernel/sched/fair.c unsigned long weight = scale_load(sched_prio_to_weight[prio]); prio 2916 kernel/sched/fair.c load->inv_weight = sched_prio_to_wmult[prio]; prio 10048 kernel/sched/fair.c if (p->prio > oldprio) prio 266 kernel/sched/rt.c return rq->rt.highest_prio.curr > prev->prio; prio 377 kernel/sched/rt.c plist_node_init(&p->pushable_tasks, p->prio); prio 381 kernel/sched/rt.c if (p->prio < rq->rt.highest_prio.next) prio 382 kernel/sched/rt.c rq->rt.highest_prio.next = p->prio; prio 393 kernel/sched/rt.c rq->rt.highest_prio.next = p->prio; prio 502 kernel/sched/rt.c if (rt_rq->highest_prio.curr < curr->prio) prio 537 kernel/sched/rt.c return p->prio != p->normal_prio; prio 905 kernel/sched/rt.c return rt_task_of(rt_se)->prio; prio 1037 kernel/sched/rt.c inc_rt_prio_smp(struct rt_rq *rt_rq, int prio, int prev_prio) prio 1048 kernel/sched/rt.c if (rq->online && prio < prev_prio) prio 1049 kernel/sched/rt.c cpupri_set(&rq->rd->cpupri, rq->cpu, prio); prio 1053 kernel/sched/rt.c dec_rt_prio_smp(struct rt_rq *rt_rq, int prio, int prev_prio) prio 1071 kernel/sched/rt.c void inc_rt_prio_smp(struct rt_rq *rt_rq, int prio, int prev_prio) {} prio 1073 kernel/sched/rt.c void dec_rt_prio_smp(struct rt_rq *rt_rq, int prio, int prev_prio) {} prio 1079 kernel/sched/rt.c inc_rt_prio(struct rt_rq *rt_rq, int prio) prio 1083 kernel/sched/rt.c if (prio < prev_prio) prio 1084 kernel/sched/rt.c rt_rq->highest_prio.curr = prio; prio 1086 kernel/sched/rt.c inc_rt_prio_smp(rt_rq, prio, prev_prio); prio 1090 kernel/sched/rt.c dec_rt_prio(struct rt_rq *rt_rq, int prio) prio 1096 kernel/sched/rt.c WARN_ON(prio < prev_prio); prio 1102 kernel/sched/rt.c if (prio == prev_prio) { prio 1112 kernel/sched/rt.c dec_rt_prio_smp(rt_rq, prio, prev_prio); prio 1117 kernel/sched/rt.c static inline void inc_rt_prio(struct rt_rq *rt_rq, int prio) {} prio 1118 kernel/sched/rt.c static inline void dec_rt_prio(struct rt_rq *rt_rq, int prio) {} prio 1184 kernel/sched/rt.c int prio = rt_se_prio(rt_se); prio 1186 kernel/sched/rt.c WARN_ON(!rt_prio(prio)); prio 1190 kernel/sched/rt.c inc_rt_prio(rt_rq, prio); prio 1428 kernel/sched/rt.c curr->prio <= p->prio)) { prio 1436 kernel/sched/rt.c p->prio < cpu_rq(target)->rt.highest_prio.curr) prio 1495 kernel/sched/rt.c if (p->prio < rq->curr->prio) { prio 1513 kernel/sched/rt.c if (p->prio == rq->curr->prio && !test_tsk_need_resched(rq->curr)) prio 1725 kernel/sched/rt.c if (lowest_rq->rt.highest_prio.curr <= task->prio) { prio 1756 kernel/sched/rt.c if (lowest_rq->rt.highest_prio.curr > task->prio) prio 1814 kernel/sched/rt.c if (unlikely(next_task->prio < rq->curr->prio)) { prio 2112 kernel/sched/rt.c if (p && (p->prio < this_rq->rt.highest_prio.curr)) { prio 2124 kernel/sched/rt.c if (p->prio < src_rq->curr->prio) prio 2158 kernel/sched/rt.c rq->curr->prio <= p->prio)) prio 2233 kernel/sched/rt.c if (p->prio < rq->curr->prio && cpu_online(cpu_of(rq))) prio 2254 kernel/sched/rt.c if (oldprio < p->prio) prio 2261 kernel/sched/rt.c if (p->prio > rq->rt.highest_prio.curr) prio 2265 kernel/sched/rt.c if (oldprio < p->prio) prio 2274 kernel/sched/rt.c if (p->prio < rq->curr->prio) prio 1875 kernel/sched/sched.h extern void reweight_task(struct task_struct *p, int prio); prio 391 kernel/trace/trace_sched_wakeup.c entry->prev_prio = prev->prio; prio 394 kernel/trace/trace_sched_wakeup.c entry->next_prio = next->prio; prio 419 kernel/trace/trace_sched_wakeup.c entry->prev_prio = curr->prio; prio 422 kernel/trace/trace_sched_wakeup.c entry->next_prio = wakee->prio; prio 550 kernel/trace/trace_sched_wakeup.c (!dl_task(p) && (p->prio >= wakeup_prio || p->prio >= current->prio))) prio 563 kernel/trace/trace_sched_wakeup.c (!dl_task(p) && p->prio >= wakeup_prio)) prio 571 kernel/trace/trace_sched_wakeup.c wakeup_prio = p->prio; prio 130 kernel/tracepoint.c int prio) prio 145 kernel/tracepoint.c if (pos < 0 && old[nr_probes].prio < prio) prio 228 kernel/tracepoint.c struct tracepoint_func *func, int prio) prio 241 kernel/tracepoint.c old = func_add(&tp_funcs, func, prio); prio 306 kernel/tracepoint.c void *data, int prio) prio 314 kernel/tracepoint.c tp_func.prio = prio; prio 315 kernel/tracepoint.c ret = tracepoint_add_func(tp, &tp_func, prio); prio 44 lib/parman.c int (*item_add)(struct parman *parman, struct parman_prio *prio, prio 46 lib/parman.c void (*item_remove)(struct parman *parman, struct parman_prio *prio, prio 87 lib/parman.c static bool parman_prio_used(struct parman_prio *prio) prio 90 lib/parman.c return !list_empty(&prio->item_list); prio 93 lib/parman.c static struct parman_item *parman_prio_first_item(struct parman_prio *prio) prio 95 lib/parman.c return list_first_entry(&prio->item_list, prio 99 lib/parman.c static unsigned long parman_prio_first_index(struct parman_prio *prio) prio 101 lib/parman.c return parman_prio_first_item(prio)->index; prio 104 lib/parman.c static struct parman_item *parman_prio_last_item(struct parman_prio *prio) prio 106 lib/parman.c return list_last_entry(&prio->item_list, prio 110 lib/parman.c static unsigned long parman_prio_last_index(struct parman_prio *prio) prio 112 lib/parman.c return parman_prio_last_item(prio)->index; prio 116 lib/parman.c struct parman_prio *prio) prio 118 lib/parman.c list_for_each_entry_from_reverse(prio, &parman->prio_list, list) { prio 119 lib/parman.c if (!parman_prio_used(prio)) prio 121 lib/parman.c return parman_prio_last_index(prio) + 1; prio 126 lib/parman.c static void __parman_prio_move(struct parman *parman, struct parman_prio *prio, prio 134 lib/parman.c struct parman_prio *prio) prio 139 lib/parman.c if (!parman_prio_used(prio)) prio 141 lib/parman.c item = parman_prio_first_item(prio); prio 142 lib/parman.c to_index = parman_prio_last_index(prio) + 1; prio 143 lib/parman.c __parman_prio_move(parman, prio, item, to_index, 1); prio 144 lib/parman.c list_move_tail(&item->list, &prio->item_list); prio 149 lib/parman.c struct parman_prio *prio) prio 154 lib/parman.c if (!parman_prio_used(prio)) prio 156 lib/parman.c item = parman_prio_last_item(prio); prio 157 lib/parman.c to_index = parman_prio_first_index(prio) - 1; prio 158 lib/parman.c __parman_prio_move(parman, prio, item, to_index, 1); prio 159 lib/parman.c list_move(&item->list, &prio->item_list); prio 164 lib/parman.c struct parman_prio *prio, prio 170 lib/parman.c last_item = parman_prio_last_item(prio); prio 176 lib/parman.c __parman_prio_move(parman, prio, last_item, to_index, 1); prio 183 lib/parman.c struct parman_prio *prio, prio 196 lib/parman.c new_index = parman_lsort_new_index_find(parman, prio); prio 198 lib/parman.c if (prio2 == prio) prio 203 lib/parman.c list_add_tail(&item->list, &prio->item_list); prio 209 lib/parman.c struct parman_prio *prio, prio 212 lib/parman.c parman_prio_item_remove(parman, prio, item); prio 213 lib/parman.c list_for_each_entry_continue(prio, &parman->prio_list, list) prio 214 lib/parman.c parman_prio_shift_up(parman, prio); prio 308 lib/parman.c void parman_prio_init(struct parman *parman, struct parman_prio *prio, prio 314 lib/parman.c INIT_LIST_HEAD(&prio->item_list); prio 315 lib/parman.c prio->priority = priority; prio 320 lib/parman.c if (prio2->priority > prio->priority) prio 323 lib/parman.c list_add_tail(&prio->list, pos); prio 333 lib/parman.c void parman_prio_fini(struct parman_prio *prio) prio 335 lib/parman.c WARN_ON(parman_prio_used(prio)); prio 336 lib/parman.c list_del(&prio->list); prio 352 lib/parman.c int parman_item_add(struct parman *parman, struct parman_prio *prio, prio 355 lib/parman.c return parman->algo->item_add(parman, prio, item); prio 367 lib/parman.c void parman_item_remove(struct parman *parman, struct parman_prio *prio, prio 370 lib/parman.c parman->algo->item_remove(parman, prio, item); prio 88 lib/plist.c if (node->prio < iter->prio) { prio 98 lib/plist.c if (!prev || prev->prio != node->prio) prio 159 lib/plist.c if (node->prio != iter->prio) prio 165 lib/plist.c if (node->prio != iter->prio) { prio 198 lib/plist.c if (node_pos->prio == prio_pos->prio) { prio 203 lib/plist.c BUG_ON(prio_pos->prio > node_pos->prio); prio 217 lib/plist.c BUG_ON(node->prio == plist_next(node)->prio); prio 235 lib/plist.c test_node[i].prio = r % 99; prio 73 lib/test_parman.c struct test_parman_prio *prio; prio 153 lib/test_parman.c struct test_parman_prio *prio = &test_parman->prios[i]; prio 155 lib/test_parman.c if (prio->priority == 0) prio 157 lib/test_parman.c if (prio->priority == priority) prio 168 lib/test_parman.c struct test_parman_prio *prio = &test_parman->prios[i]; prio 171 lib/test_parman.c prio->priority = test_parman_priority_gen(test_parman); prio 172 lib/test_parman.c parman_prio_init(test_parman->parman, &prio->parman_prio, prio 173 lib/test_parman.c prio->priority); prio 182 lib/test_parman.c struct test_parman_prio *prio = &test_parman->prios[i]; prio 184 lib/test_parman.c parman_prio_fini(&prio->parman_prio); prio 198 lib/test_parman.c item->prio = &test_parman->prios[prio_index]; prio 212 lib/test_parman.c &item->prio->parman_prio, prio 282 lib/test_parman.c &item->prio->parman_prio, prio 291 lib/test_parman.c &item->prio->parman_prio, prio 329 lib/test_parman.c if (item->prio->priority < last_priority) { prio 331 lib/test_parman.c item->prio->priority, last_priority); prio 334 lib/test_parman.c last_priority = item->prio->priority; prio 2289 mm/compaction.c gfp_t gfp_mask, enum compact_priority prio, prio 2299 mm/compaction.c .mode = (prio == COMPACT_PRIO_ASYNC) ? prio 2304 mm/compaction.c .whole_zone = (prio == MIN_COMPACT_PRIORITY), prio 2305 mm/compaction.c .ignore_skip_hint = (prio == MIN_COMPACT_PRIORITY), prio 2306 mm/compaction.c .ignore_block_suitable = (prio == MIN_COMPACT_PRIORITY) prio 2341 mm/compaction.c enum compact_priority prio, struct page **capture) prio 2355 mm/compaction.c trace_mm_compaction_try_to_compact_pages(order, gfp_mask, prio); prio 2362 mm/compaction.c if (prio > MIN_COMPACT_PRIORITY prio 2368 mm/compaction.c status = compact_zone_order(zone, order, gfp_mask, prio, prio 2385 mm/compaction.c if (prio != COMPACT_PRIO_ASYNC && (status == COMPACT_COMPLETE || prio 2399 mm/compaction.c if ((prio == COMPACT_PRIO_ASYNC && need_resched()) prio 3887 mm/page_alloc.c enum compact_priority prio, enum compact_result *compact_result) prio 3900 mm/page_alloc.c prio, &page); prio 4020 mm/page_alloc.c enum compact_priority prio, enum compact_result *compact_result) prio 2421 mm/swapfile.c static void setup_swap_info(struct swap_info_struct *p, int prio, prio 2427 mm/swapfile.c if (prio >= 0) prio 2428 mm/swapfile.c p->prio = prio; prio 2430 mm/swapfile.c p->prio = --least_priority; prio 2435 mm/swapfile.c p->list.prio = -p->prio; prio 2437 mm/swapfile.c if (p->prio >= 0) prio 2438 mm/swapfile.c p->avail_lists[i].prio = -p->prio; prio 2441 mm/swapfile.c p->avail_lists[i].prio = 1; prio 2443 mm/swapfile.c p->avail_lists[i].prio = -p->prio; prio 2471 mm/swapfile.c static void enable_swap_info(struct swap_info_struct *p, int prio, prio 2479 mm/swapfile.c setup_swap_info(p, prio, swap_map, cluster_info); prio 2498 mm/swapfile.c setup_swap_info(p, p->prio, p->swap_map, p->cluster_info); prio 2566 mm/swapfile.c if (p->prio < 0) { prio 2571 mm/swapfile.c si->prio++; prio 2572 mm/swapfile.c si->list.prio--; prio 2574 mm/swapfile.c if (si->avail_lists[nid].prio != 1) prio 2575 mm/swapfile.c si->avail_lists[nid].prio--; prio 2774 mm/swapfile.c si->prio); prio 3104 mm/swapfile.c int prio; prio 3291 mm/swapfile.c prio = -1; prio 3293 mm/swapfile.c prio = prio 3295 mm/swapfile.c enable_swap_info(p, prio, swap_map, cluster_info, frontswap_map); prio 3298 mm/swapfile.c p->pages<<(PAGE_SHIFT-10), name->name, p->prio, prio 323 mm/vmpressure.c void vmpressure_prio(gfp_t gfp, struct mem_cgroup *memcg, int prio) prio 329 mm/vmpressure.c if (prio > vmpressure_level_critical_prio) prio 386 net/batman-adv/main.c u32 prio; prio 402 net/batman-adv/main.c prio = ntohs(vhdr->h_vlan_TCI) & VLAN_PRIO_MASK; prio 403 net/batman-adv/main.c prio = prio >> VLAN_PRIO_SHIFT; prio 410 net/batman-adv/main.c prio = (ipv4_get_dsfield(ip_hdr) & 0xfc) >> 5; prio 417 net/batman-adv/main.c prio = (ipv6_get_dsfield(ip6_hdr) & 0xfc) >> 5; prio 423 net/batman-adv/main.c skb->priority = prio + 256; prio 47 net/bluetooth/l2cap_sock.c int proto, gfp_t prio, int kern); prio 1588 net/bluetooth/l2cap_sock.c int proto, gfp_t prio, int kern) prio 1593 net/bluetooth/l2cap_sock.c sk = sk_alloc(net, PF_BLUETOOTH, prio, &l2cap_proto, kern); prio 303 net/bluetooth/rfcomm/core.c struct rfcomm_dlc *rfcomm_dlc_alloc(gfp_t prio) prio 305 net/bluetooth/rfcomm/core.c struct rfcomm_dlc *d = kzalloc(sizeof(*d), prio); prio 274 net/bluetooth/rfcomm/sock.c static struct sock *rfcomm_sock_alloc(struct net *net, struct socket *sock, int proto, gfp_t prio, int kern) prio 279 net/bluetooth/rfcomm/sock.c sk = sk_alloc(net, PF_BLUETOOTH, prio, &rfcomm_proto, kern); prio 286 net/bluetooth/rfcomm/sock.c d = rfcomm_dlc_alloc(prio); prio 470 net/bluetooth/sco.c int proto, gfp_t prio, int kern) prio 474 net/bluetooth/sco.c sk = sk_alloc(net, PF_BLUETOOTH, prio, &sco_proto, kern); prio 437 net/bridge/br_device.c br->bridge_id.prio[0] = 0x80; prio 438 net/bridge/br_device.c br->bridge_id.prio[1] = 0x00; prio 1395 net/bridge/br_netlink.c u16 priority = (br->bridge_id.prio[0] << 8) | br->bridge_id.prio[1]; prio 56 net/bridge/br_private.h unsigned char prio[2]; prio 92 net/bridge/br_stp_bpdu.c buf[5] = bpdu->root.prio[0]; prio 93 net/bridge/br_stp_bpdu.c buf[6] = bpdu->root.prio[1]; prio 104 net/bridge/br_stp_bpdu.c buf[17] = bpdu->bridge_id.prio[0]; prio 105 net/bridge/br_stp_bpdu.c buf[18] = bpdu->bridge_id.prio[1]; prio 196 net/bridge/br_stp_bpdu.c bpdu.root.prio[0] = buf[2]; prio 197 net/bridge/br_stp_bpdu.c bpdu.root.prio[1] = buf[3]; prio 209 net/bridge/br_stp_bpdu.c bpdu.bridge_id.prio[0] = buf[14]; prio 210 net/bridge/br_stp_bpdu.c bpdu.bridge_id.prio[1] = buf[15]; prio 284 net/bridge/br_stp_if.c p->designated_bridge.prio[0] = (newprio >> 8) & 0xFF; prio 285 net/bridge/br_stp_if.c p->designated_bridge.prio[1] = newprio & 0xFF; prio 290 net/bridge/br_stp_if.c br->bridge_id.prio[0] = (newprio >> 8) & 0xFF; prio 291 net/bridge/br_stp_if.c br->bridge_id.prio[1] = newprio & 0xFF; prio 339 net/bridge/br_stp_if.c id->prio[0], id->prio[1], prio 58 net/bridge/br_stp_timer.c id->prio[0], id->prio[1], &id->addr); prio 174 net/bridge/br_sysfs_br.c (br->bridge_id.prio[0] << 8) | br->bridge_id.prio[1]); prio 32 net/bridge/netfilter/ebt_vlan.c unsigned char prio; /* user_priority, given from frame TCI */ prio 60 net/bridge/netfilter/ebt_vlan.c prio = (TCI >> 13) & 0x7; prio 68 net/bridge/netfilter/ebt_vlan.c EXIT_ON_MISMATCH(prio, EBT_VLAN_PRIO); prio 130 net/bridge/netfilter/ebt_vlan.c if ((unsigned char) info->prio > 7) { prio 132 net/bridge/netfilter/ebt_vlan.c info->prio); prio 378 net/caif/cfctrl.c u8 prio; prio 395 net/caif/cfctrl.c prio = tmp >> 3; prio 397 net/caif/cfctrl.c linkparam.priority = prio; prio 378 net/caif/cfpkt_skbuff.c void cfpkt_set_prio(struct cfpkt *pkt, int prio) prio 380 net/caif/cfpkt_skbuff.c pkt_to_skb(pkt)->priority = prio; prio 34 net/can/j1939/socket.c static inline u32 j1939_to_sk_priority(priority_t prio) prio 36 net/can/j1939/socket.c return 7 - prio; prio 113 net/core/netprio_cgroup.c struct net_device *dev, u32 prio) prio 121 net/core/netprio_cgroup.c if (!prio && (!map || map->priomap_len <= id)) prio 129 net/core/netprio_cgroup.c map->priomap[id] = prio; prio 163 net/core/netprio_cgroup.c u32 prio = netprio_prio(parent_css, dev); prio 165 net/core/netprio_cgroup.c ret = netprio_set_prio(css, dev, prio); prio 200 net/core/netprio_cgroup.c u32 prio; prio 203 net/core/netprio_cgroup.c if (sscanf(buf, "%"__stringify(IFNAMSIZ)"s %u", devname, &prio) != 2) prio 214 net/core/netprio_cgroup.c ret = netprio_set_prio(of_css(of), dev, prio); prio 2608 net/core/pktgen.c unsigned int prio) prio 2610 net/core/pktgen.c return htons(id | (cfi << 12) | (prio << 13)); prio 557 net/dcb/dcbnl.c u8 prio, pgid, tc_pct, up_map; prio 603 net/dcb/dcbnl.c prio = DCB_ATTR_VALUE_UNDEFINED; prio 610 net/dcb/dcbnl.c i - DCB_PG_ATTR_TC_0, &prio, prio 615 net/dcb/dcbnl.c i - DCB_PG_ATTR_TC_0, &prio, prio 636 net/dcb/dcbnl.c DCB_TC_ATTR_PARAM_STRICT_PRIO, prio); prio 775 net/dcb/dcbnl.c u8 prio; prio 805 net/dcb/dcbnl.c prio = DCB_ATTR_VALUE_UNDEFINED; prio 810 net/dcb/dcbnl.c prio = prio 828 net/dcb/dcbnl.c prio, pgid, tc_pct, up_map); prio 833 net/dcb/dcbnl.c prio, pgid, tc_pct, up_map); prio 1177 net/dcb/dcbnl.c u8 pgid, up_map, prio, tc_pct; prio 1192 net/dcb/dcbnl.c prio = DCB_ATTR_VALUE_UNDEFINED; prio 1198 net/dcb/dcbnl.c &prio, &pgid, &tc_pct, &up_map); prio 1201 net/dcb/dcbnl.c &prio, &pgid, &tc_pct, &up_map); prio 1205 net/dcb/dcbnl.c nla_put_u8(skb, DCB_TC_ATTR_PARAM_STRICT_PRIO, prio) || prio 1790 net/dcb/dcbnl.c int ifindex, int prio) prio 1798 net/dcb/dcbnl.c ((prio == -1) || itr->app.priority == prio)) prio 1830 net/dcb/dcbnl.c u8 prio = 0; prio 1835 net/dcb/dcbnl.c prio = itr->app.priority; prio 1838 net/dcb/dcbnl.c return prio; prio 1893 net/dcb/dcbnl.c u8 prio = 0; prio 1898 net/dcb/dcbnl.c prio |= 1 << itr->app.priority; prio 1901 net/dcb/dcbnl.c return prio; prio 1981 net/dcb/dcbnl.c u8 prio; prio 1991 net/dcb/dcbnl.c prio = itr->app.priority; prio 1992 net/dcb/dcbnl.c p_map->map[prio] |= 1ULL << itr->app.protocol; prio 671 net/dccp/output.c const gfp_t prio = active ? GFP_KERNEL : GFP_ATOMIC; prio 673 net/dccp/output.c skb = alloc_skb(sk->sk_prot->max_header, prio); prio 983 net/ipv4/fib_trie.c u8 tos, u32 prio, u32 tb_id) prio 1001 net/ipv4/fib_trie.c if (fa->fa_info->fib_priority >= prio || fa->fa_tos < tos) prio 4603 net/ipv6/addrconf.c u32 prio; prio 4611 net/ipv6/addrconf.c prio = ifp->rt_priority ? : IP6_RT_PRIO_ADDRCONF; prio 4612 net/ipv6/addrconf.c if (f6i->fib6_metric != prio) { prio 254 net/ipv6/ip6_gre.c int prio = 0; prio 257 net/ipv6/ip6_gre.c prio |= 1; prio 259 net/ipv6/ip6_gre.c prio |= 2; prio 263 net/ipv6/ip6_gre.c return &ign->tunnels[prio][h]; prio 1529 net/ipv6/ip6_gre.c int prio; prio 1537 net/ipv6/ip6_gre.c for (prio = 0; prio < 4; prio++) { prio 1542 net/ipv6/ip6_gre.c t = rtnl_dereference(ign->tunnels[prio][h]); prio 196 net/ipv6/ip6_tunnel.c int prio = 0; prio 199 net/ipv6/ip6_tunnel.c prio = 1; prio 202 net/ipv6/ip6_tunnel.c return &ip6n->tnls[prio][h]; prio 85 net/ipv6/ip6_udp_tunnel.c __u8 prio, __u8 ttl, __be32 label, prio 107 net/ipv6/ip6_udp_tunnel.c ip6_flow_hdr(ip6h, prio, label); prio 142 net/ipv6/ip6_vti.c int prio = 0; prio 145 net/ipv6/ip6_vti.c prio = 1; prio 148 net/ipv6/ip6_vti.c return &ip6n->tnls[prio][h]; prio 134 net/ipv6/sit.c int prio = 0; prio 137 net/ipv6/sit.c prio |= 2; prio 141 net/ipv6/sit.c prio |= 1; prio 144 net/ipv6/sit.c return &sitn->tunnels[prio][h]; prio 1815 net/ipv6/sit.c int prio; prio 1821 net/ipv6/sit.c for (prio = 1; prio < 4; prio++) { prio 1826 net/ipv6/sit.c t = rtnl_dereference(sitn->tunnels[prio][h]); prio 567 net/iucv/af_iucv.c static struct sock *iucv_sock_alloc(struct socket *sock, int proto, gfp_t prio, int kern) prio 572 net/iucv/af_iucv.c sk = sk_alloc(&init_net, PF_IUCV, prio, &iucv_proto, kern); prio 168 net/netfilter/core.c int prio = INT_MIN; prio 177 net/netfilter/core.c WARN_ON(orig_ops[i]->priority < prio); prio 179 net/netfilter/core.c if (orig_ops[i]->priority > prio) prio 180 net/netfilter/core.c prio = orig_ops[i]->priority; prio 134 net/netfilter/nf_tables_offload.c common->prio = priority; prio 53 net/sched/cls_api.c return jhash_3words(tp->chain->index, tp->prio, prio 72 net/sched/cls_api.c tp1->prio == tp2->prio && prio 217 net/sched/cls_api.c first = tp->prio - 1; prio 251 net/sched/cls_api.c u32 prio, struct tcf_chain *chain, prio 269 net/sched/cls_api.c tp->prio = prio; prio 990 net/sched/cls_api.c u32 prio = 0; prio 1002 net/sched/cls_api.c prio = tp->prio + 1; prio 1006 net/sched/cls_api.c if (!tp->deleting && tp->prio >= prio) prio 1613 net/sched/cls_api.c tp->prio & 0xffff, prio 1665 net/sched/cls_api.c u32 protocol, u32 prio, prio 1675 net/sched/cls_api.c u32 protocol, u32 prio, prio 1691 net/sched/cls_api.c protocol, prio, false); prio 1750 net/sched/cls_api.c u32 protocol, u32 prio, prio 1760 net/sched/cls_api.c if (tp->prio >= prio) { prio 1761 net/sched/cls_api.c if (tp->prio == prio) { prio 1805 net/sched/cls_api.c tcm->tcm_info = TC_H_MAKE(tp->prio, tp->protocol); prio 1927 net/sched/cls_api.c u32 prio; prio 1955 net/sched/cls_api.c prio = TC_H_MAJ(t->tcm_info); prio 1962 net/sched/cls_api.c if (prio == 0) { prio 1967 net/sched/cls_api.c prio = TC_H_MAKE(0x80000000U, 0U); prio 2025 net/sched/cls_api.c prio, prio_allocate); prio 2055 net/sched/cls_api.c prio = tcf_auto_prio(tcf_chain_tp_prev(chain, prio 2059 net/sched/cls_api.c tp_new = tcf_proto_create(name, protocol, prio, chain, prio 2067 net/sched/cls_api.c tp = tcf_chain_tp_insert_unique(chain, tp_new, protocol, prio, prio 2154 net/sched/cls_api.c u32 prio; prio 2177 net/sched/cls_api.c prio = TC_H_MAJ(t->tcm_info); prio 2180 net/sched/cls_api.c if (prio == 0 && (protocol || t->tcm_handle || tca[TCA_KIND])) { prio 2200 net/sched/cls_api.c if (!prio || prio 2229 net/sched/cls_api.c if (prio == 0) { prio 2238 net/sched/cls_api.c if (prio == 0) { prio 2248 net/sched/cls_api.c prio, false); prio 2314 net/sched/cls_api.c u32 prio; prio 2334 net/sched/cls_api.c prio = TC_H_MAJ(t->tcm_info); prio 2337 net/sched/cls_api.c if (prio == 0) { prio 2389 net/sched/cls_api.c prio, false); prio 2467 net/sched/cls_api.c TC_H_MAJ(tcm->tcm_info) != tp->prio) prio 3479 net/sched/cls_api.c entry->vlan.prio = tcf_vlan_push_prio(act); prio 3488 net/sched/cls_api.c entry->vlan.prio = tcf_vlan_push_prio(act); prio 72 net/sched/cls_u32.c u32 prio; prio 362 net/sched/cls_u32.c root_ht->prio = tp->prio; prio 481 net/sched/cls_u32.c cls_u32.hnode.prio = h->prio; prio 499 net/sched/cls_u32.c cls_u32.hnode.prio = h->prio; prio 961 net/sched/cls_u32.c ht->prio = tp->prio; prio 1125 net/sched/cls_u32.c if (ht->prio != tp->prio) prio 1163 net/sched/cls_u32.c cls_u32.hnode.prio = ht->prio; prio 1223 net/sched/cls_u32.c if (ht->prio != tp->prio) prio 211 net/sched/sch_cbq.c u32 prio = skb->priority; prio 218 net/sched/sch_cbq.c if (TC_H_MAJ(prio ^ sch->handle) == 0 && prio 219 net/sched/sch_cbq.c (cl = cbq_class_lookup(q, prio)) != NULL) prio 277 net/sched/sch_cbq.c if (TC_H_MAJ(prio) == 0 && prio 278 net/sched/sch_cbq.c !(cl = head->defaults[prio & TC_PRIO_MAX]) && prio 294 net/sched/sch_cbq.c int prio = cl->cpriority; prio 297 net/sched/sch_cbq.c cl_tail = q->active[prio]; prio 298 net/sched/sch_cbq.c q->active[prio] = cl; prio 305 net/sched/sch_cbq.c q->activemask |= (1<<prio); prio 318 net/sched/sch_cbq.c int prio = this->cpriority; prio 320 net/sched/sch_cbq.c struct cbq_class *cl_prev = q->active[prio]; prio 328 net/sched/sch_cbq.c if (cl == q->active[prio]) { prio 329 net/sched/sch_cbq.c q->active[prio] = cl_prev; prio 330 net/sched/sch_cbq.c if (cl == q->active[prio]) { prio 331 net/sched/sch_cbq.c q->active[prio] = NULL; prio 332 net/sched/sch_cbq.c q->activemask &= ~(1<<prio); prio 338 net/sched/sch_cbq.c } while ((cl_prev = cl) != q->active[prio]); prio 444 net/sched/sch_cbq.c static psched_tdiff_t cbq_undelay_prio(struct cbq_sched_data *q, int prio, prio 448 net/sched/sch_cbq.c struct cbq_class *cl_prev = q->active[prio]; prio 463 net/sched/sch_cbq.c if (cl == q->active[prio]) { prio 464 net/sched/sch_cbq.c q->active[prio] = cl_prev; prio 465 net/sched/sch_cbq.c if (cl == q->active[prio]) { prio 466 net/sched/sch_cbq.c q->active[prio] = NULL; prio 474 net/sched/sch_cbq.c } while ((cl_prev = cl) != q->active[prio]); prio 494 net/sched/sch_cbq.c int prio = ffz(~pmask); prio 497 net/sched/sch_cbq.c pmask &= ~(1<<prio); prio 499 net/sched/sch_cbq.c tmp = cbq_undelay_prio(q, prio, now); prio 501 net/sched/sch_cbq.c q->pmask |= 1<<prio; prio 679 net/sched/sch_cbq.c cbq_dequeue_prio(struct Qdisc *sch, int prio) prio 686 net/sched/sch_cbq.c cl_tail = cl_prev = q->active[prio]; prio 733 net/sched/sch_cbq.c q->active[prio] = cl; prio 740 net/sched/sch_cbq.c if (cl->q->q.qlen == 0 || prio != cl->cpriority) { prio 755 net/sched/sch_cbq.c q->active[prio] = NULL; prio 756 net/sched/sch_cbq.c q->activemask &= ~(1<<prio); prio 762 net/sched/sch_cbq.c q->active[prio] = cl_tail; prio 776 net/sched/sch_cbq.c q->active[prio] = cl_prev; prio 790 net/sched/sch_cbq.c int prio = ffz(~activemask); prio 791 net/sched/sch_cbq.c activemask &= ~(1<<prio); prio 792 net/sched/sch_cbq.c skb = cbq_dequeue_prio(sch, prio); prio 884 net/sched/sch_cbq.c static void cbq_normalize_quanta(struct cbq_sched_data *q, int prio) prio 889 net/sched/sch_cbq.c if (q->quanta[prio] == 0) prio 897 net/sched/sch_cbq.c if (cl->priority == prio) { prio 898 net/sched/sch_cbq.c cl->quantum = (cl->weight*cl->allot*q->nclasses[prio])/ prio 899 net/sched/sch_cbq.c q->quanta[prio]; prio 1030 net/sched/sch_cbq.c int prio; prio 1042 net/sched/sch_cbq.c for (prio = 0; prio <= TC_CBQ_MAXPRIO; prio++) prio 1043 net/sched/sch_cbq.c q->active[prio] = NULL; prio 729 net/sched/sch_generic.c int prio; prio 735 net/sched/sch_generic.c for (prio = 0; prio < PFIFO_FAST_BANDS; prio++) { prio 736 net/sched/sch_generic.c struct skb_array *q = band2list(priv, prio); prio 752 net/sched/sch_generic.c int prio; prio 754 net/sched/sch_generic.c for (prio = 0; prio < PFIFO_FAST_BANDS; prio++) { prio 755 net/sched/sch_generic.c struct skb_array *q = band2list(priv, prio); prio 774 net/sched/sch_generic.c int prio; prio 776 net/sched/sch_generic.c for (prio = 0; prio < PFIFO_FAST_BANDS; prio++) { prio 777 net/sched/sch_generic.c struct skb_array *q = band2list(priv, prio); prio 779 net/sched/sch_generic.c bands[prio] = q; prio 40 net/sched/sch_gred.c u8 prio; /* the prio of this vq */ prio 105 net/sched/sch_gred.c if (table->tab[n] && table->tab[n]->prio == q->prio) prio 198 net/sched/sch_gred.c if (t->tab[i] && t->tab[i]->prio < q->prio && prio 338 net/sched/sch_gred.c opt.set.tab[i].prio = q->prio; prio 475 net/sched/sch_gred.c struct tc_gred_qopt *ctl, int prio, prio 497 net/sched/sch_gred.c q->prio = prio; prio 641 net/sched/sch_gred.c int err, prio = GRED_DEF_PRIO; prio 685 net/sched/sch_gred.c if (ctl->prio == 0) { prio 689 net/sched/sch_gred.c def_prio = table->tab[table->def]->prio; prio 694 net/sched/sch_gred.c prio = def_prio; prio 696 net/sched/sch_gred.c prio = ctl->prio; prio 702 net/sched/sch_gred.c err = gred_change_vq(sch, ctl->DP, ctl, prio, stab, max_P, &prealloc, prio 814 net/sched/sch_gred.c opt.prio = q->prio; prio 100 net/sched/sch_htb.c u32 prio; /* these two are used only by leaves... */ prio 275 net/sched/sch_htb.c struct htb_class *cl, int prio) prio 282 net/sched/sch_htb.c c = rb_entry(parent, struct htb_class, node[prio]); prio 289 net/sched/sch_htb.c rb_link_node(&cl->node[prio], parent, p); prio 290 net/sched/sch_htb.c rb_insert_color(&cl->node[prio], root); prio 348 net/sched/sch_htb.c int prio = ffz(~mask); prio 349 net/sched/sch_htb.c mask &= ~(1 << prio); prio 350 net/sched/sch_htb.c htb_add_to_id_tree(&q->hlevel[cl->level].hprio[prio].row, cl, prio); prio 379 net/sched/sch_htb.c int prio = ffz(~mask); prio 380 net/sched/sch_htb.c struct htb_prio *hprio = &hlevel->hprio[prio]; prio 382 net/sched/sch_htb.c mask &= ~(1 << prio); prio 383 net/sched/sch_htb.c if (hprio->ptr == cl->node + prio) prio 386 net/sched/sch_htb.c htb_safe_rb_erase(cl->node + prio, &hprio->row); prio 388 net/sched/sch_htb.c m |= 1 << prio; prio 408 net/sched/sch_htb.c int prio = ffz(~m); prio 409 net/sched/sch_htb.c m &= ~(1 << prio); prio 411 net/sched/sch_htb.c if (p->inner.clprio[prio].feed.rb_node) prio 415 net/sched/sch_htb.c mask &= ~(1 << prio); prio 417 net/sched/sch_htb.c htb_add_to_id_tree(&p->inner.clprio[prio].feed, cl, prio); prio 444 net/sched/sch_htb.c int prio = ffz(~m); prio 445 net/sched/sch_htb.c m &= ~(1 << prio); prio 447 net/sched/sch_htb.c if (p->inner.clprio[prio].ptr == cl->node + prio) { prio 452 net/sched/sch_htb.c p->inner.clprio[prio].last_ptr_id = cl->common.classid; prio 453 net/sched/sch_htb.c p->inner.clprio[prio].ptr = NULL; prio 456 net/sched/sch_htb.c htb_safe_rb_erase(cl->node + prio, prio 457 net/sched/sch_htb.c &p->inner.clprio[prio].feed); prio 459 net/sched/sch_htb.c if (!p->inner.clprio[prio].feed.rb_node) prio 460 net/sched/sch_htb.c mask |= 1 << prio; prio 560 net/sched/sch_htb.c cl->prio_activity = 1 << cl->prio; prio 741 net/sched/sch_htb.c static struct rb_node *htb_id_find_next_upper(int prio, struct rb_node *n, prio 747 net/sched/sch_htb.c rb_entry(n, struct htb_class, node[prio]); prio 766 net/sched/sch_htb.c static struct htb_class *htb_lookup_leaf(struct htb_prio *hprio, const int prio) prio 786 net/sched/sch_htb.c htb_id_find_next_upper(prio, sp->root, *sp->pid); prio 807 net/sched/sch_htb.c cl = rb_entry(*sp->pptr, struct htb_class, node[prio]); prio 810 net/sched/sch_htb.c clp = &cl->inner.clprio[prio]; prio 823 net/sched/sch_htb.c static struct sk_buff *htb_dequeue_tree(struct htb_sched *q, const int prio, prio 829 net/sched/sch_htb.c struct htb_prio *hprio = &hlevel->hprio[prio]; prio 832 net/sched/sch_htb.c start = cl = htb_lookup_leaf(hprio, prio); prio 849 net/sched/sch_htb.c if ((q->row_mask[level] & (1 << prio)) == 0) prio 852 net/sched/sch_htb.c next = htb_lookup_leaf(hprio, prio); prio 865 net/sched/sch_htb.c htb_next_rb_node(level ? &cl->parent->inner.clprio[prio].ptr: prio 866 net/sched/sch_htb.c &q->hlevel[0].hprio[prio].ptr); prio 867 net/sched/sch_htb.c cl = htb_lookup_leaf(hprio, prio); prio 876 net/sched/sch_htb.c htb_next_rb_node(level ? &cl->parent->inner.clprio[prio].ptr : prio 877 net/sched/sch_htb.c &q->hlevel[0].hprio[prio].ptr); prio 931 net/sched/sch_htb.c int prio = ffz(m); prio 933 net/sched/sch_htb.c m |= 1 << prio; prio 934 net/sched/sch_htb.c skb = htb_dequeue_tree(q, prio, level); prio 1100 net/sched/sch_htb.c opt.prio = cl->prio; prio 1340 net/sched/sch_htb.c int prio; prio 1392 net/sched/sch_htb.c for (prio = 0; prio < TC_HTB_NUMPRIO; prio++) prio 1393 net/sched/sch_htb.c RB_CLEAR_NODE(&cl->node[prio]); prio 1476 net/sched/sch_htb.c if ((cl->prio = hopt->prio) >= TC_HTB_NUMPRIO) prio 1477 net/sched/sch_htb.c cl->prio = TC_HTB_NUMPRIO - 1; prio 100 net/sched/sch_prio.c int prio; prio 102 net/sched/sch_prio.c for (prio = 0; prio < q->bands; prio++) { prio 103 net/sched/sch_prio.c struct Qdisc *qdisc = q->queues[prio]; prio 114 net/sched/sch_prio.c int prio; prio 116 net/sched/sch_prio.c for (prio = 0; prio < q->bands; prio++) { prio 117 net/sched/sch_prio.c struct Qdisc *qdisc = q->queues[prio]; prio 133 net/sched/sch_prio.c int prio; prio 136 net/sched/sch_prio.c for (prio = 0; prio < q->bands; prio++) prio 137 net/sched/sch_prio.c qdisc_reset(q->queues[prio]); prio 169 net/sched/sch_prio.c int prio; prio 174 net/sched/sch_prio.c for (prio = 0; prio < q->bands; prio++) prio 175 net/sched/sch_prio.c qdisc_put(q->queues[prio]); prio 375 net/sched/sch_prio.c int prio; prio 380 net/sched/sch_prio.c for (prio = 0; prio < q->bands; prio++) { prio 385 net/sched/sch_prio.c if (arg->fn(sch, prio + 1, arg) < 0) { prio 42 net/sched/sch_skbprio.c int prio; prio 44 net/sched/sch_skbprio.c for (prio = q->highest_prio - 1; prio >= q->lowest_prio; prio--) { prio 45 net/sched/sch_skbprio.c if (!skb_queue_empty(&q->qdiscs[prio])) prio 46 net/sched/sch_skbprio.c return prio; prio 55 net/sched/sch_skbprio.c int prio; prio 57 net/sched/sch_skbprio.c for (prio = q->lowest_prio + 1; prio <= q->highest_prio; prio++) { prio 58 net/sched/sch_skbprio.c if (!skb_queue_empty(&q->qdiscs[prio])) prio 59 net/sched/sch_skbprio.c return prio; prio 76 net/sched/sch_skbprio.c u16 prio, lp; prio 79 net/sched/sch_skbprio.c prio = min(skb->priority, max_priority); prio 81 net/sched/sch_skbprio.c qdisc = &q->qdiscs[prio]; prio 85 net/sched/sch_skbprio.c q->qstats[prio].backlog += qdisc_pkt_len(skb); prio 88 net/sched/sch_skbprio.c if (prio > q->highest_prio) prio 89 net/sched/sch_skbprio.c q->highest_prio = prio; prio 91 net/sched/sch_skbprio.c if (prio < q->lowest_prio) prio 92 net/sched/sch_skbprio.c q->lowest_prio = prio; prio 100 net/sched/sch_skbprio.c if (prio <= lp) { prio 101 net/sched/sch_skbprio.c q->qstats[prio].drops++; prio 102 net/sched/sch_skbprio.c q->qstats[prio].overlimits++; prio 108 net/sched/sch_skbprio.c q->qstats[prio].backlog += qdisc_pkt_len(skb); prio 126 net/sched/sch_skbprio.c q->lowest_prio = prio; prio 127 net/sched/sch_skbprio.c q->highest_prio = prio; prio 133 net/sched/sch_skbprio.c if (prio > q->highest_prio) prio 134 net/sched/sch_skbprio.c q->highest_prio = prio; prio 183 net/sched/sch_skbprio.c int prio; prio 186 net/sched/sch_skbprio.c for (prio = 0; prio < SKBPRIO_MAX_PRIORITY; prio++) prio 187 net/sched/sch_skbprio.c __skb_queue_head_init(&q->qdiscs[prio]); prio 214 net/sched/sch_skbprio.c int prio; prio 219 net/sched/sch_skbprio.c for (prio = 0; prio < SKBPRIO_MAX_PRIORITY; prio++) prio 220 net/sched/sch_skbprio.c __skb_queue_purge(&q->qdiscs[prio]); prio 230 net/sched/sch_skbprio.c int prio; prio 232 net/sched/sch_skbprio.c for (prio = 0; prio < SKBPRIO_MAX_PRIORITY; prio++) prio 233 net/sched/sch_skbprio.c __skb_queue_purge(&q->qdiscs[prio]); prio 460 net/sched/sch_taprio.c int prio; prio 473 net/sched/sch_taprio.c prio = skb->priority; prio 474 net/sched/sch_taprio.c tc = netdev_get_prio_tc_map(dev, prio); prio 546 net/sched/sch_taprio.c int prio; prio 564 net/sched/sch_taprio.c prio = skb->priority; prio 565 net/sched/sch_taprio.c tc = netdev_get_prio_tc_map(dev, prio); prio 29 net/sctp/stream_sched_prio.c struct sctp_stream *stream, int prio, gfp_t gfp) prio 40 net/sctp/stream_sched_prio.c p->prio = prio; prio 46 net/sctp/stream_sched_prio.c struct sctp_stream *stream, int prio, gfp_t gfp) prio 55 net/sctp/stream_sched_prio.c if (p->prio == prio) prio 57 net/sctp/stream_sched_prio.c if (p->prio > prio) prio 72 net/sctp/stream_sched_prio.c if (p->prio == prio) prio 77 net/sctp/stream_sched_prio.c return sctp_sched_prio_new_head(stream, prio, gfp); prio 120 net/sctp/stream_sched_prio.c struct sctp_stream_priorities *prio, *prio_head; prio 140 net/sctp/stream_sched_prio.c list_for_each_entry(prio, &stream->prio_list, prio_sched) { prio 141 net/sctp/stream_sched_prio.c if (prio->prio > prio_head->prio) { prio 142 net/sctp/stream_sched_prio.c list_add(&prio_head->prio_sched, prio->prio_sched.prev); prio 151 net/sctp/stream_sched_prio.c __u16 prio, gfp_t gfp) prio 159 net/sctp/stream_sched_prio.c prio_head = sctp_sched_prio_get_head(stream, prio, gfp); prio 189 net/sctp/stream_sched_prio.c *value = SCTP_SO(stream, sid)->ext->prio_head->prio; prio 209 net/sctp/stream_sched_prio.c struct sctp_stream_priorities *prio, *n; prio 223 net/sctp/stream_sched_prio.c prio = SCTP_SO(stream, i)->ext->prio_head; prio 224 net/sctp/stream_sched_prio.c if (prio && list_empty(&prio->prio_sched)) prio 225 net/sctp/stream_sched_prio.c list_add(&prio->prio_sched, &list); prio 227 net/sctp/stream_sched_prio.c list_for_each_entry_safe(prio, n, &list, prio_sched) { prio 228 net/sctp/stream_sched_prio.c list_del_init(&prio->prio_sched); prio 229 net/sctp/stream_sched_prio.c kfree(prio); prio 249 net/sctp/stream_sched_prio.c struct sctp_stream_priorities *prio; prio 263 net/sctp/stream_sched_prio.c prio = list_entry(stream->prio_list.next, prio 265 net/sctp/stream_sched_prio.c soute = prio->next; prio 277 net/sctp/stream_sched_prio.c struct sctp_stream_priorities *prio; prio 286 net/sctp/stream_sched_prio.c prio = soute->prio_head; prio 288 net/sctp/stream_sched_prio.c sctp_sched_prio_next_stream(prio); prio 66 net/sctp/stream_sched_rr.c __u16 prio, gfp_t gfp) prio 235 net/tipc/bearer.c u32 disc_domain, u32 prio, prio 253 net/tipc/bearer.c if (prio > TIPC_MAX_LINK_PRI && prio != TIPC_MEDIA_LINK_PRI) { prio 264 net/tipc/bearer.c if (prio == TIPC_MEDIA_LINK_PRI) prio 265 net/tipc/bearer.c prio = m->priority; prio 277 net/tipc/bearer.c if (b->priority != prio) prio 282 net/tipc/bearer.c name, prio); prio 283 net/tipc/bearer.c if (prio == TIPC_MIN_LINK_PRI) { prio 288 net/tipc/bearer.c prio--; prio 316 net/tipc/bearer.c b->priority = prio; prio 335 net/tipc/bearer.c pr_info("Enabled bearer <%s>, priority %u\n", name, prio); prio 930 net/tipc/bearer.c u32 prio; prio 932 net/tipc/bearer.c prio = TIPC_MEDIA_LINK_PRI; prio 960 net/tipc/bearer.c prio = nla_get_u32(props[TIPC_NLA_PROP_PRIO]); prio 963 net/tipc/bearer.c return tipc_enable_bearer(net, bearer, domain, prio, attrs); prio 2342 net/tipc/link.c u32 prio; prio 2344 net/tipc/link.c prio = nla_get_u32(props[TIPC_NLA_PROP_PRIO]); prio 2345 net/tipc/link.c if (prio > TIPC_MAX_LINK_PRI) prio 2638 net/tipc/link.c void tipc_link_set_prio(struct tipc_link *l, u32 prio, prio 2641 net/tipc/link.c l->priority = prio; prio 2642 net/tipc/link.c tipc_link_build_proto_msg(l, STATE_MSG, 0, 0, 0, 0, prio, xmitq); prio 124 net/tipc/link.h void tipc_link_set_prio(struct tipc_link *l, u32 prio, prio 811 net/tipc/node.c int i, highest = 0, prio; prio 836 net/tipc/node.c prio = tipc_link_prio(_l); prio 837 net/tipc/node.c if (prio < highest) prio 839 net/tipc/node.c if (prio > highest) { prio 840 net/tipc/node.c highest = prio; prio 2138 net/tipc/node.c u32 prio; prio 2140 net/tipc/node.c prio = nla_get_u32(props[TIPC_NLA_PROP_PRIO]); prio 2141 net/tipc/node.c tipc_link_set_prio(link, prio, &xmitq); prio 30 samples/bpf/test_cgrp2_sock.c static int prog_load(__u32 idx, __u32 mark, __u32 prio) prio 69 samples/bpf/test_cgrp2_sock.c BPF_MOV64_IMM(BPF_REG_3, prio), prio 86 samples/bpf/test_cgrp2_sock.c if (prio) prio 108 samples/bpf/test_cgrp2_sock.c if (prio) { prio 154 samples/bpf/test_cgrp2_sock.c unsigned int prio = 0; prio 155 samples/bpf/test_cgrp2_sock.c socklen_t optlen = sizeof(prio); prio 158 samples/bpf/test_cgrp2_sock.c rc = getsockopt(sd, SOL_SOCKET, SO_PRIORITY, &prio, &optlen); prio 162 samples/bpf/test_cgrp2_sock.c return prio; prio 167 samples/bpf/test_cgrp2_sock.c unsigned int mark, prio; prio 181 samples/bpf/test_cgrp2_sock.c prio = get_priority(sd); prio 185 samples/bpf/test_cgrp2_sock.c printf("sd %d: dev %s, mark %u, priority %u\n", sd, name, mark, prio); prio 206 samples/bpf/test_cgrp2_sock.c __u32 idx = 0, mark = 0, prio = 0; prio 232 samples/bpf/test_cgrp2_sock.c prio = strtoumax(optarg, NULL, 0); prio 251 samples/bpf/test_cgrp2_sock.c if (do_attach && !idx && !mark && !prio) { prio 264 samples/bpf/test_cgrp2_sock.c prog_fd = prog_load(idx, mark, prio); prio 298 tools/include/uapi/linux/if_link.h __u8 prio[2]; prio 340 tools/include/uapi/linux/pkt_sched.h __u8 prio; /* prio of this VQ */ prio 396 tools/include/uapi/linux/pkt_sched.h __u32 prio; prio 1934 tools/lib/traceevent/event-parse.c arg->op.prio = 0; prio 1936 tools/lib/traceevent/event-parse.c arg->op.prio = get_op_prio(arg->op.op); prio 1938 tools/lib/traceevent/event-parse.c return arg->op.prio; prio 2000 tools/lib/traceevent/event-parse.c arg->op.prio = 0; prio 2109 tools/lib/traceevent/event-parse.c arg->op.prio = 0; prio 2122 tools/lib/traceevent/event-parse.c int prio; prio 2125 tools/lib/traceevent/event-parse.c prio = get_op_prio(*tok); prio 2127 tools/lib/traceevent/event-parse.c if (prio > arg->op.prio) prio 225 tools/lib/traceevent/event-parse.h int prio; prio 234 tools/perf/util/mmap.c int delta_max, i, prio, ret; prio 278 tools/perf/util/mmap.c prio = delta_max - i; prio 279 tools/perf/util/mmap.c map->aio.cblocks[i].aio_reqprio = prio >= 0 ? prio : 0; prio 110 tools/power/cpupower/bench/main.c config->prio = string_to_prio(optarg); prio 133 tools/power/cpupower/bench/parse.c config->prio = SCHED_HIGH; prio 218 tools/power/cpupower/bench/parse.c config->prio = string_to_prio(val); prio 26 tools/power/cpupower/bench/parse.h } prio; prio 159 tools/power/cpupower/bench/system.c switch (config->prio) { prio 71 tools/testing/selftests/futex/functional/futex_requeue_pi.c int policy, int prio) prio 92 tools/testing/selftests/futex/functional/futex_requeue_pi.c schedp.sched_priority = prio; prio 50 tools/testing/selftests/futex/functional/futex_requeue_pi_signal_restart.c int policy, int prio) prio 71 tools/testing/selftests/futex/functional/futex_requeue_pi_signal_restart.c schedp.sched_priority = prio; prio 376 tools/testing/selftests/mqueue/mq_perf_tests.c void const_prio(int *prio) prio 381 tools/testing/selftests/mqueue/mq_perf_tests.c void inc_prio(int *prio) prio 383 tools/testing/selftests/mqueue/mq_perf_tests.c if (++*prio == mq_prio_max) prio 384 tools/testing/selftests/mqueue/mq_perf_tests.c *prio = 0; prio 387 tools/testing/selftests/mqueue/mq_perf_tests.c void dec_prio(int *prio) prio 389 tools/testing/selftests/mqueue/mq_perf_tests.c if (--*prio < 0) prio 390 tools/testing/selftests/mqueue/mq_perf_tests.c *prio = mq_prio_max - 1; prio 393 tools/testing/selftests/mqueue/mq_perf_tests.c void random_prio(int *prio) prio 395 tools/testing/selftests/mqueue/mq_perf_tests.c *prio = random() % mq_prio_max; prio 788 virt/kvm/arm/vgic/vgic.c u8 prio = 0xff; prio 808 virt/kvm/arm/vgic/vgic.c if (multi_sgi && irq->priority > prio) { prio 817 virt/kvm/arm/vgic/vgic.c prio = irq->priority;