next 310 arch/alpha/include/asm/core_marvel.h struct io7 *next; next 24 arch/alpha/include/asm/gct.h gct6_handle next; next 124 arch/alpha/include/asm/mmu_context.h unsigned long next = asn + 1; next 129 arch/alpha/include/asm/mmu_context.h next = (asn & ~HARDWARE_ASN_MASK) + ASN_FIRST_VERSION; next 131 arch/alpha/include/asm/mmu_context.h cpu_last_asn(cpu) = next; next 132 arch/alpha/include/asm/mmu_context.h return next; next 137 arch/alpha/include/asm/mmu_context.h struct task_struct *next) next 162 arch/alpha/include/asm/mmu_context.h task_thread_info(next)->pcb.asn = mmc & HARDWARE_ASN_MASK; next 167 arch/alpha/include/asm/mmu_context.h struct task_struct *next) next 182 arch/alpha/include/asm/mmu_context.h ev5_switch_mm(prev_mm, next_mm, next); next 22 arch/alpha/include/asm/pci.h struct pci_controller *next; next 84 arch/alpha/kernel/console.c for (hose = hose_head; hose; hose = hose->next) { next 97 arch/alpha/kernel/core_marvel.c return (prev ? prev->next : io7_head); next 105 arch/alpha/kernel/core_marvel.c for (io7 = io7_head; io7 && io7->pe != pe; io7 = io7->next) next 142 arch/alpha/kernel/core_marvel.c io7->next = io7_head; next 145 arch/alpha/kernel/core_marvel.c for (insp = io7_head; insp; insp = insp->next) { next 152 arch/alpha/kernel/core_marvel.c if (NULL == insp->next || next 153 arch/alpha/kernel/core_marvel.c insp->next->pe > io7->pe) { /* insert here */ next 154 arch/alpha/kernel/core_marvel.c io7->next = insp->next; next 155 arch/alpha/kernel/core_marvel.c insp->next = io7; next 163 arch/alpha/kernel/core_marvel.c io7->next = io7_head; next 704 arch/alpha/kernel/core_marvel.c for (hose = hose_head; hose; hose = hose->next) { next 440 arch/alpha/kernel/core_mcpcia.c for (hose = hose_head; hose; hose = hose->next) next 551 arch/alpha/kernel/core_mcpcia.c for (hose = hose_head; hose; hose = hose->next, iodpp++) { next 595 arch/alpha/kernel/core_mcpcia.c for (hose = hose_head; hose; hose = hose->next) next 480 arch/alpha/kernel/core_titan.c for (hose = hose_head; hose; hose = hose->next) next 754 arch/alpha/kernel/core_titan.c for (hose = hose_head; hose; hose = hose->next) next 159 arch/alpha/kernel/err_common.c struct el_subpacket *next = NULL; next 162 arch/alpha/kernel/err_common.c for (; h && h->class != header->class; h = h->next); next 163 arch/alpha/kernel/err_common.c if (h) next = h->handler(header); next 165 arch/alpha/kernel/err_common.c return next; next 197 arch/alpha/kernel/err_common.c struct el_subpacket *next = NULL; next 204 arch/alpha/kernel/err_common.c next = el_process_header_subpacket(header); next 207 arch/alpha/kernel/err_common.c if (NULL == (next = el_process_subpacket_reg(header))) { next 216 arch/alpha/kernel/err_common.c return next; next 225 arch/alpha/kernel/err_common.c for (a = subpacket_annotation_list; a; a = a->next) { next 286 arch/alpha/kernel/err_common.c for (; a->next != NULL; a = a->next) { next 294 arch/alpha/kernel/err_common.c a->next = new; next 296 arch/alpha/kernel/err_common.c new->next = NULL; next 308 arch/alpha/kernel/err_common.c for (; h->next != NULL; h = h->next) { next 315 arch/alpha/kernel/err_common.c h->next = new; next 317 arch/alpha/kernel/err_common.c new->next = NULL; next 18 arch/alpha/kernel/err_impl.h struct el_subpacket_annotation *next; next 28 arch/alpha/kernel/err_impl.h struct el_subpacket_handler *next; next 40 arch/alpha/kernel/gct.c if (node->next) next 41 arch/alpha/kernel/gct.c status |= gct6_find_nodes(GCT_NODE_PTR(node->next), search); next 23 arch/alpha/kernel/module.c struct got_entry *next; next 40 arch/alpha/kernel/module.c for (g = chains + r_sym; g ; g = g->next) next 50 arch/alpha/kernel/module.c g->next = chains[r_sym].next; next 54 arch/alpha/kernel/module.c chains[r_sym].next = g; next 122 arch/alpha/kernel/module.c for (g = chains[i].next; g ; g = n) { next 123 arch/alpha/kernel/module.c n = g->next; next 42 arch/alpha/kernel/pci-noop.c hose_tail = &hose->next; next 66 arch/alpha/kernel/pci-noop.c for (hose = hose_head; hose; hose = hose->next) next 204 arch/alpha/kernel/pci.c struct pdev_srm_saved_conf *next; next 228 arch/alpha/kernel/pci.c tmp->next = srm_saved_configs; next 246 arch/alpha/kernel/pci.c for (tmp = srm_saved_configs; tmp; tmp = tmp->next) { next 335 arch/alpha/kernel/pci.c for (next_busno = 0, hose = hose_head; hose; hose = hose->next) { next 383 arch/alpha/kernel/pci.c for (hose = hose_head; hose; hose = hose->next) { next 401 arch/alpha/kernel/pci.c hose_tail = &hose->next; next 430 arch/alpha/kernel/pci.c for(hose = hose_head; hose; hose = hose->next) next 248 arch/alpha/kernel/setup.c for (hose = hose_head; hose; hose = hose->next) next 1427 arch/alpha/kernel/setup.c .next = c_next, next 2342 arch/alpha/kernel/smc37c669.c int __init smcc669_open( struct FILE *fp, char *info, char *next, char *mode ) next 173 arch/alpha/kernel/sys_rawhide.c for (hose = hose_head; hose; hose = hose->next) { next 129 arch/arc/include/asm/mmu_context.h static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, next 145 arch/arc/include/asm/mmu_context.h cpumask_set_cpu(cpu, mm_cpumask(next)); next 149 arch/arc/include/asm/mmu_context.h write_aux_reg(ARC_REG_SCRATCH_DATA0, next->pgd); next 152 arch/arc/include/asm/mmu_context.h get_new_mmu_context(next); next 162 arch/arc/include/asm/mmu_context.h #define activate_mm(prev, next) switch_mm(prev, next, NULL) next 36 arch/arc/include/asm/switch_to.h #define switch_to(prev, next, last) \ next 38 arch/arc/include/asm/switch_to.h ARC_EZNPS_DP_PREV(prev, next); \ next 39 arch/arc/include/asm/switch_to.h ARC_FPU_PREV(prev, next); \ next 40 arch/arc/include/asm/switch_to.h last = __switch_to(prev, next);\ next 41 arch/arc/include/asm/switch_to.h ARC_FPU_NEXT(next); \ next 28 arch/arc/kernel/ctx_sw.c unsigned int next = (unsigned int)next_task; next 117 arch/arc/kernel/ctx_sw.c : "n"(KSP_WORD_OFF), "r"(next), "r"(prev) next 30 arch/arc/kernel/fpu.c void fpu_save_restore(struct task_struct *prev, struct task_struct *next) next 33 arch/arc/kernel/fpu.c unsigned int *readfrom = &next->thread.fpu.aux_dpfp[0].l; next 685 arch/arc/kernel/setup.c .next = c_next, next 690 arch/arc/mm/tlb.c pmd_huge_pte(mm, pmdp) = (pgtable_t) lh->next; next 10 arch/arc/plat-eznps/ctop.c void dp_save_restore(struct task_struct *prev, struct task_struct *next) next 13 arch/arc/plat-eznps/ctop.c struct eznps_dp *next_task_dp = &next->thread.dp; next 140 arch/arm/include/asm/ecard.h struct expansion_card *next; next 23 arch/arm/include/asm/fiq.h struct fiq_handler *next; next 105 arch/arm/include/asm/mmu_context.h #define activate_mm(prev,next) switch_mm(prev, next, NULL) next 128 arch/arm/include/asm/mmu_context.h switch_mm(struct mm_struct *prev, struct mm_struct *next, next 140 arch/arm/include/asm/mmu_context.h !cpumask_empty(mm_cpumask(next)) && next 141 arch/arm/include/asm/mmu_context.h !cpumask_test_cpu(cpu, mm_cpumask(next))) next 144 arch/arm/include/asm/mmu_context.h if (!cpumask_test_and_set_cpu(cpu, mm_cpumask(next)) || prev != next) { next 145 arch/arm/include/asm/mmu_context.h check_and_switch_context(next, tsk); next 73 arch/arm/include/asm/spinlock.h while (lockval.tickets.next != lockval.tickets.owner) { next 116 arch/arm/include/asm/spinlock.h return lock.tickets.owner == lock.tickets.next; next 127 arch/arm/include/asm/spinlock.h return (tickets.next - tickets.owner) > 1; next 16 arch/arm/include/asm/spinlock_types.h u16 next; next 20 arch/arm/include/asm/spinlock_types.h u16 next; next 26 arch/arm/include/asm/switch_to.h #define switch_to(prev,next,last) \ next 29 arch/arm/include/asm/switch_to.h last = __switch_to(prev,task_thread_info(prev), task_thread_info(next)); \ next 118 arch/arm/kernel/fiq.c f->next = current_fiq; next 135 arch/arm/kernel/fiq.c current_fiq = current_fiq->next; next 86 arch/arm/kernel/hibernate.c for (pbe = restore_pblist; pbe; pbe = pbe->next) next 1322 arch/arm/kernel/setup.c .next = c_next, next 72 arch/arm/mach-ep93xx/timer-ep93xx.c static int ep93xx_clkevt_set_next_event(unsigned long next, next 83 arch/arm/mach-ep93xx/timer-ep93xx.c writel(next, EP93XX_TIMER3_LOAD); next 800 arch/arm/mach-omap1/clock.c if (clk->node.next || clk->node.prev) next 164 arch/arm/mach-omap2/powerdomain.c int prev, next, state, trace_state = 0; next 185 arch/arm/mach-omap2/powerdomain.c next = pwrdm_read_next_pwrst(pwrdm); next 186 arch/arm/mach-omap2/powerdomain.c if (next != prev) { next 188 arch/arm/mach-omap2/powerdomain.c ((next & OMAP_POWERSTATE_MASK) << 8) | next 499 arch/arm/mach-rpc/ecard.c for (ec = cards; ec; ec = ec->next) { next 562 arch/arm/mach-rpc/ecard.c for (ec = cards; ec; ec = ec->next) { next 656 arch/arm/mach-rpc/ecard.c ec = ec->next; next 966 arch/arm/mach-rpc/ecard.c for (ecp = &cards; *ecp; ecp = &(*ecp)->next); next 28 arch/arm/mm/idmap.c unsigned long next; next 49 arch/arm/mm/idmap.c next = pmd_addr_end(addr, end); next 52 arch/arm/mm/idmap.c } while (pmd++, addr = next, addr != end); next 72 arch/arm/mm/idmap.c unsigned long next; next 75 arch/arm/mm/idmap.c next = pud_addr_end(addr, end); next 76 arch/arm/mm/idmap.c idmap_add_pmd(pud, addr, next, prot); next 77 arch/arm/mm/idmap.c } while (pud++, addr = next, addr != end); next 84 arch/arm/mm/idmap.c unsigned long next; next 97 arch/arm/mm/idmap.c next = pgd_addr_end(addr, end); next 98 arch/arm/mm/idmap.c idmap_add_pud(pgd, addr, next, prot); next 99 arch/arm/mm/idmap.c } while (pgd++, addr = next, addr != end); next 804 arch/arm/mm/mmu.c unsigned long next; next 811 arch/arm/mm/mmu.c next = pmd_addr_end(addr, end); next 818 arch/arm/mm/mmu.c ((addr | next | phys) & ~SECTION_MASK) == 0) { next 819 arch/arm/mm/mmu.c __map_init_section(pmd, addr, next, phys, type, ng); next 821 arch/arm/mm/mmu.c alloc_init_pte(pmd, addr, next, next 825 arch/arm/mm/mmu.c phys += next - addr; next 827 arch/arm/mm/mmu.c } while (pmd++, addr = next, addr != end); next 836 arch/arm/mm/mmu.c unsigned long next; next 839 arch/arm/mm/mmu.c next = pud_addr_end(addr, end); next 840 arch/arm/mm/mmu.c alloc_init_pmd(pud, addr, next, phys, type, alloc, ng); next 841 arch/arm/mm/mmu.c phys += next - addr; next 842 arch/arm/mm/mmu.c } while (pud++, addr = next, addr != end); next 941 arch/arm/mm/mmu.c unsigned long next = pgd_addr_end(addr, end); next 943 arch/arm/mm/mmu.c alloc_init_pud(pgd, addr, next, phys, type, alloc, ng); next 945 arch/arm/mm/mmu.c phys += next - addr; next 946 arch/arm/mm/mmu.c addr = next; next 1061 arch/arm/mm/mmu.c unsigned long addr, next = 0; next 1067 arch/arm/mm/mmu.c if (addr < next) next 1094 arch/arm/mm/mmu.c next = (addr + PMD_SIZE - 1) & PMD_MASK; next 127 arch/arm/plat-samsung/adc.c struct s3c_adc_client *next = adc->ts_pend; next 129 arch/arm/plat-samsung/adc.c if (!next && !list_empty(&adc_pending)) { next 130 arch/arm/plat-samsung/adc.c next = list_first_entry(&adc_pending, next 132 arch/arm/plat-samsung/adc.c list_del(&next->pend); next 136 arch/arm/plat-samsung/adc.c if (next) { next 137 arch/arm/plat-samsung/adc.c adc_dbg(adc, "new client is %p\n", next); next 138 arch/arm/plat-samsung/adc.c adc->cur = next; next 139 arch/arm/plat-samsung/adc.c s3c_adc_select(adc, next); next 417 arch/arm/probes/decode.c const struct decode_header *next; next 445 arch/arm/probes/decode.c for (;; h = next) { next 452 arch/arm/probes/decode.c next = (struct decode_header *) next 465 arch/arm/probes/decode.c next = (struct decode_header *)d->table.table; next 13 arch/arm64/include/asm/exec.h void uao_thread_switch(struct task_struct *next); next 40 arch/arm64/include/asm/fpsimd.h extern void fpsimd_thread_switch(struct task_struct *next); next 128 arch/arm64/include/asm/hw_breakpoint.h extern void hw_breakpoint_thread_switch(struct task_struct *next); next 131 arch/arm64/include/asm/hw_breakpoint.h static inline void hw_breakpoint_thread_switch(struct task_struct *next) next 29 arch/arm64/include/asm/mmu_context.h static inline void contextidr_thread_switch(struct task_struct *next) next 34 arch/arm64/include/asm/mmu_context.h write_sysreg(task_pid_nr(next), contextidr_el1); next 213 arch/arm64/include/asm/mmu_context.h static inline void __switch_mm(struct mm_struct *next) next 221 arch/arm64/include/asm/mmu_context.h if (next == &init_mm) { next 226 arch/arm64/include/asm/mmu_context.h check_and_switch_context(next, cpu); next 230 arch/arm64/include/asm/mmu_context.h switch_mm(struct mm_struct *prev, struct mm_struct *next, next 233 arch/arm64/include/asm/mmu_context.h if (prev != next) next 234 arch/arm64/include/asm/mmu_context.h __switch_mm(next); next 242 arch/arm64/include/asm/mmu_context.h update_saved_ttbr0(tsk, next); next 246 arch/arm64/include/asm/mmu_context.h #define activate_mm(prev,next) switch_mm(prev, next, current) next 252 arch/arm64/include/asm/processor.h struct task_struct *next); next 120 arch/arm64/kernel/asm-offsets.c DEFINE(HIBERN_PBE_NEXT, offsetof(struct pbe, next)); next 202 arch/arm64/kernel/cpuinfo.c .next = c_next, next 983 arch/arm64/kernel/fpsimd.c void fpsimd_thread_switch(struct task_struct *next) next 1001 arch/arm64/kernel/fpsimd.c &next->thread.uw.fpsimd_state; next 1002 arch/arm64/kernel/fpsimd.c wrong_cpu = next->thread.fpsimd_cpu != smp_processor_id(); next 1004 arch/arm64/kernel/fpsimd.c update_tsk_thread_flag(next, TIF_FOREIGN_FPSTATE, next 395 arch/arm64/kernel/hibernate.c unsigned long next; next 410 arch/arm64/kernel/hibernate.c next = pmd_addr_end(addr, end); next 414 arch/arm64/kernel/hibernate.c if (copy_pte(dst_pmdp, src_pmdp, addr, next)) next 420 arch/arm64/kernel/hibernate.c } while (dst_pmdp++, src_pmdp++, addr = next, addr != end); next 430 arch/arm64/kernel/hibernate.c unsigned long next; next 445 arch/arm64/kernel/hibernate.c next = pud_addr_end(addr, end); next 449 arch/arm64/kernel/hibernate.c if (copy_pmd(dst_pudp, src_pudp, addr, next)) next 455 arch/arm64/kernel/hibernate.c } while (dst_pudp++, src_pudp++, addr = next, addr != end); next 463 arch/arm64/kernel/hibernate.c unsigned long next; next 469 arch/arm64/kernel/hibernate.c next = pgd_addr_end(addr, end); next 472 arch/arm64/kernel/hibernate.c if (copy_pud(dst_pgdp, src_pgdp, addr, next)) next 474 arch/arm64/kernel/hibernate.c } while (dst_pgdp++, src_pgdp++, addr = next, addr != end); next 901 arch/arm64/kernel/hw_breakpoint.c void hw_breakpoint_thread_switch(struct task_struct *next) next 915 arch/arm64/kernel/hw_breakpoint.c next_debug_info = &next->thread.debug; next 431 arch/arm64/kernel/process.c static void tls_thread_switch(struct task_struct *next) next 435 arch/arm64/kernel/process.c if (is_compat_thread(task_thread_info(next))) next 436 arch/arm64/kernel/process.c write_sysreg(next->thread.uw.tp_value, tpidrro_el0); next 440 arch/arm64/kernel/process.c write_sysreg(*task_user_tls(next), tpidr_el0); next 444 arch/arm64/kernel/process.c void uao_thread_switch(struct task_struct *next) next 447 arch/arm64/kernel/process.c if (task_thread_info(next)->addr_limit == KERNEL_DS) next 458 arch/arm64/kernel/process.c static void ssbs_thread_switch(struct task_struct *next) next 460 arch/arm64/kernel/process.c struct pt_regs *regs = task_pt_regs(next); next 466 arch/arm64/kernel/process.c if (unlikely(next->flags & PF_KTHREAD)) next 478 arch/arm64/kernel/process.c test_tsk_thread_flag(next, TIF_SSBD)) next 496 arch/arm64/kernel/process.c static void entry_task_switch(struct task_struct *next) next 498 arch/arm64/kernel/process.c __this_cpu_write(__entry_task, next); next 505 arch/arm64/kernel/process.c struct task_struct *next) next 509 arch/arm64/kernel/process.c fpsimd_thread_switch(next); next 510 arch/arm64/kernel/process.c tls_thread_switch(next); next 511 arch/arm64/kernel/process.c hw_breakpoint_thread_switch(next); next 512 arch/arm64/kernel/process.c contextidr_thread_switch(next); next 513 arch/arm64/kernel/process.c entry_task_switch(next); next 514 arch/arm64/kernel/process.c uao_thread_switch(next); next 515 arch/arm64/kernel/process.c ptrauth_thread_switch(next); next 516 arch/arm64/kernel/process.c ssbs_thread_switch(next); next 527 arch/arm64/kernel/process.c last = cpu_switch_to(prev, next); next 629 arch/arm64/kernel/smp.c goto next; next 634 arch/arm64/kernel/smp.c goto next; next 647 arch/arm64/kernel/smp.c goto next; next 663 arch/arm64/kernel/smp.c goto next; next 669 arch/arm64/kernel/smp.c next: next 309 arch/arm64/mm/dump.c unsigned long next, addr = start; next 314 arch/arm64/mm/dump.c next = pmd_addr_end(addr, end); next 320 arch/arm64/mm/dump.c walk_pte(st, pmdp, addr, next); next 322 arch/arm64/mm/dump.c } while (pmdp++, addr = next, addr != end); next 328 arch/arm64/mm/dump.c unsigned long next, addr = start; next 333 arch/arm64/mm/dump.c next = pud_addr_end(addr, end); next 339 arch/arm64/mm/dump.c walk_pmd(st, pudp, addr, next); next 341 arch/arm64/mm/dump.c } while (pudp++, addr = next, addr != end); next 348 arch/arm64/mm/dump.c unsigned long next, addr = start; next 353 arch/arm64/mm/dump.c next = pgd_addr_end(addr, end); next 359 arch/arm64/mm/dump.c walk_pud(st, pgdp, addr, next); next 361 arch/arm64/mm/dump.c } while (pgdp++, addr = next, addr != end); next 103 arch/arm64/mm/kasan_init.c unsigned long next; next 112 arch/arm64/mm/kasan_init.c next = addr + PAGE_SIZE; next 114 arch/arm64/mm/kasan_init.c } while (ptep++, addr = next, addr != end && pte_none(READ_ONCE(*ptep))); next 120 arch/arm64/mm/kasan_init.c unsigned long next; next 124 arch/arm64/mm/kasan_init.c next = pmd_addr_end(addr, end); next 125 arch/arm64/mm/kasan_init.c kasan_pte_populate(pmdp, addr, next, node, early); next 126 arch/arm64/mm/kasan_init.c } while (pmdp++, addr = next, addr != end && pmd_none(READ_ONCE(*pmdp))); next 132 arch/arm64/mm/kasan_init.c unsigned long next; next 136 arch/arm64/mm/kasan_init.c next = pud_addr_end(addr, end); next 137 arch/arm64/mm/kasan_init.c kasan_pmd_populate(pudp, addr, next, node, early); next 138 arch/arm64/mm/kasan_init.c } while (pudp++, addr = next, addr != end && pud_none(READ_ONCE(*pudp))); next 144 arch/arm64/mm/kasan_init.c unsigned long next; next 149 arch/arm64/mm/kasan_init.c next = pgd_addr_end(addr, end); next 150 arch/arm64/mm/kasan_init.c kasan_pud_populate(pgdp, addr, next, node, early); next 151 arch/arm64/mm/kasan_init.c } while (pgdp++, addr = next, addr != end); next 170 arch/arm64/mm/mmu.c unsigned long next; next 186 arch/arm64/mm/mmu.c next = pte_cont_addr_end(addr, end); next 189 arch/arm64/mm/mmu.c if ((((addr | next | phys) & ~CONT_PTE_MASK) == 0) && next 193 arch/arm64/mm/mmu.c init_pte(pmdp, addr, next, phys, __prot); next 195 arch/arm64/mm/mmu.c phys += next - addr; next 196 arch/arm64/mm/mmu.c } while (addr = next, addr != end); next 203 arch/arm64/mm/mmu.c unsigned long next; next 210 arch/arm64/mm/mmu.c next = pmd_addr_end(addr, end); next 213 arch/arm64/mm/mmu.c if (((addr | next | phys) & ~SECTION_MASK) == 0 && next 224 arch/arm64/mm/mmu.c alloc_init_cont_pte(pmdp, addr, next, phys, prot, next 230 arch/arm64/mm/mmu.c phys += next - addr; next 231 arch/arm64/mm/mmu.c } while (pmdp++, addr = next, addr != end); next 241 arch/arm64/mm/mmu.c unsigned long next; next 260 arch/arm64/mm/mmu.c next = pmd_cont_addr_end(addr, end); next 263 arch/arm64/mm/mmu.c if ((((addr | next | phys) & ~CONT_PMD_MASK) == 0) && next 267 arch/arm64/mm/mmu.c init_pmd(pudp, addr, next, phys, __prot, pgtable_alloc, flags); next 269 arch/arm64/mm/mmu.c phys += next - addr; next 270 arch/arm64/mm/mmu.c } while (addr = next, addr != end); next 273 arch/arm64/mm/mmu.c static inline bool use_1G_block(unsigned long addr, unsigned long next, next 279 arch/arm64/mm/mmu.c if (((addr | next | phys) & ~PUD_MASK) != 0) next 290 arch/arm64/mm/mmu.c unsigned long next; next 307 arch/arm64/mm/mmu.c next = pud_addr_end(addr, end); next 312 arch/arm64/mm/mmu.c if (use_1G_block(addr, next, phys) && next 323 arch/arm64/mm/mmu.c alloc_init_cont_pmd(pudp, addr, next, phys, prot, next 329 arch/arm64/mm/mmu.c phys += next - addr; next 330 arch/arm64/mm/mmu.c } while (pudp++, addr = next, addr != end); next 341 arch/arm64/mm/mmu.c unsigned long addr, length, end, next; next 357 arch/arm64/mm/mmu.c next = pgd_addr_end(addr, end); next 358 arch/arm64/mm/mmu.c alloc_init_pud(pgdp, addr, next, phys, prot, pgtable_alloc, next 360 arch/arm64/mm/mmu.c phys += next - addr; next 361 arch/arm64/mm/mmu.c } while (pgdp++, addr = next, addr != end); next 740 arch/arm64/mm/mmu.c unsigned long next; next 746 arch/arm64/mm/mmu.c next = pmd_addr_end(addr, end); next 766 arch/arm64/mm/mmu.c vmemmap_verify((pte_t *)pmdp, node, addr, next); next 767 arch/arm64/mm/mmu.c } while (addr = next, addr != end); next 1024 arch/arm64/mm/mmu.c unsigned long next, end; next 1035 arch/arm64/mm/mmu.c next = addr; next 1038 arch/arm64/mm/mmu.c pmd_free_pte_page(pmdp, next); next 1039 arch/arm64/mm/mmu.c } while (pmdp++, next += PMD_SIZE, next != end); next 18 arch/c6x/include/asm/switch_to.h struct thread_struct *next, next 21 arch/c6x/include/asm/switch_to.h #define switch_to(prev, next, last) \ next 25 arch/c6x/include/asm/switch_to.h &(next)->thread, (prev)); \ next 27 arch/csky/include/asm/mmu_context.h #define activate_mm(prev,next) switch_mm(prev, next, current) next 36 arch/csky/include/asm/mmu_context.h switch_mm(struct mm_struct *prev, struct mm_struct *next, next 41 arch/csky/include/asm/mmu_context.h if (prev != next) next 42 arch/csky/include/asm/mmu_context.h check_and_switch_context(next, cpu); next 44 arch/csky/include/asm/mmu_context.h TLBMISS_HANDLER_SETUP_PGD(next->pgd); next 45 arch/csky/include/asm/mmu_context.h write_mmu_entryhi(next->context.asid.counter); next 31 arch/csky/include/asm/spinlock.h while (lockval.tickets.next != lockval.tickets.owner) next 73 arch/csky/include/asm/spinlock.h return lock.tickets.owner == lock.tickets.next; next 85 arch/csky/include/asm/spinlock.h return (tickets.next - tickets.owner) > 1; next 18 arch/csky/include/asm/spinlock_types.h u16 next; next 11 arch/csky/include/asm/switch_to.h struct task_struct *next) next 14 arch/csky/include/asm/switch_to.h restore_from_user_fp(&next->thread.user_fp); next 18 arch/csky/include/asm/switch_to.h struct task_struct *next) next 28 arch/csky/include/asm/switch_to.h #define switch_to(prev, next, last) \ next 31 arch/csky/include/asm/switch_to.h struct task_struct *__next = (next); \ next 33 arch/csky/include/asm/switch_to.h ((last) = __switch_to((prev), (next))); \ next 14 arch/csky/kernel/cpu-probe.c unsigned int cur, next, i; next 24 arch/csky/kernel/cpu-probe.c next = mfcr("cr13"); next 27 arch/csky/kernel/cpu-probe.c if (cur == next) next 30 arch/csky/kernel/cpu-probe.c cur = next; next 33 arch/csky/kernel/cpu-probe.c if (!(next >> 28)) { next 76 arch/csky/kernel/cpu-probe.c .next = c_next, next 35 arch/h8300/include/asm/switch_to.h #define switch_to(prev, next, last) \ next 46 arch/h8300/include/asm/switch_to.h "r" (&(next->thread)), \ next 241 arch/h8300/kernel/ptrace_h.c unsigned short *next; next 243 arch/h8300/kernel/ptrace_h.c next = nextpc(child, (unsigned short *)h8300_get_reg(child, PT_PC)); next 244 arch/h8300/kernel/ptrace_h.c child->thread.breakinfo.addr = next; next 245 arch/h8300/kernel/ptrace_h.c child->thread.breakinfo.inst = *next; next 246 arch/h8300/kernel/ptrace_h.c *next = BREAKINST; next 165 arch/h8300/kernel/setup.c .next = c_next, next 55 arch/hexagon/include/asm/mmu_context.h static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, next 64 arch/hexagon/include/asm/mmu_context.h if (next->context.generation < prev->context.generation) { next 66 arch/hexagon/include/asm/mmu_context.h next->pgd[l1] = init_mm.pgd[l1]; next 68 arch/hexagon/include/asm/mmu_context.h next->context.generation = prev->context.generation; next 71 arch/hexagon/include/asm/mmu_context.h __vmnewmap((void *)next->context.ptbase); next 77 arch/hexagon/include/asm/mmu_context.h static inline void activate_mm(struct mm_struct *prev, struct mm_struct *next) next 82 arch/hexagon/include/asm/mmu_context.h switch_mm(prev, next, current_thread_info()->task); next 134 arch/hexagon/kernel/setup.c .next = &c_next, next 231 arch/ia64/hp/common/sba_iommu.c struct ioc *next; /* list of IOC's in system */ next 1787 arch/ia64/hp/common/sba_iommu.c ioc->next = ioc_list; next 1847 arch/ia64/hp/common/sba_iommu.c for (ioc = ioc_list; ioc; ioc = ioc->next) next 1860 arch/ia64/hp/common/sba_iommu.c return ioc->next; next 1912 arch/ia64/hp/common/sba_iommu.c .next = ioc_next, next 1953 arch/ia64/hp/common/sba_iommu.c for (ioc = ioc_list; ioc; ioc = ioc->next) next 1987 arch/ia64/hp/common/sba_iommu.c ioc_found = ioc->next; next 2039 arch/ia64/hp/common/sba_iommu.c ioc->next = ioc_found; next 38 arch/ia64/include/asm/mmu_context.h unsigned int next; /* next context number to use */ next 93 arch/ia64/include/asm/mmu_context.h if (ia64_ctx.next >= ia64_ctx.limit) { next 94 arch/ia64/include/asm/mmu_context.h ia64_ctx.next = find_next_zero_bit(ia64_ctx.bitmap, next 95 arch/ia64/include/asm/mmu_context.h ia64_ctx.max_ctx, ia64_ctx.next); next 97 arch/ia64/include/asm/mmu_context.h ia64_ctx.max_ctx, ia64_ctx.next); next 98 arch/ia64/include/asm/mmu_context.h if (ia64_ctx.next >= ia64_ctx.max_ctx) next 101 arch/ia64/include/asm/mmu_context.h mm->context = context = ia64_ctx.next++; next 187 arch/ia64/include/asm/mmu_context.h activate_mm (struct mm_struct *prev, struct mm_struct *next) next 193 arch/ia64/include/asm/mmu_context.h ia64_set_kr(IA64_KR_PT_BASE, __pa(next->pgd)); next 194 arch/ia64/include/asm/mmu_context.h activate_context(next); next 45 arch/ia64/include/asm/switch_to.h #define __switch_to(prev,next,last) do { \ next 48 arch/ia64/include/asm/switch_to.h if (IA64_HAS_EXTRA_STATE(next)) \ next 49 arch/ia64/include/asm/switch_to.h ia64_load_extra(next); \ next 50 arch/ia64/include/asm/switch_to.h ia64_psr(task_pt_regs(next))->dfh = !ia64_is_local_fpu_owner(next); \ next 51 arch/ia64/include/asm/switch_to.h (last) = ia64_switch_to((next)); \ next 61 arch/ia64/include/asm/switch_to.h # define switch_to(prev,next,last) do { \ next 67 arch/ia64/include/asm/switch_to.h __switch_to(prev, next, last); \ next 76 arch/ia64/include/asm/switch_to.h # define switch_to(prev,next,last) __switch_to(prev, next, last) next 527 arch/ia64/kernel/iosapic.c list_move(iosapic_intr_info[irq].rtes.next, next 224 arch/ia64/kernel/mca.c unsigned long next = (mlogbuf_end + 1) % MLOGBUF_SIZE; next 225 arch/ia64/kernel/mca.c if (next != mlogbuf_start) { next 227 arch/ia64/kernel/mca.c mlogbuf_end = next; next 108 arch/ia64/kernel/mca_drv.h (((head)->next != (head)) ? list_entry((head)->next, typeof(slidx_list_t), list) : NULL) next 766 arch/ia64/kernel/perfmon.c int idx, next; next 768 arch/ia64/kernel/perfmon.c next = (ctx->ctx_msgq_tail+1) % PFM_MAX_MSGS; next 771 arch/ia64/kernel/perfmon.c if (next == ctx->ctx_msgq_head) return NULL; next 774 arch/ia64/kernel/perfmon.c ctx->ctx_msgq_tail = next; next 5657 arch/ia64/kernel/perfmon.c .next = pfm_proc_next, next 764 arch/ia64/kernel/setup.c .next = c_next, next 626 arch/ia64/kernel/unwind.c sr->curr.next = rs; next 632 arch/ia64/kernel/unwind.c struct unw_reg_state *rs = sr->curr.next; next 656 arch/ia64/kernel/unwind.c prev->next = copy; next 659 arch/ia64/kernel/unwind.c rs = rs->next; next 669 arch/ia64/kernel/unwind.c struct unw_reg_state *p, *next; next 671 arch/ia64/kernel/unwind.c for (p = rs->next; p != NULL; p = next) { next 672 arch/ia64/kernel/unwind.c next = p->next; next 675 arch/ia64/kernel/unwind.c rs->next = NULL; next 1037 arch/ia64/kernel/unwind.c for (ls = sr->labeled_states; ls; ls = ls->next) { next 1041 arch/ia64/kernel/unwind.c sr->curr.next = dup_state_stack(ls->saved_state.next); next 1060 arch/ia64/kernel/unwind.c ls->saved_state.next = dup_state_stack(sr->curr.next); next 1063 arch/ia64/kernel/unwind.c ls->next = sr->labeled_states; next 1531 arch/ia64/kernel/unwind.c struct unw_labeled_state *ls, *next; next 1564 arch/ia64/kernel/unwind.c for (table = unw.tables; table; table = table->next) { next 1575 arch/ia64/kernel/unwind.c prev->next = table->next; next 1576 arch/ia64/kernel/unwind.c table->next = unw.tables->next; next 1577 arch/ia64/kernel/unwind.c unw.tables->next = table; next 1701 arch/ia64/kernel/unwind.c for (ls = sr.labeled_states; ls; ls = next) { next 1702 arch/ia64/kernel/unwind.c next = ls->next; next 2119 arch/ia64/kernel/unwind.c table->next = unw.tables->next; next 2120 arch/ia64/kernel/unwind.c unw.tables->next = table; next 2152 arch/ia64/kernel/unwind.c for (prev = (struct unw_table *) &unw.tables; prev; prev = prev->next) next 2153 arch/ia64/kernel/unwind.c if (prev->next == table) next 2161 arch/ia64/kernel/unwind.c prev->next = table->next; next 50 arch/ia64/kernel/unwind_i.h struct unw_table *next; /* must be first member! */ next 84 arch/ia64/kernel/unwind_i.h struct unw_reg_state *next; /* next (outer) element on state stack */ next 89 arch/ia64/kernel/unwind_i.h struct unw_labeled_state *next; /* next labeled state (or NULL) */ next 45 arch/ia64/mm/tlb.c .next = 1, next 90 arch/ia64/mm/tlb.c ia64_ctx.next = find_next_zero_bit(ia64_ctx.bitmap, next 93 arch/ia64/mm/tlb.c ia64_ctx.max_ctx, ia64_ctx.next); next 185 arch/m68k/emu/nfblock.c struct nfhd_device *dev, *next; next 187 arch/m68k/emu/nfblock.c list_for_each_entry_safe(dev, next, &nfhd_list, list) { next 111 arch/m68k/include/asm/dvma.h struct Linux_SBus_DMA *next; next 140 arch/m68k/include/asm/mac_iop.h struct iop_msg *next; /* next message in queue or NULL */ next 75 arch/m68k/include/asm/mmu_context.h static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, next 79 arch/m68k/include/asm/mmu_context.h set_context(tsk->mm->context, next->pgd); next 95 arch/m68k/include/asm/mmu_context.h #define prepare_arch_switch(next) load_ksp_mmu(next) next 197 arch/m68k/include/asm/mmu_context.h static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, next 283 arch/m68k/include/asm/mmu_context.h static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, struct task_struct *tsk) next 285 arch/m68k/include/asm/mmu_context.h if (prev != next) { next 287 arch/m68k/include/asm/mmu_context.h switch_mm_0230(next); next 289 arch/m68k/include/asm/mmu_context.h switch_mm_0460(next); next 316 arch/m68k/include/asm/mmu_context.h static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, struct task_struct *tsk) next 31 arch/m68k/include/asm/switch_to.h #define switch_to(prev,next,last) do { \ next 33 arch/m68k/include/asm/switch_to.h register void *_next __asm__ ("a1") = (next); \ next 206 arch/m68k/include/asm/tlbflush.h goto next; next 213 arch/m68k/include/asm/tlbflush.h next: next 494 arch/m68k/kernel/setup_mm.c .next = c_next, next 218 arch/m68k/kernel/setup_no.c .next = c_next, next 417 arch/m68k/mac/iop.c msg = msg->next; next 483 arch/m68k/mac/iop.c msg->next = NULL; next 494 arch/m68k/mac/iop.c while (q->next) q = q->next; next 495 arch/m68k/mac/iop.c q->next = msg; next 66 arch/m68k/mm/kmap.c for (p = &iolist; (tmp = *p) ; p = &tmp->next) { next 77 arch/m68k/mm/kmap.c area->next = *p; next 89 arch/m68k/mm/kmap.c for (p = &iolist ; (tmp = *p) ; p = &tmp->next) { next 91 arch/m68k/mm/kmap.c *p = tmp->next; next 62 arch/m68k/mm/memory.c ptable_desc *dp = ptable_list.next; next 118 arch/m68k/mm/memory.c } else if (ptable_list.next != dp) { next 145 arch/m68k/sun3/sun3dvma.c ret = list_entry(hole_cache.next, struct hole, list); next 120 arch/microblaze/include/asm/mmu_context_mm.h static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, next 123 arch/microblaze/include/asm/mmu_context_mm.h tsk->thread.pgdir = next->pgd; next 124 arch/microblaze/include/asm/mmu_context_mm.h get_mmu_context(next); next 125 arch/microblaze/include/asm/mmu_context_mm.h set_context(next->context, next->pgd); next 16 arch/microblaze/include/asm/switch_to.h struct thread_info *next); next 18 arch/microblaze/include/asm/switch_to.h #define switch_to(prev, next, last) \ next 21 arch/microblaze/include/asm/switch_to.h task_thread_info(next)); \ next 155 arch/microblaze/kernel/cpu/mb.c .next = c_next, next 541 arch/microblaze/pci/pci-common.c unsigned int next = isa_hole + 1; next 543 arch/microblaze/pci/pci-common.c if (next < memno) next 545 arch/microblaze/pci/pci-common.c &hose->mem_resources[next], next 546 arch/microblaze/pci/pci-common.c sizeof(struct resource) * (memno - next)); next 998 arch/mips/alchemy/common/clock.c goto next; next 1022 arch/mips/alchemy/common/clock.c next: next 81 arch/mips/bcm63xx/irq.c unsigned int *next = &i[cpu]; \ next 102 arch/mips/bcm63xx/irq.c unsigned int to_call = *next; \ next 104 arch/mips/bcm63xx/irq.c *next = (*next + 1) & (width - 1); \ next 110 arch/mips/cavium-octeon/executive/cvmx-bootmem.c static void cvmx_bootmem_phy_set_next(uint64_t addr, uint64_t next) next 112 arch/mips/cavium-octeon/executive/cvmx-bootmem.c cvmx_write64_uint64((addr + NEXT_OFFSET) | (1ull << 63), next); next 161 arch/mips/emma/markeins/irq.c .next = NULL, next 114 arch/mips/include/asm/mach-cavium-octeon/kernel-entry-init.h # This is the variable where the next core to boot is stored next 116 arch/mips/include/asm/mach-cavium-octeon/kernel-entry-init.h # Get the core id of the next to be booted next 158 arch/mips/include/asm/mmu_context.h static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, next 166 arch/mips/include/asm/mmu_context.h check_switch_mmu_context(next); next 173 arch/mips/include/asm/mmu_context.h cpumask_set_cpu(cpu, mm_cpumask(next)); next 188 arch/mips/include/asm/mmu_context.h #define activate_mm(prev, next) switch_mm(prev, next, current) next 33 arch/mips/include/asm/switch_to.h struct task_struct *next, struct thread_info *next_ti); next 62 arch/mips/include/asm/switch_to.h next->thread.emulated_fp = 0; \ next 88 arch/mips/include/asm/switch_to.h # define __sanitize_fcr31(next) \ next 90 arch/mips/include/asm/switch_to.h unsigned long fcr31 = mask_fcr31_x(next->thread.fpu.fcr31); \ next 94 arch/mips/include/asm/switch_to.h pc = (void __user *)task_pt_regs(next)->cp0_epc; \ next 95 arch/mips/include/asm/switch_to.h next->thread.fpu.fcr31 &= ~fcr31; \ next 96 arch/mips/include/asm/switch_to.h force_fcr31_sig(fcr31, pc, next); \ next 100 arch/mips/include/asm/switch_to.h # define __sanitize_fcr31(next) next 109 arch/mips/include/asm/switch_to.h #define switch_to(prev, next, last) \ next 113 arch/mips/include/asm/switch_to.h if (tsk_used_math(next)) \ next 114 arch/mips/include/asm/switch_to.h __sanitize_fcr31(next); \ next 117 arch/mips/include/asm/switch_to.h __restore_dsp(next); \ next 126 arch/mips/include/asm/switch_to.h if (KSTK_STATUS(next) & ST0_CU2 && \ next 128 arch/mips/include/asm/switch_to.h cop2_restore(next); \ next 135 arch/mips/include/asm/switch_to.h write_c0_userlocal(task_thread_info(next)->tp_value); \ next 136 arch/mips/include/asm/switch_to.h __restore_watch(next); \ next 137 arch/mips/include/asm/switch_to.h (last) = resume(prev, next, task_thread_info(next)); \ next 151 arch/mips/include/asm/txx9/tx4939.h __u32 next; next 170 arch/mips/include/asm/txx9/tx4939.h __u32 next; next 327 arch/mips/kernel/asm-offsets.c OFFSET(PBE_NEXT, pbe, next); next 27 arch/mips/kernel/module.c struct mips_hi16 *next; next 101 arch/mips/kernel/module.c n->next = me->arch.r_mips_hi16_list; next 109 arch/mips/kernel/module.c struct mips_hi16 *next; next 112 arch/mips/kernel/module.c next = l->next; next 114 arch/mips/kernel/module.c l = next; next 136 arch/mips/kernel/module.c struct mips_hi16 *next; next 164 arch/mips/kernel/module.c next = l->next; next 166 arch/mips/kernel/module.c l = next; next 186 arch/mips/kernel/proc.c .next = c_next, next 214 arch/mips/kernel/vpe-mt.c t = list_entry(v->tc.next, struct tc, tc); next 233 arch/mips/kernel/vpe-mt.c t = list_entry(v->tc.next, struct tc, tc); next 220 arch/mips/kernel/vpe.c struct mips_hi16 *next; next 327 arch/mips/kernel/vpe.c n->next = mips_hi16_list; next 338 arch/mips/kernel/vpe.c struct mips_hi16 *l, *next; next 376 arch/mips/kernel/vpe.c next = l->next; next 378 arch/mips/kernel/vpe.c l = next; next 395 arch/mips/kernel/vpe.c next = l->next; next 397 arch/mips/kernel/vpe.c l = next; next 145 arch/mips/sibyte/common/sb_tbprof.c u64 next = (1ULL << 40) - tb_period; next 178 arch/mips/sibyte/common/sb_tbprof.c __raw_writeq(next, IOADDR(A_SCD_PERF_CNT_1)); next 54 arch/nds32/include/asm/mmu_context.h static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, next 59 arch/nds32/include/asm/mmu_context.h if (!cpumask_test_and_set_cpu(cpu, mm_cpumask(next)) || prev != next) { next 60 arch/nds32/include/asm/mmu_context.h check_context(next); next 61 arch/nds32/include/asm/mmu_context.h cpu_switch_mm(next); next 66 arch/nds32/include/asm/mmu_context.h #define activate_mm(prev,next) switch_mm(prev, next, NULL) next 205 arch/nds32/kernel/process.c struct task_struct *_switch_fpu(struct task_struct *prev, struct task_struct *next) next 210 arch/nds32/kernel/process.c if (!(next->flags & PF_KTHREAD)) next 211 arch/nds32/kernel/process.c clear_fpu(task_pt_regs(next)); next 379 arch/nds32/kernel/setup.c .next = c_next, next 54 arch/nios2/include/asm/mmu_context.h void switch_mm(struct mm_struct *prev, struct mm_struct *next, next 66 arch/nios2/include/asm/mmu_context.h void activate_mm(struct mm_struct *prev, struct mm_struct *next); next 17 arch/nios2/include/asm/switch_to.h #define switch_to(prev, next, last) \ next 26 arch/nios2/include/asm/switch_to.h : "r" (prev), "r" (next) \ next 190 arch/nios2/kernel/cpuinfo.c .next = cpuinfo_next, next 80 arch/nios2/mm/mmu_context.c void switch_mm(struct mm_struct *prev, struct mm_struct *next, next 89 arch/nios2/mm/mmu_context.c if (unlikely(CTX_VERSION(next->context) != next 91 arch/nios2/mm/mmu_context.c next->context = get_new_context(); next 94 arch/nios2/mm/mmu_context.c pgd_current = next->pgd; next 97 arch/nios2/mm/mmu_context.c set_context(next->context); next 106 arch/nios2/mm/mmu_context.c void activate_mm(struct mm_struct *prev, struct mm_struct *next) next 108 arch/nios2/mm/mmu_context.c next->context = get_new_context(); next 109 arch/nios2/mm/mmu_context.c set_context(next->context); next 110 arch/nios2/mm/mmu_context.c pgd_current = next->pgd; next 22 arch/openrisc/include/asm/mmu_context.h extern void switch_mm(struct mm_struct *prev, struct mm_struct *next, next 27 arch/openrisc/include/asm/mmu_context.h #define activate_mm(prev, next) switch_mm((prev), (next), NULL) next 27 arch/openrisc/kernel/dma.c unsigned long next, struct mm_walk *walk) next 41 arch/openrisc/kernel/dma.c for (cl = __pa(addr); cl < __pa(next); cl += cpuinfo->dcache_block_size) next 53 arch/openrisc/kernel/dma.c unsigned long next, struct mm_walk *walk) next 403 arch/openrisc/kernel/setup.c .next = c_next, next 138 arch/openrisc/mm/tlb.c void switch_mm(struct mm_struct *prev, struct mm_struct *next, next 147 arch/openrisc/mm/tlb.c current_pgd[smp_processor_id()] = next->pgd; next 153 arch/openrisc/mm/tlb.c if (prev != next) next 54 arch/parisc/include/asm/mmu_context.h struct mm_struct *next, struct task_struct *tsk) next 56 arch/parisc/include/asm/mmu_context.h if (prev != next) { next 57 arch/parisc/include/asm/mmu_context.h mtctl(__pa(next->pgd), 25); next 58 arch/parisc/include/asm/mmu_context.h load_context(next->context); next 63 arch/parisc/include/asm/mmu_context.h struct mm_struct *next, struct task_struct *tsk) next 67 arch/parisc/include/asm/mmu_context.h if (prev == next) next 71 arch/parisc/include/asm/mmu_context.h switch_mm_irqs_off(prev, next, tsk); next 78 arch/parisc/include/asm/mmu_context.h static inline void activate_mm(struct mm_struct *prev, struct mm_struct *next) next 88 arch/parisc/include/asm/mmu_context.h BUG_ON(next == &init_mm); /* Should never happen */ next 90 arch/parisc/include/asm/mmu_context.h if (next->context == 0) next 91 arch/parisc/include/asm/mmu_context.h next->context = alloc_sid(); next 93 arch/parisc/include/asm/mmu_context.h switch_mm(prev,next,current); next 34 arch/parisc/include/asm/parisc-device.h struct parisc_driver *next; next 74 arch/parisc/include/asm/ropes.h struct sba_device *next; /* list of SBA's in system */ next 9 arch/parisc/include/asm/switch_to.h #define switch_to(prev, next, last) do { \ next 10 arch/parisc/include/asm/switch_to.h (last) = _switch_to(prev, next); \ next 230 arch/parisc/kernel/irq.c while ((action = action->next)) next 233 arch/parisc/kernel/irq.c for ( ;action; action = action->next) { next 192 arch/parisc/kernel/setup.c .next = c_next, next 29 arch/powerpc/include/asm/bootx.h u32 next; next 45 arch/powerpc/include/asm/bootx.h u32 next; next 185 arch/powerpc/include/asm/iommu.h struct list_head next; next 43 arch/powerpc/include/asm/kvm_book3s_64.h struct kvm_nested_guest *next; next 107 arch/powerpc/include/asm/kvm_book3s_64.h (pos)->list.next)), true); \ next 188 arch/powerpc/include/asm/kvm_host.h struct list_head next; next 5 arch/powerpc/include/asm/membarrier.h struct mm_struct *next, next 15 arch/powerpc/include/asm/membarrier.h if (likely(!(atomic_read(&next->membarrier_state) & next 62 arch/powerpc/include/asm/mmu_context.h struct mm_struct *next); next 64 arch/powerpc/include/asm/mmu_context.h struct mm_struct *next, next 68 arch/powerpc/include/asm/mmu_context.h return radix__switch_mmu_context(prev, next); next 69 arch/powerpc/include/asm/mmu_context.h return switch_slb(tsk, next); next 104 arch/powerpc/include/asm/mmu_context.h extern void switch_mmu_context(struct mm_struct *prev, struct mm_struct *next, next 129 arch/powerpc/include/asm/mmu_context.h extern void switch_cop(struct mm_struct *next); next 196 arch/powerpc/include/asm/mmu_context.h extern void switch_mm_irqs_off(struct mm_struct *prev, struct mm_struct *next, next 199 arch/powerpc/include/asm/mmu_context.h static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, next 205 arch/powerpc/include/asm/mmu_context.h switch_mm_irqs_off(prev, next, tsk); next 217 arch/powerpc/include/asm/mmu_context.h static inline void activate_mm(struct mm_struct *prev, struct mm_struct *next) next 219 arch/powerpc/include/asm/mmu_context.h switch_mm(prev, next, current); next 332 arch/powerpc/include/asm/mpic.h struct mpic *next; next 982 arch/powerpc/include/asm/opal-api.h __be64 next; next 16 arch/powerpc/include/asm/switch_to.h #define switch_to(prev, next, last) ((last) = __switch_to((prev), (next))) next 19 arch/powerpc/include/asm/switch_to.h struct thread_struct *next); next 371 arch/powerpc/kernel/asm-offsets.c OFFSET(pbe_next, pbe, next); next 41 arch/powerpc/kernel/cacheinfo.c struct cache_index_dir *next; /* next index in parent directory */ next 192 arch/powerpc/kernel/cacheinfo.c struct cache *next = cache; next 194 arch/powerpc/kernel/cacheinfo.c while (next) { next 195 arch/powerpc/kernel/cacheinfo.c WARN_ONCE(cpumask_test_cpu(cpu, &next->shared_cpu_map), next 197 arch/powerpc/kernel/cacheinfo.c cpu, next->ofnode, next 198 arch/powerpc/kernel/cacheinfo.c cache_type_string(next)); next 199 arch/powerpc/kernel/cacheinfo.c cpumask_set_cpu(cpu, &next->shared_cpu_map); next 200 arch/powerpc/kernel/cacheinfo.c next = next->next_local; next 768 arch/powerpc/kernel/cacheinfo.c index_dir->next = cache_dir->index; next 832 arch/powerpc/kernel/cacheinfo.c struct cache_index_dir *next; next 834 arch/powerpc/kernel/cacheinfo.c next = index->next; next 836 arch/powerpc/kernel/cacheinfo.c index = next; next 855 arch/powerpc/kernel/cacheinfo.c struct cache *next = cache->next_local; next 869 arch/powerpc/kernel/cacheinfo.c cache = next; next 52 arch/powerpc/kernel/eeh_event.c event = list_entry(eeh_eventlist.next, next 184 arch/powerpc/kernel/eeh_pe.c struct list_head *next = pe->child_list.next; next 186 arch/powerpc/kernel/eeh_pe.c if (next == &pe->child_list) { next 190 arch/powerpc/kernel/eeh_pe.c next = pe->child.next; next 191 arch/powerpc/kernel/eeh_pe.c if (next != &pe->parent->child_list) next 197 arch/powerpc/kernel/eeh_pe.c return list_entry(next, struct eeh_pe, child); next 444 arch/powerpc/kernel/pci_dn.c struct list_head *next = pdn->child_list.next; next 446 arch/powerpc/kernel/pci_dn.c if (next != &pdn->child_list) next 447 arch/powerpc/kernel/pci_dn.c return list_entry(next, struct pci_dn, list); next 453 arch/powerpc/kernel/pci_dn.c next = pdn->list.next; next 454 arch/powerpc/kernel/pci_dn.c if (next != &pdn->parent->child_list) next 460 arch/powerpc/kernel/pci_dn.c return list_entry(next, struct pci_dn, list); next 109 arch/powerpc/kernel/rtas_flash.c struct flash_block_list *next; next 171 arch/powerpc/kernel/rtas_flash.c for (f = flist; f; f = f->next) { next 196 arch/powerpc/kernel/rtas_flash.c struct flash_block_list *next; next 202 arch/powerpc/kernel/rtas_flash.c next = f->next; next 204 arch/powerpc/kernel/rtas_flash.c f = next; next 334 arch/powerpc/kernel/rtas_flash.c while (fl->next) next 335 arch/powerpc/kernel/rtas_flash.c fl = fl->next; /* seek to last block_list for append */ next 339 arch/powerpc/kernel/rtas_flash.c fl->next = kmem_cache_zalloc(flash_block_cache, GFP_KERNEL); next 340 arch/powerpc/kernel/rtas_flash.c if (!fl->next) next 342 arch/powerpc/kernel/rtas_flash.c fl = fl->next; next 560 arch/powerpc/kernel/rtas_flash.c struct flash_block_list *f, *next, *flist; next 595 arch/powerpc/kernel/rtas_flash.c flist->next = rtas_firmware_flash_list; next 607 arch/powerpc/kernel/rtas_flash.c for (f = flist; f; f = next) { next 614 arch/powerpc/kernel/rtas_flash.c next = f->next; next 616 arch/powerpc/kernel/rtas_flash.c if (f->next) next 617 arch/powerpc/kernel/rtas_flash.c f->next = (struct flash_block_list *)cpu_to_be64(__pa(f->next)); next 619 arch/powerpc/kernel/rtas_flash.c f->next = NULL; next 367 arch/powerpc/kernel/setup-common.c .next = c_next, next 64 arch/powerpc/kvm/book3s_64_vio.c list_del_rcu(&stit->next); next 83 arch/powerpc/kvm/book3s_64_vio.c list_for_each_entry_safe(stit, tmp, &stt->iommu_tables, next) { next 147 arch/powerpc/kvm/book3s_64_vio.c list_for_each_entry_rcu(stit, &stt->iommu_tables, next) { next 172 arch/powerpc/kvm/book3s_64_vio.c list_add_rcu(&stit->next, &stt->iommu_tables); next 248 arch/powerpc/kvm/book3s_64_vio.c list_for_each_entry_safe(stit, tmp, &stt->iommu_tables, next) { next 367 arch/powerpc/kvm/book3s_64_vio.c list_for_each_entry_rcu(stit, &stt->iommu_tables, next) { next 575 arch/powerpc/kvm/book3s_64_vio.c list_for_each_entry_lockless(stit, &stt->iommu_tables, next) { next 671 arch/powerpc/kvm/book3s_64_vio.c list_for_each_entry_lockless(stit, &stt->iommu_tables, next) { next 687 arch/powerpc/kvm/book3s_64_vio.c list_for_each_entry_lockless(stit, &stt->iommu_tables, next) next 717 arch/powerpc/kvm/book3s_64_vio.c list_for_each_entry_lockless(stit, &stt->iommu_tables, next) { next 739 arch/powerpc/kvm/book3s_64_vio.c list_for_each_entry_lockless(stit, &stt->iommu_tables, next) next 123 arch/powerpc/kvm/book3s_64_vio_hv.c list_for_each_entry_lockless(stit, &stt->iommu_tables, next) { next 420 arch/powerpc/kvm/book3s_64_vio_hv.c list_for_each_entry_lockless(stit, &stt->iommu_tables, next) { next 574 arch/powerpc/kvm/book3s_64_vio_hv.c list_for_each_entry_lockless(stit, &stt->iommu_tables, next) { next 590 arch/powerpc/kvm/book3s_64_vio_hv.c list_for_each_entry_lockless(stit, &stt->iommu_tables, next) next 624 arch/powerpc/kvm/book3s_64_vio_hv.c list_for_each_entry_lockless(stit, &stt->iommu_tables, next) { next 646 arch/powerpc/kvm/book3s_64_vio_hv.c list_for_each_entry_lockless(stit, &stt->iommu_tables, next) next 661 arch/powerpc/kvm/book3s_hv_nested.c gp->next = freelist; next 668 arch/powerpc/kvm/book3s_hv_nested.c freelist = gp->next; next 785 arch/powerpc/kvm/book3s_hv_nested.c (*n_rmap)->list.next = (struct llist_node *) rmap; next 164 arch/powerpc/kvm/book3s_hv_rm_mmu.c struct revmap_entry *next, *prev; next 179 arch/powerpc/kvm/book3s_hv_rm_mmu.c next = real_vmalloc_addr(&kvm->arch.hpt.rev[rev->forw]); next 181 arch/powerpc/kvm/book3s_hv_rm_mmu.c next->back = rev->back; next 405 arch/powerpc/kvm/book3s_xive_template.c goto next; next 408 arch/powerpc/kvm/book3s_xive_template.c goto next; next 413 arch/powerpc/kvm/book3s_xive_template.c goto next; next 431 arch/powerpc/kvm/book3s_xive_template.c next: next 143 arch/powerpc/kvm/mpic.c int next; next 295 arch/powerpc/kvm/mpic.c int next = -1; next 307 arch/powerpc/kvm/mpic.c next = irq; next 312 arch/powerpc/kvm/mpic.c q->next = next; next 321 arch/powerpc/kvm/mpic.c return q->next; next 387 arch/powerpc/kvm/mpic.c __func__, n_IRQ, dst->servicing.next, n_CPU); next 390 arch/powerpc/kvm/mpic.c __func__, n_CPU, n_IRQ, dst->raised.next); next 398 arch/powerpc/kvm/mpic.c __func__, n_IRQ, dst->raised.next, next 549 arch/powerpc/kvm/mpic.c opp->dst[i].raised.next = -1; next 551 arch/powerpc/kvm/mpic.c opp->dst[i].servicing.next = -1; next 1067 arch/powerpc/kvm/mpic.c __func__, idx, dst->raised.next); next 35 arch/powerpc/lib/rheap.c pp = (unsigned long *)&l->next; next 133 arch/powerpc/lib/rheap.c blk = list_entry(info->empty_list.next, rh_block_t, list); next 156 arch/powerpc/lib/rheap.c rh_block_t *next; next 170 arch/powerpc/lib/rheap.c next = NULL; next 178 arch/powerpc/lib/rheap.c if (next == NULL && s >= bs) next 179 arch/powerpc/lib/rheap.c next = blk; next 202 arch/powerpc/lib/rheap.c if (next != NULL) next 203 arch/powerpc/lib/rheap.c list_add(&blkn->list, &next->list); next 28 arch/powerpc/mm/book3s64/iommu_api.c struct list_head next; next 150 arch/powerpc/mm/book3s64/iommu_api.c list_for_each_entry_rcu(mem2, &mm->context.iommu_group_mem_list, next) { next 161 arch/powerpc/mm/book3s64/iommu_api.c list_add_rcu(&mem->next, &mm->context.iommu_group_mem_list); next 241 arch/powerpc/mm/book3s64/iommu_api.c list_del_rcu(&mem->next); next 289 arch/powerpc/mm/book3s64/iommu_api.c list_for_each_entry_rcu(mem, &mm->context.iommu_group_mem_list, next) { next 308 arch/powerpc/mm/book3s64/iommu_api.c next) { next 327 arch/powerpc/mm/book3s64/iommu_api.c list_for_each_entry_rcu(mem, &mm->context.iommu_group_mem_list, next) { next 421 arch/powerpc/mm/book3s64/iommu_api.c list_for_each_entry_rcu(mem, &mm->context.iommu_group_mem_list, next) { next 304 arch/powerpc/mm/book3s64/mmu_context.c void radix__switch_mmu_context(struct mm_struct *prev, struct mm_struct *next) next 306 arch/powerpc/mm/book3s64/mmu_context.c mtspr(SPRN_PID, next->context.id); next 719 arch/powerpc/mm/book3s64/radix_pgtable.c unsigned long next; next 723 arch/powerpc/mm/book3s64/radix_pgtable.c for (; addr < end; addr = next, pte++) { next 724 arch/powerpc/mm/book3s64/radix_pgtable.c next = (addr + PAGE_SIZE) & PAGE_MASK; next 725 arch/powerpc/mm/book3s64/radix_pgtable.c if (next > end) next 726 arch/powerpc/mm/book3s64/radix_pgtable.c next = end; next 731 arch/powerpc/mm/book3s64/radix_pgtable.c if (!PAGE_ALIGNED(addr) || !PAGE_ALIGNED(next)) { next 793 arch/powerpc/mm/book3s64/radix_pgtable.c unsigned long next; next 798 arch/powerpc/mm/book3s64/radix_pgtable.c for (; addr < end; addr = next, pmd++) { next 799 arch/powerpc/mm/book3s64/radix_pgtable.c next = pmd_addr_end(addr, end); next 810 arch/powerpc/mm/book3s64/radix_pgtable.c remove_pte_table(pte_base, addr, next); next 818 arch/powerpc/mm/book3s64/radix_pgtable.c unsigned long next; next 823 arch/powerpc/mm/book3s64/radix_pgtable.c for (; addr < end; addr = next, pud++) { next 824 arch/powerpc/mm/book3s64/radix_pgtable.c next = pud_addr_end(addr, end); next 835 arch/powerpc/mm/book3s64/radix_pgtable.c remove_pmd_table(pmd_base, addr, next); next 842 arch/powerpc/mm/book3s64/radix_pgtable.c unsigned long addr, next; next 848 arch/powerpc/mm/book3s64/radix_pgtable.c for (addr = start; addr < end; addr = next) { next 849 arch/powerpc/mm/book3s64/radix_pgtable.c next = pgd_addr_end(addr, end); next 861 arch/powerpc/mm/book3s64/radix_pgtable.c remove_pud_table(pud_base, addr, next); next 998 arch/powerpc/mm/book3s64/radix_pgtable.c pmd_huge_pte(mm, pmdp) = (pgtable_t) lh->next; next 93 arch/powerpc/mm/book3s64/subpage_prot.c unsigned long next, limit; next 104 arch/powerpc/mm/book3s64/subpage_prot.c for (; addr < limit; addr = next) { next 105 arch/powerpc/mm/book3s64/subpage_prot.c next = pmd_addr_end(addr, limit); next 120 arch/powerpc/mm/book3s64/subpage_prot.c if (addr + (nw << PAGE_SHIFT) > next) next 121 arch/powerpc/mm/book3s64/subpage_prot.c nw = (next - addr) >> PAGE_SHIFT; next 196 arch/powerpc/mm/book3s64/subpage_prot.c unsigned long next, limit; next 237 arch/powerpc/mm/book3s64/subpage_prot.c for (limit = addr + len; addr < limit; addr = next) { next 238 arch/powerpc/mm/book3s64/subpage_prot.c next = pmd_addr_end(addr, limit); next 267 arch/powerpc/mm/book3s64/subpage_prot.c if (addr + (nw << PAGE_SHIFT) > next) next 268 arch/powerpc/mm/book3s64/subpage_prot.c nw = (next - addr) >> PAGE_SHIFT; next 345 arch/powerpc/mm/hugetlbpage.c unsigned long next; next 353 arch/powerpc/mm/hugetlbpage.c next = pmd_addr_end(addr, end); next 369 arch/powerpc/mm/hugetlbpage.c if (more > next) next 370 arch/powerpc/mm/hugetlbpage.c next = more; next 373 arch/powerpc/mm/hugetlbpage.c addr, next, floor, ceiling); next 374 arch/powerpc/mm/hugetlbpage.c } while (addr = next, addr != end); next 398 arch/powerpc/mm/hugetlbpage.c unsigned long next; next 404 arch/powerpc/mm/hugetlbpage.c next = pud_addr_end(addr, end); next 408 arch/powerpc/mm/hugetlbpage.c hugetlb_free_pmd_range(tlb, pud, addr, next, floor, next 419 arch/powerpc/mm/hugetlbpage.c if (more > next) next 420 arch/powerpc/mm/hugetlbpage.c next = more; next 423 arch/powerpc/mm/hugetlbpage.c addr, next, floor, ceiling); next 425 arch/powerpc/mm/hugetlbpage.c } while (addr = next, addr != end); next 452 arch/powerpc/mm/hugetlbpage.c unsigned long next; next 472 arch/powerpc/mm/hugetlbpage.c next = pgd_addr_end(addr, end); next 477 arch/powerpc/mm/hugetlbpage.c hugetlb_free_pud_range(tlb, pgd, addr, next, floor, ceiling); next 487 arch/powerpc/mm/hugetlbpage.c if (more > next) next 488 arch/powerpc/mm/hugetlbpage.c next = more; next 491 arch/powerpc/mm/hugetlbpage.c addr, next, floor, ceiling); next 493 arch/powerpc/mm/hugetlbpage.c } while (addr = next, addr != end); next 117 arch/powerpc/mm/init_64.c static struct vmemmap_backing *next; next 135 arch/powerpc/mm/init_64.c vmem_back = next; next 136 arch/powerpc/mm/init_64.c next = next->list; next 143 arch/powerpc/mm/init_64.c next = vmemmap_alloc_block(PAGE_SIZE, node); next 144 arch/powerpc/mm/init_64.c if (unlikely(!next)) { next 153 arch/powerpc/mm/init_64.c return next++; next 264 arch/powerpc/mm/init_64.c vmem_back->list = next; next 265 arch/powerpc/mm/init_64.c next = vmem_back; next 207 arch/powerpc/mm/kasan/kasan_init_32.c unsigned long next; next 215 arch/powerpc/mm/kasan/kasan_init_32.c next = pgd_addr_end(addr, end); next 217 arch/powerpc/mm/kasan/kasan_init_32.c } while (pmd++, addr = next, addr != end); next 34 arch/powerpc/mm/mmu_context.c void switch_mm_irqs_off(struct mm_struct *prev, struct mm_struct *next, next 40 arch/powerpc/mm/mmu_context.c if (!cpumask_test_cpu(smp_processor_id(), mm_cpumask(next))) { next 41 arch/powerpc/mm/mmu_context.c cpumask_set_cpu(smp_processor_id(), mm_cpumask(next)); next 42 arch/powerpc/mm/mmu_context.c inc_mm_active_cpus(next); next 71 arch/powerpc/mm/mmu_context.c switch_mm_pgdir(tsk, next); next 74 arch/powerpc/mm/mmu_context.c if (prev == next) next 85 arch/powerpc/mm/mmu_context.c radix_kvm_prefetch_workaround(next); next 87 arch/powerpc/mm/mmu_context.c membarrier_arch_switch_mm(prev, next, tsk); next 93 arch/powerpc/mm/mmu_context.c switch_mmu_context(prev, next, tsk); next 21 arch/powerpc/mm/nohash/book3e_hugetlbpage.c int this, next; next 26 arch/powerpc/mm/nohash/book3e_hugetlbpage.c next = this + 1; next 27 arch/powerpc/mm/nohash/book3e_hugetlbpage.c if (next >= tcd->esel_max) next 28 arch/powerpc/mm/nohash/book3e_hugetlbpage.c next = tcd->esel_first; next 30 arch/powerpc/mm/nohash/book3e_hugetlbpage.c tcd->esel_next = next; next 261 arch/powerpc/mm/nohash/mmu_context.c void switch_mmu_context(struct mm_struct *prev, struct mm_struct *next, next 274 arch/powerpc/mm/nohash/mmu_context.c cpu, next, next->context.active, next->context.id); next 278 arch/powerpc/mm/nohash/mmu_context.c next->context.active++; next 289 arch/powerpc/mm/nohash/mmu_context.c id = next->context.id; next 292 arch/powerpc/mm/nohash/mmu_context.c if (context_mm[id] != next) next 294 arch/powerpc/mm/nohash/mmu_context.c next, id, id, context_mm[id]); next 331 arch/powerpc/mm/nohash/mmu_context.c context_mm[id] = next; next 332 arch/powerpc/mm/nohash/mmu_context.c next->context.id = id; next 347 arch/powerpc/mm/nohash/mmu_context.c local_flush_tlb_mm(next); next 360 arch/powerpc/mm/nohash/mmu_context.c set_context(id, next->pgd); next 1091 arch/powerpc/mm/numa.c struct topology_update_data *next; next 1268 arch/powerpc/mm/numa.c for (update = data; update; update = update->next) { next 1296 arch/powerpc/mm/numa.c for (update = data; update; update = update->next) { next 1368 arch/powerpc/mm/numa.c ud->next = &updates[i]; next 1382 arch/powerpc/mm/numa.c updates[i-1].next = NULL; next 1386 arch/powerpc/mm/numa.c for (ud = &updates[0]; ud; ud = ud->next) { next 1425 arch/powerpc/mm/numa.c for (ud = &updates[0]; ud; ud = ud->next) { next 41 arch/powerpc/oprofile/cell/pr_util.h struct vma_to_fileoffset_map *next; /* list pointer */ next 26 arch/powerpc/oprofile/cell/vma_map.c struct vma_to_fileoffset_map *next = map->next; next 28 arch/powerpc/oprofile/cell/vma_map.c map = next; next 46 arch/powerpc/oprofile/cell/vma_map.c for (; map; map = map->next) { next 77 arch/powerpc/oprofile/cell/vma_map.c new->next = map; next 1183 arch/powerpc/perf/hv-24x7.c struct hv_24x7_result **next) next 1199 arch/powerpc/perf/hv-24x7.c if (next) next 1200 arch/powerpc/perf/hv-24x7.c *next = (struct hv_24x7_result *) res->elements; next 1241 arch/powerpc/perf/hv-24x7.c if (next) next 1242 arch/powerpc/perf/hv-24x7.c *next = element_data - data_offset; next 108 arch/powerpc/platforms/83xx/mpc832x_rdb.c goto next; next 113 arch/powerpc/platforms/83xx/mpc832x_rdb.c next: next 554 arch/powerpc/platforms/cell/iommu.c window = list_entry(iommu->windows.next, struct iommu_window, list); next 309 arch/powerpc/platforms/cell/spufs/inode.c aff_supp = !list_empty(&(list_entry(cbe_spu_info[0].spus.next, next 336 arch/powerpc/platforms/cell/spufs/inode.c !list_entry(neighbor->aff_list.next, struct spu_context, next 387 arch/powerpc/platforms/cell/spufs/inode.c || list_entry(neighbor->aff_list.next, struct spu_context, next 382 arch/powerpc/platforms/cell/spufs/sched.c list_for_each_entry_reverse(spu, ref->aff_list.next, aff_list) { next 1953 arch/powerpc/platforms/cell/spufs/switch.c static void restore_lscsa(struct spu_state *next, struct spu *spu) next 1961 arch/powerpc/platforms/cell/spufs/switch.c set_watchdog_timer(next, spu); /* Step 26. */ next 1962 arch/powerpc/platforms/cell/spufs/switch.c setup_spu_status_part1(next, spu); /* Step 27. */ next 1963 arch/powerpc/platforms/cell/spufs/switch.c setup_spu_status_part2(next, spu); /* Step 28. */ next 1964 arch/powerpc/platforms/cell/spufs/switch.c restore_mfc_rag(next, spu); /* Step 29. */ next 1966 arch/powerpc/platforms/cell/spufs/switch.c setup_mfc_slbs(next, spu, spu_restore_code, sizeof(spu_restore_code)); next 1967 arch/powerpc/platforms/cell/spufs/switch.c set_spu_npc(next, spu); /* Step 31. */ next 1968 arch/powerpc/platforms/cell/spufs/switch.c set_signot1(next, spu); /* Step 32. */ next 1969 arch/powerpc/platforms/cell/spufs/switch.c set_signot2(next, spu); /* Step 33. */ next 1970 arch/powerpc/platforms/cell/spufs/switch.c setup_decr(next, spu); /* Step 34. */ next 1971 arch/powerpc/platforms/cell/spufs/switch.c setup_ppu_mb(next, spu); /* Step 35. */ next 1972 arch/powerpc/platforms/cell/spufs/switch.c setup_ppuint_mb(next, spu); /* Step 36. */ next 1973 arch/powerpc/platforms/cell/spufs/switch.c send_restore_code(next, spu); /* Step 37. */ next 1974 arch/powerpc/platforms/cell/spufs/switch.c set_ppu_querymask(next, spu); /* Step 38. */ next 1975 arch/powerpc/platforms/cell/spufs/switch.c wait_tag_complete(next, spu); /* Step 39. */ next 1976 arch/powerpc/platforms/cell/spufs/switch.c wait_spu_stopped(next, spu); /* Step 40. */ next 1979 arch/powerpc/platforms/cell/spufs/switch.c static void restore_csa(struct spu_state *next, struct spu *spu) next 1986 arch/powerpc/platforms/cell/spufs/switch.c restore_spu_privcntl(next, spu); /* Step 41. */ next 1987 arch/powerpc/platforms/cell/spufs/switch.c restore_status_part1(next, spu); /* Step 42. */ next 1988 arch/powerpc/platforms/cell/spufs/switch.c restore_status_part2(next, spu); /* Step 43. */ next 1989 arch/powerpc/platforms/cell/spufs/switch.c restore_ls_16kb(next, spu); /* Step 44. */ next 1990 arch/powerpc/platforms/cell/spufs/switch.c wait_tag_complete(next, spu); /* Step 45. */ next 1991 arch/powerpc/platforms/cell/spufs/switch.c suspend_mfc(next, spu); /* Step 46. */ next 1992 arch/powerpc/platforms/cell/spufs/switch.c wait_suspend_mfc_complete(next, spu); /* Step 47. */ next 1993 arch/powerpc/platforms/cell/spufs/switch.c issue_mfc_tlbie(next, spu); /* Step 48. */ next 1994 arch/powerpc/platforms/cell/spufs/switch.c clear_interrupts(next, spu); /* Step 49. */ next 1995 arch/powerpc/platforms/cell/spufs/switch.c restore_mfc_queues(next, spu); /* Step 50. */ next 1996 arch/powerpc/platforms/cell/spufs/switch.c restore_ppu_querymask(next, spu); /* Step 51. */ next 1997 arch/powerpc/platforms/cell/spufs/switch.c restore_ppu_querytype(next, spu); /* Step 52. */ next 1998 arch/powerpc/platforms/cell/spufs/switch.c restore_mfc_csr_tsq(next, spu); /* Step 53. */ next 1999 arch/powerpc/platforms/cell/spufs/switch.c restore_mfc_csr_cmd(next, spu); /* Step 54. */ next 2000 arch/powerpc/platforms/cell/spufs/switch.c restore_mfc_csr_ato(next, spu); /* Step 55. */ next 2001 arch/powerpc/platforms/cell/spufs/switch.c restore_mfc_tclass_id(next, spu); /* Step 56. */ next 2002 arch/powerpc/platforms/cell/spufs/switch.c set_llr_event(next, spu); /* Step 57. */ next 2003 arch/powerpc/platforms/cell/spufs/switch.c restore_decr_wrapped(next, spu); /* Step 58. */ next 2004 arch/powerpc/platforms/cell/spufs/switch.c restore_ch_part1(next, spu); /* Step 59. */ next 2005 arch/powerpc/platforms/cell/spufs/switch.c restore_ch_part2(next, spu); /* Step 60. */ next 2006 arch/powerpc/platforms/cell/spufs/switch.c restore_spu_lslr(next, spu); /* Step 61. */ next 2007 arch/powerpc/platforms/cell/spufs/switch.c restore_spu_cfg(next, spu); /* Step 62. */ next 2008 arch/powerpc/platforms/cell/spufs/switch.c restore_pm_trace(next, spu); /* Step 63. */ next 2009 arch/powerpc/platforms/cell/spufs/switch.c restore_spu_npc(next, spu); /* Step 64. */ next 2010 arch/powerpc/platforms/cell/spufs/switch.c restore_spu_mb(next, spu); /* Step 65. */ next 2011 arch/powerpc/platforms/cell/spufs/switch.c check_ppu_mb_stat(next, spu); /* Step 66. */ next 2012 arch/powerpc/platforms/cell/spufs/switch.c check_ppuint_mb_stat(next, spu); /* Step 67. */ next 2014 arch/powerpc/platforms/cell/spufs/switch.c restore_mfc_sr1(next, spu); /* Step 69. */ next 2015 arch/powerpc/platforms/cell/spufs/switch.c set_int_route(next, spu); /* NEW */ next 2016 arch/powerpc/platforms/cell/spufs/switch.c restore_other_spu_access(next, spu); /* Step 70. */ next 2017 arch/powerpc/platforms/cell/spufs/switch.c restore_spu_runcntl(next, spu); /* Step 71. */ next 2018 arch/powerpc/platforms/cell/spufs/switch.c restore_mfc_cntl(next, spu); /* Step 72. */ next 2019 arch/powerpc/platforms/cell/spufs/switch.c enable_user_access(next, spu); /* Step 73. */ next 2020 arch/powerpc/platforms/cell/spufs/switch.c reset_switch_active(next, spu); /* Step 74. */ next 2021 arch/powerpc/platforms/cell/spufs/switch.c reenable_interrupts(next, spu); /* Step 75. */ next 2056 arch/powerpc/platforms/cell/spufs/switch.c static int __do_spu_restore(struct spu_state *next, struct spu *spu) next 2071 arch/powerpc/platforms/cell/spufs/switch.c restore_lscsa(next, spu); /* Steps 24-39. */ next 2072 arch/powerpc/platforms/cell/spufs/switch.c rc = check_restore_status(next, spu); /* Step 40. */ next 2082 arch/powerpc/platforms/cell/spufs/switch.c restore_csa(next, spu); next 104 arch/powerpc/platforms/powermac/bootx_init.c ppp = &pp->next; next 256 arch/powerpc/platforms/powermac/bootx_init.c goto next; next 261 arch/powerpc/platforms/powermac/bootx_init.c next: next 262 arch/powerpc/platforms/powermac/bootx_init.c ppp = &pp->next; next 316 arch/powerpc/platforms/powermac/bootx_init.c goto next; next 319 arch/powerpc/platforms/powermac/bootx_init.c goto next; next 325 arch/powerpc/platforms/powermac/bootx_init.c next: next 326 arch/powerpc/platforms/powermac/bootx_init.c ppp = &pp->next; next 666 arch/powerpc/platforms/powermac/pci.c unsigned long base, end, next = -1; next 682 arch/powerpc/platforms/powermac/pci.c if (base != next) { next 697 arch/powerpc/platforms/powermac/pci.c next = end + 1; next 688 arch/powerpc/platforms/powermac/pfunc_core.c for (pp = dev->node->properties; pp != 0; pp = pp->next) { next 798 arch/powerpc/platforms/powermac/pfunc_core.c list_entry(dev->functions.next, typeof(*func), link); next 242 arch/powerpc/platforms/powernv/opal-async.c .next = NULL, next 276 arch/powerpc/platforms/powernv/opal-hmi.c msg_node = list_entry(opal_hmi_evt_list.next, next 357 arch/powerpc/platforms/powernv/opal-hmi.c .next = NULL, next 62 arch/powerpc/platforms/powernv/opal-memory-errors.c msg_node = list_entry(opal_memory_err_list.next, next 114 arch/powerpc/platforms/powernv/opal-memory-errors.c .next = NULL, next 119 arch/powerpc/platforms/powernv/opal-power.c .next = NULL, next 126 arch/powerpc/platforms/powernv/opal-power.c .next = NULL, next 133 arch/powerpc/platforms/powernv/opal-power.c .next = NULL, next 371 arch/powerpc/platforms/powernv/opal-prd.c .next = NULL, next 1067 arch/powerpc/platforms/powernv/opal.c struct opal_sg_list *next; next 1069 arch/powerpc/platforms/powernv/opal.c next = kzalloc(PAGE_SIZE, GFP_KERNEL); next 1070 arch/powerpc/platforms/powernv/opal.c if (!next) next 1076 arch/powerpc/platforms/powernv/opal.c sg->next = cpu_to_be64(__pa(next)); next 1077 arch/powerpc/platforms/powernv/opal.c sg = next; next 1097 arch/powerpc/platforms/powernv/opal.c uint64_t next = be64_to_cpu(sg->next); next 1101 arch/powerpc/platforms/powernv/opal.c if (next) next 1102 arch/powerpc/platforms/powernv/opal.c sg = __va(next); next 363 arch/powerpc/platforms/powernv/pci-ioda-tce.c list_for_each_entry_rcu(tgl, &tbl->it_group_list, next) { next 365 arch/powerpc/platforms/powernv/pci-ioda-tce.c list_del_rcu(&tgl->next); next 402 arch/powerpc/platforms/powernv/pci-ioda-tce.c list_add_rcu(&tgl->next, &tbl->it_group_list); next 1906 arch/powerpc/platforms/powernv/pci-ioda.c next); next 2050 arch/powerpc/platforms/powernv/pci-ioda.c list_for_each_entry_lockless(tgl, &tbl->it_group_list, next) { next 575 arch/powerpc/platforms/ps3/mm.c last = list_entry(r->chunk_list.head.next, next 421 arch/powerpc/platforms/ps3/os-area.c next: next 441 arch/powerpc/platforms/ps3/os-area.c goto next; next 444 arch/powerpc/platforms/ps3/os-area.c goto next; next 83 arch/powerpc/platforms/pseries/cmm.c struct cmm_page_array *next; next 180 arch/powerpc/platforms/pseries/cmm.c npa->next = pa; next 227 arch/powerpc/platforms/pseries/cmm.c pa = pa->next; next 512 arch/powerpc/platforms/pseries/cmm.c pa = pa->next; next 579 arch/powerpc/platforms/pseries/cmm.c pa_curr = pa_last->next; next 580 arch/powerpc/platforms/pseries/cmm.c pa_last = pa_last->next; next 585 arch/powerpc/platforms/pseries/cmm.c pa_curr = pa_curr->next; next 608 arch/powerpc/platforms/pseries/cmm.c pa_last->next = npa; next 615 arch/powerpc/platforms/pseries/cmm.c pa_curr = pa_curr->next; next 105 arch/powerpc/platforms/pseries/dlpar.c dn->properties = prop->next; next 210 arch/powerpc/platforms/pseries/dlpar.c last_property->next = property; next 75 arch/powerpc/platforms/pseries/hvCall_inst.c .next = hc_next, next 71 arch/powerpc/platforms/pseries/hvcserver.c element = head->next; next 360 arch/powerpc/platforms/pseries/iommu.c u64 tce_size, num_tce, dma_offset, next; next 366 arch/powerpc/platforms/pseries/iommu.c next = start_pfn << PAGE_SHIFT; next 370 arch/powerpc/platforms/pseries/iommu.c num_tce += next & (tce_size - 1); next 371 arch/powerpc/platforms/pseries/iommu.c next &= ~(tce_size - 1); next 383 arch/powerpc/platforms/pseries/iommu.c dma_offset = next + be64_to_cpu(maprange->dma_base); next 388 arch/powerpc/platforms/pseries/iommu.c next += limit * tce_size; next 399 arch/powerpc/platforms/pseries/iommu.c u64 tce_size, num_tce, dma_offset, next, proto_tce, liobn; next 435 arch/powerpc/platforms/pseries/iommu.c next = start_pfn << PAGE_SHIFT; next 439 arch/powerpc/platforms/pseries/iommu.c num_tce += next & (tce_size - 1); next 440 arch/powerpc/platforms/pseries/iommu.c next &= ~(tce_size - 1); next 453 arch/powerpc/platforms/pseries/iommu.c dma_offset = next + be64_to_cpu(maprange->dma_base); next 456 arch/powerpc/platforms/pseries/iommu.c tcep[l] = cpu_to_be64(proto_tce | next); next 457 arch/powerpc/platforms/pseries/iommu.c next += tce_size; next 92 arch/powerpc/platforms/pseries/reconfig.c struct property *next; next 93 arch/powerpc/platforms/pseries/reconfig.c for (; prop; prop = next) { next 94 arch/powerpc/platforms/pseries/reconfig.c next = prop->next; next 184 arch/powerpc/platforms/pseries/reconfig.c new->next = last; next 573 arch/powerpc/sysdev/mpic.c goto next; next 577 arch/powerpc/sysdev/mpic.c goto next; next 582 arch/powerpc/sysdev/mpic.c next: next 1519 arch/powerpc/sysdev/mpic.c mpic->next = mpics; next 1953 arch/powerpc/sysdev/mpic.c mpic = mpic->next; next 1995 arch/powerpc/sysdev/mpic.c mpic = mpic->next; next 32 arch/riscv/include/asm/mmu_context.h void switch_mm(struct mm_struct *prev, struct mm_struct *next, next 36 arch/riscv/include/asm/mmu_context.h struct mm_struct *next) next 38 arch/riscv/include/asm/mmu_context.h switch_mm(prev, next, NULL); next 48 arch/riscv/include/asm/switch_to.h struct task_struct *next) next 55 arch/riscv/include/asm/switch_to.h fstate_restore(next, task_pt_regs(next)); next 69 arch/riscv/include/asm/switch_to.h #define switch_to(prev, next, last) \ next 72 arch/riscv/include/asm/switch_to.h struct task_struct *__next = (next); \ next 151 arch/riscv/kernel/cpu.c .next = c_next, next 43 arch/riscv/mm/context.c void switch_mm(struct mm_struct *prev, struct mm_struct *next, next 48 arch/riscv/mm/context.c if (unlikely(prev == next)) next 59 arch/riscv/mm/context.c cpumask_set_cpu(cpu, mm_cpumask(next)); next 61 arch/riscv/mm/context.c csr_write(CSR_SATP, virt_to_pfn(next->pgd) | SATP_MODE); next 64 arch/riscv/mm/context.c flush_icache_deferred(next); next 34 arch/s390/include/asm/debug.h struct debug_info *next; next 75 arch/s390/include/asm/gmap.h struct gmap_rmap *next; next 80 arch/s390/include/asm/gmap.h for (pos = (head); pos; pos = pos->next) next 83 arch/s390/include/asm/gmap.h for (pos = (head); n = pos ? pos->next : NULL, pos; pos = n) next 825 arch/s390/include/asm/kvm_host.h int next; next 88 arch/s390/include/asm/mmu_context.h static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, next 93 arch/s390/include/asm/mmu_context.h S390_lowcore.user_asce = next->context.asce; next 94 arch/s390/include/asm/mmu_context.h cpumask_set_cpu(cpu, &next->context.cpu_attach_mask); next 104 arch/s390/include/asm/mmu_context.h if (prev != next) next 129 arch/s390/include/asm/mmu_context.h struct mm_struct *next) next 131 arch/s390/include/asm/mmu_context.h switch_mm(prev, next, current); next 132 arch/s390/include/asm/mmu_context.h cpumask_set_cpu(smp_processor_id(), mm_cpumask(next)); next 133 arch/s390/include/asm/mmu_context.h set_user_asce(next); next 33 arch/s390/include/asm/switch_to.h #define switch_to(prev, next, last) do { \ next 42 arch/s390/include/asm/switch_to.h update_cr_regs(next); \ next 43 arch/s390/include/asm/switch_to.h restore_access_regs(&next->thread.acrs[0]); \ next 44 arch/s390/include/asm/switch_to.h restore_ri_cb(next->thread.ri_cb, prev->thread.ri_cb); \ next 45 arch/s390/include/asm/switch_to.h restore_gs_cb(next->thread.gs_cb); \ next 46 arch/s390/include/asm/switch_to.h prev = __switch_to(prev, next); \ next 341 arch/s390/kernel/debug.c debug_area_last->next = rc; next 345 arch/s390/kernel/debug.c rc->next = NULL; next 418 arch/s390/kernel/debug.c debug_area_first = db_info->next; next 422 arch/s390/kernel/debug.c db_info->prev->next = db_info->next; next 423 arch/s390/kernel/debug.c if (db_info->next) next 424 arch/s390/kernel/debug.c db_info->next->prev = db_info->prev; next 102 arch/s390/kernel/diag.c .next = show_diag_stat_next, next 207 arch/s390/kernel/processor.c .next = c_next, next 36 arch/s390/kernel/suspend.c struct page_key_data *next; next 68 arch/s390/kernel/suspend.c page_key_data = pkd->next; next 89 arch/s390/kernel/suspend.c pk->next = page_key_data; next 124 arch/s390/kernel/suspend.c page_key_wp = page_key_wp->next; next 148 arch/s390/kernel/suspend.c page_key_rp = page_key_rp->next; next 333 arch/s390/kernel/sysinfo.c struct service_level *ptr, *next; next 337 arch/s390/kernel/sysinfo.c list_for_each_entry_safe(ptr, next, &service_level_list, list) { next 376 arch/s390/kernel/sysinfo.c .next = service_level_next, next 43 arch/s390/kernel/topology.c struct mask_info *next; next 80 arch/s390/kernel/topology.c info = info->next; next 154 arch/s390/kernel/topology.c info = info->next; next 159 arch/s390/kernel/topology.c info = info->next; next 164 arch/s390/kernel/topology.c info = info->next; next 188 arch/s390/kernel/topology.c drawer = drawer->next; next 192 arch/s390/kernel/topology.c book = book->next; next 196 arch/s390/kernel/topology.c socket = socket->next; next 523 arch/s390/kernel/topology.c mask->next = memblock_alloc(sizeof(*mask->next), 8); next 524 arch/s390/kernel/topology.c if (!mask->next) next 526 arch/s390/kernel/topology.c __func__, sizeof(*mask->next), 8); next 527 arch/s390/kernel/topology.c mask = mask->next; next 1251 arch/s390/kvm/vsie.c page = kvm->arch.vsie.pages[kvm->arch.vsie.next]; next 1255 arch/s390/kvm/vsie.c kvm->arch.vsie.next++; next 1256 arch/s390/kvm/vsie.c kvm->arch.vsie.next %= nr_vcpus; next 40 arch/s390/lib/spinlock.c struct spin_wait *next, *prev; next 119 arch/s390/lib/spinlock.c struct spin_wait *node, *next; next 126 arch/s390/lib/spinlock.c node->prev = node->next = NULL; next 157 arch/s390/lib/spinlock.c WRITE_ONCE(node->prev->next, node); next 202 arch/s390/lib/spinlock.c while ((next = READ_ONCE(node->next)) == NULL) next 204 arch/s390/lib/spinlock.c next->prev = NULL; next 41 arch/s390/mm/cmm.c struct cmm_page_array *next; next 89 arch/s390/mm/cmm.c npa->next = pa; next 117 arch/s390/mm/cmm.c pa = pa->next; next 780 arch/s390/mm/fault.c struct thread_struct *thread, *next; next 784 arch/s390/mm/fault.c list_for_each_entry_safe(thread, next, &pfault_list, list) { next 185 arch/s390/mm/gmap.c struct page *page, *next; next 191 arch/s390/mm/gmap.c list_for_each_entry_safe(page, next, &gmap->crst_list, lru) next 199 arch/s390/mm/gmap.c list_for_each_entry_safe(page, next, &gmap->pt_list, lru) next 241 arch/s390/mm/gmap.c struct gmap *sg, *next; next 247 arch/s390/mm/gmap.c list_for_each_entry_safe(sg, next, &gmap->children, list) { next 1185 arch/s390/mm/gmap.c rmap->next = radix_tree_deref_slot_protected(slot, next 1189 arch/s390/mm/gmap.c rmap->next = NULL; next 2240 arch/s390/mm/gmap.c struct gmap *gmap, *sg, *next; next 2257 arch/s390/mm/gmap.c list_for_each_entry_safe(sg, next, next 2561 arch/s390/mm/gmap.c unsigned long next, struct mm_walk *walk) next 2569 arch/s390/mm/gmap.c unsigned long hmask, unsigned long next, next 2631 arch/s390/mm/gmap.c unsigned long next, struct mm_walk *walk) next 106 arch/s390/mm/page-states.c unsigned long next; next 112 arch/s390/mm/page-states.c next = pmd_addr_end(addr, end); next 117 arch/s390/mm/page-states.c } while (pmd++, addr = next, addr != end); next 122 arch/s390/mm/page-states.c unsigned long next; next 129 arch/s390/mm/page-states.c next = pud_addr_end(addr, end); next 137 arch/s390/mm/page-states.c mark_kernel_pmd(pud, addr, next); next 138 arch/s390/mm/page-states.c } while (pud++, addr = next, addr != end); next 143 arch/s390/mm/page-states.c unsigned long next; next 150 arch/s390/mm/page-states.c next = p4d_addr_end(addr, end); next 158 arch/s390/mm/page-states.c mark_kernel_pud(p4d, addr, next); next 159 arch/s390/mm/page-states.c } while (p4d++, addr = next, addr != end); next 164 arch/s390/mm/page-states.c unsigned long addr, next; next 172 arch/s390/mm/page-states.c next = pgd_addr_end(addr, MODULES_END); next 180 arch/s390/mm/page-states.c mark_kernel_p4d(pgd, addr, next); next 181 arch/s390/mm/page-states.c } while (pgd++, addr = next, addr != MODULES_END); next 158 arch/s390/mm/pageattr.c unsigned long next; next 166 arch/s390/mm/pageattr.c next = pmd_addr_end(addr, end); next 168 arch/s390/mm/pageattr.c if (addr & ~PMD_MASK || addr + PMD_SIZE > next) { next 176 arch/s390/mm/pageattr.c rc = walk_pte_level(pmdp, addr, next, flags); next 181 arch/s390/mm/pageattr.c addr = next; next 235 arch/s390/mm/pageattr.c unsigned long next; next 243 arch/s390/mm/pageattr.c next = pud_addr_end(addr, end); next 245 arch/s390/mm/pageattr.c if (addr & ~PUD_MASK || addr + PUD_SIZE > next) { next 253 arch/s390/mm/pageattr.c rc = walk_pmd_level(pudp, addr, next, flags); next 256 arch/s390/mm/pageattr.c addr = next; next 265 arch/s390/mm/pageattr.c unsigned long next; next 273 arch/s390/mm/pageattr.c next = p4d_addr_end(addr, end); next 274 arch/s390/mm/pageattr.c rc = walk_pud_level(p4dp, addr, next, flags); next 276 arch/s390/mm/pageattr.c addr = next; next 287 arch/s390/mm/pageattr.c unsigned long next; next 300 arch/s390/mm/pageattr.c next = pgd_addr_end(addr, end); next 301 arch/s390/mm/pageattr.c rc = walk_p4d_level(pgdp, addr, next, flags); next 305 arch/s390/mm/pageattr.c } while (pgdp++, addr = next, addr < end && !rc); next 370 arch/s390/mm/pgalloc.c unsigned long next = (addr + (SIZE)) & ~((SIZE) - 1); \ next 372 arch/s390/mm/pgalloc.c return (next - 1) < (end - 1) ? next : end; \ next 394 arch/s390/mm/pgalloc.c unsigned long *pte, next; next 401 arch/s390/mm/pgalloc.c next = base_page_addr_end(addr, end); next 403 arch/s390/mm/pgalloc.c } while (pte++, addr = next, addr < end); next 410 arch/s390/mm/pgalloc.c unsigned long *ste, next, table; next 416 arch/s390/mm/pgalloc.c next = base_segment_addr_end(addr, end); next 426 arch/s390/mm/pgalloc.c rc = base_page_walk(table, addr, next, alloc); next 432 arch/s390/mm/pgalloc.c } while (ste++, addr = next, addr < end); next 439 arch/s390/mm/pgalloc.c unsigned long *rtte, next, table; next 445 arch/s390/mm/pgalloc.c next = base_region3_addr_end(addr, end); next 455 arch/s390/mm/pgalloc.c rc = base_segment_walk(table, addr, next, alloc); next 460 arch/s390/mm/pgalloc.c } while (rtte++, addr = next, addr < end); next 467 arch/s390/mm/pgalloc.c unsigned long *rste, next, table; next 473 arch/s390/mm/pgalloc.c next = base_region2_addr_end(addr, end); next 483 arch/s390/mm/pgalloc.c rc = base_region3_walk(table, addr, next, alloc); next 488 arch/s390/mm/pgalloc.c } while (rste++, addr = next, addr < end); next 495 arch/s390/mm/pgalloc.c unsigned long *rfte, next, table; next 501 arch/s390/mm/pgalloc.c next = base_region1_addr_end(addr, end); next 511 arch/s390/mm/pgalloc.c rc = base_region2_walk(table, addr, next, alloc); next 516 arch/s390/mm/pgalloc.c } while (rfte++, addr = next, addr < end); next 546 arch/s390/mm/pgtable.c pmd_huge_pte(mm, pmdp) = (pgtable_t) lh->next; next 117 arch/sh/drivers/pci/pci.c hose_tail = &hose->next; next 157 arch/sh/drivers/pci/pci.c for (hose = hose_head; hose; hose = hose->next) next 264 arch/sh/drivers/pci/pci.c for (hose = hose_head; hose; hose = hose->next) { next 273 arch/sh/include/asm/dwarf.h struct dwarf_frame *prev, *next; next 122 arch/sh/include/asm/mmu_context.h struct mm_struct *next, next 127 arch/sh/include/asm/mmu_context.h if (likely(prev != next)) { next 128 arch/sh/include/asm/mmu_context.h cpumask_set_cpu(cpu, mm_cpumask(next)); next 129 arch/sh/include/asm/mmu_context.h set_TTB(next->pgd); next 130 arch/sh/include/asm/mmu_context.h activate_context(next, cpu); next 132 arch/sh/include/asm/mmu_context.h if (!cpumask_test_and_set_cpu(cpu, mm_cpumask(next))) next 133 arch/sh/include/asm/mmu_context.h activate_context(next, cpu); next 136 arch/sh/include/asm/mmu_context.h #define activate_mm(prev, next) switch_mm((prev),(next),NULL) next 18 arch/sh/include/asm/pci.h struct pci_channel *next; next 66 arch/sh/include/asm/processor.h unsigned long long next; next 65 arch/sh/include/asm/switch_to_32.h struct task_struct *next); next 70 arch/sh/include/asm/switch_to_32.h #define switch_to(prev, next, last) \ next 82 arch/sh/include/asm/switch_to_32.h if (is_dsp_enabled(next)) \ next 83 arch/sh/include/asm/switch_to_32.h __restore_dsp(next); \ next 88 arch/sh/include/asm/switch_to_32.h __ts5 = (u32 *)next; \ next 89 arch/sh/include/asm/switch_to_32.h __ts6 = (u32 *)&next->thread.sp; \ next 90 arch/sh/include/asm/switch_to_32.h __ts7 = next->thread.pc; \ next 18 arch/sh/include/asm/switch_to_64.h struct task_struct *next, next 21 arch/sh/include/asm/switch_to_64.h #define switch_to(prev,next,last) \ next 23 arch/sh/include/asm/switch_to_64.h if (last_task_used_math != next) { \ next 24 arch/sh/include/asm/switch_to_64.h struct pt_regs *regs = next->thread.uregs; \ next 27 arch/sh/include/asm/switch_to_64.h last = sh64_switch_to(prev, &prev->thread, next, \ next 28 arch/sh/include/asm/switch_to_64.h &next->thread); \ next 33 arch/sh/kernel/asm-offsets.c DEFINE(PBE_NEXT, offsetof(struct pbe, next)); next 148 arch/sh/kernel/cpu/proc.c .next = c_next, next 34 arch/sh/kernel/cpu/sh4/sq.c struct sq_mapping *next; next 78 arch/sh/kernel/cpu/sh4/sq.c p = &tmp->next; next 80 arch/sh/kernel/cpu/sh4/sq.c map->next = tmp; next 92 arch/sh/kernel/cpu/sh4/sq.c for (p = &sq_mapping_list; (tmp = *p); p = &tmp->next) next 94 arch/sh/kernel/cpu/sh4/sq.c *p = tmp->next; next 211 arch/sh/kernel/cpu/sh4/sq.c for (p = &sq_mapping_list; (map = *p); p = &map->next) next 293 arch/sh/kernel/cpu/sh4/sq.c for (list = &sq_mapping_list; (entry = *list); list = &entry->next) next 173 arch/sh/kernel/process_32.c __switch_to(struct task_struct *prev, struct task_struct *next) next 175 arch/sh/kernel/process_32.c struct thread_struct *next_t = &next->thread; next 178 arch/sh/kernel/process_32.c __stack_chk_guard = next->stack_canary; next 184 arch/sh/kernel/process_32.c if (next->thread.fpu_counter > 5) next 194 arch/sh/kernel/process_32.c : "r" (task_thread_info(next))); next 202 arch/sh/kernel/process_32.c if (next->thread.fpu_counter > 5) next 34 arch/sh/kernel/unwinder.c .next = &unwinder_list, next 52 arch/sh/kernel/unwinder.c .next = &stack_reader.list, next 73 arch/sh/kernel/unwinder.c best = list_entry(unwinder_list.next, struct unwinder, list); next 27 arch/sh/mm/tlb-sh5.c cpu_data->dtlb.next = cpu_data->dtlb.first; next 38 arch/sh/mm/tlb-sh5.c cpu_data->itlb.next = cpu_data->itlb.first; next 51 arch/sh/mm/tlb-sh5.c return cpu_data->dtlb.next; next 62 arch/sh/mm/tlb-sh5.c cpu_data->dtlb.next += cpu_data->dtlb.step; next 104 arch/sh/mm/tlb-sh5.c cpu_data->dtlb.next = entry; next 191 arch/sh/mm/tlb-sh5.c unsigned long long next; next 214 arch/sh/mm/tlb-sh5.c next = tlbp->next; next 215 arch/sh/mm/tlb-sh5.c __flush_tlb_slot(next); next 218 arch/sh/mm/tlb-sh5.c : : "r" (next), "r" (pteh), "r" (ptel) ); next 220 arch/sh/mm/tlb-sh5.c next += TLB_STEP; next 221 arch/sh/mm/tlb-sh5.c if (next > tlbp->last) next 222 arch/sh/mm/tlb-sh5.c next = tlbp->first; next 223 arch/sh/mm/tlb-sh5.c tlbp->next = next; next 25 arch/sparc/include/asm/cpudata_32.h int next; next 161 arch/sparc/include/asm/leon_amba.h struct amba_confarea_type *next;/* next bus in chain */ next 72 arch/sparc/include/asm/mdesc.h struct mdesc_notifier_client *next; next 28 arch/sparc/include/asm/switch_to_32.h #define SWITCH_DO_LAZY_FPU(next) /* */ next 38 arch/sparc/include/asm/switch_to_32.h #define prepare_arch_switch(next) do { \ next 58 arch/sparc/include/asm/switch_to_32.h #define switch_to(prev, next, last) do { \ next 60 arch/sparc/include/asm/switch_to_32.h SWITCH_DO_LAZY_FPU(next); \ next 61 arch/sparc/include/asm/switch_to_32.h cpumask_set_cpu(smp_processor_id(), mm_cpumask(next->active_mm)); \ next 93 arch/sparc/include/asm/switch_to_32.h "r" (task_thread_info(next)), \ next 7 arch/sparc/include/asm/switch_to_64.h #define prepare_arch_switch(next) \ next 21 arch/sparc/include/asm/switch_to_64.h #define switch_to(prev, next, last) \ next 26 arch/sparc/include/asm/switch_to_64.h : : "r" (task_thread_info(next)->current_ds));\ next 28 arch/sparc/include/asm/switch_to_64.h task_thread_info(next); \ next 59 arch/sparc/include/asm/switch_to_64.h : "0" (task_thread_info(next)), \ next 434 arch/sparc/kernel/cpu.c .next = c_next, next 216 arch/sparc/kernel/ds.c struct ds_info *next; next 752 arch/sparc/kernel/ds.c for (dp = ds_info_list; dp; dp = dp->next) { next 762 arch/sparc/kernel/ds.c for (dp = ds_info_list; dp; dp = dp->next) { next 1217 arch/sparc/kernel/ds.c dp->next = ds_info_list; next 7 arch/sparc/kernel/irq.h struct irq_bucket *next; next 170 arch/sparc/kernel/irq_32.c p->next = irq_map[pil]; next 189 arch/sparc/kernel/irq_32.c pnext = &(*pnext)->next; next 190 arch/sparc/kernel/irq_32.c *pnext = p->next; next 229 arch/sparc/kernel/irq_32.c struct irq_bucket *next = p->next; next 232 arch/sparc/kernel/irq_32.c p = next; next 162 arch/sparc/kernel/leon_pci_grpci2.c unsigned int next; /* 0x0C Next Data Descriptor in chan */ next 242 arch/sparc/kernel/leon_smp.c prev = &cpu_data(i).next; next 309 arch/sparc/kernel/mdesc.c client->next = client_list; next 496 arch/sparc/kernel/mdesc.c p = p->next; next 446 arch/sparc/kernel/of_device_64.c goto next; next 454 arch/sparc/kernel/of_device_64.c next: next 453 arch/sparc/kernel/pci_fire.c pbm->next = pci_pbm_root; next 62 arch/sparc/kernel/pci_impl.h struct pci_pbm_info *next; next 499 arch/sparc/kernel/pci_psycho.c for (pbm = pci_pbm_root; pbm; pbm = pbm->next) { next 576 arch/sparc/kernel/pci_psycho.c pbm->next = pci_pbm_root; next 571 arch/sparc/kernel/pci_sabre.c pbm->next = pci_pbm_root; next 1348 arch/sparc/kernel/pci_schizo.c pbm->next = pci_pbm_root; next 1408 arch/sparc/kernel/pci_schizo.c for (pbm = pci_pbm_root; pbm; pbm = pbm->next) { next 1216 arch/sparc/kernel/pci_sun4v.c pbm->next = pci_pbm_root; next 89 arch/sparc/kernel/prom_common.c prevp = &(*prevp)->next; next 41 arch/sparc/kernel/starfire.c struct starfire_irqinfo *next; next 73 arch/sparc/kernel/starfire.c p->next = sflist; next 85 arch/sparc/kernel/starfire.c for (p = sflist; p != NULL; p = p->next) next 136 arch/sparc/kernel/sun4d_irq.c struct irq_bucket *next; next 138 arch/sparc/kernel/sun4d_irq.c next = p->next; next 140 arch/sparc/kernel/sun4d_irq.c p = next; next 175 arch/sparc/kernel/sun4d_irq.c struct irq_bucket *next; next 177 arch/sparc/kernel/sun4d_irq.c next = p->next; next 179 arch/sparc/kernel/sun4d_irq.c p = next; next 173 arch/sparc/kernel/sun4d_smp.c prev = &cpu_data(i).next; next 131 arch/sparc/kernel/sun4m_smp.c prev = &cpu_data(i).next; next 206 arch/sparc/kernel/time_32.c unsigned int next = (unsigned int)delta; next 208 arch/sparc/kernel/time_32.c sparc_config.load_profile_irq(cpu, next); next 420 arch/sparc/mm/hugetlbpage.c unsigned long next; next 426 arch/sparc/mm/hugetlbpage.c next = pmd_addr_end(addr, end); next 433 arch/sparc/mm/hugetlbpage.c } while (pmd++, addr = next, addr != end); next 457 arch/sparc/mm/hugetlbpage.c unsigned long next; next 463 arch/sparc/mm/hugetlbpage.c next = pud_addr_end(addr, end); next 469 arch/sparc/mm/hugetlbpage.c hugetlb_free_pmd_range(tlb, pud, addr, next, floor, next 471 arch/sparc/mm/hugetlbpage.c } while (pud++, addr = next, addr != end); next 495 arch/sparc/mm/hugetlbpage.c unsigned long next; next 515 arch/sparc/mm/hugetlbpage.c next = pgd_addr_end(addr, end); next 518 arch/sparc/mm/hugetlbpage.c hugetlb_free_pud_range(tlb, pgd, addr, next, floor, ceiling); next 519 arch/sparc/mm/hugetlbpage.c } while (pgd++, addr = next, addr != end); next 112 arch/sparc/mm/io-unit.c next: j = (i & 15); next 126 arch/sparc/mm/io-unit.c goto next; next 406 arch/sparc/mm/srmmu.c struct ctx_list *next; next 421 arch/sparc/mm/srmmu.c entry->next->prev = entry->prev; next 422 arch/sparc/mm/srmmu.c entry->prev->next = entry->next; next 427 arch/sparc/mm/srmmu.c entry->next = head; next 428 arch/sparc/mm/srmmu.c (entry->prev = head->prev)->next = entry; next 439 arch/sparc/mm/srmmu.c ctxp = ctx_free.next; next 447 arch/sparc/mm/srmmu.c ctxp = ctx_used.next; next 449 arch/sparc/mm/srmmu.c ctxp = ctxp->next; next 487 arch/sparc/mm/srmmu.c ctx_free.next = ctx_free.prev = &ctx_free; next 488 arch/sparc/mm/srmmu.c ctx_used.next = ctx_used.prev = &ctx_used; next 293 arch/sparc/mm/tlb.c pmd_huge_pte(mm, pmdp) = (pgtable_t) lh->next; next 329 arch/um/drivers/chan_kern.c struct list_head *ele, *next; next 332 arch/um/drivers/chan_kern.c list_for_each_safe(ele, next, chans) { next 707 arch/um/drivers/line.c struct list_head *ele, *next; next 713 arch/um/drivers/line.c list_for_each_safe(ele, next, &winch_handlers) { next 727 arch/um/drivers/line.c struct list_head *ele, *next; next 732 arch/um/drivers/line.c list_for_each_safe(ele, next, &winch_handlers) { next 65 arch/um/drivers/mconsole_kern.c req = list_entry(mc_requests.next, struct mconsole_entry, list); next 322 arch/um/drivers/mconsole_kern.c unplugged = list_entry(unplugged_pages.next, next 348 arch/um/drivers/mconsole_kern.c struct list_head *entry = unplugged_pages.next; next 827 arch/um/drivers/mconsole_kern.c .next = NULL, next 587 arch/um/drivers/net_kern.c struct list_head *ele, *next; next 598 arch/um/drivers/net_kern.c list_for_each_safe(ele, next, ð_cmd_line) { next 238 arch/um/drivers/port_kern.c conn = list_entry(port->connections.next, struct connection, next 60 arch/um/include/asm/mmu_context.h static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, next 65 arch/um/include/asm/mmu_context.h if(prev != next){ next 67 arch/um/include/asm/mmu_context.h cpumask_set_cpu(cpu, mm_cpumask(next)); next 68 arch/um/include/asm/mmu_context.h if(next != &init_mm) next 69 arch/um/include/asm/mmu_context.h __switch_mm(&next->context.id); next 13 arch/um/include/shared/irq_user.h struct irq_fd *next; next 36 arch/um/include/shared/mem_user.h struct iomem_region *next; next 33 arch/um/kernel/irq.c struct irq_entry *next; next 147 arch/um/kernel/irq.c irq_entry != NULL; irq_entry = irq_entry->next) { next 165 arch/um/kernel/irq.c irq_entry->next = active_fds; next 243 arch/um/kernel/irq.c active_fds = walk->next; next 245 arch/um/kernel/irq.c previous->next = walk->next; next 250 arch/um/kernel/irq.c walk = walk->next; next 266 arch/um/kernel/irq.c walk = walk->next; next 350 arch/um/kernel/irq.c to_free = to_free->next; next 402 arch/um/kernel/irq.c to_free = to_free->next; next 136 arch/um/kernel/physmem.c region = region->next; next 191 arch/um/kernel/physmem.c region = region->next; next 216 arch/um/kernel/physmem.c region = region->next; next 208 arch/um/kernel/process.c unsigned long long next = time_travel_time + duration; next 219 arch/um/kernel/process.c time_travel_timer_expiry < next) { next 232 arch/um/kernel/process.c time_travel_set_time(next); next 239 arch/um/kernel/process.c os_timer_one_shot(time_travel_timer_expiry - next); next 263 arch/um/kernel/tlb.c unsigned long next; next 268 arch/um/kernel/tlb.c next = pmd_addr_end(addr, end); next 271 arch/um/kernel/tlb.c ret = add_munmap(addr, next - addr, hvc); next 275 arch/um/kernel/tlb.c else ret = update_pte_range(pmd, addr, next, hvc); next 276 arch/um/kernel/tlb.c } while (pmd++, addr = next, ((addr < end) && !ret)); next 285 arch/um/kernel/tlb.c unsigned long next; next 290 arch/um/kernel/tlb.c next = pud_addr_end(addr, end); next 293 arch/um/kernel/tlb.c ret = add_munmap(addr, next - addr, hvc); next 297 arch/um/kernel/tlb.c else ret = update_pmd_range(pud, addr, next, hvc); next 298 arch/um/kernel/tlb.c } while (pud++, addr = next, ((addr < end) && !ret)); next 307 arch/um/kernel/tlb.c unsigned long addr = start_addr, next; next 313 arch/um/kernel/tlb.c next = pgd_addr_end(addr, end_addr); next 316 arch/um/kernel/tlb.c ret = add_munmap(addr, next - addr, &hvc); next 320 arch/um/kernel/tlb.c else ret = update_pud_range(pgd, addr, next, &hvc); next 321 arch/um/kernel/tlb.c } while (pgd++, addr = next, ((addr < end_addr) && !ret)); next 95 arch/um/kernel/um_arch.c .next = c_next, next 224 arch/um/kernel/um_arch.c .next = NULL, next 378 arch/um/os-Linux/start_up.c *new = ((struct iomem_region) { .next = iomem_regions, next 46 arch/unicore32/include/asm/mmu_context.h switch_mm(struct mm_struct *prev, struct mm_struct *next, next 51 arch/unicore32/include/asm/mmu_context.h if (!cpumask_test_and_set_cpu(cpu, mm_cpumask(next)) || prev != next) next 52 arch/unicore32/include/asm/mmu_context.h cpu_switch_mm(next->pgd, next); next 56 arch/unicore32/include/asm/mmu_context.h #define activate_mm(prev, next) switch_mm(prev, next, NULL) next 21 arch/unicore32/include/asm/switch_to.h #define switch_to(prev, next, last) \ next 24 arch/unicore32/include/asm/switch_to.h task_thread_info(next)); \ next 96 arch/unicore32/kernel/asm-offsets.c DEFINE(PBE_NEXT, offsetof(struct pbe, next)); next 351 arch/unicore32/kernel/setup.c .next = c_next, next 36 arch/unicore32/kernel/time.c unsigned long next, oscr; next 39 arch/unicore32/kernel/time.c next = readl(OST_OSCR) + delta; next 40 arch/unicore32/kernel/time.c writel(next, OST_OSMR0); next 43 arch/unicore32/kernel/time.c return (signed)(next - oscr) <= MIN_OSCR_DELTA ? -ETIME : 0; next 239 arch/unicore32/mm/mmu.c unsigned long next = pgd_addr_end(addr, end); next 241 arch/unicore32/mm/mmu.c alloc_init_section(pgd, addr, next, phys, type); next 243 arch/unicore32/mm/mmu.c phys += next - addr; next 244 arch/unicore32/mm/mmu.c addr = next; next 80 arch/x86/boot/compressed/acpi.c pa_data = data->next; next 84 arch/x86/boot/compressed/eboot.c rom->data.next = 0; next 164 arch/x86/boot/compressed/eboot.c while (data && data->next) next 165 arch/x86/boot/compressed/eboot.c data = (struct setup_data *)(unsigned long)data->next; next 184 arch/x86/boot/compressed/eboot.c data->next = (unsigned long)rom; next 233 arch/x86/boot/compressed/eboot.c new->next = 0; next 239 arch/x86/boot/compressed/eboot.c while (data->next) next 240 arch/x86/boot/compressed/eboot.c data = (struct setup_data *)(unsigned long)data->next; next 241 arch/x86/boot/compressed/eboot.c data->next = (unsigned long)new; next 498 arch/x86/boot/compressed/eboot.c e820ext->next = 0; next 502 arch/x86/boot/compressed/eboot.c while (data && data->next) next 503 arch/x86/boot/compressed/eboot.c data = (struct setup_data *)(unsigned long)data->next; next 506 arch/x86/boot/compressed/eboot.c data->next = (unsigned long)e820ext; next 462 arch/x86/boot/compressed/kaslr.c ptr = (struct setup_data *)(unsigned long)ptr->next; next 1815 arch/x86/events/core.c for (quirk = x86_pmu.quirks; quirk; quirk = quirk->next) next 788 arch/x86/events/intel/pt.c buf->cur = list_entry(buf->cur->list.next, struct topa, next 529 arch/x86/events/perf_event.h struct x86_pmu_quirk *next; next 726 arch/x86/events/perf_event.h __quirk.next = x86_pmu.quirks; \ next 139 arch/x86/include/asm/mce.h unsigned next; next 22 arch/x86/include/asm/mmu_context.h struct mm_struct *next) next 152 arch/x86/include/asm/mmu_context.h static inline void switch_ldt(struct mm_struct *prev, struct mm_struct *next) next 173 arch/x86/include/asm/mmu_context.h (unsigned long)next->context.ldt)) next 174 arch/x86/include/asm/mmu_context.h load_mm_ldt(next); next 210 arch/x86/include/asm/mmu_context.h extern void switch_mm(struct mm_struct *prev, struct mm_struct *next, next 213 arch/x86/include/asm/mmu_context.h extern void switch_mm_irqs_off(struct mm_struct *prev, struct mm_struct *next, next 217 arch/x86/include/asm/mmu_context.h #define activate_mm(prev, next) \ next 219 arch/x86/include/asm/mmu_context.h paravirt_activate_mm((prev), (next)); \ next 220 arch/x86/include/asm/mmu_context.h switch_mm((prev), (next), NULL); \ next 303 arch/x86/include/asm/paravirt.h struct mm_struct *next) next 305 arch/x86/include/asm/paravirt.h PVOP_VCALL2(mmu.activate_mm, prev, next); next 609 arch/x86/include/asm/paravirt.h static inline void arch_end_context_switch(struct task_struct *next) next 611 arch/x86/include/asm/paravirt.h PVOP_VCALL1(cpu.end_context_switch, next); next 179 arch/x86/include/asm/paravirt_types.h void (*end_context_switch)(struct task_struct *next); next 226 arch/x86/include/asm/paravirt_types.h struct mm_struct *next); next 657 arch/x86/include/asm/paravirt_types.h void paravirt_end_context_switch(struct task_struct *next); next 10 arch/x86/include/asm/switch_to.h struct task_struct *next); next 13 arch/x86/include/asm/switch_to.h struct task_struct *next); next 16 arch/x86/include/asm/switch_to.h static inline void prepare_switch_to(struct task_struct *next) next 32 arch/x86/include/asm/switch_to.h READ_ONCE(*(unsigned char *)next->thread.sp); next 68 arch/x86/include/asm/switch_to.h #define switch_to(prev, next, last) \ next 70 arch/x86/include/asm/switch_to.h prepare_switch_to(next); \ next 72 arch/x86/include/asm/switch_to.h ((last) = __switch_to_asm((prev), (next))); \ next 46 arch/x86/include/uapi/asm/bootparam.h __u64 next; next 485 arch/x86/kernel/alternative.c struct list_head next; next 520 arch/x86/kernel/alternative.c list_add_tail(&smp->next, &smp_alt_modules); next 532 arch/x86/kernel/alternative.c list_for_each_entry(item, &smp_alt_modules, next) { next 535 arch/x86/kernel/alternative.c list_del(&item->next); next 556 arch/x86/kernel/alternative.c list_for_each_entry(mod, &smp_alt_modules, next) next 577 arch/x86/kernel/alternative.c list_for_each_entry(mod, &smp_alt_modules, next) { next 333 arch/x86/kernel/apm_32.c struct apm_user *next; next 1183 arch/x86/kernel/apm_32.c for (as = user_list; as != NULL; as = as->next) { next 1271 arch/x86/kernel/apm_32.c for (as = user_list; as != NULL; as = as->next) { next 1586 arch/x86/kernel/apm_32.c user_list = as->next; next 1591 arch/x86/kernel/apm_32.c (as1 != NULL) && (as1->next != as); next 1592 arch/x86/kernel/apm_32.c as1 = as1->next) next 1597 arch/x86/kernel/apm_32.c as1->next = as->next; next 1627 arch/x86/kernel/apm_32.c as->next = user_list; next 49 arch/x86/kernel/asm-offsets.c OFFSET(pbe_next, pbe, next); next 48 arch/x86/kernel/cpu/mce/dev-mcelog.c entry = mcelog.next; next 59 arch/x86/kernel/cpu/mce/dev-mcelog.c mcelog.next = entry + 1; next 204 arch/x86/kernel/cpu/mce/dev-mcelog.c unsigned next; next 220 arch/x86/kernel/cpu/mce/dev-mcelog.c next = mcelog.next; next 223 arch/x86/kernel/cpu/mce/dev-mcelog.c for (i = 0; i < next; i++) { next 230 arch/x86/kernel/cpu/mce/dev-mcelog.c memset(mcelog.entry, 0, next * sizeof(struct mce)); next 231 arch/x86/kernel/cpu/mce/dev-mcelog.c mcelog.next = 0; next 245 arch/x86/kernel/cpu/mce/dev-mcelog.c if (READ_ONCE(mcelog.next)) next 373 arch/x86/kernel/cpu/mce/severity.c .next = s_next, next 630 arch/x86/kernel/cpu/microcode/amd.c __list_del(p->plist.prev, p->plist.next); next 354 arch/x86/kernel/cpu/microcode/intel.c goto next; next 363 arch/x86/kernel/cpu/microcode/intel.c goto next; next 372 arch/x86/kernel/cpu/microcode/intel.c goto next; next 378 arch/x86/kernel/cpu/microcode/intel.c next: next 164 arch/x86/kernel/cpu/proc.c .next = c_next, next 268 arch/x86/kernel/cpu/resctrl/ctrlmondata.c next: next 299 arch/x86/kernel/cpu/resctrl/ctrlmondata.c goto next; next 176 arch/x86/kernel/cpu/resctrl/pseudo_lock.c struct pseudo_lock_pm_req *pm_req, *next; next 178 arch/x86/kernel/cpu/resctrl/pseudo_lock.c list_for_each_entry_safe(pm_req, next, &plr->pm_reqs, list) { next 240 arch/x86/kernel/dumpstack.c goto next; next 263 arch/x86/kernel/dumpstack.c next: next 1002 arch/x86/kernel/e820.c pa_data = data->next; next 102 arch/x86/kernel/hpet.c char *next = strchr(str, ','); next 104 arch/x86/kernel/hpet.c if (next) next 105 arch/x86/kernel/hpet.c *next++ = 0; next 112 arch/x86/kernel/hpet.c str = next; next 177 arch/x86/kernel/jailhouse.c pa_data = header.next; next 115 arch/x86/kernel/kdebugfs.c pa_data = data->next; next 154 arch/x86/kernel/kexec-bzimage64.c sd->next = params->hdr.setup_data; next 85 arch/x86/kernel/ksysfs.c pa_data = data->next; next 108 arch/x86/kernel/ksysfs.c pa_data = data->next; next 257 arch/x86/kernel/ksysfs.c pa_data = data->next; next 177 arch/x86/kernel/kvm.c struct hlist_node *p, *next; next 180 arch/x86/kernel/kvm.c hlist_for_each_safe(p, next, &b->list) { next 262 arch/x86/kernel/paravirt.c void paravirt_end_context_switch(struct task_struct *next) next 268 arch/x86/kernel/paravirt.c if (test_and_clear_ti_thread_flag(task_thread_info(next), TIF_LAZY_MMU_UPDATES)) next 273 arch/x86/kernel/process.c struct thread_struct *next, next 283 arch/x86/kernel/process.c memcpy(tss->io_bitmap, next->io_bitmap_ptr, next 284 arch/x86/kernel/process.c max(prev->io_bitmap_max, next->io_bitmap_max)); next 492 arch/x86/kernel/process.c struct thread_struct *prev, *next; next 496 arch/x86/kernel/process.c next = &next_p->thread; next 500 arch/x86/kernel/process.c switch_to_bitmap(prev, next, tifp, tifn); next 14 arch/x86/kernel/process.h struct task_struct *next) next 16 arch/x86/kernel/process.h unsigned long next_tif = task_thread_info(next)->flags; next 38 arch/x86/kernel/process.h __switch_to_xtra(prev, next); next 231 arch/x86/kernel/process_32.c *next = &next_p->thread; next 233 arch/x86/kernel/process_32.c struct fpu *next_fpu = &next->fpu; next 256 arch/x86/kernel/process_32.c load_TLS(next, cpu); next 264 arch/x86/kernel/process_32.c if (get_kernel_rpl() && unlikely(prev->iopl != next->iopl)) next 265 arch/x86/kernel/process_32.c set_iopl_mask(next->iopl); next 282 arch/x86/kernel/process_32.c refresh_sysenter_cs(next); next 290 arch/x86/kernel/process_32.c if (prev->gs | next->gs) next 291 arch/x86/kernel/process_32.c lazy_load_gs(next->gs); next 281 arch/x86/kernel/process_64.c struct thread_struct *next) next 284 arch/x86/kernel/process_64.c next->fsindex, next->fsbase, FS); next 286 arch/x86/kernel/process_64.c next->gsindex, next->gsbase, GS); next 507 arch/x86/kernel/process_64.c struct thread_struct *next = &next_p->thread; next 509 arch/x86/kernel/process_64.c struct fpu *next_fpu = &next->fpu; next 529 arch/x86/kernel/process_64.c load_TLS(next, cpu); next 553 arch/x86/kernel/process_64.c if (unlikely(next->es | prev->es)) next 554 arch/x86/kernel/process_64.c loadsegment(es, next->es); next 557 arch/x86/kernel/process_64.c if (unlikely(next->ds | prev->ds)) next 558 arch/x86/kernel/process_64.c loadsegment(ds, next->ds); next 560 arch/x86/kernel/process_64.c x86_fsgsbase_load(prev, next); next 582 arch/x86/kernel/process_64.c prev->iopl != next->iopl)) next 583 arch/x86/kernel/process_64.c xen_set_iopl_mask(next->iopl); next 190 arch/x86/kernel/rtc.c for (id = dev->id; id; id = id->next) { next 412 arch/x86/kernel/setup.c pa_next = data->next; next 441 arch/x86/kernel/setup.c pa_data = data->next; next 24 arch/x86/mm/ident_map.c unsigned long next; next 26 arch/x86/mm/ident_map.c for (; addr < end; addr = next) { next 30 arch/x86/mm/ident_map.c next = (addr & PUD_MASK) + PUD_SIZE; next 31 arch/x86/mm/ident_map.c if (next > end) next 32 arch/x86/mm/ident_map.c next = end; next 48 arch/x86/mm/ident_map.c ident_pmd_init(info, pmd, addr, next); next 54 arch/x86/mm/ident_map.c ident_pmd_init(info, pmd, addr, next); next 64 arch/x86/mm/ident_map.c unsigned long next; next 66 arch/x86/mm/ident_map.c for (; addr < end; addr = next) { next 70 arch/x86/mm/ident_map.c next = (addr & P4D_MASK) + P4D_SIZE; next 71 arch/x86/mm/ident_map.c if (next > end) next 72 arch/x86/mm/ident_map.c next = end; next 76 arch/x86/mm/ident_map.c ident_pud_init(info, pud, addr, next); next 82 arch/x86/mm/ident_map.c ident_pud_init(info, pud, addr, next); next 94 arch/x86/mm/ident_map.c unsigned long next; next 104 arch/x86/mm/ident_map.c for (; addr < end; addr = next) { next 108 arch/x86/mm/ident_map.c next = (addr & PGDIR_MASK) + PGDIR_SIZE; next 109 arch/x86/mm/ident_map.c if (next > end) next 110 arch/x86/mm/ident_map.c next = end; next 114 arch/x86/mm/ident_map.c result = ident_p4d_init(info, p4d, addr, next); next 123 arch/x86/mm/ident_map.c result = ident_p4d_init(info, p4d, addr, next); next 618 arch/x86/mm/init.c unsigned long next, start; next 634 arch/x86/mm/init.c next = round_up(start + 1, step_size); next 635 arch/x86/mm/init.c if (next > map_end) next 636 arch/x86/mm/init.c next = map_end; next 638 arch/x86/mm/init.c next = map_end; next 641 arch/x86/mm/init.c mapped_ram_size += init_range_memory_mapping(start, next); next 642 arch/x86/mm/init.c start = next; next 961 arch/x86/mm/init_64.c unsigned long next, pages = 0; next 967 arch/x86/mm/init_64.c for (; addr < end; addr = next, pte++) { next 968 arch/x86/mm/init_64.c next = (addr + PAGE_SIZE) & PAGE_MASK; next 969 arch/x86/mm/init_64.c if (next > end) next 970 arch/x86/mm/init_64.c next = end; next 984 arch/x86/mm/init_64.c if (PAGE_ALIGNED(addr) && PAGE_ALIGNED(next)) { next 1009 arch/x86/mm/init_64.c memset((void *)addr, PAGE_INUSE, next - addr); next 1032 arch/x86/mm/init_64.c unsigned long next, pages = 0; next 1038 arch/x86/mm/init_64.c for (; addr < end; addr = next, pmd++) { next 1039 arch/x86/mm/init_64.c next = pmd_addr_end(addr, end); next 1046 arch/x86/mm/init_64.c IS_ALIGNED(next, PMD_SIZE)) { next 1057 arch/x86/mm/init_64.c memset((void *)addr, PAGE_INUSE, next - addr); next 1075 arch/x86/mm/init_64.c remove_pte_table(pte_base, addr, next, direct); next 1088 arch/x86/mm/init_64.c unsigned long next, pages = 0; next 1094 arch/x86/mm/init_64.c for (; addr < end; addr = next, pud++) { next 1095 arch/x86/mm/init_64.c next = pud_addr_end(addr, end); next 1102 arch/x86/mm/init_64.c IS_ALIGNED(next, PUD_SIZE)) { next 1113 arch/x86/mm/init_64.c memset((void *)addr, PAGE_INUSE, next - addr); next 1131 arch/x86/mm/init_64.c remove_pmd_table(pmd_base, addr, next, direct, altmap); next 1143 arch/x86/mm/init_64.c unsigned long next, pages = 0; next 1148 arch/x86/mm/init_64.c for (; addr < end; addr = next, p4d++) { next 1149 arch/x86/mm/init_64.c next = p4d_addr_end(addr, end); next 1157 arch/x86/mm/init_64.c remove_pud_table(pud_base, addr, next, altmap, direct); next 1176 arch/x86/mm/init_64.c unsigned long next; next 1181 arch/x86/mm/init_64.c for (addr = start; addr < end; addr = next) { next 1182 arch/x86/mm/init_64.c next = pgd_addr_end(addr, end); next 1189 arch/x86/mm/init_64.c remove_p4d_table(p4d, addr, next, altmap, direct); next 1453 arch/x86/mm/init_64.c unsigned long next; next 1459 arch/x86/mm/init_64.c for (addr = start; addr < end; addr = next) { next 1460 arch/x86/mm/init_64.c next = pmd_addr_end(addr, end); next 1505 arch/x86/mm/init_64.c vmemmap_verify((pte_t *)pmd, node, addr, next); next 1508 arch/x86/mm/init_64.c if (vmemmap_populate_basepages(addr, next, node)) next 1540 arch/x86/mm/init_64.c unsigned long next; next 1548 arch/x86/mm/init_64.c for (; addr < end; addr = next) { next 1553 arch/x86/mm/init_64.c next = (addr + PAGE_SIZE) & PAGE_MASK; next 1560 arch/x86/mm/init_64.c next = (addr + PAGE_SIZE) & PAGE_MASK; next 1567 arch/x86/mm/init_64.c next = (addr + PAGE_SIZE) & PAGE_MASK; next 1573 arch/x86/mm/init_64.c next = (addr + PAGE_SIZE) & PAGE_MASK; next 1586 arch/x86/mm/init_64.c next = pmd_addr_end(addr, end); next 647 arch/x86/mm/ioremap.c paddr_next = data->next; next 680 arch/x86/mm/ioremap.c paddr_next = data->next; next 79 arch/x86/mm/kasan_init_64.c unsigned long next; next 100 arch/x86/mm/kasan_init_64.c next = pmd_addr_end(addr, end); next 102 arch/x86/mm/kasan_init_64.c kasan_populate_pmd(pmd, addr, next, nid); next 103 arch/x86/mm/kasan_init_64.c } while (pmd++, addr = next, addr != end); next 110 arch/x86/mm/kasan_init_64.c unsigned long next; next 120 arch/x86/mm/kasan_init_64.c next = pud_addr_end(addr, end); next 122 arch/x86/mm/kasan_init_64.c kasan_populate_pud(pud, addr, next, nid); next 123 arch/x86/mm/kasan_init_64.c } while (pud++, addr = next, addr != end); next 131 arch/x86/mm/kasan_init_64.c unsigned long next; next 140 arch/x86/mm/kasan_init_64.c next = p4d_addr_end(addr, end); next 141 arch/x86/mm/kasan_init_64.c kasan_populate_p4d(p4d, addr, next, nid); next 142 arch/x86/mm/kasan_init_64.c } while (p4d++, addr = next, addr != end); next 149 arch/x86/mm/kasan_init_64.c unsigned long next; next 155 arch/x86/mm/kasan_init_64.c next = pgd_addr_end(addr, end); next 156 arch/x86/mm/kasan_init_64.c kasan_populate_pgd(pgd, addr, next, nid); next 157 arch/x86/mm/kasan_init_64.c } while (pgd++, addr = next, addr != end); next 213 arch/x86/mm/kasan_init_64.c unsigned long next; next 223 arch/x86/mm/kasan_init_64.c next = p4d_addr_end(addr, end); next 231 arch/x86/mm/kasan_init_64.c } while (p4d++, addr = next, addr != end && p4d_none(*p4d)); next 239 arch/x86/mm/kasan_init_64.c unsigned long next; next 243 arch/x86/mm/kasan_init_64.c next = pgd_addr_end(addr, end); next 244 arch/x86/mm/kasan_init_64.c kasan_early_p4d_populate(pgd, addr, next); next 245 arch/x86/mm/kasan_init_64.c } while (pgd++, addr = next, addr != end); next 486 arch/x86/mm/kmmio.c struct kmmio_fault_page *next = f->release_next; next 489 arch/x86/mm/kmmio.c f = next; next 777 arch/x86/mm/mpx.c struct vm_area_struct *next; next 794 arch/x86/mm/mpx.c next = find_vma_prev(mm, start, &prev); next 803 arch/x86/mm/mpx.c while (next && (next->vm_flags & VM_MPX)) next 804 arch/x86/mm/mpx.c next = next->vm_next; next 814 arch/x86/mm/mpx.c next = find_vma_prev(mm, start, &prev); next 816 arch/x86/mm/mpx.c (!next || next->vm_start >= bta_end_vaddr)) { next 1156 arch/x86/mm/pat.c .next = memtype_seq_next, next 76 arch/x86/mm/tlb.c static void choose_new_asid(struct mm_struct *next, u64 next_tlb_gen, next 92 arch/x86/mm/tlb.c next->context.ctx_id) next 154 arch/x86/mm/tlb.c void switch_mm(struct mm_struct *prev, struct mm_struct *next, next 160 arch/x86/mm/tlb.c switch_mm_irqs_off(prev, next, tsk); next 192 arch/x86/mm/tlb.c static inline unsigned long mm_mangle_tif_spec_ib(struct task_struct *next) next 194 arch/x86/mm/tlb.c unsigned long next_tif = task_thread_info(next)->flags; next 197 arch/x86/mm/tlb.c return (unsigned long)next->mm | ibpb; next 200 arch/x86/mm/tlb.c static void cond_ibpb(struct task_struct *next) next 202 arch/x86/mm/tlb.c if (!next || !next->mm) next 248 arch/x86/mm/tlb.c next_mm = mm_mangle_tif_spec_ib(next); next 268 arch/x86/mm/tlb.c if (this_cpu_read(cpu_tlbstate.last_user_mm) != next->mm) { next 270 arch/x86/mm/tlb.c this_cpu_write(cpu_tlbstate.last_user_mm, next->mm); next 275 arch/x86/mm/tlb.c void switch_mm_irqs_off(struct mm_struct *prev, struct mm_struct *next, next 333 arch/x86/mm/tlb.c if (real_prev == next) { next 335 arch/x86/mm/tlb.c next->context.ctx_id); next 343 arch/x86/mm/tlb.c !cpumask_test_cpu(cpu, mm_cpumask(next)))) next 344 arch/x86/mm/tlb.c cpumask_set_cpu(cpu, mm_cpumask(next)); next 361 arch/x86/mm/tlb.c next_tlb_gen = atomic64_read(&next->context.tlb_gen); next 386 arch/x86/mm/tlb.c sync_current_stack_to_mm(next); next 403 arch/x86/mm/tlb.c if (next != &init_mm) next 404 arch/x86/mm/tlb.c cpumask_set_cpu(cpu, mm_cpumask(next)); next 405 arch/x86/mm/tlb.c next_tlb_gen = atomic64_read(&next->context.tlb_gen); next 407 arch/x86/mm/tlb.c choose_new_asid(next, next_tlb_gen, &new_asid, &need_flush); next 415 arch/x86/mm/tlb.c this_cpu_write(cpu_tlbstate.ctxs[new_asid].ctx_id, next->context.ctx_id); next 417 arch/x86/mm/tlb.c load_new_mm_cr3(next->pgd, new_asid, true); next 430 arch/x86/mm/tlb.c load_new_mm_cr3(next->pgd, new_asid, false); next 439 arch/x86/mm/tlb.c this_cpu_write(cpu_tlbstate.loaded_mm, next); next 442 arch/x86/mm/tlb.c if (next != real_prev) { next 443 arch/x86/mm/tlb.c load_mm_cr4_irqsoff(next); next 444 arch/x86/mm/tlb.c switch_ldt(real_prev, next); next 697 arch/x86/pci/common.c pa_data = data->next; next 120 arch/x86/pci/i386.c struct pcibios_fwaddrmap *entry, *next; next 123 arch/x86/pci/i386.c list_for_each_entry_safe(entry, next, &pcibios_fwaddrmappings, list) { next 334 arch/x86/pci/mmconfig-shared.c if (cfg->list.next == &pci_mmcfg_list) next 337 arch/x86/pci/mmconfig-shared.c cfgx = list_entry(cfg->list.next, typeof(*cfg), list); next 1656 arch/x86/platform/uv/tlb_uv.c .next = ptc_seq_next, next 331 arch/x86/xen/enlighten_pv.c static void xen_end_context_switch(struct task_struct *next) next 334 arch/x86/xen/enlighten_pv.c paravirt_end_context_switch(next); next 977 arch/x86/xen/mmu_pv.c static void xen_activate_mm(struct mm_struct *prev, struct mm_struct *next) next 979 arch/x86/xen/mmu_pv.c spin_lock(&next->page_table_lock); next 980 arch/x86/xen/mmu_pv.c xen_pgd_pin(next); next 981 arch/x86/xen/mmu_pv.c spin_unlock(&next->page_table_lock); next 125 arch/xtensa/include/asm/mmu_context.h static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, next 129 arch/xtensa/include/asm/mmu_context.h int migrated = next->context.cpu != cpu; next 133 arch/xtensa/include/asm/mmu_context.h next->context.cpu = cpu; next 135 arch/xtensa/include/asm/mmu_context.h if (migrated || prev != next) next 136 arch/xtensa/include/asm/mmu_context.h activate_context(next, cpu); next 139 arch/xtensa/include/asm/mmu_context.h #define activate_mm(prev, next) switch_mm((prev), (next), NULL) next 23 arch/xtensa/include/asm/nommu_context.h static inline void activate_mm(struct mm_struct *prev, struct mm_struct *next) next 27 arch/xtensa/include/asm/nommu_context.h static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, next 35 arch/xtensa/include/asm/pci-bridge.h struct pci_controller *next; next 15 arch/xtensa/include/asm/switch_to.h extern void *_switch_to(void *last, void *next); next 17 arch/xtensa/include/asm/switch_to.h #define switch_to(prev,next,last) \ next 19 arch/xtensa/include/asm/switch_to.h (last) = _switch_to(prev, next); \ next 730 arch/xtensa/kernel/setup.c .next = c_next, next 64 arch/xtensa/kernel/time.c unsigned long flags, next; next 68 arch/xtensa/kernel/time.c next = get_ccount() + delta; next 69 arch/xtensa/kernel/time.c set_linux_timer(next); next 70 arch/xtensa/kernel/time.c if (next - get_ccount() > delta) next 666 arch/xtensa/platforms/iss/network.c struct list_head *ele, *next; next 670 arch/xtensa/platforms/iss/network.c list_for_each_safe(ele, next, ð_cmd_line) { next 177 block/badblocks.c sector_t next = s + sectors; next 180 block/badblocks.c next += (1<<bb->shift) - 1; next 181 block/badblocks.c next >>= bb->shift; next 182 block/badblocks.c sectors = next - s; next 881 block/bfq-iosched.c rq = rq_entry_fifo(bfqq->fifo.next); next 896 block/bfq-iosched.c struct request *next, *prev = NULL; next 899 block/bfq-iosched.c next = bfq_check_fifo(bfqq, last); next 900 block/bfq-iosched.c if (next) next 901 block/bfq-iosched.c return next; next 907 block/bfq-iosched.c next = rb_entry_rq(rbnext); next 911 block/bfq-iosched.c next = rb_entry_rq(rbnext); next 914 block/bfq-iosched.c return bfq_choose_req(bfqd, next, prev, blk_rq_pos(last)); next 2321 block/bfq-iosched.c struct request *next) next 2324 block/bfq-iosched.c *next_bfqq = bfq_init_rq(next); next 2339 block/bfq-iosched.c !list_empty(&rq->queuelist) && !list_empty(&next->queuelist) && next 2340 block/bfq-iosched.c next->fifo_time < rq->fifo_time) { next 2342 block/bfq-iosched.c list_replace_init(&next->queuelist, &rq->queuelist); next 2343 block/bfq-iosched.c rq->fifo_time = next->fifo_time; next 2346 block/bfq-iosched.c if (bfqq->next_rq == next) next 2349 block/bfq-iosched.c bfqg_stats_update_io_merged(bfqq_group(bfqq), next->cmd_flags); next 4905 block/bfq-iosched.c struct bfq_queue *__bfqq, *next; next 4916 block/bfq-iosched.c next = __bfqq->new_bfqq; next 4918 block/bfq-iosched.c __bfqq = next; next 348 block/bfq-wf2q.c struct rb_node *next; next 351 block/bfq-wf2q.c next = rb_next(&entity->rb_node); next 352 block/bfq-wf2q.c st->first_idle = bfq_entity_of(next); next 356 block/bfq-wf2q.c next = rb_prev(&entity->rb_node); next 357 block/bfq-wf2q.c st->last_idle = bfq_entity_of(next); next 1718 block/bio.c struct bio *bio, *next; next 1721 block/bio.c next = bio_dirty_list; next 1725 block/bio.c while ((bio = next) != NULL) { next 1726 block/bio.c next = bio->bi_private; next 165 block/blk-integrity.c struct request *next) next 167 block/blk-integrity.c if (blk_integrity_rq(req) == 0 && blk_integrity_rq(next) == 0) next 170 block/blk-integrity.c if (blk_integrity_rq(req) == 0 || blk_integrity_rq(next) == 0) next 174 block/blk-integrity.c bio_integrity(next->bio)->bip_flags) next 177 block/blk-integrity.c if (req->nr_integrity_segments + next->nr_integrity_segments > next 181 block/blk-integrity.c if (integrity_req_gap_back_merge(req, next->bio)) next 192 block/blk-integrity.c struct bio *next = bio->bi_next; next 205 block/blk-integrity.c bio->bi_next = next; next 218 block/blk-ioc.c struct io_cq *icq = list_entry(icq_list->next, next 680 block/blk-iolatency.c goto next; next 686 block/blk-iolatency.c goto next; next 711 block/blk-iolatency.c next: next 16 block/blk-merge.c struct request *prev_rq, struct bio *prev, struct bio *next) next 45 block/blk-merge.c bio_get_first_bvec(next, &nb); next 605 block/blk-merge.c struct request *next) next 611 block/blk-merge.c if (blk_rq_sectors(req) + bio_sectors(next->bio) > next 615 block/blk-merge.c req->nr_phys_segments = segments + blk_rq_nr_discard_segments(next); next 623 block/blk-merge.c struct request *next) next 627 block/blk-merge.c if (req_gap_back_merge(req, next->bio)) next 633 block/blk-merge.c if ((blk_rq_sectors(req) + blk_rq_sectors(next)) > next 637 block/blk-merge.c total_phys_segments = req->nr_phys_segments + next->nr_phys_segments; next 641 block/blk-merge.c if (blk_integrity_merge_rq(q, req, next) == false) next 710 block/blk-merge.c struct request *next) next 714 block/blk-merge.c else if (blk_rq_pos(req) + blk_rq_sectors(req) == blk_rq_pos(next)) next 725 block/blk-merge.c struct request *req, struct request *next) next 727 block/blk-merge.c if (!rq_mergeable(req) || !rq_mergeable(next)) next 730 block/blk-merge.c if (req_op(req) != req_op(next)) next 733 block/blk-merge.c if (rq_data_dir(req) != rq_data_dir(next) next 734 block/blk-merge.c || req->rq_disk != next->rq_disk) next 738 block/blk-merge.c !blk_write_same_mergeable(req->bio, next->bio)) next 745 block/blk-merge.c if (req->write_hint != next->write_hint) next 748 block/blk-merge.c if (req->ioprio != next->ioprio) next 759 block/blk-merge.c switch (blk_try_req_merge(req, next)) { next 761 block/blk-merge.c if (!req_attempt_discard_merge(q, req, next)) next 765 block/blk-merge.c if (!ll_merge_requests_fn(q, req, next)) next 778 block/blk-merge.c if (((req->rq_flags | next->rq_flags) & RQF_MIXED_MERGE) || next 780 block/blk-merge.c (next->cmd_flags & REQ_FAILFAST_MASK)) { next 782 block/blk-merge.c blk_rq_set_mixed_merge(next); next 790 block/blk-merge.c if (next->start_time_ns < req->start_time_ns) next 791 block/blk-merge.c req->start_time_ns = next->start_time_ns; next 793 block/blk-merge.c req->biotail->bi_next = next->bio; next 794 block/blk-merge.c req->biotail = next->biotail; next 796 block/blk-merge.c req->__data_len += blk_rq_bytes(next); next 799 block/blk-merge.c elv_merge_requests(q, req, next); next 804 block/blk-merge.c blk_account_io_merge(next); next 810 block/blk-merge.c next->bio = NULL; next 811 block/blk-merge.c return next; next 816 block/blk-merge.c struct request *next = elv_latter_request(q, rq); next 818 block/blk-merge.c if (next) next 819 block/blk-merge.c return attempt_merge(q, rq, next); next 835 block/blk-merge.c struct request *next) next 839 block/blk-merge.c free = attempt_merge(q, rq, next); next 70 block/blk-mq-debugfs.c .next = queue_requeue_list_next, next 383 block/blk-mq-debugfs.c .next = hctx_dispatch_next, next 653 block/blk-mq-debugfs.c .next = ctx_##name##_rq_list_next, \ next 852 block/blk-mq-debugfs.c rqos = rqos->next; next 746 block/blk-mq.c struct request *rq, *next; next 752 block/blk-mq.c list_for_each_entry_safe(rq, next, &rq_list, queuelist) { next 770 block/blk-mq.c rq = list_entry(rq_list.next, struct request, queuelist); next 869 block/blk-mq.c static bool blk_mq_req_expired(struct request *rq, unsigned long *next) next 882 block/blk-mq.c if (*next == 0) next 883 block/blk-mq.c *next = deadline; next 884 block/blk-mq.c else if (time_after(*next, deadline)) next 885 block/blk-mq.c *next = deadline; next 892 block/blk-mq.c unsigned long *next = priv; next 898 block/blk-mq.c if (!blk_mq_req_expired(rq, next)) next 919 block/blk-mq.c if (blk_mq_req_expired(rq, next)) next 934 block/blk-mq.c unsigned long next = 0; next 954 block/blk-mq.c blk_mq_queue_tag_busy_iter(q, blk_mq_check_expired, &next); next 956 block/blk-mq.c if (next != 0) { next 957 block/blk-mq.c mod_timer(&q->timeout, next); next 1023 block/blk-mq.c dispatch_data->rq = list_entry_rq(ctx->rq_lists[type].next); next 1751 block/blk-mq.c rq = list_entry_rq(list.next); next 2683 block/blk-mq.c struct blk_mq_hw_ctx *hctx, *next; next 2690 block/blk-mq.c list_for_each_entry_safe(hctx, next, &q->unused_hctx_list, hctx_list) { next 37 block/blk-rq-qos.c rqos = rqos->next; next 46 block/blk-rq-qos.c rqos = rqos->next; next 55 block/blk-rq-qos.c rqos = rqos->next; next 64 block/blk-rq-qos.c rqos = rqos->next; next 73 block/blk-rq-qos.c rqos = rqos->next; next 82 block/blk-rq-qos.c rqos = rqos->next; next 91 block/blk-rq-qos.c rqos = rqos->next; next 100 block/blk-rq-qos.c rqos = rqos->next; next 109 block/blk-rq-qos.c rqos = rqos->next; next 301 block/blk-rq-qos.c q->rq_qos = rqos->next; next 30 block/blk-rq-qos.h struct rq_qos *next; next 64 block/blk-rq-qos.h for (rqos = q->rq_qos; rqos; rqos = rqos->next) { next 102 block/blk-rq-qos.h rqos->next = q->rq_qos; next 113 block/blk-rq-qos.h for (cur = &q->rq_qos; *cur; cur = &(*cur)->next) { next 115 block/blk-rq-qos.h *cur = rqos->next; next 35 block/blk-softirq.c rq = list_entry(local_list.next, struct request, ipi_list); next 52 block/blk-softirq.c if (list->next == &rq->ipi_list) next 139 block/blk-softirq.c if (list->next == &req->ipi_list) next 134 block/blk.h struct bio *next) next 137 block/blk.h struct bio_integrity_payload *bip_next = bio_integrity(next); next 154 block/blk.h struct bio *next) next 238 block/blk.h struct request *next); next 43 block/cmdline-parser.c char *next = strchr(++partdef, ')'); next 45 block/cmdline-parser.c if (!next) { next 51 block/cmdline-parser.c length = min_t(int, next - partdef, next 56 block/cmdline-parser.c partdef = ++next; next 93 block/cmdline-parser.c char *next; next 105 block/cmdline-parser.c next = strchr(bdevdef, ':'); next 106 block/cmdline-parser.c if (!next) { next 111 block/cmdline-parser.c length = min_t(int, next - bdevdef, sizeof(newparts->name) - 1); next 118 block/cmdline-parser.c while (next && *(++next)) { next 119 block/cmdline-parser.c bdevdef = next; next 120 block/cmdline-parser.c next = strchr(bdevdef, ','); next 122 block/cmdline-parser.c length = (!next) ? (sizeof(buf) - 1) : next 123 block/cmdline-parser.c min_t(int, next - bdevdef, sizeof(buf) - 1); next 169 block/cmdline-parser.c char *next; next 174 block/cmdline-parser.c next = pbuf = buf = kstrdup(cmdline, GFP_KERNEL); next 180 block/cmdline-parser.c while (next && *pbuf) { next 181 block/cmdline-parser.c next = strchr(pbuf, ';'); next 182 block/cmdline-parser.c if (next) next 183 block/cmdline-parser.c *next = '\0'; next 189 block/cmdline-parser.c if (next) next 190 block/cmdline-parser.c pbuf = ++next; next 233 block/elevator.c struct hlist_node *next; next 236 block/elevator.c hash_for_each_possible_safe(e->hash, rq, next, hash, offset) { next 405 block/elevator.c struct request *next) next 410 block/elevator.c e->type->ops.requests_merged(q, rq, next); next 310 block/genhd.c struct blk_major_name *next; next 327 block/genhd.c for (dp = major_names[major_to_index(offset)]; dp; dp = dp->next) next 394 block/genhd.c p->next = NULL; next 397 block/genhd.c for (n = &major_names[index]; *n; n = &(*n)->next) { next 425 block/genhd.c for (n = &major_names[index]; *n; n = &(*n)->next) next 432 block/genhd.c *n = p->next; next 1079 block/genhd.c .next = disk_seqf_next, next 1415 block/genhd.c .next = disk_seqf_next, next 591 block/kyber-iosched.c struct request *rq, *next; next 593 block/kyber-iosched.c list_for_each_entry_safe(rq, next, rq_list, queuelist) { next 936 block/kyber-iosched.c .next = kyber_##name##_rqs_next, \ next 141 block/mq-deadline.c struct request *next) next 147 block/mq-deadline.c if (!list_empty(&req->queuelist) && !list_empty(&next->queuelist)) { next 148 block/mq-deadline.c if (time_before((unsigned long)next->fifo_time, next 150 block/mq-deadline.c list_move(&req->queuelist, &next->queuelist); next 151 block/mq-deadline.c req->fifo_time = next->fifo_time; next 158 block/mq-deadline.c deadline_remove_request(q, next); next 185 block/mq-deadline.c struct request *rq = rq_entry_fifo(dd->fifo_list[ddir].next); next 212 block/mq-deadline.c rq = rq_entry_fifo(dd->fifo_list[data_dir].next); next 688 block/mq-deadline.c .next = deadline_##name##_fifo_next, \ next 755 block/mq-deadline.c .next = deadline_dispatch_next, next 530 block/partitions/acorn.c sector_t next; next 535 block/partitions/acorn.c next = le32_to_cpu(p->start); next 537 block/partitions/acorn.c put_partition(state, slot++, start, next - start); next 538 block/partitions/acorn.c start = next; next 20 block/partitions/check.h int next; next 140 block/partitions/msdos.c if (state->next == state->limit) next 164 block/partitions/msdos.c sector_t offs, size, next; next 173 block/partitions/msdos.c next = this_sector + offs; next 177 block/partitions/msdos.c if (next < first_sector) next 179 block/partitions/msdos.c if (next + size > first_sector + first_size) next 183 block/partitions/msdos.c put_partition(state, state->next, next, size); next 184 block/partitions/msdos.c set_info(state, state->next, disksig); next 186 block/partitions/msdos.c state->parts[state->next].flags = ADDPART_FLAG_RAID; next 188 block/partitions/msdos.c if (++state->next == state->limit) next 249 block/partitions/msdos.c for (i = 0; i < max_nparts && state->next < state->limit; i++) { next 259 block/partitions/msdos.c put_partition(state, state->next++, next 298 block/partitions/msdos.c if (state->next == state->limit) next 315 block/partitions/msdos.c put_partition(state, state->next++, bsd_start, bsd_size); next 381 block/partitions/msdos.c if (state->next == state->limit) next 385 block/partitions/msdos.c put_partition(state, state->next++, next 425 block/partitions/msdos.c if (state->next == state->limit) next 429 block/partitions/msdos.c put_partition(state, state->next++, next 532 block/partitions/msdos.c state->next = 5; next 2006 block/sed-opal.c struct opal_suspend_data *suspend, *next; next 2009 block/sed-opal.c list_for_each_entry_safe(suspend, next, &dev->unlk_lst, node) { next 67 block/t10-pi.c goto next; next 80 block/t10-pi.c goto next; next 93 block/t10-pi.c next: next 99 crypto/algapi.c n->alg = (n->list.next == stack) ? alg : next 175 crypto/algapi.c if (spawns->next == NULL) next 937 crypto/algapi.c queue->backlog = queue->backlog->next; next 939 crypto/algapi.c request = queue->list.next; next 479 crypto/asymmetric_keys/asymmetric_type.c char *next; next 494 crypto/asymmetric_keys/asymmetric_type.c next = parse_buf; next 495 crypto/asymmetric_keys/asymmetric_type.c restrict_method = strsep(&next, ":"); next 497 crypto/asymmetric_keys/asymmetric_type.c if ((strcmp(restrict_method, "key_or_keyring") == 0) && next) { next 505 crypto/asymmetric_keys/asymmetric_type.c key_text = strsep(&next, ":"); next 507 crypto/asymmetric_keys/asymmetric_type.c if (next) { next 508 crypto/asymmetric_keys/asymmetric_type.c if (strcmp(next, "chain") != 0) next 65 crypto/asymmetric_keys/pkcs7_parser.c pkcs7->certs = cert->next; next 70 crypto/asymmetric_keys/pkcs7_parser.c pkcs7->crl = cert->next; next 75 crypto/asymmetric_keys/pkcs7_parser.c pkcs7->signed_infos = sinfo->next; next 100 crypto/asymmetric_keys/pkcs7_parser.c for (sinfo = sinfo->next; sinfo; sinfo = sinfo->next) next 158 crypto/asymmetric_keys/pkcs7_parser.c ctx->certs = cert->next; next 409 crypto/asymmetric_keys/pkcs7_parser.c ctx->ppcerts = &x509->next; next 673 crypto/asymmetric_keys/pkcs7_parser.c ctx->ppsinfo = &sinfo->next; next 18 crypto/asymmetric_keys/pkcs7_parser.h struct pkcs7_signed_info *next; next 166 crypto/asymmetric_keys/pkcs7_trust.c for (p = pkcs7->certs; p; p = p->next) next 169 crypto/asymmetric_keys/pkcs7_trust.c for (sinfo = pkcs7->signed_infos; sinfo; sinfo = sinfo->next) { next 134 crypto/asymmetric_keys/pkcs7_verify.c if (sinfo == NULL || sinfo->next != NULL) next 167 crypto/asymmetric_keys/pkcs7_verify.c for (x509 = pkcs7->certs; x509; x509 = x509->next, certix++) { next 210 crypto/asymmetric_keys/pkcs7_verify.c for (p = pkcs7->certs; p; p = p->next) next 261 crypto/asymmetric_keys/pkcs7_verify.c for (p = pkcs7->certs; p; p = p->next) { next 270 crypto/asymmetric_keys/pkcs7_verify.c for (p = pkcs7->certs; p; p = p->next) { next 451 crypto/asymmetric_keys/pkcs7_verify.c for (sinfo = pkcs7->signed_infos; sinfo; sinfo = sinfo->next) { next 13 crypto/asymmetric_keys/x509_parser.h struct x509_certificate *next; next 87 crypto/proc.c .next = c_next, next 162 crypto/rsa-pkcs1pad.c struct scatterlist *next) next 164 crypto/rsa-pkcs1pad.c int nsegs = next ? 2 : 1; next 169 crypto/rsa-pkcs1pad.c if (next) next 170 crypto/rsa-pkcs1pad.c sg_chain(sg, nsegs, next); next 66 crypto/xor.c tmpl->next = template_list; next 134 crypto/xor.c for (f = fastest; f; f = f->next) next 2004 drivers/acpi/acpi_video.c struct acpi_video_device *dev, *next; next 2007 drivers/acpi/acpi_video.c list_for_each_entry_safe(dev, next, &video->video_device_list, entry) { next 245 drivers/acpi/acpica/acdebug.h char **next, acpi_object_type *return_type); next 193 drivers/acpi/acpica/aclocal.h struct acpi_new_table_desc *next; next 412 drivers/acpi/acpica/aclocal.h struct acpi_sci_handler_info *next; next 431 drivers/acpi/acpica/aclocal.h struct acpi_gpe_notify_info *next; next 476 drivers/acpi/acpica/aclocal.h struct acpi_gpe_block_info *next; next 492 drivers/acpi/acpica/aclocal.h struct acpi_gpe_xrupt_info *next; next 552 drivers/acpi/acpica/aclocal.h void *next; \ next 680 drivers/acpi/acpica/aclocal.h struct acpi_address_range *next; next 736 drivers/acpi/acpica/aclocal.h union acpi_parse_object *next; /* Next op */\ next 795 drivers/acpi/acpica/aclocal.h struct acpi_comment_node *next; next 800 drivers/acpi/acpica/aclocal.h struct acpi_comment_addr_node *next; next 812 drivers/acpi/acpica/aclocal.h struct acpi_file_node *next; next 1040 drivers/acpi/acpica/aclocal.h struct acpi_interface_info *next; next 1137 drivers/acpi/acpica/aclocal.h struct acpi_external_list *next; next 1155 drivers/acpi/acpica/aclocal.h struct acpi_external_file *next; next 1160 drivers/acpi/acpica/aclocal.h struct acpi_parse_object_list *next; next 1231 drivers/acpi/acpica/aclocal.h struct acpi_debug_mem_block *next; \ next 129 drivers/acpi/acpica/acobject.h union acpi_operand_object *next; /* Link for list of acquired mutexes */ next 138 drivers/acpi/acpica/acobject.h union acpi_operand_object *next; next 277 drivers/acpi/acpica/acobject.h union acpi_operand_object *next[2]; /* Device and System handler lists */ next 288 drivers/acpi/acpica/acobject.h union acpi_operand_object *next; next 357 drivers/acpi/acpica/acobject.h ACPI_OBJECT_COMMON_HEADER union acpi_operand_object *next; /* Link for object cache and internal lists */ next 38 drivers/acpi/acpica/acstruct.h struct acpi_walk_state *next; /* Next walk_state in list */ next 86 drivers/acpi/acpica/amlresrc.h struct asl_resource_node *next; next 429 drivers/acpi/acpica/dbcmds.c next_interface = next_interface->next; next 164 drivers/acpi/acpica/dbconvert.c char *next; next 176 drivers/acpi/acpica/dbconvert.c this = acpi_db_get_next_token(this, &next, &type); next 190 drivers/acpi/acpica/dbconvert.c this = next; next 122 drivers/acpi/acpica/dbdisply.c acpi_os_printf("%20.20s : %p\n", "NextOp", op->common.next); next 537 drivers/acpi/acpica/dbdisply.c frame = frame->results.next; next 575 drivers/acpi/acpica/dbdisply.c walk_state = walk_state->next; next 883 drivers/acpi/acpica/dbdisply.c notify = notify->next; next 910 drivers/acpi/acpica/dbdisply.c gpe_block = gpe_block->next; next 913 drivers/acpi/acpica/dbdisply.c gpe_xrupt_info = gpe_xrupt_info->next; next 992 drivers/acpi/acpica/dbdisply.c handler_obj = handler_obj->address_space.next; next 1091 drivers/acpi/acpica/dbdisply.c handler_obj = handler_obj->address_space.next; next 114 drivers/acpi/acpica/dbfileio.c table_list_head = table_list_head->next; next 389 drivers/acpi/acpica/dbinput.c const struct acpi_db_command_help *next; next 392 drivers/acpi/acpica/dbinput.c next = acpi_gbl_db_command_help; next 393 drivers/acpi/acpica/dbinput.c while (next->invocation) { next 394 drivers/acpi/acpica/dbinput.c matched = acpi_db_match_command_help(command, next); next 399 drivers/acpi/acpica/dbinput.c next++; next 419 drivers/acpi/acpica/dbinput.c const struct acpi_db_command_help *next = acpi_gbl_db_command_help; next 427 drivers/acpi/acpica/dbinput.c while (next->invocation) { next 428 drivers/acpi/acpica/dbinput.c acpi_os_printf("%-38s%s", next->invocation, next 429 drivers/acpi/acpica/dbinput.c next->description); next 430 drivers/acpi/acpica/dbinput.c next++; next 455 drivers/acpi/acpica/dbinput.c char **next, acpi_object_type *return_type) next 559 drivers/acpi/acpica/dbinput.c *next = NULL; next 562 drivers/acpi/acpica/dbinput.c *next = string + 1; next 586 drivers/acpi/acpica/dbinput.c char *next; next 600 drivers/acpi/acpica/dbinput.c acpi_gbl_db_args[i] = acpi_db_get_next_token(this, &next, next 607 drivers/acpi/acpica/dbinput.c this = next; next 370 drivers/acpi/acpica/dbutils.c start_op = (method->common.value.arg)->common.next; next 205 drivers/acpi/acpica/dbxface.c union acpi_parse_object *next; next 291 drivers/acpi/acpica/dbxface.c next = op->common.next; next 292 drivers/acpi/acpica/dbxface.c op->common.next = NULL; next 325 drivers/acpi/acpica/dbxface.c op->common.next = next; next 93 drivers/acpi/acpica/dsdebug.c union acpi_parse_object *next; next 161 drivers/acpi/acpica/dsdebug.c next = op->common.next; next 162 drivers/acpi/acpica/dsdebug.c op->common.next = NULL; next 175 drivers/acpi/acpica/dsdebug.c op->common.next = next; next 189 drivers/acpi/acpica/dsdebug.c next_walk_state = next_walk_state->next; next 453 drivers/acpi/acpica/dsfield.c arg = arg->common.next; next 511 drivers/acpi/acpica/dsfield.c arg = arg->common.next; next 520 drivers/acpi/acpica/dsfield.c status = acpi_ds_get_field_names(&info, walk_state, arg->common.next); next 648 drivers/acpi/acpica/dsfield.c arg = arg->common.next; next 703 drivers/acpi/acpica/dsfield.c arg = arg->common.next; next 720 drivers/acpi/acpica/dsfield.c arg = arg->common.next; next 724 drivers/acpi/acpica/dsfield.c arg = arg->common.next; next 743 drivers/acpi/acpica/dsfield.c status = acpi_ds_get_field_names(&info, walk_state, arg->common.next); next 788 drivers/acpi/acpica/dsfield.c arg = arg->common.next; next 802 drivers/acpi/acpica/dsfield.c arg = arg->common.next; next 810 drivers/acpi/acpica/dsfield.c status = acpi_ds_get_field_names(&info, walk_state, arg->common.next); next 187 drivers/acpi/acpica/dsobject.c byte_list = arg->named.next; next 376 drivers/acpi/acpica/dsopcode.c next_op = next_op->common.next; next 701 drivers/acpi/acpica/dsopcode.c next_op = next_op->common.next; next 705 drivers/acpi/acpica/dsopcode.c next_op = next_op->common.next; next 754 drivers/acpi/acpica/dsopcode.c arg = arg->common.next; next 130 drivers/acpi/acpica/dspkginit.c arg = arg->common.next; next 246 drivers/acpi/acpica/dspkginit.c arg = arg->common.next; next 280 drivers/acpi/acpica/dspkginit.c arg = arg->common.next; next 688 drivers/acpi/acpica/dsutils.c arg = arg->common.next; next 498 drivers/acpi/acpica/dswexec.c next_op = next_op->common.next; next 517 drivers/acpi/acpica/dswload.c (u8)op->common.value.arg->common.next->common.value.integer; next 40 drivers/acpi/acpica/dswscope.c walk_state->scope_info = scope_info->scope.next; next 446 drivers/acpi/acpica/dswstate.c walk_state->next = thread->walk_state_list; next 478 drivers/acpi/acpica/dswstate.c thread->walk_state_list = walk_state->next; next 698 drivers/acpi/acpica/dswstate.c walk_state->control_state = state->common.next; next 707 drivers/acpi/acpica/dswstate.c walk_state->scope_info = state->common.next; next 716 drivers/acpi/acpica/dswstate.c walk_state->results = state->common.next; next 432 drivers/acpi/acpica/evgpe.c gpe_block = gpe_block->next; next 485 drivers/acpi/acpica/evgpe.c notify = notify->next; next 67 drivers/acpi/acpica/evgpeblk.c while (next_gpe_block->next) { next 68 drivers/acpi/acpica/evgpeblk.c next_gpe_block = next_gpe_block->next; next 71 drivers/acpi/acpica/evgpeblk.c next_gpe_block->next = gpe_block; next 114 drivers/acpi/acpica/evgpeblk.c if (!gpe_block->previous && !gpe_block->next) { next 127 drivers/acpi/acpica/evgpeblk.c gpe_block->previous->next = gpe_block->next; next 130 drivers/acpi/acpica/evgpeblk.c gpe_block->next; next 133 drivers/acpi/acpica/evgpeblk.c if (gpe_block->next) { next 134 drivers/acpi/acpica/evgpeblk.c gpe_block->next->previous = gpe_block->previous; next 243 drivers/acpi/acpica/evgpeinit.c gpe_block = gpe_block->next; next 246 drivers/acpi/acpica/evgpeinit.c gpe_xrupt_info = gpe_xrupt_info->next; next 64 drivers/acpi/acpica/evgpeutil.c gpe_block = gpe_block->next; next 67 drivers/acpi/acpica/evgpeutil.c gpe_xrupt_info = gpe_xrupt_info->next; next 150 drivers/acpi/acpica/evgpeutil.c next_gpe_xrupt = next_gpe_xrupt->next; next 167 drivers/acpi/acpica/evgpeutil.c while (next_gpe_xrupt->next) { next 168 drivers/acpi/acpica/evgpeutil.c next_gpe_xrupt = next_gpe_xrupt->next; next 171 drivers/acpi/acpica/evgpeutil.c next_gpe_xrupt->next = gpe_xrupt; next 237 drivers/acpi/acpica/evgpeutil.c gpe_xrupt->previous->next = gpe_xrupt->next; next 241 drivers/acpi/acpica/evgpeutil.c acpi_gbl_gpe_xrupt_list_head = gpe_xrupt->next; next 244 drivers/acpi/acpica/evgpeutil.c if (gpe_xrupt->next) { next 245 drivers/acpi/acpica/evgpeutil.c gpe_xrupt->next->previous = gpe_xrupt->previous; next 276 drivers/acpi/acpica/evgpeutil.c struct acpi_gpe_notify_info *next; next 311 drivers/acpi/acpica/evgpeutil.c next = notify->next; next 313 drivers/acpi/acpica/evgpeutil.c notify = next; next 140 drivers/acpi/acpica/evhandler.c handler_obj = handler_obj->address_space.next; next 304 drivers/acpi/acpica/evhandler.c handler_obj = handler_obj->address_space.next; next 502 drivers/acpi/acpica/evhandler.c handler_obj->address_space.next = obj_desc->common_notify.handler; next 185 drivers/acpi/acpica/evmisc.c handler_obj->notify.next[info->notify.handler_list_id]; next 353 drivers/acpi/acpica/evregion.c *last_obj_ptr = obj_desc->region.next; next 354 drivers/acpi/acpica/evregion.c obj_desc->region.next = NULL; /* Must clear field */ next 435 drivers/acpi/acpica/evregion.c last_obj_ptr = &obj_desc->region.next; next 436 drivers/acpi/acpica/evregion.c obj_desc = obj_desc->region.next; next 495 drivers/acpi/acpica/evregion.c region_obj->region.next = handler_obj->address_space.region_list; next 56 drivers/acpi/acpica/evsci.c sci_handler = sci_handler->next; next 206 drivers/acpi/acpica/evsci.c acpi_gbl_sci_handler_list = sci_handler->next; next 152 drivers/acpi/acpica/evxface.c handler_obj = handler_obj->notify.next[i]; next 174 drivers/acpi/acpica/evxface.c handler_obj->notify.next[i] = next 296 drivers/acpi/acpica/evxface.c handler_obj = handler_obj->notify.next[i]; next 307 drivers/acpi/acpica/evxface.c previous_handler_obj->notify.next[i] = next 308 drivers/acpi/acpica/evxface.c handler_obj->notify.next[i]; next 312 drivers/acpi/acpica/evxface.c handler_obj->notify.next[i]; next 430 drivers/acpi/acpica/evxface.c sci_handler = sci_handler->next; next 435 drivers/acpi/acpica/evxface.c new_sci_handler->next = acpi_gbl_sci_handler_list; next 493 drivers/acpi/acpica/evxface.c prev_sci_handler->next = next_sci_handler->next; next 496 drivers/acpi/acpica/evxface.c next_sci_handler->next; next 505 drivers/acpi/acpica/evxface.c next_sci_handler = next_sci_handler->next; next 448 drivers/acpi/acpica/evxfgpe.c notify = notify->next; next 454 drivers/acpi/acpica/evxfgpe.c new_notify->next = gpe_event_info->dispatch.notify_list; next 200 drivers/acpi/acpica/evxfregn.c *last_obj_ptr = handler_obj->address_space.next; next 210 drivers/acpi/acpica/evxfregn.c last_obj_ptr = &handler_obj->address_space.next; next 211 drivers/acpi/acpica/evxfregn.c handler_obj = handler_obj->address_space.next; next 122 drivers/acpi/acpica/exdump.c {ACPI_EXD_POINTER, ACPI_EXD_OFFSET(region.next), "Next"} next 210 drivers/acpi/acpica/exdump.c {ACPI_EXD_HDLR_LIST, ACPI_EXD_OFFSET(address_space.next), "Next"}, next 223 drivers/acpi/acpica/exdump.c {ACPI_EXD_POINTER, ACPI_EXD_OFFSET(notify.next[0]), next 225 drivers/acpi/acpica/exdump.c {ACPI_EXD_POINTER, ACPI_EXD_OFFSET(notify.next[1]), "Next Device Notify"} next 338 drivers/acpi/acpica/exdump.c union acpi_operand_object *next; next 440 drivers/acpi/acpica/exdump.c next = start; next 442 drivers/acpi/acpica/exdump.c acpi_os_printf("%20s : %p ", name, next); next 443 drivers/acpi/acpica/exdump.c if (next) { next 446 drivers/acpi/acpica/exdump.c (next), next->common.type); next 448 drivers/acpi/acpica/exdump.c while (next->common.next_object) { next 449 drivers/acpi/acpica/exdump.c if ((next->common.type == next 451 drivers/acpi/acpica/exdump.c data = next; next 454 drivers/acpi/acpica/exdump.c next = next->common.next_object; next 455 drivers/acpi/acpica/exdump.c acpi_os_printf("->%p(%s %2.2X)", next, next 457 drivers/acpi/acpica/exdump.c (next), next 458 drivers/acpi/acpica/exdump.c next->common.type); next 460 drivers/acpi/acpica/exdump.c if ((next == start) || (next == data)) { next 476 drivers/acpi/acpica/exdump.c next = start; next 478 drivers/acpi/acpica/exdump.c acpi_os_printf("%20s : %p", name, next); next 479 drivers/acpi/acpica/exdump.c if (next) { next 482 drivers/acpi/acpica/exdump.c (next), next 483 drivers/acpi/acpica/exdump.c next->address_space.space_id); next 485 drivers/acpi/acpica/exdump.c while (next->address_space.next) { next 486 drivers/acpi/acpica/exdump.c if ((next->common.type == next 488 drivers/acpi/acpica/exdump.c data = next; next 491 drivers/acpi/acpica/exdump.c next = next->address_space.next; next 492 drivers/acpi/acpica/exdump.c acpi_os_printf("->%p(%s %2.2X)", next, next 494 drivers/acpi/acpica/exdump.c (next), next 495 drivers/acpi/acpica/exdump.c next->address_space. next 498 drivers/acpi/acpica/exdump.c if ((next == start) || (next == data)) { next 512 drivers/acpi/acpica/exdump.c next = start; next 514 drivers/acpi/acpica/exdump.c acpi_os_printf("%20s : %p", name, next); next 515 drivers/acpi/acpica/exdump.c if (next) { next 518 drivers/acpi/acpica/exdump.c (next), next->common.type); next 520 drivers/acpi/acpica/exdump.c while (next->region.next) { next 521 drivers/acpi/acpica/exdump.c if ((next->common.type == next 523 drivers/acpi/acpica/exdump.c data = next; next 526 drivers/acpi/acpica/exdump.c next = next->region.next; next 527 drivers/acpi/acpica/exdump.c acpi_os_printf("->%p(%s %2.2X)", next, next 529 drivers/acpi/acpica/exdump.c (next), next 530 drivers/acpi/acpica/exdump.c next->common.type); next 532 drivers/acpi/acpica/exdump.c if ((next == start) || (next == data)) { next 45 drivers/acpi/acpica/exmutex.c if (obj_desc->mutex.next) { next 46 drivers/acpi/acpica/exmutex.c (obj_desc->mutex.next)->mutex.prev = obj_desc->mutex.prev; next 50 drivers/acpi/acpica/exmutex.c (obj_desc->mutex.prev)->mutex.next = obj_desc->mutex.next; next 61 drivers/acpi/acpica/exmutex.c thread->acquired_mutex_list = obj_desc->mutex.next; next 89 drivers/acpi/acpica/exmutex.c obj_desc->mutex.next = list_head; next 463 drivers/acpi/acpica/exmutex.c union acpi_operand_object *next = thread->acquired_mutex_list; next 470 drivers/acpi/acpica/exmutex.c while (next) { next 471 drivers/acpi/acpica/exmutex.c obj_desc = next; next 496 drivers/acpi/acpica/exmutex.c next = obj_desc->mutex.next; next 499 drivers/acpi/acpica/exmutex.c obj_desc->mutex.next = NULL; next 24 drivers/acpi/acpica/hwpci.c struct acpi_pci_device *next; next 179 drivers/acpi/acpica/hwpci.c list_element->next = *return_list_head; next 237 drivers/acpi/acpica/hwpci.c info = info->next; next 264 drivers/acpi/acpica/hwpci.c struct acpi_pci_device *next; next 267 drivers/acpi/acpica/hwpci.c next = list_head; next 268 drivers/acpi/acpica/hwpci.c while (next) { next 269 drivers/acpi/acpica/hwpci.c previous = next; next 270 drivers/acpi/acpica/hwpci.c next = previous->next; next 758 drivers/acpi/acpica/psargs.c prev->common.next = field; next 104 drivers/acpi/acpica/psparse.c union acpi_parse_object *next; next 235 drivers/acpi/acpica/psparse.c replacement_op->common.next = op->common.next; next 238 drivers/acpi/acpica/psparse.c op->common.next; next 249 drivers/acpi/acpica/psparse.c next = prev->common.next; next 250 drivers/acpi/acpica/psparse.c if (next == op) { next 258 drivers/acpi/acpica/psparse.c prev->common.next = next 260 drivers/acpi/acpica/psparse.c replacement_op->common.next = next 261 drivers/acpi/acpica/psparse.c op->common.next; next 262 drivers/acpi/acpica/psparse.c next = NULL; next 264 drivers/acpi/acpica/psparse.c prev->common.next = next 265 drivers/acpi/acpica/psparse.c op->common.next; next 266 drivers/acpi/acpica/psparse.c next = NULL; next 269 drivers/acpi/acpica/psparse.c prev = next; next 174 drivers/acpi/acpica/psscope.c if (scope->common.next) { next 74 drivers/acpi/acpica/pstree.c arg = arg->common.next; next 133 drivers/acpi/acpica/pstree.c while (prev_arg->common.next) { next 134 drivers/acpi/acpica/pstree.c prev_arg = prev_arg->common.next; next 136 drivers/acpi/acpica/pstree.c prev_arg->common.next = arg; next 147 drivers/acpi/acpica/pstree.c arg = arg->common.next; next 172 drivers/acpi/acpica/pstree.c union acpi_parse_object *next = NULL; next 184 drivers/acpi/acpica/pstree.c next = acpi_ps_get_arg(op, 0); next 185 drivers/acpi/acpica/pstree.c if (next) { next 186 drivers/acpi/acpica/pstree.c ASL_CV_LABEL_FILENODE(next); next 187 drivers/acpi/acpica/pstree.c return (next); next 192 drivers/acpi/acpica/pstree.c next = op->common.next; next 193 drivers/acpi/acpica/pstree.c if (next) { next 194 drivers/acpi/acpica/pstree.c ASL_CV_LABEL_FILENODE(next); next 195 drivers/acpi/acpica/pstree.c return (next); next 207 drivers/acpi/acpica/pstree.c arg = arg->common.next; next 217 drivers/acpi/acpica/pstree.c if (parent->common.next) { next 221 drivers/acpi/acpica/pstree.c ASL_CV_LABEL_FILENODE(parent->common.next); next 222 drivers/acpi/acpica/pstree.c return (parent->common.next); next 229 drivers/acpi/acpica/pstree.c ASL_CV_LABEL_FILENODE(next); next 230 drivers/acpi/acpica/pstree.c return (next); next 32 drivers/acpi/acpica/pswalk.c union acpi_parse_object *next = NULL; next 72 drivers/acpi/acpica/pswalk.c next = acpi_ps_get_arg(op, 0); next 73 drivers/acpi/acpica/pswalk.c if (next) { next 77 drivers/acpi/acpica/pswalk.c op = next; next 85 drivers/acpi/acpica/pswalk.c next = op->common.next; next 96 drivers/acpi/acpica/pswalk.c if (next) { next 97 drivers/acpi/acpica/pswalk.c op = next; next 65 drivers/acpi/acpica/utaddress.c range_info->next = acpi_gbl_address_range_list[space_id]; next 115 drivers/acpi/acpica/utaddress.c range_info->next; next 117 drivers/acpi/acpica/utaddress.c prev->next = range_info->next; next 134 drivers/acpi/acpica/utaddress.c range_info = range_info->next; next 218 drivers/acpi/acpica/utaddress.c range_info = range_info->next; next 239 drivers/acpi/acpica/utaddress.c struct acpi_address_range *next; next 246 drivers/acpi/acpica/utaddress.c next = acpi_gbl_address_range_list[i]; next 248 drivers/acpi/acpica/utaddress.c while (next) { next 249 drivers/acpi/acpica/utaddress.c range_info = next; next 250 drivers/acpi/acpica/utaddress.c next = range_info->next; next 76 drivers/acpi/acpica/utcache.c void *next; next 96 drivers/acpi/acpica/utcache.c next = ACPI_GET_DESCRIPTOR_PTR(cache->list_head); next 99 drivers/acpi/acpica/utcache.c cache->list_head = next; next 124 drivers/acpi/acpica/utdelete.c next_desc = handler_desc->address_space.next; next 216 drivers/acpi/acpica/utdelete.c next_desc->region.next; next 222 drivers/acpi/acpica/utdelete.c last_obj_ptr = &next_desc->region.next; next 223 drivers/acpi/acpica/utdelete.c next_desc = next_desc->region.next; next 512 drivers/acpi/acpica/utdelete.c prev_object->notify.next[i]; next 57 drivers/acpi/acpica/utinit.c next_gpe_block = gpe_block->next; next 64 drivers/acpi/acpica/utinit.c next_gpe_xrupt_info = gpe_xrupt_info->next; next 124 drivers/acpi/acpica/utosi.c acpi_default_supported_interfaces[i].next = next 157 drivers/acpi/acpica/utosi.c acpi_gbl_supported_interfaces = next_interface->next; next 217 drivers/acpi/acpica/utosi.c interface_info->next = acpi_gbl_supported_interfaces; next 254 drivers/acpi/acpica/utosi.c next_interface->next; next 256 drivers/acpi/acpica/utosi.c previous_interface->next = next 257 drivers/acpi/acpica/utosi.c next_interface->next; next 278 drivers/acpi/acpica/utosi.c next_interface = next_interface->next; next 323 drivers/acpi/acpica/utosi.c next_interface = next_interface->next; next 352 drivers/acpi/acpica/utosi.c next_interface = next_interface->next; next 34 drivers/acpi/acpica/utstate.c state->common.next = *list_head; next 65 drivers/acpi/acpica/utstate.c *list_head = state->common.next; next 299 drivers/acpi/acpica/uttrack.c if (!element->next) { next 303 drivers/acpi/acpica/uttrack.c element = element->next; next 383 drivers/acpi/acpica/uttrack.c allocation->next = mem_list->list_head; next 390 drivers/acpi/acpica/uttrack.c allocation->next = element->next; next 393 drivers/acpi/acpica/uttrack.c if (element->next) { next 394 drivers/acpi/acpica/uttrack.c (element->next)->previous = allocation; next 397 drivers/acpi/acpica/uttrack.c element->next = allocation; next 452 drivers/acpi/acpica/uttrack.c (allocation->previous)->next = allocation->next; next 454 drivers/acpi/acpica/uttrack.c mem_list->list_head = allocation->next; next 457 drivers/acpi/acpica/uttrack.c if (allocation->next) { next 458 drivers/acpi/acpica/uttrack.c (allocation->next)->previous = allocation->previous; next 687 drivers/acpi/acpica/uttrack.c element = element->next; next 337 drivers/acpi/apei/apei-base.c res1 = list_entry(res_list1->next, struct apei_res, list); next 338 drivers/acpi/apei/apei-base.c resn1 = list_entry(res1->list.next, struct apei_res, list); next 367 drivers/acpi/apei/apei-base.c resn1 = list_entry(resn1->list.next, struct apei_res, list); next 826 drivers/acpi/apei/ghes.c struct llist_node *llnode, *next; next 839 drivers/acpi/apei/ghes.c next = llnode->next; next 853 drivers/acpi/apei/ghes.c llnode = next; next 876 drivers/acpi/apei/ghes.c llnode = llnode->next; next 166 drivers/acpi/evged.c struct acpi_ged_event *event, *next; next 168 drivers/acpi/evged.c list_for_each_entry_safe(event, next, &geddev->event_list, node) { next 95 drivers/acpi/nvs.c struct nvs_page *entry, *next; next 118 drivers/acpi/nvs.c list_for_each_entry_safe(entry, next, &nvs_list, node) { next 693 drivers/acpi/pci_root.c goto next; next 701 drivers/acpi/pci_root.c goto next; next 725 drivers/acpi/pci_root.c goto next; next 729 drivers/acpi/pci_root.c next: next 438 drivers/acpi/pptt.c struct acpi_pptt_processor *next; next 446 drivers/acpi/pptt.c next = fetch_pptt_node(table_hdr, cpu->parent); next 447 drivers/acpi/pptt.c if (!(next && next->flags & ACPI_PPTT_ACPI_IDENTICAL)) next 25 drivers/acpi/proc.c struct list_head *node, *next; next 30 drivers/acpi/proc.c list_for_each_safe(node, next, &acpi_wakeup_device_list) { next 58 drivers/acpi/proc.c dev->physical_node_list.next) next 99 drivers/acpi/proc.c struct list_head *node, *next; next 112 drivers/acpi/proc.c list_for_each_safe(node, next, &acpi_wakeup_device_list) { next 435 drivers/acpi/property.c struct acpi_data_node *dn, *next; next 440 drivers/acpi/property.c list_for_each_entry_safe_reverse(dn, next, list, sibling) { next 1019 drivers/acpi/property.c struct list_head *next; next 1034 drivers/acpi/property.c next = adev->node.next; next 1035 drivers/acpi/property.c if (next == head) { next 1039 drivers/acpi/property.c child_adev = list_entry(next, struct acpi_device, node); next 1072 drivers/acpi/property.c next = dn->sibling.next; next 1073 drivers/acpi/property.c if (next == head) next 1076 drivers/acpi/property.c dn = list_entry(next, struct acpi_data_node, sibling); next 57 drivers/acpi/sleep.c .next = NULL, next 42 drivers/acpi/wakeup.c struct list_head *node, *next; next 44 drivers/acpi/wakeup.c list_for_each_safe(node, next, &acpi_wakeup_device_list) { next 69 drivers/acpi/wakeup.c struct list_head *node, *next; next 71 drivers/acpi/wakeup.c list_for_each_safe(node, next, &acpi_wakeup_device_list) { next 91 drivers/acpi/wakeup.c struct list_head *node, *next; next 94 drivers/acpi/wakeup.c list_for_each_safe(node, next, &acpi_wakeup_device_list) { next 800 drivers/android/binder.c BUG_ON(work->entry.next && !list_empty(&work->entry)); next 1943 drivers/android/binder.c struct binder_transaction *next; next 1979 drivers/android/binder.c next = t->from_parent; next 1986 drivers/android/binder.c if (next == NULL) { next 1991 drivers/android/binder.c t = next; next 52 drivers/android/binder_alloc.c return list_entry(buffer->entry.next, struct binder_buffer, entry); next 549 drivers/android/binder_alloc.c struct binder_buffer *prev, *next = NULL; next 551 drivers/android/binder_alloc.c BUG_ON(alloc->buffers.next == &buffer->entry); next 563 drivers/android/binder_alloc.c next = binder_buffer_next(buffer); next 564 drivers/android/binder_alloc.c if (buffer_start_page(next) == buffer_start_page(buffer)) { next 570 drivers/android/binder_alloc.c next->user_data); next 586 drivers/android/binder_alloc.c next ? next->user_data : NULL); next 631 drivers/android/binder_alloc.c struct binder_buffer *next = binder_buffer_next(buffer); next 633 drivers/android/binder_alloc.c if (next->free) { next 634 drivers/android/binder_alloc.c rb_erase(&next->rb_node, &alloc->free_buffers); next 635 drivers/android/binder_alloc.c binder_delete_free_buffer(alloc, next); next 638 drivers/android/binder_alloc.c if (alloc->buffers.next != &buffer->entry) { next 259 drivers/ata/libata-core.c next: next 277 drivers/ata/libata-core.c goto next; next 7033 drivers/ata/libata-core.c char *p, *cur, *next; next 7048 drivers/ata/libata-core.c for (cur = ata_force_param_buf; *cur != '\0'; cur = next) { next 7052 drivers/ata/libata-core.c next = cur; next 7053 drivers/ata/libata-core.c if (ata_parse_force_one(&next, &te, &reason)) { next 190 drivers/atm/adummy.c struct adummy_dev *adummy_dev, *next; next 192 drivers/atm/adummy.c list_for_each_entry_safe(adummy_dev, next, &adummy_devs, entry) { next 666 drivers/atm/eni.c eni_dev->fast = ENI_VCC(curr)->next; next 667 drivers/atm/eni.c ENI_VCC(curr)->next = ENI_VCC_NOS; next 674 drivers/atm/eni.c eni_dev->slow = ENI_VCC(curr)->next; next 675 drivers/atm/eni.c ENI_VCC(curr)->next = ENI_VCC_NOS; next 701 drivers/atm/eni.c if (ENI_VCC(vcc)->next != ENI_VCC_NOS) { next 707 drivers/atm/eni.c ENI_VCC(vcc)->next = NULL; next 710 drivers/atm/eni.c ENI_VCC(eni_dev->last_fast)->next = vcc; next 716 drivers/atm/eni.c ENI_VCC(eni_dev->last_slow)->next = vcc; next 805 drivers/atm/eni.c eni_vcc->next = ENI_VCC_NOS; next 64 drivers/atm/eni.h struct atm_vcc *next; /* next pending RX */ next 724 drivers/atm/firestream.c td->flags, td->next, td->bsa, td->aal_bufsize, td->skb ); next 784 drivers/atm/firestream.c pe->flags, pe->next, pe->bsa, pe->aal_bufsize, next 1186 drivers/atm/firestream.c td->next = 0; next 1195 drivers/atm/firestream.c dq[qd].next = td->next; next 1494 drivers/atm/firestream.c ne->next = virt_to_bus (NULL); next 1510 drivers/atm/firestream.c qe->next = virt_to_bus(ne); next 1915 drivers/atm/firestream.c fs_dev->next = fs_boards; next 1977 drivers/atm/firestream.c nxt = bus_to_virt (fp->next); next 2012 drivers/atm/firestream.c nxtdev = dev->next; next 39 drivers/atm/firestream.h u32 next; next 455 drivers/atm/firestream.h struct fs_dev *next; /* other FS devices */ next 891 drivers/atm/fore200e.c buffer = buffer->next; next 936 drivers/atm/fore200e.c bsq->freebuf = buffer->next; next 1077 drivers/atm/fore200e.c buffer->next = bsq->freebuf; next 2057 drivers/atm/fore200e.c buffer[ i ].next = bsq->freebuf; next 579 drivers/atm/fore200e.h struct buffer* next; /* next receive buffer */ next 394 drivers/atm/he.c he_dev->next = NULL; next 396 drivers/atm/he.c he_dev->next = he_devs; next 772 drivers/atm/he.c struct he_buff *heb, *next; next 894 drivers/atm/he.c list_for_each_entry_safe(heb, next, &he_dev->rbpl_outstanding, entry) next 1529 drivers/atm/he.c struct he_buff *heb, *next; next 1573 drivers/atm/he.c list_for_each_entry_safe(heb, next, &he_dev->rbpl_outstanding, entry) next 1651 drivers/atm/he.c struct he_buff *heb, *next; next 1777 drivers/atm/he.c list_for_each_entry_safe(heb, next, &he_vcc->buffers, entry) next 314 drivers/atm/he.h struct he_dev *next; next 1864 drivers/atm/horizon.c wr_mem (dev, &buf->next, BUFF_STATUS_EMPTY); next 1873 drivers/atm/horizon.c wr_mem (dev, &memmap->txfreebufstart.next, BUF_PTR(tx_desc) | BUFF_STATUS_EMPTY); next 1876 drivers/atm/horizon.c wr_mem (dev, &tx_desc->next, BUF_PTR(tx_desc+1) | BUFF_STATUS_EMPTY); next 1880 drivers/atm/horizon.c wr_mem (dev, &tx_desc->next, BUF_PTR(&memmap->txfreebufend) | BUFF_STATUS_EMPTY); next 1902 drivers/atm/horizon.c wr_mem (dev, &memmap->rxfreebufstart.next, BUF_PTR(rx_desc) | BUFF_STATUS_EMPTY); next 1905 drivers/atm/horizon.c wr_mem (dev, &rx_desc->next, BUF_PTR(rx_desc+1) | BUFF_STATUS_EMPTY); next 1910 drivers/atm/horizon.c wr_mem (dev, &rx_desc->next, BUF_PTR(&memmap->rxfreebufend) | BUFF_STATUS_EMPTY); next 299 drivers/atm/horizon.h HDW next; next 37 drivers/atm/idt77105.c struct idt77105_priv *next; next 91 drivers/atm/idt77105.c for (walk = idt77105_all; walk; walk = walk->next) { next 120 drivers/atm/idt77105.c for (walk = idt77105_all; walk; walk = walk->next) { next 269 drivers/atm/idt77105.c PRIV(dev)->next = idt77105_all; next 333 drivers/atm/idt77105.c prev = walk, walk = walk->next) { next 336 drivers/atm/idt77105.c prev->next = walk->next; next 338 drivers/atm/idt77105.c idt77105_all = walk->next; next 651 drivers/atm/idt77252.c scq->next = scq->base; next 662 drivers/atm/idt77252.c scq->base, scq->next, scq->last, (unsigned long long)scq->paddr); next 710 drivers/atm/idt77252.c TXPRINTK("%s: SCQ: next 0x%p\n", card->name, scq->next); next 745 drivers/atm/idt77252.c scq->next->word_1 = cpu_to_le32(tbd->word_1 | next 747 drivers/atm/idt77252.c scq->next->word_2 = cpu_to_le32(tbd->word_2); next 748 drivers/atm/idt77252.c scq->next->word_3 = cpu_to_le32(tbd->word_3); next 749 drivers/atm/idt77252.c scq->next->word_4 = cpu_to_le32(tbd->word_4); next 751 drivers/atm/idt77252.c if (scq->next == scq->last) next 752 drivers/atm/idt77252.c scq->next = scq->base; next 754 drivers/atm/idt77252.c scq->next++; next 758 drivers/atm/idt77252.c (u32)((unsigned long)scq->next - (unsigned long)scq->base)); next 772 drivers/atm/idt77252.c read_sram(card, scq->scd + 1), scq->next); next 798 drivers/atm/idt77252.c card->name, atomic_read(&scq->used), scq->next); next 982 drivers/atm/idt77252.c card->rsq.next = card->rsq.last; next 1211 drivers/atm/idt77252.c if (card->rsq.next == card->rsq.last) next 1214 drivers/atm/idt77252.c rsqe = card->rsq.next + 1; next 1224 drivers/atm/idt77252.c card->rsq.next = rsqe; next 1225 drivers/atm/idt77252.c if (card->rsq.next == card->rsq.last) next 1228 drivers/atm/idt77252.c rsqe = card->rsq.next + 1; next 1231 drivers/atm/idt77252.c writel((unsigned long) card->rsq.next - (unsigned long) card->rsq.base, next 1337 drivers/atm/idt77252.c struct sk_buff *next; next 1343 drivers/atm/idt77252.c next = sb_pool_skb(card, handle); next 1346 drivers/atm/idt77252.c if (next) { next 1347 drivers/atm/idt77252.c card->raw_cell_head = next; next 1384 drivers/atm/idt77252.c card->tsq.next = card->tsq.last; next 1389 drivers/atm/idt77252.c writel((unsigned long) card->tsq.next - (unsigned long) card->tsq.base, next 1410 drivers/atm/idt77252.c if (card->tsq.next == card->tsq.last) next 1413 drivers/atm/idt77252.c tsqe = card->tsq.next + 1; next 1416 drivers/atm/idt77252.c card->tsq.base, card->tsq.next, card->tsq.last); next 1507 drivers/atm/idt77252.c card->tsq.next = tsqe; next 1508 drivers/atm/idt77252.c if (card->tsq.next == card->tsq.last) next 1511 drivers/atm/idt77252.c tsqe = card->tsq.next + 1; next 1514 drivers/atm/idt77252.c card->tsq.base, card->tsq.next, card->tsq.last); next 1520 drivers/atm/idt77252.c writel((unsigned long)card->tsq.next - (unsigned long)card->tsq.base, next 1525 drivers/atm/idt77252.c readl(SAR_REG_TSQT), card->tsq.next); next 3703 drivers/atm/idt77252.c last = &card->next; next 3768 drivers/atm/idt77252.c idt77252_chain = card->next; next 163 drivers/atm/idt77252.h struct scqe *next; next 272 drivers/atm/idt77252.h struct rsq_entry *next; next 307 drivers/atm/idt77252.h struct tsq_entry *next; next 413 drivers/atm/idt77252.h struct idt77252_dev *next; next 98 drivers/atm/iphase.c que->next = NULL; next 104 drivers/atm/iphase.c data->next = NULL; next 105 drivers/atm/iphase.c if (que->next == NULL) next 106 drivers/atm/iphase.c que->next = que->tail = data; next 108 drivers/atm/iphase.c data->next = que->next; next 109 drivers/atm/iphase.c que->next = data; next 119 drivers/atm/iphase.c entry->next = NULL; next 120 drivers/atm/iphase.c if (que->next == NULL) next 121 drivers/atm/iphase.c que->next = que->tail = entry; next 123 drivers/atm/iphase.c que->tail->next = entry; next 124 drivers/atm/iphase.c que->tail = que->tail->next; next 131 drivers/atm/iphase.c if (que->next == NULL) next 133 drivers/atm/iphase.c tmpdata = que->next; next 134 drivers/atm/iphase.c if ( que->next == que->tail) next 135 drivers/atm/iphase.c que->next = que->tail = NULL; next 137 drivers/atm/iphase.c que->next = que->next->next; next 301 drivers/atm/iphase.h struct free_desc_q *next; next 888 drivers/atm/iphase.h struct ia_rtn_q *next, *tail; next 539 drivers/atm/nicstar.c card->tsq.next = card->tsq.base; next 558 drivers/atm/nicstar.c card->rsq.next = card->rsq.base; next 885 drivers/atm/nicstar.c scq->next = scq->base; next 1473 drivers/atm/nicstar.c scqep = scq->next; next 1484 drivers/atm/nicstar.c if (!ns_scqe_is_tsr(scqep) && scq->tail != scq->next) { next 1492 drivers/atm/nicstar.c scqi = scq->next - scq->base; next 1496 drivers/atm/nicstar.c *scq->next = tsr; next 1499 drivers/atm/nicstar.c if (scq->next == scq->last) next 1500 drivers/atm/nicstar.c scq->next = scq->base; next 1502 drivers/atm/nicstar.c scq->next++; next 1503 drivers/atm/nicstar.c data = scq_virt_to_bus(scq, scq->next); next 1556 drivers/atm/nicstar.c card->tsq.base, card->tsq.next, next 1560 drivers/atm/nicstar.c card->rsq.base, card->rsq.next, next 1728 drivers/atm/nicstar.c while (scq->tail == scq->next) { next 1737 drivers/atm/nicstar.c scq->tail != scq->next, next 1748 drivers/atm/nicstar.c *scq->next = *tbd; next 1749 drivers/atm/nicstar.c index = (int)(scq->next - scq->base); next 1756 drivers/atm/nicstar.c scq->next); next 1757 drivers/atm/nicstar.c if (scq->next == scq->last) next 1758 drivers/atm/nicstar.c scq->next = scq->base; next 1760 drivers/atm/nicstar.c scq->next++; next 1773 drivers/atm/nicstar.c while (scq->tail == scq->next) { next 1775 drivers/atm/nicstar.c data = scq_virt_to_bus(scq, scq->next); next 1787 drivers/atm/nicstar.c scq->tail != scq->next, next 1798 drivers/atm/nicstar.c scqi = scq->next - scq->base; next 1803 drivers/atm/nicstar.c *scq->next = tsr; next 1810 drivers/atm/nicstar.c le32_to_cpu(tsr.word_4), scq->next); next 1811 drivers/atm/nicstar.c if (scq->next == scq->last) next 1812 drivers/atm/nicstar.c scq->next = scq->base; next 1814 drivers/atm/nicstar.c scq->next++; next 1821 drivers/atm/nicstar.c data = scq_virt_to_bus(scq, scq->next); next 1838 drivers/atm/nicstar.c if (card->tsq.next == card->tsq.last) next 1841 drivers/atm/nicstar.c one_ahead = card->tsq.next + 1; next 1848 drivers/atm/nicstar.c while (!ns_tsi_isempty(card->tsq.next) || !ns_tsi_isempty(one_ahead) || next 1855 drivers/atm/nicstar.c while (ns_tsi_isempty(card->tsq.next)) { next 1856 drivers/atm/nicstar.c if (card->tsq.next == card->tsq.last) next 1857 drivers/atm/nicstar.c card->tsq.next = card->tsq.base; next 1859 drivers/atm/nicstar.c card->tsq.next++; next 1862 drivers/atm/nicstar.c if (!ns_tsi_tmrof(card->tsq.next)) { next 1863 drivers/atm/nicstar.c scdi = ns_tsi_getscdindex(card->tsq.next); next 1871 drivers/atm/nicstar.c ns_tsi_init(card->tsq.next); next 1876 drivers/atm/nicstar.c drain_scq(card, scq, ns_tsi_getscqpos(card->tsq.next)); next 1881 drivers/atm/nicstar.c ns_tsi_init(card->tsq.next); next 1882 drivers/atm/nicstar.c previous = card->tsq.next; next 1883 drivers/atm/nicstar.c if (card->tsq.next == card->tsq.last) next 1884 drivers/atm/nicstar.c card->tsq.next = card->tsq.base; next 1886 drivers/atm/nicstar.c card->tsq.next++; next 1888 drivers/atm/nicstar.c if (card->tsq.next == card->tsq.last) next 1891 drivers/atm/nicstar.c one_ahead = card->tsq.next + 1; next 1950 drivers/atm/nicstar.c if (!ns_rsqe_valid(card->rsq.next)) next 1953 drivers/atm/nicstar.c dequeue_rx(card, card->rsq.next); next 1954 drivers/atm/nicstar.c ns_rsqe_init(card->rsq.next); next 1955 drivers/atm/nicstar.c previous = card->rsq.next; next 1956 drivers/atm/nicstar.c if (card->rsq.next == card->rsq.last) next 1957 drivers/atm/nicstar.c card->rsq.next = card->rsq.base; next 1959 drivers/atm/nicstar.c card->rsq.next++; next 1960 drivers/atm/nicstar.c } while (ns_rsqe_valid(card->rsq.next)); next 658 drivers/atm/nicstar.h ns_tsi *next; next 667 drivers/atm/nicstar.h ns_scqe *next; next 684 drivers/atm/nicstar.h ns_rsqe *next; next 63 drivers/atm/suni.c for (walk = sunis; walk; walk = walk->next) { next 315 drivers/atm/suni.c PRIV(dev)->next = sunis; next 333 drivers/atm/suni.c (unsigned long) poll_timer.list.next); next 349 drivers/atm/suni.c walk = &PRIV((*walk)->dev)->next); next 350 drivers/atm/suni.c *walk = PRIV((*walk)->dev)->next; next 236 drivers/atm/suni.h struct suni_priv *next; /* next SUNI */ next 437 drivers/base/devres.c cur = cur->next; next 447 drivers/base/devres.c first = first->next; next 468 drivers/base/devres.c cur = cur->next; next 526 drivers/base/devres.c if (WARN_ON(dev->devres_head.next == NULL)) next 529 drivers/base/devres.c return release_nodes(dev, dev->devres_head.next, &dev->devres_head, next 676 drivers/base/devres.c end = grp->node[1].entry.next; next 42 drivers/base/devtmpfs.c struct req *next; next 123 drivers/base/devtmpfs.c req.next = requests; next 153 drivers/base/devtmpfs.c req.next = requests; next 410 drivers/base/devtmpfs.c struct req *next = req->next; next 414 drivers/base/devtmpfs.c req = next; next 110 drivers/base/firmware_loader/fallback.c struct fw_priv *next; next 113 drivers/base/firmware_loader/fallback.c list_for_each_entry_safe(fw_priv, next, &pending_fw_head, next 1257 drivers/base/firmware_loader/main.c fce = list_entry(fwc->fw_names.next, next 20 drivers/base/isa.c struct device *next; next 107 drivers/base/isa.c struct device *tmp = to_isa_dev(dev)->next; next 155 drivers/base/isa.c isa_dev->next = isa_driver->devices; next 21 drivers/base/map.c struct probe *next; next 60 drivers/base/map.c s = &(*s)->next; next 61 drivers/base/map.c p->next = *s; next 81 drivers/base/map.c for (s = &domain->probes[index % 255]; *s; s = &(*s)->next) { next 84 drivers/base/map.c *s = p->next; next 103 drivers/base/map.c for (p = domain->probes[MAJOR(dev) % 255]; p; p = p->next) { next 339 drivers/base/node.c struct node_cache_info *info, *next; next 344 drivers/base/node.c list_for_each_entry_safe(info, next, &node->cache_attrs, node) { next 1326 drivers/base/platform.c if (!epdrv->list.next) { next 1383 drivers/base/platform.c if (!dev->devres_head.next) { next 766 drivers/base/power/main.c dev = to_device(dpm_noirq_list.next); next 916 drivers/base/power/main.c dev = to_device(dpm_late_early_list.next); next 1077 drivers/base/power/main.c dev = to_device(dpm_suspended_list.next); next 2008 drivers/base/power/main.c struct device *dev = to_device(dpm_list.next); next 1102 drivers/base/power/wakeup.c .next = wakeup_sources_stats_seq_next, next 215 drivers/base/regmap/regcache-rbtree.c struct rb_node *next; next 225 drivers/base/regmap/regcache-rbtree.c next = rb_first(&rbtree_ctx->root); next 226 drivers/base/regmap/regcache-rbtree.c while (next) { next 227 drivers/base/regmap/regcache-rbtree.c rbtree_node = rb_entry(next, struct regcache_rbtree_node, node); next 228 drivers/base/regmap/regcache-rbtree.c next = rb_next(&rbtree_node->node); next 537 drivers/base/regmap/regmap-debugfs.c struct rb_node *next; next 623 drivers/base/regmap/regmap-debugfs.c next = rb_first(&map->range_tree); next 624 drivers/base/regmap/regmap-debugfs.c while (next) { next 625 drivers/base/regmap/regmap-debugfs.c range_node = rb_entry(next, struct regmap_range_node, node); next 632 drivers/base/regmap/regmap-debugfs.c next = rb_next(&range_node->node); next 578 drivers/base/regmap/regmap.c struct rb_node *next; next 581 drivers/base/regmap/regmap.c next = rb_first(&map->range_tree); next 582 drivers/base/regmap/regmap.c while (next) { next 583 drivers/base/regmap/regmap.c range_node = rb_entry(next, struct regmap_range_node, node); next 584 drivers/base/regmap/regmap.c next = rb_next(&range_node->node); next 157 drivers/block/aoe/aoe.h struct aoedev *next; next 137 drivers/block/aoe/aoeblk.c list_empty(&(*t)->ffree) ? NULL : (*t)->ffree.next); next 221 drivers/block/aoe/aoecmd.c pos = t->ffree.next; next 769 drivers/block/aoe/aoecmd.c pos = flist.next; next 1205 drivers/block/aoe/aoecmd.c pos = iocq[id].head.next; next 1665 drivers/block/aoe/aoecmd.c pos = flist.next; next 329 drivers/block/aoe/aoedev.c for (d = devlist; d; d = d->next) { next 360 drivers/block/aoe/aoedev.c for (d = devlist; d; d = d->next) { next 378 drivers/block/aoe/aoedev.c *dd = d->next; next 381 drivers/block/aoe/aoedev.c dd = &d->next; next 449 drivers/block/aoe/aoedev.c for (d=devlist; d; d=d->next) next 491 drivers/block/aoe/aoedev.c d->next = devlist; next 494 drivers/block/brd.c struct brd_device *brd, *next; next 542 drivers/block/brd.c list_for_each_entry_safe(brd, next, &brd_devices, brd_list) { next 554 drivers/block/brd.c struct brd_device *brd, *next; next 556 drivers/block/brd.c list_for_each_entry_safe(brd, next, &brd_devices, brd_list) next 742 drivers/block/drbd/drbd_actlog.c int next = (device->rs_last_mark + 1) % DRBD_SYNC_MARKS; next 747 drivers/block/drbd/drbd_actlog.c device->rs_mark_time[next] = now; next 748 drivers/block/drbd/drbd_actlog.c device->rs_mark_left[next] = still_to_go; next 749 drivers/block/drbd/drbd_actlog.c device->rs_last_mark = next; next 3772 drivers/block/drbd/drbd_nl.c if (list_empty(connections) || connections->next->next != connections) next 3976 drivers/block/drbd/drbd_nl.c pos = list_entry_rcu(resource->resources.next, next 4942 drivers/block/drbd/drbd_nl.c goto next; next 4948 drivers/block/drbd/drbd_nl.c goto next; next 4954 drivers/block/drbd/drbd_nl.c goto next; next 4960 drivers/block/drbd/drbd_nl.c goto next; next 4963 drivers/block/drbd/drbd_nl.c next: next 4966 drivers/block/drbd/drbd_nl.c list_entry(state_change->list.next, next 5015 drivers/block/drbd/drbd_nl.c list_entry(head.next, struct drbd_state_change, list); next 1398 drivers/block/drbd/drbd_receiver.c next_epoch = list_entry(epoch->list.next, struct drbd_epoch, list); next 3048 drivers/block/floppy.c if (raw_cmd->next && next 3053 drivers/block/floppy.c raw_cmd = raw_cmd->next; next 3074 drivers/block/floppy.c cmd.next = NULL; next 3090 drivers/block/floppy.c ptr = ptr->next; next 3098 drivers/block/floppy.c struct floppy_raw_cmd *next; next 3109 drivers/block/floppy.c next = this->next; next 3111 drivers/block/floppy.c this = next; next 3130 drivers/block/floppy.c ptr->next = NULL; next 3167 drivers/block/floppy.c rcmd = &(ptr->next); next 607 drivers/block/nbd.c struct bio *next = bio->bi_next; next 612 drivers/block/nbd.c bool is_last = !next && bio_iter_last(bvec, iter); next 652 drivers/block/nbd.c bio = next; next 986 drivers/block/null_blk_main.c goto next; next 991 drivers/block/null_blk_main.c next: next 1490 drivers/block/null_blk_main.c cmd->ll_list.next = NULL; next 1828 drivers/block/null_blk_main.c nullb = list_entry(nullb_list.next, struct nullb, list); next 1854 drivers/block/null_blk_main.c nullb = list_entry(nullb_list.next, struct nullb, list); next 593 drivers/block/pktcdvd.c struct packet_data *pkt, *next; next 597 drivers/block/pktcdvd.c list_for_each_entry_safe(pkt, next, &pd->cdrw.pkt_free_list, list) { next 645 drivers/block/pktcdvd.c struct rb_node *next; next 656 drivers/block/pktcdvd.c next = n->rb_left; next 658 drivers/block/pktcdvd.c next = n->rb_right; next 659 drivers/block/pktcdvd.c if (!next) next 661 drivers/block/pktcdvd.c n = next; next 1065 drivers/block/pktcdvd.c if (pkt->sector == zone || pkt->list.next == &pd->cdrw.pkt_free_list) { next 1381 drivers/block/pktcdvd.c struct packet_data *pkt, *next; next 1397 drivers/block/pktcdvd.c list_for_each_entry_safe(pkt, next, &pd->cdrw.pkt_active_list, list) { next 547 drivers/block/ps3vram.c struct bio *next; next 580 drivers/block/ps3vram.c next = bio_list_peek(&priv->list); next 585 drivers/block/ps3vram.c return next; next 409 drivers/block/rsxx/dma.c dma = list_entry(ctrl->queue.next, struct rsxx_dma, list); next 477 drivers/block/xen-blkback/blkback.c req = list_entry(ring->pending_free.next, struct pending_req, next 880 drivers/block/xen-blkback/blkback.c goto next; next 900 drivers/block/xen-blkback/blkback.c goto next; next 909 drivers/block/xen-blkback/blkback.c goto next; next 915 drivers/block/xen-blkback/blkback.c goto next; next 926 drivers/block/xen-blkback/blkback.c next: next 682 drivers/block/zram/zram_drv.c goto next; next 687 drivers/block/zram/zram_drv.c goto next; next 691 drivers/block/zram/zram_drv.c goto next; next 694 drivers/block/zram/zram_drv.c goto next; next 746 drivers/block/zram/zram_drv.c goto next; next 759 drivers/block/zram/zram_drv.c next: next 887 drivers/block/zram/zram_drv.c goto next; next 905 drivers/block/zram/zram_drv.c next: next 278 drivers/bluetooth/btsdio.c tuple = tuple->next; next 92 drivers/bus/brcmstb_gisb.c struct list_head next; next 232 drivers/bus/brcmstb_gisb.c list_for_each_entry(gdev, &brcmstb_gisb_arb_device_list, next) { next 286 drivers/bus/brcmstb_gisb.c list_for_each_entry(gdev, &brcmstb_gisb_arb_device_list, next) next 333 drivers/bus/brcmstb_gisb.c INIT_LIST_HEAD(&gdev->next); next 390 drivers/bus/brcmstb_gisb.c list_add_tail(&gdev->next, &brcmstb_gisb_arb_device_list); next 555 drivers/bus/fsl-mc/fsl-mc-allocator.c struct fsl_mc_resource *next; next 561 drivers/bus/fsl-mc/fsl-mc-allocator.c list_for_each_entry_safe(resource, next, &res_pool->free_list, node) { next 3255 drivers/cdrom/cdrom.c long next = 0; next 3257 drivers/cdrom/cdrom.c ret = cdrom_get_next_writable(cdi, &next); next 3260 drivers/cdrom/cdrom.c if (copy_to_user((long __user *)arg, &next, sizeof(next))) next 59 drivers/char/agp/frontend.c curr = curr->next; next 69 drivers/char/agp/frontend.c struct agp_memory *next; next 75 drivers/char/agp/frontend.c next = temp->next; next 79 drivers/char/agp/frontend.c prev->next = next; next 80 drivers/char/agp/frontend.c if (next != NULL) next 81 drivers/char/agp/frontend.c next->prev = prev; next 85 drivers/char/agp/frontend.c if (next != NULL) next 86 drivers/char/agp/frontend.c next->prev = NULL; next 88 drivers/char/agp/frontend.c agp_fe.current_controller->pool = next; next 207 drivers/char/agp/frontend.c temp->next = prev; next 224 drivers/char/agp/frontend.c curr = curr->next; next 238 drivers/char/agp/frontend.c priv->next = prev; next 244 drivers/char/agp/frontend.c struct agp_file_private *next; next 247 drivers/char/agp/frontend.c next = priv->next; next 251 drivers/char/agp/frontend.c prev->next = next; next 253 drivers/char/agp/frontend.c if (next != NULL) next 254 drivers/char/agp/frontend.c next->prev = prev; next 257 drivers/char/agp/frontend.c if (next != NULL) next 258 drivers/char/agp/frontend.c next->prev = NULL; next 260 drivers/char/agp/frontend.c agp_fe.file_priv_list = next; next 302 drivers/char/agp/frontend.c controller = controller->next; next 325 drivers/char/agp/frontend.c controller->next = prev_controller; next 353 drivers/char/agp/frontend.c client = client->next; next 367 drivers/char/agp/frontend.c memory = memory->next; next 378 drivers/char/agp/frontend.c next_controller = controller->next; next 381 drivers/char/agp/frontend.c prev_controller->next = next_controller; next 419 drivers/char/agp/frontend.c clients = clients->next; next 441 drivers/char/agp/frontend.c clients = clients->next; next 467 drivers/char/agp/frontend.c client = client->next; next 482 drivers/char/agp/frontend.c controller = controller->next; next 504 drivers/char/agp/frontend.c client->next = prev_client; next 542 drivers/char/agp/frontend.c next_client = client->next; next 545 drivers/char/agp/frontend.c prev_client->next = next_client; next 40 drivers/char/agp/isoch.c struct list_head *pos, *tmp, *head = &list->list, *start = head->next; next 53 drivers/char/agp/isoch.c pos = pos->next; next 304 drivers/char/agp/isoch.c for (pos=head->next; cdev<ndevs; cdev++, pos=pos->next) { next 462 drivers/char/agp/isoch.c for (pos=head->next; pos!=head; ) { next 465 drivers/char/agp/isoch.c pos = pos->next; next 305 drivers/char/hw_random/core.c new_rng = list_entry(rng_list.next, struct hwrng, list); next 126 drivers/char/ipmi/ipmi_devintf.c struct ipmi_recv_msg *msg, *next; next 132 drivers/char/ipmi/ipmi_devintf.c list_for_each_entry_safe(msg, next, &priv->recv_msgs, link) next 226 drivers/char/ipmi/ipmi_devintf.c entry = priv->recv_msgs.next; next 28 drivers/char/ipmi/ipmi_dmi.c struct ipmi_dmi_info *next; next 85 drivers/char/ipmi/ipmi_dmi.c info->next = ipmi_dmi_infos; next 110 drivers/char/ipmi/ipmi_dmi.c info = info->next; next 255 drivers/char/ipmi/ipmi_msghandler.c struct cmd_rcvr *next; next 1346 drivers/char/ipmi/ipmi_msghandler.c rcvr->next = rcvrs; next 1354 drivers/char/ipmi/ipmi_msghandler.c rcvrs = rcvr->next; next 1715 drivers/char/ipmi/ipmi_msghandler.c rcvr->next = rcvrs; next 1726 drivers/char/ipmi/ipmi_msghandler.c rcvrs = rcvr->next; next 3569 drivers/char/ipmi/ipmi_msghandler.c entry = tmplist.next; next 4437 drivers/char/ipmi/ipmi_msghandler.c smi_msg = list_entry(intf->waiting_rcv_msgs.next, next 4508 drivers/char/ipmi/ipmi_msghandler.c entry = intf->hp_xmit_msgs.next; next 4510 drivers/char/ipmi/ipmi_msghandler.c entry = intf->xmit_msgs.next; next 5137 drivers/char/ipmi/ipmi_msghandler.c .next = NULL, next 188 drivers/char/ipmi/ipmi_si_hotmod.c char *str = kstrdup(val, GFP_KERNEL), *curr, *next; next 205 drivers/char/ipmi/ipmi_si_hotmod.c for (curr = str; curr; curr = next) { next 208 drivers/char/ipmi/ipmi_si_hotmod.c next = strchr(curr, ':'); next 209 drivers/char/ipmi/ipmi_si_hotmod.c if (next) { next 210 drivers/char/ipmi/ipmi_si_hotmod.c *next = '\0'; next 211 drivers/char/ipmi/ipmi_si_hotmod.c next++; next 1166 drivers/char/ipmi/ipmi_watchdog.c .next = NULL, next 94 drivers/char/misc.c .next = misc_seq_next, next 286 drivers/char/tpm/eventlog/tpm1.c .next = tpm1_bios_measurements_next, next 293 drivers/char/tpm/eventlog/tpm1.c .next = tpm1_bios_measurements_next, next 158 drivers/char/tpm/eventlog/tpm2.c .next = tpm2_bios_measurements_next, next 4851 drivers/clk/clk.c struct clock_provider *clk_provider, *next; next 4868 drivers/clk/clk.c list_for_each_entry_safe(clk_provider, next, next 4885 drivers/clk/clk.c list_for_each_entry_safe(clk_provider, next, next 116 drivers/clocksource/em_sti.c static u64 em_sti_set_next(struct em_sti_priv *p, u64 next) next 126 drivers/clocksource/em_sti.c em_sti_write(p, STI_COMPA_H, next >> 32); next 127 drivers/clocksource/em_sti.c em_sti_write(p, STI_COMPA_L, next & 0xffffffff); next 137 drivers/clocksource/em_sti.c return next; next 253 drivers/clocksource/em_sti.c u64 next; next 256 drivers/clocksource/em_sti.c next = em_sti_set_next(p, em_sti_count(p) + delta); next 257 drivers/clocksource/em_sti.c safe = em_sti_count(p) < (next - 1); next 69 drivers/clocksource/ingenic-timer.c static int ingenic_tcu_cevt_set_next(unsigned long next, next 74 drivers/clocksource/ingenic-timer.c if (next > 0xffff) next 77 drivers/clocksource/ingenic-timer.c regmap_write(tcu->map, TCU_REG_TDFRc(tcu->timer_channel), next); next 62 drivers/clocksource/mps2-timer.c static int mps2_timer_set_next_event(unsigned long next, struct clock_event_device *ce) next 64 drivers/clocksource/mps2-timer.c clockevent_mps2_writel(next, ce, TIMER_VALUE); next 340 drivers/clocksource/timer-cadence-ttc.c ttccs->ttc.clk_rate_change_nb.next = NULL; next 421 drivers/clocksource/timer-cadence-ttc.c ttcce->ttc.clk_rate_change_nb.next = NULL; next 60 drivers/clocksource/timer-imx-sysctr.c u64 next; next 64 drivers/clocksource/timer-imx-sysctr.c next = sysctr_read_counter(); next 66 drivers/clocksource/timer-imx-sysctr.c next += delta; next 68 drivers/clocksource/timer-imx-sysctr.c cmp_hi = (next >> 32) & 0x00fffff; next 69 drivers/clocksource/timer-imx-sysctr.c cmp_lo = next & 0xffffffff; next 86 drivers/clocksource/timer-imx-tpm.c unsigned long next, now; next 88 drivers/clocksource/timer-imx-tpm.c next = tpm_read_counter(); next 89 drivers/clocksource/timer-imx-tpm.c next += delta; next 90 drivers/clocksource/timer-imx-tpm.c writel(next, timer_base + TPM_C0V); next 99 drivers/clocksource/timer-imx-tpm.c return (int)(next - now) <= 0 ? -ETIME : 0; next 103 drivers/clocksource/timer-integrator-ap.c static int clkevt_set_next_event(unsigned long next, struct clock_event_device *evt) next 108 drivers/clocksource/timer-integrator-ap.c writel(next, clkevt_base + TIMER_LOAD); next 91 drivers/clocksource/timer-prima2.c unsigned long now, next; next 96 drivers/clocksource/timer-prima2.c next = now + delta; next 97 drivers/clocksource/timer-prima2.c writel_relaxed(next, sirfsoc_timer_base + SIRFSOC_TIMER_MATCH_0); next 102 drivers/clocksource/timer-prima2.c return next - now > delta ? -ETIME : 0; next 81 drivers/clocksource/timer-pxa.c unsigned long next, oscr; next 84 drivers/clocksource/timer-pxa.c next = timer_readl(OSCR) + delta; next 85 drivers/clocksource/timer-pxa.c timer_writel(next, OSMR0); next 88 drivers/clocksource/timer-pxa.c return (signed)(next - oscr) <= MIN_OSCR_DELTA ? -ETIME : 0; next 147 drivers/clocksource/timer-sp804.c static int sp804_set_next_event(unsigned long next, next 153 drivers/clocksource/timer-sp804.c writel(next, clkevt_base + TIMER_LOAD); next 130 drivers/clocksource/timer-stm32.c unsigned long now, next; next 132 drivers/clocksource/timer-stm32.c next = readl_relaxed(timer_of_base(to) + TIM_CNT) + evt; next 133 drivers/clocksource/timer-stm32.c writel_relaxed(next, timer_of_base(to) + TIM_CCR1); next 136 drivers/clocksource/timer-stm32.c if ((next - now) > evt) next 214 drivers/clocksource/timer-sun5i.c cs->timer.clk_rate_cb.next = NULL; next 293 drivers/clocksource/timer-sun5i.c ce->timer.clk_rate_cb.next = NULL; next 466 drivers/cpufreq/cpufreq.c for (nb = cpufreq_transition_notifier_list.head; nb; nb = nb->next) next 264 drivers/cpufreq/powernv-cpufreq.c goto next; next 270 drivers/cpufreq/powernv-cpufreq.c goto next; next 278 drivers/cpufreq/powernv-cpufreq.c next: next 569 drivers/cpufreq/powernv-cpufreq.c goto next; next 590 drivers/cpufreq/powernv-cpufreq.c next: next 999 drivers/cpufreq/powernv-cpufreq.c .next = NULL, next 158 drivers/cpufreq/pxa3xx-cpufreq.c struct pxa3xx_freq_info *next; next 164 drivers/cpufreq/pxa3xx-cpufreq.c next = &pxa3xx_freqs[index]; next 167 drivers/cpufreq/pxa3xx-cpufreq.c __update_core_freq(next); next 168 drivers/cpufreq/pxa3xx-cpufreq.c __update_bus_freq(next); next 432 drivers/crypto/axis/artpec6_crypto.c struct artpec6_crypto_bounce_buffer *next; next 434 drivers/crypto/axis/artpec6_crypto.c list_for_each_entry_safe(b, next, &dma->bounce_buffers, list) { next 1063 drivers/crypto/axis/artpec6_crypto.c struct artpec6_crypto_bounce_buffer *next; next 1065 drivers/crypto/axis/artpec6_crypto.c list_for_each_entry_safe(b, next, &dma->bounce_buffers, list) { next 5369 drivers/crypto/caam/caamalg_qi2.c if (hash_list.next) { next 1928 drivers/crypto/caam/caamhash.c if (!hash_list.next) next 124 drivers/crypto/ccp/ccp-crypto-main.c req_queue.backlog = crypto_cmd->entry.next; next 128 drivers/crypto/ccp/ccp-crypto-main.c req_queue.backlog = req_queue.backlog->next; next 132 drivers/crypto/ccp/ccp-crypto-main.c req_queue.backlog = crypto_cmd->entry.next; next 147 drivers/crypto/ccp/ccp-crypto-main.c struct ccp_crypto_cmd *held, *next, *backlog; next 198 drivers/crypto/ccp/ccp-crypto-main.c next = ccp_crypto_cmd_complete(held, &backlog); next 205 drivers/crypto/ccp/ccp-crypto-main.c held = next; next 70 drivers/crypto/chelsio/chtls/chtls.h struct listen_info *next; /* Link to next entry */ next 226 drivers/crypto/chelsio/chtls/chtls.h struct sk_buff *next; next 291 drivers/crypto/chelsio/chtls/chtls_cm.c int next = (int)new_state[sk->sk_state]; next 293 drivers/crypto/chelsio/chtls/chtls_cm.c tcp_set_state(sk, next & TCP_STATE_MASK); next 294 drivers/crypto/chelsio/chtls/chtls_cm.c return next & TCP_ACTION_FIN; next 512 drivers/crypto/chelsio/chtls/chtls_cm.c p->next = cdev->listen_hash_tab[key]; next 529 drivers/crypto/chelsio/chtls/chtls_cm.c for (p = cdev->listen_hash_tab[key]; p; p = p->next) next 549 drivers/crypto/chelsio/chtls/chtls_cm.c for (p = *prev; p; prev = &p->next, p = p->next) next 552 drivers/crypto/chelsio/chtls/chtls_cm.c *prev = p->next; next 164 drivers/crypto/hisilicon/sec/sec_algs.c sgl_next = sgl_current->next; next 205 drivers/crypto/hisilicon/sec/sec_algs.c sgl_current->next = sgl_next; next 388 drivers/crypto/hisilicon/sec/sec_drv.h struct sec_hw_sgl *next; next 54 drivers/crypto/hisilicon/sgl.c struct hisi_acc_hw_sgl *next; next 105 drivers/crypto/ixp4xx_crypto.c struct buffer_desc *next; next 330 drivers/crypto/ixp4xx_crypto.c buf1 = buf->next; next 643 drivers/crypto/ixp4xx_crypto.c buf->next = 0; next 799 drivers/crypto/ixp4xx_crypto.c buf->next = next_buf; next 807 drivers/crypto/ixp4xx_crypto.c buf->next = NULL; next 917 drivers/crypto/ixp4xx_crypto.c req_ctx->dst = dst_hook.next; next 927 drivers/crypto/ixp4xx_crypto.c req_ctx->src = src_hook.next; next 1027 drivers/crypto/ixp4xx_crypto.c req_ctx->src = src_hook.next; next 1047 drivers/crypto/ixp4xx_crypto.c req_ctx->dst = dst_hook.next; next 309 drivers/crypto/marvell/cesa.h struct mv_cesa_tdma_desc *next; next 70 drivers/crypto/marvell/tdma.c tdma = tdma->next; next 84 drivers/crypto/marvell/tdma.c for (tdma = dreq->chain.first; tdma; tdma = tdma->next) { next 106 drivers/crypto/marvell/tdma.c last->next = dreq->chain.first; next 124 drivers/crypto/marvell/tdma.c struct mv_cesa_tdma_desc *tdma = NULL, *next = NULL; next 130 drivers/crypto/marvell/tdma.c for (tdma = engine->chain.first; tdma; tdma = next) { next 132 drivers/crypto/marvell/tdma.c next = tdma->next; next 152 drivers/crypto/marvell/tdma.c engine->chain.first = tdma->next; next 153 drivers/crypto/marvell/tdma.c tdma->next = NULL; next 203 drivers/crypto/marvell/tdma.c chain->last->next = new_tdma; next 227 drivers/crypto/marvell/tdma.c for (op_desc = chain->first; op_desc; op_desc = op_desc->next) { next 1782 drivers/crypto/picoxcell_crypto.c struct spacc_alg *alg, *next; next 1793 drivers/crypto/picoxcell_crypto.c list_for_each_entry_safe(alg, next, &engine->registered_algs, entry) { next 94 drivers/crypto/qat/qat_common/adf_cfg.c .next = qat_dev_cfg_next, next 57 drivers/crypto/qat/qat_common/adf_cfg_user.h struct adf_user_cfg_key_val *next; next 70 drivers/crypto/qat/qat_common/adf_cfg_user.h struct adf_user_cfg_section *next; next 214 drivers/crypto/qat/qat_common/adf_ctl_drv.c params_head = key_val.next; next 216 drivers/crypto/qat/qat_common/adf_ctl_drv.c section_head = section.next; next 235 drivers/crypto/qat/qat_common/adf_dev_mgr.c struct vf_id_map *next; next 241 drivers/crypto/qat/qat_common/adf_dev_mgr.c next = list_next_entry(map, list); next 242 drivers/crypto/qat/qat_common/adf_dev_mgr.c while (next && &next->list != &vfs_table) { next 243 drivers/crypto/qat/qat_common/adf_dev_mgr.c next->fake_id++; next 244 drivers/crypto/qat/qat_common/adf_dev_mgr.c next = list_next_entry(next, list); next 299 drivers/crypto/qat/qat_common/adf_dev_mgr.c struct vf_id_map *map, *next; next 308 drivers/crypto/qat/qat_common/adf_dev_mgr.c next = list_next_entry(map, list); next 309 drivers/crypto/qat/qat_common/adf_dev_mgr.c while (next && &next->list != &vfs_table) { next 310 drivers/crypto/qat/qat_common/adf_dev_mgr.c next->fake_id--; next 311 drivers/crypto/qat/qat_common/adf_dev_mgr.c next = list_next_entry(next, list); next 126 drivers/crypto/qat/qat_common/adf_transport_debug.c .next = adf_ring_next, next 237 drivers/crypto/qat/qat_common/adf_transport_debug.c .next = adf_bank_next, next 403 drivers/crypto/qat/qat_common/icp_qat_uclo.h struct icp_qat_uof_batch_init *next; next 1149 drivers/crypto/qat/qat_common/qat_hal.c plm_init = lm_init_header->next; next 1168 drivers/crypto/qat/qat_common/qat_hal.c plm_init = plm_init->next; next 210 drivers/crypto/qat/qat_common/qat_uclo.c umem_init = umem_init_header->next; next 219 drivers/crypto/qat/qat_common/qat_uclo.c umem_init = umem_init->next; next 234 drivers/crypto/qat/qat_common/qat_uclo.c umem_init = umem_init->next; next 316 drivers/crypto/qat/qat_common/qat_uclo.c while (tail_old->next) next 317 drivers/crypto/qat/qat_common/qat_uclo.c tail_old = tail_old->next; next 327 drivers/crypto/qat/qat_common/qat_uclo.c mem_init->next = NULL; next 328 drivers/crypto/qat/qat_common/qat_uclo.c tail->next = mem_init; next 336 drivers/crypto/qat/qat_common/qat_uclo.c mem_init = tail_old->next; next 134 drivers/crypto/sahara.c u32 next; next 140 drivers/crypto/sahara.c u32 next; next 418 drivers/crypto/sahara.c dev->hw_desc[i]->next); next 436 drivers/crypto/sahara.c dev->hw_link[i]->next); next 463 drivers/crypto/sahara.c dev->hw_desc[idx]->next = dev->hw_phys_desc[1]; next 506 drivers/crypto/sahara.c dev->hw_link[i]->next = 0; next 508 drivers/crypto/sahara.c dev->hw_link[i]->next = dev->hw_phys_link[i + 1]; next 520 drivers/crypto/sahara.c dev->hw_link[j]->next = 0; next 522 drivers/crypto/sahara.c dev->hw_link[j]->next = dev->hw_phys_link[j + 1]; next 531 drivers/crypto/sahara.c dev->hw_desc[idx]->next = 0; next 824 drivers/crypto/sahara.c dev->hw_link[i]->next = 0; next 826 drivers/crypto/sahara.c dev->hw_link[i]->next = dev->hw_phys_link[i + 1]; next 873 drivers/crypto/sahara.c dev->hw_link[i]->next = 0; next 901 drivers/crypto/sahara.c dev->hw_link[index]->next = 0; next 1011 drivers/crypto/sahara.c dev->hw_desc[0]->next = 0; next 1017 drivers/crypto/sahara.c dev->hw_desc[0]->next = dev->hw_phys_desc[1]; next 1019 drivers/crypto/sahara.c dev->hw_desc[1]->next = 0; next 1111 drivers/crypto/talitos.c goto next; next 1135 drivers/crypto/talitos.c next: next 194 drivers/dma-buf/sw_sync.c struct sync_pt *pt, *next; next 202 drivers/dma-buf/sw_sync.c list_for_each_entry_safe(pt, next, &obj->pt_list, link) { next 312 drivers/dma-buf/sw_sync.c struct sync_pt *pt, *next; next 316 drivers/dma-buf/sw_sync.c list_for_each_entry_safe(pt, next, &obj->pt_list, link) { next 231 drivers/dma/altera-msgdma.c struct msgdma_sw_desc *child, *next; next 235 drivers/dma/altera-msgdma.c list_for_each_entry_safe(child, next, &desc->tx_list, node) { next 249 drivers/dma/altera-msgdma.c struct msgdma_sw_desc *desc, *next; next 251 drivers/dma/altera-msgdma.c list_for_each_entry_safe(desc, next, list, node) next 534 drivers/dma/altera-msgdma.c struct msgdma_sw_desc *sdesc, *next; next 538 drivers/dma/altera-msgdma.c list_for_each_entry_safe(sdesc, next, &desc->tx_list, node) next 582 drivers/dma/altera-msgdma.c struct msgdma_sw_desc *desc, *next; next 584 drivers/dma/altera-msgdma.c list_for_each_entry_safe(desc, next, &mdev->done_list, node) { next 906 drivers/dma/amba-pl08x.c struct pl08x_dma_chan *p, *next; next 909 drivers/dma/amba-pl08x.c next = NULL; next 920 drivers/dma/amba-pl08x.c next = p; next 924 drivers/dma/amba-pl08x.c if (!next && pl08x->has_slave) { next 928 drivers/dma/amba-pl08x.c next = p; next 936 drivers/dma/amba-pl08x.c if (next) { next 943 drivers/dma/amba-pl08x.c spin_lock(&next->vc.lock); next 945 drivers/dma/amba-pl08x.c success = next->state == PL08X_CHAN_WAITING; next 947 drivers/dma/amba-pl08x.c pl08x_phy_reassign_start(plchan->phychan, next); next 948 drivers/dma/amba-pl08x.c spin_unlock(&next->vc.lock); next 2430 drivers/dma/amba-pl08x.c struct pl08x_dma_chan *next; next 2433 drivers/dma/amba-pl08x.c next, &dmadev->channels, vc.chan.device_node) { next 117 drivers/dma/bcm-sba-raid.c struct list_head next; next 232 drivers/dma/bcm-sba-raid.c INIT_LIST_HEAD(&req->next); next 303 drivers/dma/bcm-sba-raid.c list_for_each_entry(nreq, &req->next, next) next 317 drivers/dma/bcm-sba-raid.c list_add_tail(&req->next, &first->next); next 442 drivers/dma/bcm-sba-raid.c list_for_each_entry(nreq, &first->next, next) next 444 drivers/dma/bcm-sba-raid.c INIT_LIST_HEAD(&first->next); next 540 drivers/dma/bcm-sba-raid.c list_for_each_entry(nreq, &req->next, next) next 1509 drivers/dma/bcm-sba-raid.c INIT_LIST_HEAD(&req->next); next 61 drivers/dma/bcm2835-dma.c uint32_t next; next 344 drivers/dma/bcm2835-dma.c control_block->next = 0; next 359 drivers/dma/bcm2835-dma.c d->cb_list[frame - 1].cb->next = cb_entry->paddr; next 775 drivers/dma/bcm2835-dma.c d->cb_list[d->frames - 1].cb->next = d->cb_list[0].paddr; next 849 drivers/dma/bcm2835-dma.c struct bcm2835_chan *c, *next; next 851 drivers/dma/bcm2835-dma.c list_for_each_entry_safe(c, next, &od->ddev.channels, next 116 drivers/dma/coh901318_lli.c struct coh901318_lli *next; next 129 drivers/dma/coh901318_lli.c next = l->virt_link_addr; next 132 drivers/dma/coh901318_lli.c l = next; next 53 drivers/dma/dw/core.c return to_dw_desc(dwc->active_list.next); next 175 drivers/dma/dw/core.c dwc->tx_node_active = dwc->tx_node_active->next; next 231 drivers/dma/dw/core.c list_move(dwc->queue.next, &dwc->active_list); next 333 drivers/dma/dw/core.c if (active == head->next) next 445 drivers/dma/dw/core.c list_move(dwc->queue.next, dwc->active_list.prev); next 120 drivers/dma/fsldma.c struct fsl_dma_ld_hw *hw, dma_addr_t next) next 126 drivers/dma/fsldma.c hw->next_ln_addr = CPU_TO_DMA(chan, snoop_bits | next, 64); next 263 drivers/dma/img-mdc-dma.c struct mdc_hw_list_desc *curr, *next; next 269 drivers/dma/img-mdc-dma.c next = curr->next_desc; next 272 drivers/dma/img-mdc-dma.c curr = next; next 1026 drivers/dma/img-mdc-dma.c struct mdc_chan *mchan, *next; next 1031 drivers/dma/img-mdc-dma.c list_for_each_entry_safe(mchan, next, &mdma->dma_dev.channels, next 647 drivers/dma/imx-dma.c list_move_tail(imxdmac->ld_active.next, &imxdmac->ld_free); next 652 drivers/dma/imx-dma.c list_move_tail(imxdmac->ld_queue.next, &imxdmac->ld_active); next 752 drivers/dma/imx-dma.c list_move_tail(imxdmac->ld_free.next, &imxdmac->ld_queue); next 1017 drivers/dma/imx-dma.c list_move_tail(imxdmac->ld_queue.next, next 423 drivers/dma/ioat/dma.c struct ioat_ring_ent *next = ring[i+1]; next 426 drivers/dma/ioat/dma.c hw->next = next->txd.phys; next 428 drivers/dma/ioat/dma.c ring[i]->hw->next = ring[0]->txd.phys; next 225 drivers/dma/ioat/dma.h (unsigned long long) hw->next, tx->cookie, tx->flags, next 94 drivers/dma/ioat/hw.h uint64_t next; next 127 drivers/dma/ioat/hw.h uint64_t next; next 138 drivers/dma/ioat/hw.h uint64_t next; next 181 drivers/dma/ioat/hw.h uint64_t next; next 195 drivers/dma/ioat/hw.h uint64_t next; next 225 drivers/dma/ioat/hw.h uint64_t next; next 330 drivers/dma/ioat/init.c dma_chan = container_of(dma->channels.next, struct dma_chan, next 843 drivers/dma/ioat/init.c dma_chan = container_of(dma->channels.next, struct dma_chan, next 290 drivers/dma/ioat/prep.c (unsigned long long) (pq_ex ? pq_ex->next : pq->next), next 300 drivers/dma/ioat/prep.c dev_dbg(dev, "\tNEXT: %#llx\n", pq->next); next 323 drivers/dma/ioat/prep.c (unsigned long long) pq->next, next 45 drivers/dma/iop-adma.c slot = list_entry(slot->slot_node.next, next 91 drivers/dma/iop-adma.c if (desc->chain_node.next == &iop_chan->chain) next 325 drivers/dma/iop-adma.c iter = list_entry(iter->slot_node.next, next 460 drivers/dma/iop-adma.c iop_chan->last_used = list_entry(iop_chan->all_slots.next, next 858 drivers/dma/iop-adma.c dma_chan = container_of(device->common.channels.next, next 956 drivers/dma/iop-adma.c dma_chan = container_of(device->common.channels.next, next 1118 drivers/dma/iop-adma.c dma_chan = container_of(device->common.channels.next, next 1132 drivers/dma/ipu/ipu_idmac.c if ((*desc)->list.next == &ichan->queue) next 1137 drivers/dma/ipu/ipu_idmac.c *desc = list_entry((*desc)->list.next, struct idmac_tx_desc, list); next 1229 drivers/dma/ipu/ipu_idmac.c desc = list_entry(ichan->queue.next, struct idmac_tx_desc, list); next 1357 drivers/dma/ipu/ipu_idmac.c desc = list_entry(ichan->free_list.next, next 530 drivers/dma/mediatek/mtk-cqdma.c tx->next = NULL; next 535 drivers/dma/mediatek/mtk-cqdma.c prev_tx->next = tx; next 543 drivers/dma/mediatek/mtk-hsdma.c u16 next; next 560 drivers/dma/mediatek/mtk-hsdma.c next = MTK_HSDMA_NEXT_DESP_IDX(pc->ring.cur_rptr, next 562 drivers/dma/mediatek/mtk-hsdma.c rxd = &pc->ring.rxd[next]; next 572 drivers/dma/mediatek/mtk-hsdma.c cb = &pc->ring.cb[next]; next 612 drivers/dma/mediatek/mtk-hsdma.c pc->ring.cur_rptr = next; next 311 drivers/dma/mv_xor.c iter = list_entry(mv_chan->chain.next, next 321 drivers/dma/mv_xor.c iter = list_entry(iter->node.next, next 97 drivers/dma/mxs-dma.c u32 next; next 527 drivers/dma/mxs-dma.c ccw->next = mxs_chan->ccw_phys + sizeof(*ccw) * idx; next 563 drivers/dma/mxs-dma.c ccw->next = mxs_chan->ccw_phys + sizeof(*ccw) * idx; next 628 drivers/dma/mxs-dma.c ccw->next = mxs_chan->ccw_phys; next 630 drivers/dma/mxs-dma.c ccw->next = mxs_chan->ccw_phys + sizeof(*ccw) * (i + 1); next 128 drivers/dma/nbpfaxi.c u32 next; next 817 drivers/dma/nbpfaxi.c prev->hwdesc->next = (u32)ldesc->hwdesc_dma_addr; next 826 drivers/dma/nbpfaxi.c prev->hwdesc->next = 0; next 345 drivers/dma/owl-dma.c struct owl_dma_lli *next, next 349 drivers/dma/owl-dma.c list_add_tail(&next->node, &txd->lli_list); next 352 drivers/dma/owl-dma.c prev->hw.next_lli = next->phys; next 356 drivers/dma/owl-dma.c return next; next 1012 drivers/dma/owl-dma.c struct owl_dma_vchan *next; next 1015 drivers/dma/owl-dma.c next, &od->dma.channels, vc.chan.device_node) { next 68 drivers/dma/pch_dma.c u32 next; next 334 drivers/dma/pch_dma.c pd_chan->chan.chan_id, desc->regs.next); next 340 drivers/dma/pch_dma.c channel_writel(pd_chan, NEXT, desc->regs.next); next 604 drivers/dma/pch_dma.c desc->regs.next = DMA_DESC_FOLLOW_WITHOUT_IRQ; next 629 drivers/dma/pch_dma.c prev->regs.next |= desc->txd.phys; next 637 drivers/dma/pch_dma.c desc->regs.next = DMA_DESC_END_WITH_IRQ; next 639 drivers/dma/pch_dma.c desc->regs.next = DMA_DESC_END_WITHOUT_IRQ; next 756 drivers/dma/pch_dma.c pd->ch_regs[i].next = channel_readl(pd_chan, NEXT); next 779 drivers/dma/pch_dma.c channel_writel(pd_chan, NEXT, pd->ch_regs[i].next); next 2500 drivers/dma/pl330.c desc = list_entry(last->node.next, struct dma_pl330_desc, node); next 2563 drivers/dma/pl330.c desc = list_entry(pool->next, next 2702 drivers/dma/pl330.c desc = list_entry(first->node.next, next 2818 drivers/dma/pl330.c desc = list_entry(first->node.next, next 343 drivers/dma/ppc4xx/adma.c iter->hw_next = list_entry(iter->chain_node.next, next 447 drivers/dma/ppc4xx/adma.c iter = list_entry(iter->chain_node.next, next 451 drivers/dma/ppc4xx/adma.c iter = list_entry(iter->chain_node.next, next 478 drivers/dma/ppc4xx/adma.c iter->hw_next = list_entry(iter->chain_node.next, next 1453 drivers/dma/ppc4xx/adma.c slot = list_entry(slot->slot_node.next, next 1742 drivers/dma/ppc4xx/adma.c iter = list_entry(iter->slot_node.next, next 1812 drivers/dma/ppc4xx/adma.c list_entry(ppc440spe_chan->all_slots.next, next 2118 drivers/dma/ppc4xx/adma.c iter->hw_next = list_entry(iter->chain_node.next, next 2201 drivers/dma/ppc4xx/adma.c iter->hw_next = list_entry(iter->chain_node.next, next 2224 drivers/dma/ppc4xx/adma.c iter->hw_next = list_entry(iter->chain_node.next, next 2469 drivers/dma/ppc4xx/adma.c list_entry(iter->chain_node.next, next 2642 drivers/dma/ppc4xx/adma.c iter->hw_next = list_entry(iter->chain_node.next, next 2668 drivers/dma/ppc4xx/adma.c iter = list_entry(iter->chain_node.next, next 2674 drivers/dma/ppc4xx/adma.c iter->hw_next = list_entry(iter->chain_node.next, next 2995 drivers/dma/ppc4xx/adma.c iter = list_entry(iter->chain_node.next, next 3007 drivers/dma/ppc4xx/adma.c iter = list_entry(iter->chain_node.next, next 3076 drivers/dma/ppc4xx/adma.c end = list_entry(end->chain_node.next, next 3251 drivers/dma/ppc4xx/adma.c desc = list_entry(hdesc->chain_node.next, next 3372 drivers/dma/ppc4xx/adma.c desc = list_entry(desc->chain_node.next, next 3413 drivers/dma/ppc4xx/adma.c desc = list_entry(desc->chain_node.next, next 121 drivers/dma/qcom/hidma.c struct hidma_desc *next; next 133 drivers/dma/qcom/hidma.c list_for_each_entry_safe(mdesc, next, &list, node) { next 181 drivers/dma/qcom/hidma.c if (mdesc->node.next) { next 240 drivers/dma/qcom/hidma.c struct hidma_desc *qdesc, *next; next 244 drivers/dma/qcom/hidma.c list_for_each_entry_safe(qdesc, next, &mchan->queued, node) { next 518 drivers/dma/s3c24xx-dma.c txd->at = txd->dsg_list.next; next 584 drivers/dma/s3c24xx-dma.c struct s3c24xx_dma_chan *p, *next; next 587 drivers/dma/s3c24xx-dma.c next = NULL; next 592 drivers/dma/s3c24xx-dma.c next = p; next 596 drivers/dma/s3c24xx-dma.c if (!next) { next 601 drivers/dma/s3c24xx-dma.c next = p; next 609 drivers/dma/s3c24xx-dma.c if (next) { next 616 drivers/dma/s3c24xx-dma.c spin_lock(&next->vc.lock); next 618 drivers/dma/s3c24xx-dma.c success = next->state == S3C24XX_DMA_CHAN_WAITING; next 620 drivers/dma/s3c24xx-dma.c s3c24xx_dma_phy_reassign_start(s3cchan->phy, next); next 621 drivers/dma/s3c24xx-dma.c spin_unlock(&next->vc.lock); next 675 drivers/dma/s3c24xx-dma.c txd->at = txd->at->next; next 695 drivers/dma/s3c24xx-dma.c txd->at = txd->dsg_list.next; next 1135 drivers/dma/s3c24xx-dma.c struct s3c24xx_dma_chan *next; next 1138 drivers/dma/s3c24xx-dma.c next, &dmadev->channels, vc.chan.device_node) { next 330 drivers/dma/st_fdma.c hw_node->next = 0; next 485 drivers/dma/st_fdma.c hw_node->next = fdesc->node[(i + 1) % sg_len].pdesc; next 534 drivers/dma/st_fdma.c hw_node->next = fdesc->node[(i + 1) % sg_len].pdesc; next 50 drivers/dma/st_fdma.h u32 next; next 223 drivers/dma/ste_dma40_ll.c dma_addr_t next = lli_phys; next 243 drivers/dma/ste_dma40_ll.c next = cyclic ? first_phys : 0; next 245 drivers/dma/ste_dma40_ll.c next = ALIGN(next + sizeof(struct d40_phy_lli), next 248 drivers/dma/ste_dma40_ll.c err = d40_phy_fill_lli(lli, addr, size_seg, next, next 312 drivers/dma/ste_dma40_ll.c int next, unsigned int flags) next 318 drivers/dma/ste_dma40_ll.c if (next != -EINVAL) { next 319 drivers/dma/ste_dma40_ll.c slos = next * 2; next 320 drivers/dma/ste_dma40_ll.c dlos = next * 2 + 1; next 338 drivers/dma/ste_dma40_ll.c int next, unsigned int flags) next 340 drivers/dma/ste_dma40_ll.c d40_log_lli_link(lli_dst, lli_src, next, flags); next 351 drivers/dma/ste_dma40_ll.c int next, unsigned int flags) next 353 drivers/dma/ste_dma40_ll.c d40_log_lli_link(lli_dst, lli_src, next, flags); next 463 drivers/dma/ste_dma40_ll.h int next, unsigned int flags); next 468 drivers/dma/ste_dma40_ll.h int next, unsigned int flags); next 698 drivers/dma/stm32-mdma.c u32 next = count + 1; next 720 drivers/dma/stm32-mdma.c hwdesc->clar = desc->node[next].hwdesc_phys; next 366 drivers/dma/sun6i-dma.c struct sun6i_dma_lli *next, next 370 drivers/dma/sun6i-dma.c if ((!prev && !txd) || !next) next 375 drivers/dma/sun6i-dma.c txd->v_lli = next; next 378 drivers/dma/sun6i-dma.c prev->v_lli_next = next; next 381 drivers/dma/sun6i-dma.c next->p_lli_next = LLI_LAST_ITEM; next 382 drivers/dma/sun6i-dma.c next->v_lli_next = NULL; next 384 drivers/dma/sun6i-dma.c return next; next 188 drivers/dma/timb_dma.c td_desc = list_entry(td_chan->active_list.next, struct timb_dma_desc, next 229 drivers/dma/timb_dma.c td_desc = list_entry(td_chan->active_list.next, struct timb_dma_desc, next 267 drivers/dma/timb_dma.c list_entry(td_chan->active_list.next, next 284 drivers/dma/timb_dma.c td_desc = list_entry(td_chan->queue.next, struct timb_dma_desc, next 165 drivers/dma/txx9dmac.c return list_entry(dc->active_list.next, next 177 drivers/dma/txx9dmac.c return list_entry(dc->queue.next, struct txx9dmac_desc, desc_node); next 948 drivers/dma/txx9dmac.c desc = list_entry(list.next, struct txx9dmac_desc, desc_node); next 694 drivers/dma/xilinx/xilinx_dma.c struct xilinx_vdma_tx_segment *segment, *next; next 702 drivers/dma/xilinx/xilinx_dma.c list_for_each_entry_safe(segment, next, &desc->segments, node) { next 733 drivers/dma/xilinx/xilinx_dma.c struct xilinx_dma_tx_descriptor *desc, *next; next 735 drivers/dma/xilinx/xilinx_dma.c list_for_each_entry_safe(desc, next, list, node) { next 820 drivers/dma/xilinx/xilinx_dma.c struct xilinx_dma_tx_descriptor *desc, *next; next 825 drivers/dma/xilinx/xilinx_dma.c list_for_each_entry_safe(desc, next, &chan->done_list, node) { next 1395 drivers/dma/xilinx/xilinx_dma.c struct xilinx_dma_tx_descriptor *desc, *next; next 1401 drivers/dma/xilinx/xilinx_dma.c list_for_each_entry_safe(desc, next, &chan->active_list, node) { next 432 drivers/dma/xilinx/zynqmp_dma.c struct zynqmp_dma_desc_sw *child, *next; next 436 drivers/dma/xilinx/zynqmp_dma.c list_for_each_entry_safe(child, next, &sdesc->tx_list, node) { next 450 drivers/dma/xilinx/zynqmp_dma.c struct zynqmp_dma_desc_sw *desc, *next; next 452 drivers/dma/xilinx/zynqmp_dma.c list_for_each_entry_safe(desc, next, list, node) next 599 drivers/dma/xilinx/zynqmp_dma.c struct zynqmp_dma_desc_sw *desc, *next; next 601 drivers/dma/xilinx/zynqmp_dma.c list_for_each_entry_safe(desc, next, &chan->done_list, node) { next 1658 drivers/edac/altera_edac.c list_add(&altdev->next, &altdev->edac->a10_ecc_devices); next 2067 drivers/edac/altera_edac.c list_add(&altdev->next, &edac->a10_ecc_devices); next 2144 drivers/edac/altera_edac.c next); next 374 drivers/edac/altera_edac.h struct list_head next; next 747 drivers/edac/sb_edac.c sbridge_dev = list_entry(prev ? prev->list.next next 748 drivers/edac/sb_edac.c : sbridge_edac_list.next, struct sbridge_dev, list); next 1713 drivers/edac/sb_edac.c goto next; next 1728 drivers/edac/sb_edac.c next: next 132 drivers/edac/xgene_edac.c struct list_head next; next 421 drivers/edac/xgene_edac.c list_add(&ctx->next, &edac->mcus); next 509 drivers/edac/xgene_edac.c struct list_head next; next 963 drivers/edac/xgene_edac.c list_add(&ctx->next, &edac->pmds); next 1017 drivers/edac/xgene_edac.c struct list_head next; next 1246 drivers/edac/xgene_edac.c list_add(&ctx->next, &edac->l3s); next 1787 drivers/edac/xgene_edac.c list_add(&ctx->next, &edac->socs); next 1829 drivers/edac/xgene_edac.c list_for_each_entry(mcu, &ctx->mcus, next) next 1833 drivers/edac/xgene_edac.c list_for_each_entry(pmd, &ctx->pmds, next) { next 1838 drivers/edac/xgene_edac.c list_for_each_entry(node, &ctx->l3s, next) next 1841 drivers/edac/xgene_edac.c list_for_each_entry(node, &ctx->socs, next) next 1976 drivers/edac/xgene_edac.c list_for_each_entry_safe(mcu, temp_mcu, &edac->mcus, next) next 1979 drivers/edac/xgene_edac.c list_for_each_entry_safe(pmd, temp_pmd, &edac->pmds, next) next 1982 drivers/edac/xgene_edac.c list_for_each_entry_safe(node, temp_node, &edac->l3s, next) next 1985 drivers/edac/xgene_edac.c list_for_each_entry_safe(node, temp_node, &edac->socs, next) next 1050 drivers/firewire/core-cdev.c struct fw_cdev_iso_packet __user *p, *end, *next; next 1119 drivers/firewire/core-cdev.c next = (struct fw_cdev_iso_packet __user *) next 1121 drivers/firewire/core-cdev.c if (next > end) next 1136 drivers/firewire/core-cdev.c p = next; next 261 drivers/firewire/core-topology.c child = fw_node(child->link.next); next 316 drivers/firewire/core-topology.c struct fw_node *node, *next, *child, *parent; next 342 drivers/firewire/core-topology.c list_for_each_entry_safe(node, next, &list, link) next 420 drivers/firewire/core-topology.c node0 = fw_node(list0.next); next 421 drivers/firewire/core-topology.c node1 = fw_node(list1.next); next 483 drivers/firewire/core-topology.c node0 = fw_node(node0->link.next); next 484 drivers/firewire/core-topology.c next1 = fw_node(node1->link.next); next 309 drivers/firewire/net.c fi2 = list_entry(fi->fi_link.next, next 445 drivers/firewire/net.c fi = list_entry(pd->fi_list.next, struct fwnet_fragment_info, fi_link); next 41 drivers/firewire/nosy.c __le32 next; next 575 drivers/firewire/nosy.c lynx->rcv_start_pcl->next = cpu_to_le32(lynx->rcv_pcl_bus); next 576 drivers/firewire/nosy.c lynx->rcv_pcl->next = cpu_to_le32(PCL_NEXT_INVALID); next 898 drivers/firewire/ohci.c void *next; next 901 drivers/firewire/ohci.c next = handle_ar_packet(ctx, p); next 902 drivers/firewire/ohci.c if (!next) next 904 drivers/firewire/ohci.c p = next; next 1060 drivers/firewire/ohci.c desc = list_entry(ctx->buffer_list.next, next 1074 drivers/firewire/ohci.c desc = list_entry(desc->list.next, next 1145 drivers/firewire/ohci.c ctx->buffer_tail = list_entry(ctx->buffer_list.next, next 1192 drivers/firewire/ohci.c if (desc->list.next == &ctx->buffer_list) { next 1198 drivers/firewire/ohci.c desc = list_entry(desc->list.next, next 307 drivers/firewire/sbp2.c struct sbp2_pointer next; next 514 drivers/firewire/sbp2.c struct sbp2_orb *orb, *next; next 523 drivers/firewire/sbp2.c list_for_each_entry_safe(orb, next, &list, link) { next 1214 drivers/firewire/sbp2.c struct sbp2_logical_unit *lu, *next; next 1222 drivers/firewire/sbp2.c list_for_each_entry_safe(lu, next, &tgt->lu_list, link) { next 1457 drivers/firewire/sbp2.c orb->request.next.high = cpu_to_be32(SBP2_ORB_NULL); next 641 drivers/firmware/dmi-sysfs.c struct dmi_sysfs_entry *entry, *next; next 644 drivers/firmware/dmi-sysfs.c list_for_each_entry_safe(entry, next, &entry_list, list) { next 951 drivers/firmware/dmi_scan.c for (d = head->next; d != &dmi_devices; d = d->next) { next 193 drivers/firmware/efi/apple-properties.c pa_data = data->next; next 110 drivers/firmware/efi/efi-pstore.c struct efivar_entry *next, next 114 drivers/firmware/efi/efi-pstore.c if (&next->list != head) next 115 drivers/firmware/efi/efi-pstore.c next->scanning = true; next 146 drivers/firmware/efi/efi-pstore.c struct efivar_entry *next, next 155 drivers/firmware/efi/efi-pstore.c ret = __efi_pstore_scan_sysfs_exit(next, &next->list != head); next 617 drivers/firmware/efi/efi.c prsv = rsv->next; next 1007 drivers/firmware/efi/efi.c for (prsv = efi_memreserve_root->next; prsv; prsv = rsv->next) { next 1043 drivers/firmware/efi/efi.c rsv->next = efi_memreserve_root->next; next 1044 drivers/firmware/efi/efi.c efi_memreserve_root->next = __pa(rsv); next 369 drivers/firmware/efi/esrt.c struct esre_entry *entry, *next; next 371 drivers/firmware/efi/esrt.c list_for_each_entry_safe(entry, next, &entry_list, list) { next 82 drivers/firmware/efi/libstub/arm-stub.c rsv->next = 0; next 387 drivers/firmware/qemu_fw_cfg.c struct fw_cfg_sysfs_entry *entry, *next; next 389 drivers/firmware/qemu_fw_cfg.c list_for_each_entry_safe(entry, next, &fw_cfg_entry_cache, list) { next 564 drivers/firmware/qemu_fw_cfg.c struct kobject *k, *next; next 566 drivers/firmware/qemu_fw_cfg.c list_for_each_entry_safe(k, next, &kset->list, entry) next 211 drivers/fpga/fpga-bridge.c struct fpga_bridge *bridge, *next; next 214 drivers/fpga/fpga-bridge.c list_for_each_entry_safe(bridge, next, bridge_list, node) { next 257 drivers/gpio/gpiolib.c struct gpio_device *prev, *next; next 265 drivers/gpio/gpiolib.c next = list_entry(gpio_devices.next, struct gpio_device, list); next 266 drivers/gpio/gpiolib.c if (gdev->base + gdev->ngpio <= next->base) { next 279 drivers/gpio/gpiolib.c list_for_each_entry_safe(prev, next, &gpio_devices, list) { next 281 drivers/gpio/gpiolib.c if (&next->list == &gpio_devices) next 286 drivers/gpio/gpiolib.c && gdev->base + gdev->ngpio <= next->base) { next 5019 drivers/gpio/gpiolib.c ret = list_entry(gdev->list.next, struct gpio_device, list); next 5068 drivers/gpio/gpiolib.c .next = gpiolib_seq_next, next 3394 drivers/gpu/drm/amd/amdgpu/amdgpu_device.c struct dma_fence *fence = NULL, *next = NULL; next 3413 drivers/gpu/drm/amd/amdgpu/amdgpu_device.c r = amdgpu_bo_restore_shadow(shadow, &next); next 3420 drivers/gpu/drm/amd/amdgpu/amdgpu_device.c fence = next; next 3429 drivers/gpu/drm/amd/amdgpu/amdgpu_device.c fence = next; next 714 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c for (base = robj->vm_bo; base; base = base->next) next 156 drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.c char *next; next 171 drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.c next = strchr(p, ','); next 172 drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.c if (!next) next 174 drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.c p = next + 1; next 291 drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c fault = &gmc->fault_ring[fault->next]; next 304 drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c fault->next = gmc->fault_hash[hash].idx; next 70 drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.h uint64_t next:AMDGPU_GMC_FAULT_RING_ORDER; next 451 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c head = bo->mn_list.next; next 118 drivers/gpu/drm/amd/amdgpu/amdgpu_sa.c if (sa_manager->hole->next == &sa_manager->olist) next 121 drivers/gpu/drm/amd/amdgpu/amdgpu_sa.c sa_bo = list_entry(sa_manager->hole->next, struct amdgpu_sa_bo, olist); next 145 drivers/gpu/drm/amd/amdgpu/amdgpu_sa.c if (hole->next != &sa_manager->olist) { next 146 drivers/gpu/drm/amd/amdgpu/amdgpu_sa.c return list_entry(hole->next, struct amdgpu_sa_bo, olist)->soffset; next 214 drivers/gpu/drm/amd/amdgpu/amdgpu_sa.c if (sa_manager->hole->next == &sa_manager->olist) { next 341 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c struct dma_fence *next; next 379 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c resv, &next, false, true); next 384 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c fence = next; next 297 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c base->next = NULL; next 302 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c base->next = bo->vm_bo; next 583 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c for (bo_base = abo->vm_bo; bo_base; bo_base = bo_base->next) { next 1145 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c for (base = bo->vm_bo; base; base = base->next) { next 2297 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c struct amdgpu_bo_va_mapping *before, *after, *tmp, *next; next 2349 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c list_for_each_entry_safe(tmp, next, &removed, list) { next 2445 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c struct amdgpu_bo_va_mapping *mapping, *next; next 2455 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c base = &(*base)->next) { next 2459 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c *base = bo_va->base.next; next 2468 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c list_for_each_entry_safe(mapping, next, &bo_va->valids, list) { next 2475 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c list_for_each_entry_safe(mapping, next, &bo_va->invalids, list) { next 2512 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c for (bo_base = bo->vm_bo; bo_base; bo_base = bo_base->next) { next 138 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h struct amdgpu_vm_bo_base *next; next 809 drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c struct device_process_node *cur, *next; next 817 drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c list_for_each_entry_safe(cur, next, &dqm->queues, list) { next 1498 drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c struct queue *q, *next; next 1506 drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c list_for_each_entry_safe(q, next, &qpd->queues_list, list) { next 1572 drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c struct queue *q, *next; next 1638 drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c list_for_each_entry_safe(q, next, &qpd->queues_list, list) { next 150 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c struct process_queue_node *pqn, *next; next 152 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c list_for_each_entry_safe(pqn, next, &pqm->queues, process_queue_list) { next 143 drivers/gpu/drm/amd/amdkfd/kfd_topology.c while (dev->mem_props.next != &dev->mem_props) { next 144 drivers/gpu/drm/amd/amdkfd/kfd_topology.c mem = container_of(dev->mem_props.next, next 150 drivers/gpu/drm/amd/amdkfd/kfd_topology.c while (dev->cache_props.next != &dev->cache_props) { next 151 drivers/gpu/drm/amd/amdkfd/kfd_topology.c cache = container_of(dev->cache_props.next, next 157 drivers/gpu/drm/amd/amdkfd/kfd_topology.c while (dev->io_link_props.next != &dev->io_link_props) { next 158 drivers/gpu/drm/amd/amdkfd/kfd_topology.c iolink = container_of(dev->io_link_props.next, next 164 drivers/gpu/drm/amd/amdkfd/kfd_topology.c while (dev->perf_props.next != &dev->perf_props) { next 165 drivers/gpu/drm/amd/amdkfd/kfd_topology.c perf = container_of(dev->perf_props.next, next 836 drivers/gpu/drm/amd/amdkfd/kfd_topology.c list_move_tail(temp_list->next, master_list); next 41 drivers/gpu/drm/amd/powerplay/inc/power_state.h struct pp_power_state *next; next 116 drivers/gpu/drm/amd/powerplay/inc/smu_ucode_xfer_cz.h uint16_t next; next 340 drivers/gpu/drm/amd/powerplay/smumgr/smu8_smumgr.c task->next = is_last ? END_OF_TASK_LIST : smu8_smu->toc_entry_used_count; next 377 drivers/gpu/drm/amd/powerplay/smumgr/smu8_smumgr.c task->next = is_last ? END_OF_TASK_LIST : smu8_smu->toc_entry_used_count; next 428 drivers/gpu/drm/arm/display/komeda/komeda_private_obj.c struct drm_private_obj *obj, *next; next 430 drivers/gpu/drm/arm/display/komeda/komeda_private_obj.c list_for_each_entry_safe(obj, next, &config->privobj_list, head) next 197 drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_dc.h dma_addr_t next; next 856 drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c dscr->next = dscr_dma; next 35 drivers/gpu/drm/bridge/thc63lvd1024.c struct drm_bridge *next; next 49 drivers/gpu/drm/bridge/thc63lvd1024.c return drm_bridge_attach(bridge->encoder, thc63->next, bridge); next 147 drivers/gpu/drm/bridge/thc63lvd1024.c thc63->next = of_drm_find_bridge(remote); next 149 drivers/gpu/drm/bridge/thc63lvd1024.c if (!thc63->next) next 141 drivers/gpu/drm/drm_bridge.c previous->next = bridge; next 201 drivers/gpu/drm/drm_bridge.c ret = ret && drm_bridge_mode_fixup(bridge->next, mode, adjusted_mode); next 236 drivers/gpu/drm/drm_bridge.c return drm_bridge_mode_valid(bridge->next, mode); next 255 drivers/gpu/drm/drm_bridge.c drm_bridge_disable(bridge->next); next 280 drivers/gpu/drm/drm_bridge.c drm_bridge_post_disable(bridge->next); next 306 drivers/gpu/drm/drm_bridge.c drm_bridge_mode_set(bridge->next, mode, adjusted_mode); next 326 drivers/gpu/drm/drm_bridge.c drm_bridge_pre_enable(bridge->next); next 351 drivers/gpu/drm/drm_bridge.c drm_bridge_enable(bridge->next); next 373 drivers/gpu/drm/drm_bridge.c drm_atomic_bridge_disable(bridge->next, state); next 406 drivers/gpu/drm/drm_bridge.c drm_atomic_bridge_post_disable(bridge->next, state); next 429 drivers/gpu/drm/drm_bridge.c drm_atomic_bridge_pre_enable(bridge->next, state); next 461 drivers/gpu/drm/drm_bridge.c drm_atomic_bridge_enable(bridge->next, state); next 812 drivers/gpu/drm/drm_bufs.c buf->next = NULL; next 1011 drivers/gpu/drm/drm_bufs.c buf->next = NULL; next 1175 drivers/gpu/drm/drm_bufs.c buf->next = NULL; next 683 drivers/gpu/drm/drm_connector.c if (lhead->next == &config->connector_list) { next 688 drivers/gpu/drm/drm_connector.c lhead = lhead->next; next 893 drivers/gpu/drm/drm_dp_mst_topology.c list_del(&txmsg->next); next 1120 drivers/gpu/drm/drm_dp_mst_topology.c list_for_each_entry_safe(port, tmp, &mstb->ports, next) { next 1121 drivers/gpu/drm/drm_dp_mst_topology.c list_del(&port->next); next 1259 drivers/gpu/drm/drm_dp_mst_topology.c list_add(&port->next, &mgr->destroy_connector_list); next 1354 drivers/gpu/drm/drm_dp_mst_topology.c list_for_each_entry(port, &mstb->ports, next) { next 1389 drivers/gpu/drm/drm_dp_mst_topology.c list_for_each_entry(port, &mstb->ports, next) { next 1426 drivers/gpu/drm/drm_dp_mst_topology.c list_for_each_entry(port, &mstb->ports, next) { next 1670 drivers/gpu/drm/drm_dp_mst_topology.c list_add(&port->next, &mstb->ports); next 1704 drivers/gpu/drm/drm_dp_mst_topology.c list_del(&port->next); next 1781 drivers/gpu/drm/drm_dp_mst_topology.c list_for_each_entry(port, &mstb->ports, next) { next 1812 drivers/gpu/drm/drm_dp_mst_topology.c list_for_each_entry(port, &mstb->ports, next) { next 1854 drivers/gpu/drm/drm_dp_mst_topology.c list_for_each_entry(port, &mstb->ports, next) { next 2062 drivers/gpu/drm/drm_dp_mst_topology.c txmsg = list_first_entry(&mgr->tx_msg_downq, struct drm_dp_sideband_msg_tx, next); next 2066 drivers/gpu/drm/drm_dp_mst_topology.c list_del(&txmsg->next); next 2069 drivers/gpu/drm/drm_dp_mst_topology.c list_del(&txmsg->next); next 2100 drivers/gpu/drm/drm_dp_mst_topology.c list_add_tail(&txmsg->next, &mgr->tx_msg_downq); next 3254 drivers/gpu/drm/drm_dp_mst_topology.c list_for_each_entry(pos, &topology_state->vcpis, next) { next 3290 drivers/gpu/drm/drm_dp_mst_topology.c list_add(&vcpi->next, &topology_state->vcpis); next 3337 drivers/gpu/drm/drm_dp_mst_topology.c list_for_each_entry(pos, &topology_state->vcpis, next) { next 3620 drivers/gpu/drm/drm_dp_mst_topology.c list_for_each_entry(port, &mstb->ports, next) { next 3751 drivers/gpu/drm/drm_dp_mst_topology.c port = list_first_entry_or_null(&mgr->destroy_connector_list, struct drm_dp_mst_port, next); next 3756 drivers/gpu/drm/drm_dp_mst_topology.c list_del(&port->next); next 3759 drivers/gpu/drm/drm_dp_mst_topology.c INIT_LIST_HEAD(&port->next); next 3788 drivers/gpu/drm/drm_dp_mst_topology.c list_for_each_entry(pos, &old_state->vcpis, next) { next 3798 drivers/gpu/drm/drm_dp_mst_topology.c list_add(&vcpi->next, &state->vcpis); next 3804 drivers/gpu/drm/drm_dp_mst_topology.c list_for_each_entry_safe(pos, vcpi, &state->vcpis, next) { next 3820 drivers/gpu/drm/drm_dp_mst_topology.c list_for_each_entry_safe(pos, tmp, &mst_state->vcpis, next) { next 3837 drivers/gpu/drm/drm_dp_mst_topology.c list_for_each_entry(vcpi, &mst_state->vcpis, next) { next 170 drivers/gpu/drm/drm_encoder.c struct drm_bridge *next; next 173 drivers/gpu/drm/drm_encoder.c next = bridge->next; next 175 drivers/gpu/drm/drm_encoder.c bridge = next; next 344 drivers/gpu/drm/drm_file.c struct pci_bus *b = list_entry(pci_root_buses.next, next 206 drivers/gpu/drm/drm_syncobj.c if (!wait->node.next) next 370 drivers/gpu/drm/drm_syncobj.c if (wait.node.next) next 112 drivers/gpu/drm/exynos/exynos_dp.c bridge->next = NULL; next 508 drivers/gpu/drm/gma500/mmu.c unsigned long next; next 531 drivers/gpu/drm/gma500/mmu.c next = psb_pd_addr_end(addr, end); next 538 drivers/gpu/drm/gma500/mmu.c (addr & clflush_mask) < next); next 541 drivers/gpu/drm/gma500/mmu.c } while (addr = next, next != end); next 561 drivers/gpu/drm/gma500/mmu.c unsigned long next; next 570 drivers/gpu/drm/gma500/mmu.c next = psb_pd_addr_end(addr, end); next 577 drivers/gpu/drm/gma500/mmu.c } while (addr += PAGE_SIZE, addr < next); next 580 drivers/gpu/drm/gma500/mmu.c } while (addr = next, next != end); next 603 drivers/gpu/drm/gma500/mmu.c unsigned long next; next 626 drivers/gpu/drm/gma500/mmu.c next = psb_pd_addr_end(addr, end); next 634 drivers/gpu/drm/gma500/mmu.c } while (addr += PAGE_SIZE, addr < next); next 637 drivers/gpu/drm/gma500/mmu.c } while (addr = next, next != end); next 658 drivers/gpu/drm/gma500/mmu.c unsigned long next; next 668 drivers/gpu/drm/gma500/mmu.c next = psb_pd_addr_end(addr, end); next 678 drivers/gpu/drm/gma500/mmu.c } while (addr += PAGE_SIZE, addr < next); next 681 drivers/gpu/drm/gma500/mmu.c } while (addr = next, next != end); next 707 drivers/gpu/drm/gma500/mmu.c unsigned long next; next 732 drivers/gpu/drm/gma500/mmu.c next = psb_pd_addr_end(addr, end); next 741 drivers/gpu/drm/gma500/mmu.c } while (addr += PAGE_SIZE, addr < next); next 744 drivers/gpu/drm/gma500/mmu.c } while (addr = next, next != end); next 13864 drivers/gpu/drm/i915/display/intel_display.c struct intel_atomic_state *state, *next; next 13868 drivers/gpu/drm/i915/display/intel_display.c llist_for_each_entry_safe(state, next, freed, freed) next 499 drivers/gpu/drm/i915/display/intel_dsi_vbt.c const u8 *next = data + operation_size; next 504 drivers/gpu/drm/i915/display/intel_dsi_vbt.c if (operation_size && data != next) { next 1859 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c capture->next = eb->request->capture_list; next 16 drivers/gpu/drm/i915/gem/i915_gem_pm.c struct llist_node *node, *next; next 18 drivers/gpu/drm/i915/gem/i915_gem_pm.c llist_for_each_safe(node, next, llist_del_all(&engine->barrier_tasks)) { next 426 drivers/gpu/drm/i915/gem/i915_gem_shrinker.c struct i915_vma *vma, *next; next 442 drivers/gpu/drm/i915/gem/i915_gem_shrinker.c list_for_each_entry_safe(vma, next, next 136 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c struct list_head *pos, *next; next 147 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c list_for_each_safe(pos, next, &ce->signals) { next 187 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c list_for_each_safe(pos, next, &signal) { next 254 drivers/gpu/drm/i915/gt/intel_engine.h u32 next, u32 prev) next 256 drivers/gpu/drm/i915/gt/intel_engine.h typecheck(typeof(ring->size), next); next 258 drivers/gpu/drm/i915/gt/intel_engine.h return (next - prev) << ring->wrap; next 79 drivers/gpu/drm/i915/gt/intel_engine_user.c struct llist_node *pos, *next; next 81 drivers/gpu/drm/i915/gt/intel_engine_user.c llist_for_each_safe(pos, next, get_engines(i915)) { next 197 drivers/gpu/drm/i915/gt/intel_engine_user.c struct list_head *it, *next; next 205 drivers/gpu/drm/i915/gt/intel_engine_user.c list_for_each_safe(it, next, &engines) { next 360 drivers/gpu/drm/i915/gt/intel_lrc.c struct i915_request *next; next 363 drivers/gpu/drm/i915/gt/intel_lrc.c next = READ_ONCE(ve->request); next 364 drivers/gpu/drm/i915/gt/intel_lrc.c if (next) next 365 drivers/gpu/drm/i915/gt/intel_lrc.c preempt = rq_prio(next) > last_prio; next 388 drivers/gpu/drm/i915/gt/intel_lrc.c const struct i915_request *next) next 400 drivers/gpu/drm/i915/gt/intel_lrc.c return rq_prio(prev) >= rq_prio(next); next 597 drivers/gpu/drm/i915/gt/intel_lrc.c struct i915_request *next = READ_ONCE(ve->request); next 599 drivers/gpu/drm/i915/gt/intel_lrc.c if (next && next->execution_mask & ~rq->execution_mask) next 804 drivers/gpu/drm/i915/gt/intel_lrc.c const struct intel_context *next) next 806 drivers/gpu/drm/i915/gt/intel_lrc.c if (prev != next) next 816 drivers/gpu/drm/i915/gt/intel_lrc.c const struct i915_request *next) next 818 drivers/gpu/drm/i915/gt/intel_lrc.c GEM_BUG_ON(prev == next); next 819 drivers/gpu/drm/i915/gt/intel_lrc.c GEM_BUG_ON(!assert_priority_queue(prev, next)); next 829 drivers/gpu/drm/i915/gt/intel_lrc.c if (i915_request_completed(next)) next 832 drivers/gpu/drm/i915/gt/intel_lrc.c if (!can_merge_ctx(prev->hw_context, next->hw_context)) next 94 drivers/gpu/drm/i915/gvt/debugfs.c struct diff_mmio *node, *next; next 113 drivers/gpu/drm/i915/gvt/debugfs.c list_for_each_entry_safe(node, next, ¶m.diff_mmio_list, node) { next 406 drivers/gpu/drm/i915/gvt/execlist.c struct list_head *next = workload_q_head(vgpu, ring_id)->next; next 419 drivers/gpu/drm/i915/gvt/execlist.c next_workload = container_of(next, next 1531 drivers/gpu/drm/i915/gvt/gtt.c oos_page = container_of(gtt->oos_page_use_list_head.next, next 1540 drivers/gpu/drm/i915/gvt/gtt.c oos_page = container_of(gtt->oos_page_free_list_head.next, next 2501 drivers/gpu/drm/i915/gvt/gtt.c struct intel_gvt_partial_pte *pos, *next; next 2503 drivers/gpu/drm/i915/gvt/gtt.c list_for_each_entry_safe(pos, next, next 622 drivers/gpu/drm/i915/gvt/handlers.c unsigned int next, unsigned int end, i915_reg_t i915_end) next 624 drivers/gpu/drm/i915/gvt/handlers.c unsigned int range = next - start; next 391 drivers/gpu/drm/i915/gvt/mmio_context.c static void switch_mocs(struct intel_vgpu *pre, struct intel_vgpu *next, next 407 drivers/gpu/drm/i915/gvt/mmio_context.c dev_priv = pre ? pre->gvt->dev_priv : next->gvt->dev_priv; next 423 drivers/gpu/drm/i915/gvt/mmio_context.c if (next) next 424 drivers/gpu/drm/i915/gvt/mmio_context.c new_v = vgpu_vreg_t(next, offset); next 441 drivers/gpu/drm/i915/gvt/mmio_context.c if (next) next 442 drivers/gpu/drm/i915/gvt/mmio_context.c new_v = vgpu_vreg_t(next, l3_offset); next 468 drivers/gpu/drm/i915/gvt/mmio_context.c struct intel_vgpu *next, next 476 drivers/gpu/drm/i915/gvt/mmio_context.c dev_priv = pre ? pre->gvt->dev_priv : next->gvt->dev_priv; next 478 drivers/gpu/drm/i915/gvt/mmio_context.c switch_mocs(pre, next, ring_id); next 503 drivers/gpu/drm/i915/gvt/mmio_context.c if (next) { next 504 drivers/gpu/drm/i915/gvt/mmio_context.c s = &next->submission; next 515 drivers/gpu/drm/i915/gvt/mmio_context.c new_v = vgpu_vreg_t(next, mmio->reg) | next 518 drivers/gpu/drm/i915/gvt/mmio_context.c new_v = vgpu_vreg_t(next, mmio->reg); next 531 drivers/gpu/drm/i915/gvt/mmio_context.c next ? next->id : 0, next 537 drivers/gpu/drm/i915/gvt/mmio_context.c if (next) next 538 drivers/gpu/drm/i915/gvt/mmio_context.c handle_tlb_pending_event(next, ring_id); next 551 drivers/gpu/drm/i915/gvt/mmio_context.c struct intel_vgpu *next, int ring_id) next 555 drivers/gpu/drm/i915/gvt/mmio_context.c if (WARN_ON(!pre && !next)) next 559 drivers/gpu/drm/i915/gvt/mmio_context.c pre ? "vGPU" : "host", next ? "vGPU" : "HOST"); next 561 drivers/gpu/drm/i915/gvt/mmio_context.c dev_priv = pre ? pre->gvt->dev_priv : next->gvt->dev_priv; next 569 drivers/gpu/drm/i915/gvt/mmio_context.c switch_mmio(pre, next, ring_id); next 48 drivers/gpu/drm/i915/gvt/mmio_context.h struct intel_vgpu *next, int ring_id); next 785 drivers/gpu/drm/i915/gvt/scheduler.c workload_q_head(scheduler->current_vgpu, ring_id)->next, next 265 drivers/gpu/drm/i915/i915_active.c struct llist_node *pos, *next; next 284 drivers/gpu/drm/i915/i915_active.c llist_for_each_safe(pos, next, llist_del_all(&engine->barrier_tasks)) { next 290 drivers/gpu/drm/i915/i915_active.c pos->next = head; next 590 drivers/gpu/drm/i915/i915_active.c struct llist_node *pos, *next; next 646 drivers/gpu/drm/i915/i915_active.c llist_for_each_safe(pos, next, take_preallocated_barriers(ref)) { next 659 drivers/gpu/drm/i915/i915_active.c struct llist_node *pos, *next; next 670 drivers/gpu/drm/i915/i915_active.c llist_for_each_safe(pos, next, take_preallocated_barriers(ref)) { next 700 drivers/gpu/drm/i915/i915_active.c struct llist_node *node, *next; next 710 drivers/gpu/drm/i915/i915_active.c llist_for_each_safe(node, next, llist_del_all(&engine->barrier_tasks)) { next 101 drivers/gpu/drm/i915/i915_gem_evict.c struct i915_vma *vma, *next; next 142 drivers/gpu/drm/i915/i915_gem_evict.c list_for_each_entry_safe(vma, next, &vm->bound_list, vm_link) { next 180 drivers/gpu/drm/i915/i915_gem_evict.c list_for_each_entry_safe(vma, next, &eviction_list, evict_link) { next 226 drivers/gpu/drm/i915/i915_gem_evict.c list_for_each_entry_safe(vma, next, &eviction_list, evict_link) { next 235 drivers/gpu/drm/i915/i915_gem_evict.c list_for_each_entry_safe(vma, next, &eviction_list, evict_link) { next 268 drivers/gpu/drm/i915/i915_gem_evict.c struct i915_vma *vma, *next; next 351 drivers/gpu/drm/i915/i915_gem_evict.c list_for_each_entry_safe(vma, next, &eviction_list, evict_link) { next 375 drivers/gpu/drm/i915/i915_gem_evict.c struct i915_vma *vma, *next; next 404 drivers/gpu/drm/i915/i915_gem_evict.c list_for_each_entry_safe(vma, next, &eviction_list, evict_link) { next 57 drivers/gpu/drm/i915/i915_globals.c struct i915_global *global, *next; next 59 drivers/gpu/drm/i915/i915_globals.c list_for_each_entry_safe_reverse(global, next, &globals, link) next 684 drivers/gpu/drm/i915/i915_gpu_error.c for (ee = error->engine; ee; ee = ee->next) next 737 drivers/gpu/drm/i915/i915_gpu_error.c for (ee = error->engine; ee; ee = ee->next) next 740 drivers/gpu/drm/i915/i915_gpu_error.c for (ee = error->engine; ee; ee = ee->next) { next 929 drivers/gpu/drm/i915/i915_gpu_error.c error->engine = ee->next; next 1275 drivers/gpu/drm/i915/i915_gpu_error.c struct capture_vma *next; next 1280 drivers/gpu/drm/i915/i915_gpu_error.c capture_vma(struct capture_vma *next, next 1288 drivers/gpu/drm/i915/i915_gpu_error.c return next; next 1292 drivers/gpu/drm/i915/i915_gpu_error.c return next; next 1296 drivers/gpu/drm/i915/i915_gpu_error.c return next; next 1302 drivers/gpu/drm/i915/i915_gpu_error.c c->next = next; next 1316 drivers/gpu/drm/i915/i915_gpu_error.c for (c = request->capture_list; c; c = c->next) next 1331 drivers/gpu/drm/i915/i915_gpu_error.c for (c = request->capture_list; c; c = c->next) { next 1437 drivers/gpu/drm/i915/i915_gpu_error.c capture = this->next; next 1456 drivers/gpu/drm/i915/i915_gpu_error.c ee->next = error->engine; next 1655 drivers/gpu/drm/i915/i915_gpu_error.c for (ee = error->engine; ee; ee = ee->next) { next 163 drivers/gpu/drm/i915/i915_gpu_error.h struct drm_i915_error_engine *next; next 190 drivers/gpu/drm/i915/i915_request.c struct i915_capture_list *next = capture->next; next 193 drivers/gpu/drm/i915/i915_request.c capture = next; next 220 drivers/gpu/drm/i915/i915_request.c struct i915_active_request *active, *next; next 256 drivers/gpu/drm/i915/i915_request.c list_for_each_entry_safe(active, next, &rq->active_list, link) { next 267 drivers/gpu/drm/i915/i915_request.c prefetchw(next); next 49 drivers/gpu/drm/i915/i915_request.h struct i915_capture_list *next; next 305 drivers/gpu/drm/i915/i915_scheduler.c if (stack.dfs_link.next == stack.dfs_link.prev) next 135 drivers/gpu/drm/i915/i915_sw_fence.c wait_queue_entry_t *pos, *next; next 150 drivers/gpu/drm/i915/i915_sw_fence.c list_for_each_entry_safe(pos, next, &x->head, entry) { next 160 drivers/gpu/drm/i915/i915_sw_fence.c list_for_each_entry_safe(pos, next, &x->head, entry) { next 279 drivers/gpu/drm/i915/i915_syncmap.c struct i915_syncmap *next; next 285 drivers/gpu/drm/i915/i915_syncmap.c next = kzalloc(sizeof(*next) + KSYNCMAP * sizeof(next), next 287 drivers/gpu/drm/i915/i915_syncmap.c if (unlikely(!next)) next 293 drivers/gpu/drm/i915/i915_syncmap.c next->height = above + p->height; next 294 drivers/gpu/drm/i915/i915_syncmap.c next->prefix = __sync_branch_prefix(next, id); next 299 drivers/gpu/drm/i915/i915_syncmap.c __sync_child(p->parent)[idx] = next; next 302 drivers/gpu/drm/i915/i915_syncmap.c next->parent = p->parent; next 306 drivers/gpu/drm/i915/i915_syncmap.c __sync_set_child(next, idx, p); next 307 drivers/gpu/drm/i915/i915_syncmap.c p->parent = next; next 310 drivers/gpu/drm/i915/i915_syncmap.c p = next; next 319 drivers/gpu/drm/i915/i915_syncmap.c next = __sync_child(p)[idx]; next 320 drivers/gpu/drm/i915/i915_syncmap.c if (!next) { next 321 drivers/gpu/drm/i915/i915_syncmap.c next = __sync_alloc_leaf(p, id); next 322 drivers/gpu/drm/i915/i915_syncmap.c if (unlikely(!next)) next 325 drivers/gpu/drm/i915/i915_syncmap.c __sync_set_child(p, idx, next); next 326 drivers/gpu/drm/i915/i915_syncmap.c p = next; next 330 drivers/gpu/drm/i915/i915_syncmap.c p = next; next 26 drivers/gpu/drm/i915/i915_user_extensions.c u64 next; next 53 drivers/gpu/drm/i915/i915_user_extensions.c if (get_user(next, &ext->next_extension) || next 54 drivers/gpu/drm/i915/i915_user_extensions.c overflows_type(next, ext)) next 57 drivers/gpu/drm/i915/i915_user_extensions.c ext = u64_to_user_ptr(next); next 245 drivers/gpu/drm/i915/i915_utils.h WRITE_ONCE(head->next, first); next 837 drivers/gpu/drm/i915/i915_vma.c struct i915_vma *vma, *next; next 840 drivers/gpu/drm/i915/i915_vma.c list_for_each_entry_safe(vma, next, &i915->gt.closed_vma, closed_link) { next 390 drivers/gpu/drm/i915/selftests/i915_gem_evict.c struct reserved *next; next 442 drivers/gpu/drm/i915/selftests/i915_gem_evict.c r->next = reserved; next 516 drivers/gpu/drm/i915/selftests/i915_gem_evict.c struct reserved *next = reserved->next; next 521 drivers/gpu/drm/i915/selftests/i915_gem_evict.c reserved = next; next 147 drivers/gpu/drm/ingenic/ingenic-drm.c u32 next; next 353 drivers/gpu/drm/ingenic/ingenic-drm.c regmap_write(priv->map, JZ_REG_LCD_DA0, priv->dma_hwdesc->next); next 696 drivers/gpu/drm/ingenic/ingenic-drm.c priv->dma_hwdesc->next = priv->dma_hwdesc_phys; next 198 drivers/gpu/drm/mediatek/mtk_drm_ddp.c enum mtk_ddp_comp_id next, next 203 drivers/gpu/drm/mediatek/mtk_drm_ddp.c if (cur == DDP_COMPONENT_OVL0 && next == DDP_COMPONENT_COLOR0) { next 206 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_OVL0 && next == DDP_COMPONENT_RDMA0) { next 209 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_OD0 && next == DDP_COMPONENT_RDMA0) { next 212 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_UFOE && next == DDP_COMPONENT_DSI0) { next 215 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_OVL1 && next == DDP_COMPONENT_COLOR1) { next 218 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_GAMMA && next == DDP_COMPONENT_RDMA1) { next 221 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_OD1 && next == DDP_COMPONENT_RDMA1) { next 224 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_RDMA0 && next == DDP_COMPONENT_DPI0) { next 227 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_RDMA0 && next == DDP_COMPONENT_DPI1) { next 230 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_RDMA0 && next == DDP_COMPONENT_DSI1) { next 233 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_RDMA0 && next == DDP_COMPONENT_DSI2) { next 236 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_RDMA0 && next == DDP_COMPONENT_DSI3) { next 239 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_RDMA1 && next == DDP_COMPONENT_DSI1) { next 242 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_RDMA1 && next == DDP_COMPONENT_DSI2) { next 245 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_RDMA1 && next == DDP_COMPONENT_DSI3) { next 248 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_RDMA1 && next == DDP_COMPONENT_DPI0) { next 251 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_RDMA1 && next == DDP_COMPONENT_DPI1) { next 254 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_RDMA2 && next == DDP_COMPONENT_DPI0) { next 257 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_RDMA2 && next == DDP_COMPONENT_DPI1) { next 260 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_RDMA2 && next == DDP_COMPONENT_DSI1) { next 263 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_RDMA2 && next == DDP_COMPONENT_DSI2) { next 266 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_RDMA2 && next == DDP_COMPONENT_DSI3) { next 277 drivers/gpu/drm/mediatek/mtk_drm_ddp.c enum mtk_ddp_comp_id next, next 282 drivers/gpu/drm/mediatek/mtk_drm_ddp.c if (cur == DDP_COMPONENT_OVL0 && next == DDP_COMPONENT_COLOR0) { next 285 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_RDMA1 && next == DDP_COMPONENT_DPI0) { next 288 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_RDMA1 && next == DDP_COMPONENT_DPI1) { next 291 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_RDMA1 && next == DDP_COMPONENT_DSI0) { next 294 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_RDMA1 && next == DDP_COMPONENT_DSI1) { next 297 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_RDMA1 && next == DDP_COMPONENT_DSI2) { next 300 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_RDMA1 && next == DDP_COMPONENT_DSI3) { next 303 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_RDMA2 && next == DDP_COMPONENT_DPI0) { next 306 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_RDMA2 && next == DDP_COMPONENT_DPI1) { next 309 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_RDMA2 && next == DDP_COMPONENT_DSI0) { next 312 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_RDMA2 && next == DDP_COMPONENT_DSI1) { next 315 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_RDMA2 && next == DDP_COMPONENT_DSI2) { next 318 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_RDMA2 && next == DDP_COMPONENT_DSI3) { next 321 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_OVL1 && next == DDP_COMPONENT_COLOR1) { next 324 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_BLS && next == DDP_COMPONENT_DSI0) { next 336 drivers/gpu/drm/mediatek/mtk_drm_ddp.c enum mtk_ddp_comp_id next) next 338 drivers/gpu/drm/mediatek/mtk_drm_ddp.c if (cur == DDP_COMPONENT_BLS && next == DDP_COMPONENT_DSI0) { next 341 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_BLS && next == DDP_COMPONENT_DPI0) { next 353 drivers/gpu/drm/mediatek/mtk_drm_ddp.c enum mtk_ddp_comp_id next) next 357 drivers/gpu/drm/mediatek/mtk_drm_ddp.c value = mtk_ddp_mout_en(cur, next, &addr); next 363 drivers/gpu/drm/mediatek/mtk_drm_ddp.c mtk_ddp_sout_sel(config_regs, cur, next); next 365 drivers/gpu/drm/mediatek/mtk_drm_ddp.c value = mtk_ddp_sel_in(cur, next, &addr); next 374 drivers/gpu/drm/mediatek/mtk_drm_ddp.c enum mtk_ddp_comp_id next) next 378 drivers/gpu/drm/mediatek/mtk_drm_ddp.c value = mtk_ddp_mout_en(cur, next, &addr); next 384 drivers/gpu/drm/mediatek/mtk_drm_ddp.c value = mtk_ddp_sel_in(cur, next, &addr); next 17 drivers/gpu/drm/mediatek/mtk_drm_ddp.h enum mtk_ddp_comp_id next); next 20 drivers/gpu/drm/mediatek/mtk_drm_ddp.h enum mtk_ddp_comp_id next); next 1245 drivers/gpu/drm/mediatek/mtk_hdmi.c if (hdmi->bridge.next) { next 1249 drivers/gpu/drm/mediatek/mtk_hdmi.c if (!drm_bridge_mode_fixup(hdmi->bridge.next, mode, next 235 drivers/gpu/drm/mga/mga_dma.c for (entry = dev_priv->head->next; entry; entry = entry->next) { next 267 drivers/gpu/drm/mga/mga_dma.c entry->next = dev_priv->head->next; next 272 drivers/gpu/drm/mga/mga_dma.c if (dev_priv->head->next != NULL) next 273 drivers/gpu/drm/mga/mga_dma.c dev_priv->head->next->prev = entry; next 274 drivers/gpu/drm/mga/mga_dma.c if (entry->next == NULL) next 281 drivers/gpu/drm/mga/mga_dma.c dev_priv->head->next = entry; next 291 drivers/gpu/drm/mga/mga_dma.c drm_mga_freelist_t *next; next 296 drivers/gpu/drm/mga/mga_dma.c next = entry->next; next 298 drivers/gpu/drm/mga/mga_dma.c entry = next; next 325 drivers/gpu/drm/mga/mga_dma.c drm_mga_freelist_t *next; next 343 drivers/gpu/drm/mga/mga_dma.c next = dev_priv->tail; next 344 drivers/gpu/drm/mga/mga_dma.c prev->next = NULL; next 345 drivers/gpu/drm/mga/mga_dma.c next->prev = next->next = NULL; next 347 drivers/gpu/drm/mga/mga_dma.c SET_AGE(&next->age, MGA_BUFFER_USED, 0); next 348 drivers/gpu/drm/mga/mga_dma.c return next->buf; next 372 drivers/gpu/drm/mga/mga_dma.c prev->next = entry; next 374 drivers/gpu/drm/mga/mga_dma.c entry->next = NULL; next 376 drivers/gpu/drm/mga/mga_dma.c prev = head->next; next 377 drivers/gpu/drm/mga/mga_dma.c head->next = entry; next 380 drivers/gpu/drm/mga/mga_dma.c entry->next = prev; next 80 drivers/gpu/drm/mga/mga_drv.h struct drm_mga_freelist *next; next 31 drivers/gpu/drm/msm/adreno/a5xx_gpu.c ring->cur = ring->next; next 60 drivers/gpu/drm/msm/adreno/a6xx_gpu.c ring->cur = ring->next; next 350 drivers/gpu/drm/msm/adreno/adreno_gpu.c ring->next = ring->start; next 492 drivers/gpu/drm/msm/adreno/adreno_gpu.c ring->cur = ring->next; next 807 drivers/gpu/drm/msm/adreno/adreno_gpu.c uint32_t wptr = ring->next - ring->start; next 937 drivers/gpu/drm/msm/msm_gem.c struct msm_gem_object *msm_obj, *next; next 943 drivers/gpu/drm/msm/msm_gem.c llist_for_each_entry_safe(msm_obj, next, next 41 drivers/gpu/drm/msm/msm_ringbuffer.c ring->next = ring->start; next 40 drivers/gpu/drm/msm/msm_ringbuffer.h uint32_t *start, *end, *cur, *next; next 64 drivers/gpu/drm/msm/msm_ringbuffer.h if (ring->next == ring->end) next 65 drivers/gpu/drm/msm/msm_ringbuffer.h ring->next = ring->start; next 66 drivers/gpu/drm/msm/msm_ringbuffer.h *(ring->next++) = data; next 111 drivers/gpu/drm/nouveau/include/nvif/list.h struct list_head *next, *prev; next 130 drivers/gpu/drm/nouveau/include/nvif/list.h list->next = list->prev = list; next 135 drivers/gpu/drm/nouveau/include/nvif/list.h struct list_head *prev, struct list_head *next) next 137 drivers/gpu/drm/nouveau/include/nvif/list.h next->prev = entry; next 138 drivers/gpu/drm/nouveau/include/nvif/list.h entry->next = next; next 140 drivers/gpu/drm/nouveau/include/nvif/list.h prev->next = entry; next 161 drivers/gpu/drm/nouveau/include/nvif/list.h __list_add(entry, head, head->next); next 186 drivers/gpu/drm/nouveau/include/nvif/list.h __list_del(struct list_head *prev, struct list_head *next) next 188 drivers/gpu/drm/nouveau/include/nvif/list.h next->prev = prev; next 189 drivers/gpu/drm/nouveau/include/nvif/list.h prev->next = next; next 209 drivers/gpu/drm/nouveau/include/nvif/list.h __list_del(entry->prev, entry->next); next 215 drivers/gpu/drm/nouveau/include/nvif/list.h __list_del(entry->prev, entry->next); next 222 drivers/gpu/drm/nouveau/include/nvif/list.h __list_del(list->prev, list->next); next 237 drivers/gpu/drm/nouveau/include/nvif/list.h return head->next == head; next 277 drivers/gpu/drm/nouveau/include/nvif/list.h list_entry((ptr)->next, type, member) next 315 drivers/gpu/drm/nouveau/include/nvif/list.h for (pos = __container_of((head)->next, pos, member); \ next 317 drivers/gpu/drm/nouveau/include/nvif/list.h pos = __container_of(pos->member.next, pos, member)) next 327 drivers/gpu/drm/nouveau/include/nvif/list.h for (pos = __container_of((head)->next, pos, member), \ next 328 drivers/gpu/drm/nouveau/include/nvif/list.h tmp = __container_of(pos->member.next, pos, member); \ next 330 drivers/gpu/drm/nouveau/include/nvif/list.h pos = tmp, tmp = __container_of(pos->member.next, tmp, member)) next 339 drivers/gpu/drm/nouveau/include/nvif/list.h for (pos = __container_of(pos->member.next, pos, member); \ next 341 drivers/gpu/drm/nouveau/include/nvif/list.h pos = __container_of(pos->member.next, pos, member)) next 351 drivers/gpu/drm/nouveau/include/nvif/list.h pos = __container_of(pos->member.next, pos, member)) next 9 drivers/gpu/drm/nouveau/include/nvkm/core/mm.h struct nvkm_mm_node *next; next 58 drivers/gpu/drm/nouveau/include/nvkm/core/mm.h return !node->next; next 75 drivers/gpu/drm/nouveau/include/nvkm/core/mm.h } while ((node = node->next)); next 136 drivers/gpu/drm/nouveau/include/nvkm/subdev/fb.h struct nvkm_ram_data *next; next 75 drivers/gpu/drm/nouveau/include/nvkm/subdev/mmu.h u64 next; next 96 drivers/gpu/drm/nouveau/nouveau_fence.c fence = list_entry(fctx->pending.next, typeof(*fence), head); next 133 drivers/gpu/drm/nouveau/nouveau_fence.c fence = list_entry(fctx->pending.next, typeof(*fence), head); next 157 drivers/gpu/drm/nouveau/nouveau_fence.c fence = list_entry(fctx->pending.next, typeof(*fence), head); next 338 drivers/gpu/drm/nouveau/nouveau_gem.c nvbo = list_entry(op->list.next, struct nouveau_bo, entry); next 183 drivers/gpu/drm/nouveau/nouveau_svm.c unsigned long next; next 190 drivers/gpu/drm/nouveau/nouveau_svm.c next = min(vma->vm_end, end); next 192 drivers/gpu/drm/nouveau/nouveau_svm.c nouveau_dmem_migrate_vma(cli->drm, vma, addr, next); next 193 drivers/gpu/drm/nouveau/nouveau_svm.c addr = next; next 54 drivers/gpu/drm/nouveau/nvkm/core/mm.c struct nvkm_mm_node *next = node(this, next); next 62 drivers/gpu/drm/nouveau/nvkm/core/mm.c if (next && next->type == NVKM_MM_TYPE_NONE) { next 63 drivers/gpu/drm/nouveau/nvkm/core/mm.c next->offset = this->offset; next 64 drivers/gpu/drm/nouveau/nvkm/core/mm.c next->length += this->length; next 114 drivers/gpu/drm/nouveau/nvkm/core/mm.c struct nvkm_mm_node *prev, *this, *next; next 133 drivers/gpu/drm/nouveau/nvkm/core/mm.c next = node(this, next); next 134 drivers/gpu/drm/nouveau/nvkm/core/mm.c if (next && next->type != type) next 150 drivers/gpu/drm/nouveau/nvkm/core/mm.c this->next = NULL; next 189 drivers/gpu/drm/nouveau/nvkm/core/mm.c struct nvkm_mm_node *prev, *this, *next; next 207 drivers/gpu/drm/nouveau/nvkm/core/mm.c next = node(this, next); next 208 drivers/gpu/drm/nouveau/nvkm/core/mm.c if (next && next->type != type) { next 210 drivers/gpu/drm/nouveau/nvkm/core/mm.c c = next->offset - e; next 229 drivers/gpu/drm/nouveau/nvkm/core/mm.c this->next = NULL; next 243 drivers/gpu/drm/nouveau/nvkm/core/mm.c u32 next; next 247 drivers/gpu/drm/nouveau/nvkm/core/mm.c next = prev->offset + prev->length; next 248 drivers/gpu/drm/nouveau/nvkm/core/mm.c if (next != offset) { next 249 drivers/gpu/drm/nouveau/nvkm/core/mm.c BUG_ON(next > offset); next 253 drivers/gpu/drm/nouveau/nvkm/core/mm.c node->offset = next; next 254 drivers/gpu/drm/nouveau/nvkm/core/mm.c node->length = offset - next; next 47 drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.c u32 next = nvkm_rd32(device, list->data[i].addr + base + 0); next 53 drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.c if (prev != next) next 54 drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.c snprintf(mods, sizeof(mods), "-> %08x", next); next 48 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c } prev, next, *chan; next 62 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c status->next.tsg = !!(stat & 0x10000000); next 63 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c status->next.id = (stat & 0x0fff0000) >> 16; next 74 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c status->chan = &status->next; next 79 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c status->chan = &status->next; next 94 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c status->next.tsg ? "tsg" : "ch", status->next.id, next 95 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c status->chan == &status->next ? "*" : " "); next 189 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c mem = fifo->runlist[runl].mem[fifo->runlist[runl].next]; next 190 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c fifo->runlist[runl].next = !fifo->runlist[runl].next; next 33 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.h int next; next 1027 drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c u32 next = init->addr + init->count * init->pitch; next 1029 drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c while (addr < next) { next 1047 drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c u32 next = init->addr + init->count * init->pitch; next 1055 drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c while (addr < next) { next 1084 drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c u32 next = init->addr + init->count * init->pitch; next 1092 drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c while (addr < next) { next 1118 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv04.c struct nv04_gr_chan *next = NULL; next 1130 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv04.c next = gr->chan[chid]; next 1131 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv04.c if (next) next 1132 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv04.c nv04_gr_load_context(next, chid); next 936 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv10.c struct nv10_gr_chan *next = NULL; next 948 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv10.c next = gr->chan[chid]; next 949 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv10.c if (next) next 950 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv10.c nv10_gr_load_context(next, chid); next 233 drivers/gpu/drm/nouveau/nvkm/engine/pm/base.c dom->func->next(pm, dom); next 256 drivers/gpu/drm/nouveau/nvkm/engine/pm/base.c dom->func->next(pm, dom); next 320 drivers/gpu/drm/nouveau/nvkm/engine/pm/base.c if (ctr->head.next) next 172 drivers/gpu/drm/nouveau/nvkm/engine/pm/gf100.c .next = gf100_perfctr_next, next 74 drivers/gpu/drm/nouveau/nvkm/engine/pm/nv40.c .next = nv40_perfctr_next, next 93 drivers/gpu/drm/nouveau/nvkm/engine/pm/priv.h void (*next)(struct nvkm_pm *, struct nvkm_perfdom *); next 75 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr3.c switch (ram->next->bios.timing_ver) { next 77 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr3.c CWL = ram->next->bios.timing_10_CWL; next 78 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr3.c CL = ram->next->bios.timing_10_CL; next 79 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr3.c WR = ram->next->bios.timing_10_WR; next 80 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr3.c DLL = !ram->next->bios.ramcfg_DLLoff; next 81 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr3.c ODT = ram->next->bios.timing_10_ODT; next 82 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr3.c RON = ram->next->bios.ramcfg_RON; next 85 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr3.c CWL = (ram->next->bios.timing[1] & 0x00000f80) >> 7; next 86 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr3.c CL = (ram->next->bios.timing[1] & 0x0000001f) >> 0; next 87 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr3.c WR = (ram->next->bios.timing[2] & 0x007f0000) >> 16; next 96 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr3.c if (ram->next->bios.timing_ver == 0x20 || next 97 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr3.c ram->next->bios.ramcfg_timing == 0xff) { next 41 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr5.c xd = !ram->next->bios.ramcfg_DLLoff; next 43 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr5.c switch (ram->next->bios.ramcfg_ver) { next 45 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr5.c pd = ram->next->bios.ramcfg_11_01_80; next 46 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr5.c lf = ram->next->bios.ramcfg_11_01_40; next 47 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr5.c vh = ram->next->bios.ramcfg_11_02_10; next 48 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr5.c vr = ram->next->bios.ramcfg_11_02_04; next 49 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr5.c vo = ram->next->bios.ramcfg_11_06; next 50 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr5.c l3 = !ram->next->bios.ramcfg_11_07_02; next 56 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr5.c switch (ram->next->bios.timing_ver) { next 58 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr5.c WL = (ram->next->bios.timing[1] & 0x00000f80) >> 7; next 59 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr5.c CL = (ram->next->bios.timing[1] & 0x0000001f); next 60 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr5.c WR = (ram->next->bios.timing[2] & 0x007f0000) >> 16; next 61 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr5.c at[0] = ram->next->bios.timing_20_2e_c0; next 62 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr5.c at[1] = ram->next->bios.timing_20_2e_30; next 63 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr5.c dt = ram->next->bios.timing_20_2e_03; next 64 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr5.c ds = ram->next->bios.timing_20_2f_03; next 82 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c struct nvkm_mm_node *next = vram->mn; next 85 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c while ((node = next)) { next 86 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c next = node->next; next 143 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c node = &r->next; next 210 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c struct nvkm_ram_data *next = ram->base.next; next 211 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c u8 v0 = next->bios.ramcfg_11_03_c0; next 212 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c u8 v1 = next->bios.ramcfg_11_03_30; next 243 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c u32 next = (prev & ~mask) | data; next 244 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c nvkm_memx_wr32(fuc->memx, addr, next); next 255 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c struct nvkm_ram_data *next = ram->base.next; next 256 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c int vc = !next->bios.ramcfg_11_02_08; next 257 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c int mv = !next->bios.ramcfg_11_02_04; next 311 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c switch (next->bios.ramcfg_11_03_c0) { next 318 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c switch (next->bios.ramcfg_11_03_30) { next 326 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if (next->bios.ramcfg_11_02_80) next 328 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if (next->bios.ramcfg_11_02_40) next 330 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if (next->bios.ramcfg_11_07_10) next 332 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if (next->bios.ramcfg_11_07_08) next 367 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if (next->bios.ramcfg_11_02_40 || next 368 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c next->bios.ramcfg_11_07_10) { next 391 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if (next->bios.ramcfg_11_07_40) next 395 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c ram_wr32(fuc, 0x10f65c, 0x00000011 * next->bios.rammap_11_11_0c); next 396 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c ram_wr32(fuc, 0x10f6b8, 0x01010101 * next->bios.ramcfg_11_09); next 397 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c ram_wr32(fuc, 0x10f6bc, 0x01010101 * next->bios.ramcfg_11_09); next 399 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if (!next->bios.ramcfg_11_07_08 && !next->bios.ramcfg_11_07_04) { next 400 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c ram_wr32(fuc, 0x10f698, 0x01010101 * next->bios.ramcfg_11_04); next 401 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c ram_wr32(fuc, 0x10f69c, 0x01010101 * next->bios.ramcfg_11_04); next 403 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if (!next->bios.ramcfg_11_07_08) { next 409 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c u32 data = 0x01000100 * next->bios.ramcfg_11_04; next 414 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if (ram->mode == 2 && next->bios.ramcfg_11_08_10) next 422 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if (!next->bios.ramcfg_11_02_80) next 424 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if (!next->bios.ramcfg_11_02_40) next 426 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if (!next->bios.ramcfg_11_07_10) next 428 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if (!next->bios.ramcfg_11_07_08) next 434 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if (next->bios.ramcfg_11_01_08) next 445 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if (next->bios.ramcfg_11_08_01) next 452 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if (next->bios.ramcfg_11_08_08) next 454 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if (next->bios.ramcfg_11_08_04) next 456 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if (next->bios.ramcfg_11_08_02) next 461 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c ram_mask(fuc, 0x10f248, 0xffffffff, next->bios.timing[10]); next 462 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c ram_mask(fuc, 0x10f290, 0xffffffff, next->bios.timing[0]); next 463 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c ram_mask(fuc, 0x10f294, 0xffffffff, next->bios.timing[1]); next 464 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c ram_mask(fuc, 0x10f298, 0xffffffff, next->bios.timing[2]); next 465 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c ram_mask(fuc, 0x10f29c, 0xffffffff, next->bios.timing[3]); next 466 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c ram_mask(fuc, 0x10f2a0, 0xffffffff, next->bios.timing[4]); next 467 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c ram_mask(fuc, 0x10f2a4, 0xffffffff, next->bios.timing[5]); next 468 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c ram_mask(fuc, 0x10f2a8, 0xffffffff, next->bios.timing[6]); next 469 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c ram_mask(fuc, 0x10f2ac, 0xffffffff, next->bios.timing[7]); next 470 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c ram_mask(fuc, 0x10f2cc, 0xffffffff, next->bios.timing[8]); next 471 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c ram_mask(fuc, 0x10f2e8, 0xffffffff, next->bios.timing[9]); next 475 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if (next->bios.ramcfg_11_08_20) next 483 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c data |= next->bios.ramcfg_11_02_03 << 8; next 487 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if (next->bios.ramcfg_11_01_10) next 495 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c data |= next->bios.timing_20_30_07 << 28; next 499 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if (next->bios.ramcfg_11_01_01) next 507 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c data |= next->bios.timing_20_30_07 << 28; next 511 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if (next->bios.ramcfg_11_01_02) next 519 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if (!next->bios.ramcfg_11_01_04) next 521 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if (!next->bios.ramcfg_11_07_80) next 526 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if (next->bios.ramcfg_11_03_f0) { next 527 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if (next->bios.rammap_11_08_0c) { next 528 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if (!next->bios.ramcfg_11_07_80) next 541 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c ram_wr32(fuc, 0x10f870, 0x11111111 * next->bios.ramcfg_11_03_0f); next 545 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c data |= next->bios.ramcfg_11_02_03; next 549 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if (next->bios.ramcfg_11_01_10) next 560 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c data = next->bios.timing_20_30_07 << 8; next 561 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if (next->bios.ramcfg_11_01_01) next 565 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c ram_mask(fuc, 0x10f250, 0x000003f0, next->bios.timing_20_2c_003f << 4); next 566 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c data = (next->bios.timing[10] & 0x7f000000) >> 24; next 567 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if (data < next->bios.timing_20_2c_1fc0) next 568 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c data = next->bios.timing_20_2c_1fc0; next 570 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c ram_mask(fuc, 0x10f224, 0x001f0000, next->bios.timing_20_30_f8 << 16); next 572 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c ram_mask(fuc, 0x10fec4, 0x041e0f07, next->bios.timing_20_31_0800 << 26 | next 573 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c next->bios.timing_20_31_0780 << 17 | next 574 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c next->bios.timing_20_31_0078 << 8 | next 575 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c next->bios.timing_20_31_0007); next 576 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c ram_mask(fuc, 0x10fec8, 0x00000027, next->bios.timing_20_31_8000 << 5 | next 577 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c next->bios.timing_20_31_7000); next 585 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if (next->bios.ramcfg_11_08_10 && (ram->mode == 2) /*XXX*/) { next 618 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if (!next->bios.ramcfg_11_07_08) { next 619 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if (!next->bios.ramcfg_11_07_04) next 633 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if (!next->bios.ramcfg_11_07_08) { next 635 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if ( next->bios.ramcfg_11_07_04) next 637 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if (!next->bios.rammap_11_08_10) next 659 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if (next->bios.ramcfg_11_07_02) next 667 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if (next->bios.rammap_11_08_01) next 706 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c struct nvkm_ram_data *next = ram->base.next; next 707 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c int vc = !next->bios.ramcfg_11_02_08; next 708 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c int mv = !next->bios.ramcfg_11_02_04; next 726 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if (next->bios.ramcfg_11_03_f0) next 731 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if (next->bios.ramcfg_DLLoff) next 755 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c switch (next->bios.ramcfg_11_03_c0) { next 762 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c switch (next->bios.ramcfg_11_03_30) { next 770 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if (next->bios.ramcfg_11_02_80) next 772 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if (next->bios.ramcfg_11_02_40) next 774 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if (next->bios.ramcfg_11_07_10) next 776 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if (next->bios.ramcfg_11_07_08) next 786 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c data |= next->bios.ramcfg_11_03_30 << 16; next 818 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if (next->bios.ramcfg_11_02_40 || next 819 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c next->bios.ramcfg_11_07_10) { next 825 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if (next->bios.ramcfg_11_07_40) next 829 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c ram_wr32(fuc, 0x10f65c, 0x00000011 * next->bios.rammap_11_11_0c); next 830 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c ram_wr32(fuc, 0x10f6b8, 0x01010101 * next->bios.ramcfg_11_09); next 831 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c ram_wr32(fuc, 0x10f6bc, 0x01010101 * next->bios.ramcfg_11_09); next 835 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if (!next->bios.ramcfg_11_02_80) next 837 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if (!next->bios.ramcfg_11_02_40) next 839 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if (!next->bios.ramcfg_11_07_10) next 841 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if (!next->bios.ramcfg_11_07_08) next 848 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if (next->bios.ramcfg_11_08_01) next 855 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c ram_mask(fuc, 0x10f248, 0xffffffff, next->bios.timing[10]); next 856 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c ram_mask(fuc, 0x10f290, 0xffffffff, next->bios.timing[0]); next 857 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c ram_mask(fuc, 0x10f294, 0xffffffff, next->bios.timing[1]); next 858 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c ram_mask(fuc, 0x10f298, 0xffffffff, next->bios.timing[2]); next 859 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c ram_mask(fuc, 0x10f29c, 0xffffffff, next->bios.timing[3]); next 860 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c ram_mask(fuc, 0x10f2a0, 0xffffffff, next->bios.timing[4]); next 861 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c ram_mask(fuc, 0x10f2a4, 0xffffffff, next->bios.timing[5]); next 862 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c ram_mask(fuc, 0x10f2a8, 0xffffffff, next->bios.timing[6]); next 863 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c ram_mask(fuc, 0x10f2ac, 0xffffffff, next->bios.timing[7]); next 864 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c ram_mask(fuc, 0x10f2cc, 0xffffffff, next->bios.timing[8]); next 865 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c ram_mask(fuc, 0x10f2e8, 0xffffffff, next->bios.timing[9]); next 869 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if (!next->bios.ramcfg_11_01_04) next 871 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if (!next->bios.ramcfg_11_07_80) next 876 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if (next->bios.ramcfg_11_03_f0) { next 877 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if (next->bios.rammap_11_08_0c) { next 878 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if (!next->bios.ramcfg_11_07_80) next 892 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c ram_wr32(fuc, 0x10f870, 0x11111111 * next->bios.ramcfg_11_03_0f); next 894 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c ram_mask(fuc, 0x10f250, 0x000003f0, next->bios.timing_20_2c_003f << 4); next 896 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c data = (next->bios.timing[10] & 0x7f000000) >> 24; next 897 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if (data < next->bios.timing_20_2c_1fc0) next 898 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c data = next->bios.timing_20_2c_1fc0; next 901 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c ram_mask(fuc, 0x10f224, 0x001f0000, next->bios.timing_20_30_f8 << 16); next 911 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if (!next->bios.ramcfg_DLLoff) { next 921 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if (!next->bios.ramcfg_DLLoff) { next 949 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if (next->bios.rammap_11_08_01) next 1041 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c gk104_ram_calc_xits(struct gk104_ram *ram, struct nvkm_ram_data *next) next 1052 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c ram->mode = (next->freq > fuc->refpll.vco1.max_freq) ? 2 : 1; next 1063 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c refclk = next->freq; next 1065 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c ret = gk104_pll_calc_hiclk(next->freq, subdev->device->crystal, next 1074 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c " (refclock: %i kHz)\n", next->freq, ret); next 1090 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c ram->base.freq = next->freq; next 1096 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c ret = gk104_ram_calc_sddr3(ram, next->freq); next 1101 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c ret = gk104_ram_calc_gddr5(ram, next->freq); next 1120 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c if (ram->base.next == NULL) { next 1143 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c ram->base.next = &ram->base.target; next 1145 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c ram->base.next = &ram->base.xition; next 1147 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c BUG_ON(ram->base.next != &ram->base.xition); next 1148 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c ram->base.next = &ram->base.target; next 1151 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c return gk104_ram_calc_xits(ram, ram->base.next); next 1236 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c struct nvkm_ram_data *next = ram->base.next; next 1240 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c return (ram->base.next == &ram->base.xition); next 1245 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c gk104_ram_prog_0(ram, next->freq); next 1247 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c return (ram->base.next == &ram->base.xition); next 1254 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c ram->base.next = NULL; next 502 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c struct nvkm_ram_data *next; next 511 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c next = &ram->base.target; next 512 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c next->freq = freq; next 513 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c ram->base.next = next; next 520 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c &next->bios); next 534 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c &ver, &hdr, &next->bios); next 541 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c if (next->bios.ramcfg_timing != 0xff) { next 542 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c data = nvbios_timingEp(bios, next->bios.ramcfg_timing, next 544 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c &next->bios); next 597 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c if (!next->bios.ramcfg_DLLoff) next 605 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c if (next->bios.ramcfg_10_02_10) { next 629 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c if (!next->bios.ramcfg_10_02_10) { next 637 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c switch (next->bios.ramcfg_DLLoff * ram->base.type) { next 646 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c if (next->bios.timing_10_ODT) next 663 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c next->bios.ramcfg_FBVDDQ) { next 671 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c gt215_ram_gpio(fuc, 0x18, !next->bios.ramcfg_FBVDDQ); next 707 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c if (next->bios.rammap_10_04_08) { next 708 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c ram_wr32(fuc, 0x1005a0, next->bios.ramcfg_10_06 << 16 | next 709 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c next->bios.ramcfg_10_05 << 8 | next 710 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c next->bios.ramcfg_10_05); next 711 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c ram_wr32(fuc, 0x1005a4, next->bios.ramcfg_10_08 << 8 | next 712 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c next->bios.ramcfg_10_07); next 713 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c ram_wr32(fuc, 0x10f804, next->bios.ramcfg_10_09_f0 << 20 | next 714 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c next->bios.ramcfg_10_03_0f << 16 | next 715 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c next->bios.ramcfg_10_09_0f | next 766 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c ram_mask(fuc, 0x100200, 0x00001000, !next->bios.ramcfg_10_02_08 << 12); next 777 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c if (!next->bios.ramcfg_10_02_04) next 780 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c if (next->bios.ramcfg_10_02_04) { next 785 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c if (next->bios.ramcfg_10_02_10) next 800 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c if (next->bios.ramcfg_10_02_10) next 814 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c unk714 |= (next->bios.ramcfg_10_04_01) << 8; next 816 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c if (next->bios.ramcfg_10_02_20) next 818 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c if (next->bios.ramcfg_10_02_02) next 820 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c if (next->bios.ramcfg_10_02_01) next 822 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c if (next->bios.timing_10_24 != 0xff) { next 824 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c unk718 |= next->bios.timing_10_24 << 28; next 826 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c if (next->bios.ramcfg_10_02_10) next 834 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c if (!next->bios.timing_10_ODT) next 838 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c if (!next->bios.ramcfg_DLLoff) next 861 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c if (next->bios.rammap_10_04_02) next 863 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c if (next->bios.ramcfg_10_02_10) { next 97 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c unkt3b = 0x19 + ram->base.next->bios.rammap_00_16_40; next 99 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c ram->base.next->bios.rammap_00_16_40) << 16 | next 227 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c struct nvkm_ram_data *next; next 235 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c next = &ram->base.target; next 236 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c next->freq = freq; next 237 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c ram->base.next = next; next 251 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c nvbios_rammapEp_from_perf(bios, data, hdr, &next->bios); next 261 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c &next->bios); next 268 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c if (next->bios.ramcfg_timing != 0xff) { next 269 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c data = nvbios_timingEp(bios, next->bios.ramcfg_timing, next 270 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c &ver, &hdr, &cnt, &len, &next->bios); next 305 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c if (subdev->device->chipset <= 0x96 && !next->bios.ramcfg_00_03_02) next 319 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c if (next->bios.timing_10_ODT) next 355 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c next->bios.rammap_00_16_40 << 14); next 363 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c nv50_ram_gpio(hwsq, 0x18, !next->bios.ramcfg_FBVDDQ); next 400 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c if (!next->bios.ramcfg_00_03_02) next 402 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c ram_mask(hwsq, 0x100200, 0x00001000, !next->bios.ramcfg_00_04_02 << 12); next 413 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c if (!next->bios.ramcfg_00_03_08) next 415 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c if (!next->bios.ramcfg_FBVDDQ) next 417 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c if ( next->bios.ramcfg_00_04_04) next 422 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c if (!next->bios.ramcfg_00_03_08) next 426 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c if ( next->bios.ramcfg_00_03_01) next 428 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c if ( next->bios.ramcfg_00_03_02) next 430 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c if (!next->bios.ramcfg_00_03_08) next 432 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c if ( next->bios.ramcfg_00_04_04) next 434 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c if ( next->bios.ramcfg_00_04_20) next 444 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c if (next->bios.rammap_00_16_20) { next 445 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c ram_wr32(hwsq, 0x1005a0, next->bios.ramcfg_00_07 << 16 | next 446 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c next->bios.ramcfg_00_06 << 8 | next 447 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c next->bios.ramcfg_00_05); next 448 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c ram_wr32(hwsq, 0x1005a4, next->bios.ramcfg_00_09 << 8 | next 449 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c next->bios.ramcfg_00_08); next 456 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c if (!next->bios.timing_10_ODT) next 460 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c if (!next->bios.ramcfg_DLLoff) next 468 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c if (next->bios.rammap_00_17_02) next 470 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c if (!next->bios.rammap_00_16_40) next 472 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c if (next->bios.ramcfg_00_03_02) next 474 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c if (subdev->device->chipset <= 0x96 && next->bios.ramcfg_00_03_02) next 65 drivers/gpu/drm/nouveau/nvkm/subdev/fb/sddr2.c switch (ram->next->bios.timing_ver) { next 67 drivers/gpu/drm/nouveau/nvkm/subdev/fb/sddr2.c CL = ram->next->bios.timing_10_CL; next 68 drivers/gpu/drm/nouveau/nvkm/subdev/fb/sddr2.c WR = ram->next->bios.timing_10_WR; next 69 drivers/gpu/drm/nouveau/nvkm/subdev/fb/sddr2.c DLL = !ram->next->bios.ramcfg_DLLoff; next 70 drivers/gpu/drm/nouveau/nvkm/subdev/fb/sddr2.c ODT = ram->next->bios.timing_10_ODT & 3; next 73 drivers/gpu/drm/nouveau/nvkm/subdev/fb/sddr2.c CL = (ram->next->bios.timing[1] & 0x0000001f); next 74 drivers/gpu/drm/nouveau/nvkm/subdev/fb/sddr2.c WR = (ram->next->bios.timing[2] & 0x007f0000) >> 16; next 80 drivers/gpu/drm/nouveau/nvkm/subdev/fb/sddr2.c if (ram->next->bios.timing_ver == 0x20 || next 81 drivers/gpu/drm/nouveau/nvkm/subdev/fb/sddr2.c ram->next->bios.ramcfg_timing == 0xff) { next 74 drivers/gpu/drm/nouveau/nvkm/subdev/fb/sddr3.c DLL = !ram->next->bios.ramcfg_DLLoff; next 76 drivers/gpu/drm/nouveau/nvkm/subdev/fb/sddr3.c switch (ram->next->bios.timing_ver) { next 78 drivers/gpu/drm/nouveau/nvkm/subdev/fb/sddr3.c if (ram->next->bios.timing_hdr < 0x17) { next 82 drivers/gpu/drm/nouveau/nvkm/subdev/fb/sddr3.c CWL = ram->next->bios.timing_10_CWL; next 83 drivers/gpu/drm/nouveau/nvkm/subdev/fb/sddr3.c CL = ram->next->bios.timing_10_CL; next 84 drivers/gpu/drm/nouveau/nvkm/subdev/fb/sddr3.c WR = ram->next->bios.timing_10_WR; next 85 drivers/gpu/drm/nouveau/nvkm/subdev/fb/sddr3.c ODT = ram->next->bios.timing_10_ODT; next 88 drivers/gpu/drm/nouveau/nvkm/subdev/fb/sddr3.c CWL = (ram->next->bios.timing[1] & 0x00000f80) >> 7; next 89 drivers/gpu/drm/nouveau/nvkm/subdev/fb/sddr3.c CL = (ram->next->bios.timing[1] & 0x0000001f) >> 0; next 90 drivers/gpu/drm/nouveau/nvkm/subdev/fb/sddr3.c WR = (ram->next->bios.timing[2] & 0x007f0000) >> 16; next 204 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c if (likely(iobj->lru.next) && iobj->map) { next 245 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c if (likely(iobj->lru.next)) next 269 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c if (likely(iobj->lru.next)) { next 271 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c iobj->lru.next = NULL; next 297 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c iobj->lru.next = NULL; /* Exclude from eviction. */ next 319 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c if (likely(iobj->lru.next)) next 450 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c bool next = pgt->pte[ptei] & NVKM_VMM_PTE_SPTES; next 451 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c if (spte != next) next 650 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c u64 next = 1ULL << page[i - 1].shift; next 651 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c u64 part = ALIGN(addr, next) - addr; next 652 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c if (size - part >= next) next 884 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c struct nvkm_vma *vma, struct nvkm_vma *next, u64 size) next 886 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c if (next) { next 888 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c vma->size += next->size; next 889 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c nvkm_vmm_node_delete(vmm, next); next 899 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c nvkm_vmm_node_remove(vmm, next); next 901 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c next->addr -= size; next 902 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c next->size += size; next 903 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c nvkm_vmm_node_insert(vmm, next); next 904 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c return next; next 1155 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c struct nvkm_vma *next = NULL; next 1162 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c if (vma->addr + vma->size == addr + size && (next = node(vma, next))) { next 1163 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c if (!next->part || next 1164 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c next->memory || next->mapped != map) next 1165 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c next = NULL; next 1168 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c if (prev || next) next 1169 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c return nvkm_vmm_node_merge(vmm, prev, vma, next, size); next 1177 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c struct nvkm_vma *next; next 1193 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c next = nvkm_vmm_pfn_split_merge(vmm, vma, start, size, 0, false); next 1194 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c if (!WARN_ON(!next)) { next 1195 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c vma = next; next 1199 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c } while ((vma = node(vma, next)) && (start = vma->addr) < limit); next 1260 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c goto next; next 1280 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c goto next; next 1311 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c next: next 1314 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c vma = node(vma, next); next 1337 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c struct nvkm_vma *next; next 1345 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c if ((next = node(vma, next)) && (!next->part || next->mapped)) next 1346 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c next = NULL; next 1347 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c nvkm_vmm_node_merge(vmm, prev, vma, next, vma->size); next 1468 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c for (; map->off; map->mem = map->mem->next) { next 1524 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c struct nvkm_vma *prev, *next; next 1532 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c if ((next = node(vma, next)) && !next->used) { next 1533 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c vma->size += next->size; next 1534 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c nvkm_vmm_free_delete(vmm, next); next 1544 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c struct nvkm_vma *next = vma; next 1550 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c const bool mem = next->memory != NULL; next 1551 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c const bool map = next->mapped; next 1552 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c const u8 refd = next->refd; next 1553 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c const u64 addr = next->addr; next 1554 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c u64 size = next->size; next 1557 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c while ((next = node(next, next)) && next->part && next 1558 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c (next->mapped == map) && next 1559 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c (next->memory != NULL) == mem && next 1560 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c (next->refd == refd)) next 1561 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c size += next->size; next 1576 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c } while (next && next->part); next 1583 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c next = vma; next 1585 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c if (next->mapped) next 1586 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c nvkm_vmm_unmap_region(vmm, next); next 1587 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c } while ((next = node(vma, next)) && next->part); next 1707 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c struct nvkm_vma *next = node(this, next); next 1716 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c if (vmm->func->page_block && next && next->page != p) next 315 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.h (MAP->mem = MAP->mem->next)) next 38 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c if (map->ctag && !(map->next & (1ULL << 44))) { next 45 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c base += map->next; next 52 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c data += map->next; next 254 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c map->next = (1 << page->shift) >> 8; next 304 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c map->next |= 1ULL << 44; next 118 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c data += map->next; next 201 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c data += map->next; next 327 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c map->next = (1ULL << page->shift) >> 4; next 375 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c map->next |= map->ctag; next 35 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c u64 next = addr + map->type, data; next 48 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c data = next | (log2blk << 7); next 49 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c next += pten * map->next; next 243 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c map->next = 1 << page->shift; next 312 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c map->next |= map->ctag; next 38 drivers/gpu/drm/nouveau/nvkm/subdev/therm/gf100.c u32 next, addr; next 41 drivers/gpu/drm/nouveau/nvkm/subdev/therm/gf100.c next = init->addr + init->count * 8; next 46 drivers/gpu/drm/nouveau/nvkm/subdev/therm/gf100.c while (addr < next) { next 32 drivers/gpu/drm/omapdrm/displays/encoder-opa362.c return omapdss_device_connect(dst->dss, dst, dst->next); next 38 drivers/gpu/drm/omapdrm/displays/encoder-opa362.c omapdss_device_disconnect(dst, dst->next); next 91 drivers/gpu/drm/omapdrm/displays/encoder-opa362.c dssdev->next = omapdss_of_find_connected_device(pdev->dev.of_node, 1); next 92 drivers/gpu/drm/omapdrm/displays/encoder-opa362.c if (IS_ERR(dssdev->next)) { next 93 drivers/gpu/drm/omapdrm/displays/encoder-opa362.c if (PTR_ERR(dssdev->next) != -EPROBE_DEFER) next 95 drivers/gpu/drm/omapdrm/displays/encoder-opa362.c return PTR_ERR(dssdev->next); next 108 drivers/gpu/drm/omapdrm/displays/encoder-opa362.c if (dssdev->next) next 109 drivers/gpu/drm/omapdrm/displays/encoder-opa362.c omapdss_device_put(dssdev->next); next 38 drivers/gpu/drm/omapdrm/displays/encoder-tpd12s015.c r = omapdss_device_connect(dst->dss, dst, dst->next); next 59 drivers/gpu/drm/omapdrm/displays/encoder-tpd12s015.c omapdss_device_disconnect(dst, dst->next); next 172 drivers/gpu/drm/omapdrm/displays/encoder-tpd12s015.c dssdev->next = omapdss_of_find_connected_device(pdev->dev.of_node, 1); next 173 drivers/gpu/drm/omapdrm/displays/encoder-tpd12s015.c if (IS_ERR(dssdev->next)) { next 174 drivers/gpu/drm/omapdrm/displays/encoder-tpd12s015.c if (PTR_ERR(dssdev->next) != -EPROBE_DEFER) next 176 drivers/gpu/drm/omapdrm/displays/encoder-tpd12s015.c return PTR_ERR(dssdev->next); next 189 drivers/gpu/drm/omapdrm/displays/encoder-tpd12s015.c if (dssdev->next) next 190 drivers/gpu/drm/omapdrm/displays/encoder-tpd12s015.c omapdss_device_put(dssdev->next); next 153 drivers/gpu/drm/omapdrm/dss/base.c (dssdev->next || dssdev->bridge || dssdev->panel)) next 241 drivers/gpu/drm/omapdrm/dss/base.c omapdss_device_pre_enable(dssdev->next); next 256 drivers/gpu/drm/omapdrm/dss/base.c omapdss_device_enable(dssdev->next); next 267 drivers/gpu/drm/omapdrm/dss/base.c omapdss_device_disable(dssdev->next); next 282 drivers/gpu/drm/omapdrm/dss/base.c omapdss_device_post_disable(dssdev->next); next 45 drivers/gpu/drm/omapdrm/dss/display.c while (output->next) next 46 drivers/gpu/drm/omapdrm/dss/display.c output = output->next; next 582 drivers/gpu/drm/omapdrm/dss/dpi.c return omapdss_device_connect(dst->dss, dst, dst->next); next 588 drivers/gpu/drm/omapdrm/dss/dpi.c omapdss_device_disconnect(dst, dst->next); next 4874 drivers/gpu/drm/omapdrm/dss/dsi.c return omapdss_device_connect(dst->dss, dst, dst->next); next 4880 drivers/gpu/drm/omapdrm/dss/dsi.c omapdss_device_disconnect(dst, dst->next); next 399 drivers/gpu/drm/omapdrm/dss/hdmi4.c return omapdss_device_connect(dst->dss, dst, dst->next); next 405 drivers/gpu/drm/omapdrm/dss/hdmi4.c omapdss_device_disconnect(dst, dst->next); next 404 drivers/gpu/drm/omapdrm/dss/hdmi5.c return omapdss_device_connect(dst->dss, dst, dst->next); next 410 drivers/gpu/drm/omapdrm/dss/hdmi5.c omapdss_device_disconnect(dst, dst->next); next 401 drivers/gpu/drm/omapdrm/dss/omapdss.h struct omap_dss_device *next; next 31 drivers/gpu/drm/omapdrm/dss/output.c out->next = omapdss_find_device_by_node(remote_node); next 39 drivers/gpu/drm/omapdrm/dss/output.c if (out->next && out->type != out->next->type) { next 41 drivers/gpu/drm/omapdrm/dss/output.c omapdss_device_put(out->next); next 42 drivers/gpu/drm/omapdrm/dss/output.c out->next = NULL; next 46 drivers/gpu/drm/omapdrm/dss/output.c return out->next || out->bridge || out->panel ? 0 : -EPROBE_DEFER; next 52 drivers/gpu/drm/omapdrm/dss/output.c if (out->next) next 53 drivers/gpu/drm/omapdrm/dss/output.c omapdss_device_put(out->next); next 237 drivers/gpu/drm/omapdrm/dss/sdi.c return omapdss_device_connect(dst->dss, dst, dst->next); next 243 drivers/gpu/drm/omapdrm/dss/sdi.c omapdss_device_disconnect(dst, dst->next); next 679 drivers/gpu/drm/omapdrm/dss/venc.c return omapdss_device_connect(dst->dss, dst, dst->next); next 685 drivers/gpu/drm/omapdrm/dss/venc.c omapdss_device_disconnect(dst, dst->next); next 41 drivers/gpu/drm/omapdrm/omap_connector.c for (dssdev = omap_connector->output; dssdev; dssdev = dssdev->next) { next 102 drivers/gpu/drm/omapdrm/omap_connector.c for (d = omap_connector->output; d; d = d->next) { next 250 drivers/gpu/drm/omapdrm/omap_connector.c for (; dssdev; dssdev = dssdev->next) { next 366 drivers/gpu/drm/omapdrm/omap_crtc.c struct omap_dss_device *dssdev = omap_crtc->pipe->output->next; next 504 drivers/gpu/drm/omapdrm/omap_crtc.c if (omap_crtc->pipe->output->next == NULL || next 505 drivers/gpu/drm/omapdrm/omap_crtc.c omap_crtc->pipe->output->next->type != OMAP_DISPLAY_TYPE_DSI) { next 557 drivers/gpu/drm/omapdrm/omap_crtc.c struct omap_dss_device *display = omap_crtc->pipe->output->next; next 331 drivers/gpu/drm/omapdrm/omap_dmm_tiler.c engine = list_entry(dmm->idle_head.next, struct refill_engine, next 210 drivers/gpu/drm/omapdrm/omap_drv.c if (output->next) { next 219 drivers/gpu/drm/omapdrm/omap_drv.c while (bridge->next) next 220 drivers/gpu/drm/omapdrm/omap_drv.c bridge = bridge->next; next 125 drivers/gpu/drm/omapdrm/omap_encoder.c for (dssdev = output; dssdev; dssdev = dssdev->next) next 128 drivers/gpu/drm/omapdrm/omap_encoder.c for (bridge = output->bridge; bridge; bridge = bridge->next) { next 142 drivers/gpu/drm/omapdrm/omap_encoder.c for (dssdev = output; dssdev; dssdev = dssdev->next) { next 170 drivers/gpu/drm/omapdrm/omap_encoder.c omapdss_device_disable(dssdev->next); next 186 drivers/gpu/drm/omapdrm/omap_encoder.c omapdss_device_post_disable(dssdev->next); next 198 drivers/gpu/drm/omapdrm/omap_encoder.c omapdss_device_pre_enable(dssdev->next); next 214 drivers/gpu/drm/omapdrm/omap_encoder.c omapdss_device_enable(dssdev->next); next 235 drivers/gpu/drm/qxl/qxl_cmd.c next_id = info->next; next 375 drivers/gpu/drm/qxl/qxl_dev.h uint64_t next; /* out */ next 166 drivers/gpu/drm/qxl/qxl_release.c entry = container_of(release->bos.next, next 799 drivers/gpu/drm/r128/r128_cce.c entry->next = dev_priv->head->next; next 800 drivers/gpu/drm/r128/r128_cce.c if (!entry->next) next 807 drivers/gpu/drm/r128/r128_cce.c dev_priv->head->next = entry; next 809 drivers/gpu/drm/r128/r128_cce.c if (dev_priv->head->next) next 810 drivers/gpu/drm/r128/r128_cce.c dev_priv->head->next->prev = entry; next 70 drivers/gpu/drm/r128/r128_drv.h struct drm_r128_freelist *next; next 43 drivers/gpu/drm/radeon/mkregtable.c struct list_head *next, *prev; next 49 drivers/gpu/drm/radeon/mkregtable.c list->next = list; next 61 drivers/gpu/drm/radeon/mkregtable.c struct list_head *prev, struct list_head *next) next 63 drivers/gpu/drm/radeon/mkregtable.c next->prev = new; next 64 drivers/gpu/drm/radeon/mkregtable.c new->next = next; next 66 drivers/gpu/drm/radeon/mkregtable.c prev->next = new; next 70 drivers/gpu/drm/radeon/mkregtable.c struct list_head *prev, struct list_head *next); next 102 drivers/gpu/drm/radeon/mkregtable.c for (pos = list_entry((head)->next, typeof(*pos), member); \ next 104 drivers/gpu/drm/radeon/mkregtable.c pos = list_entry(pos->member.next, typeof(*pos), member)) next 242 drivers/gpu/drm/radeon/radeon_mn.c head = bo->mn_list.next; next 158 drivers/gpu/drm/radeon/radeon_sa.c if (sa_manager->hole->next == &sa_manager->olist) next 161 drivers/gpu/drm/radeon/radeon_sa.c sa_bo = list_entry(sa_manager->hole->next, struct radeon_sa_bo, olist); next 184 drivers/gpu/drm/radeon/radeon_sa.c if (hole->next != &sa_manager->olist) { next 185 drivers/gpu/drm/radeon/radeon_sa.c return list_entry(hole->next, struct radeon_sa_bo, olist)->soffset; next 255 drivers/gpu/drm/radeon/radeon_sa.c if (sa_manager->hole->next == &sa_manager->olist) { next 197 drivers/gpu/drm/radeon/rv6xx_dpm.c struct rv6xx_sclk_stepping next; next 199 drivers/gpu/drm/radeon/rv6xx_dpm.c next.post_divider = cur->post_divider; next 202 drivers/gpu/drm/radeon/rv6xx_dpm.c next.vco_frequency = (cur->vco_frequency * (100 + step_size)) / 100; next 204 drivers/gpu/drm/radeon/rv6xx_dpm.c next.vco_frequency = (cur->vco_frequency * 100 + 99 + step_size) / (100 + step_size); next 206 drivers/gpu/drm/radeon/rv6xx_dpm.c return next; next 222 drivers/gpu/drm/radeon/rv6xx_dpm.c struct rv6xx_sclk_stepping next = *cur; next 224 drivers/gpu/drm/radeon/rv6xx_dpm.c while (rv6xx_can_step_post_div(rdev, &next, target)) next 225 drivers/gpu/drm/radeon/rv6xx_dpm.c next.post_divider--; next 227 drivers/gpu/drm/radeon/rv6xx_dpm.c return next; next 259 drivers/gpu/drm/radeon/rv6xx_dpm.c struct rv6xx_sclk_stepping next; next 262 drivers/gpu/drm/radeon/rv6xx_dpm.c next = rv6xx_next_post_div_step(rdev, &cur, &target); next 264 drivers/gpu/drm/radeon/rv6xx_dpm.c next = rv6xx_next_vco_step(rdev, &cur, increasing_vco, R600_VCOSTEPPCT_DFLT); next 266 drivers/gpu/drm/radeon/rv6xx_dpm.c if (rv6xx_reached_stepping_target(rdev, &next, &target, increasing_vco)) { next 269 drivers/gpu/drm/radeon/rv6xx_dpm.c tiny.post_divider = next.post_divider; next 274 drivers/gpu/drm/radeon/rv6xx_dpm.c if ((next.post_divider != target.post_divider) && next 275 drivers/gpu/drm/radeon/rv6xx_dpm.c (next.vco_frequency != target.vco_frequency)) { next 279 drivers/gpu/drm/radeon/rv6xx_dpm.c final_vco.post_divider = next.post_divider; next 287 drivers/gpu/drm/radeon/rv6xx_dpm.c rv6xx_output_stepping(rdev, step_index++, &next); next 289 drivers/gpu/drm/radeon/rv6xx_dpm.c cur = next; next 224 drivers/gpu/drm/savage/savage_bci.c dev_priv->head.next = &dev_priv->tail; next 228 drivers/gpu/drm/savage/savage_bci.c dev_priv->tail.next = NULL; next 239 drivers/gpu/drm/savage/savage_bci.c entry->next = dev_priv->head.next; next 241 drivers/gpu/drm/savage/savage_bci.c dev_priv->head.next->prev = entry; next 242 drivers/gpu/drm/savage/savage_bci.c dev_priv->head.next = entry; next 269 drivers/gpu/drm/savage/savage_bci.c drm_savage_buf_priv_t *next = tail->next; next 271 drivers/gpu/drm/savage/savage_bci.c prev->next = next; next 272 drivers/gpu/drm/savage/savage_bci.c next->prev = prev; next 273 drivers/gpu/drm/savage/savage_bci.c tail->next = tail->prev = NULL; next 284 drivers/gpu/drm/savage/savage_bci.c drm_savage_buf_priv_t *entry = buf->dev_private, *prev, *next; next 288 drivers/gpu/drm/savage/savage_bci.c if (entry->next != NULL || entry->prev != NULL) { next 294 drivers/gpu/drm/savage/savage_bci.c next = prev->next; next 295 drivers/gpu/drm/savage/savage_bci.c prev->next = entry; next 296 drivers/gpu/drm/savage/savage_bci.c next->prev = entry; next 298 drivers/gpu/drm/savage/savage_bci.c entry->next = next; next 1071 drivers/gpu/drm/savage/savage_bci.c buf_priv->next == NULL && buf_priv->prev == NULL) { next 64 drivers/gpu/drm/savage/savage_drv.h struct drm_savage_buf_priv *next; next 375 drivers/gpu/drm/selftests/test-drm_mm.c struct drm_mm_node tmp, *nodes, *node, *next; next 489 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_for_each_node_safe(node, next, &mm) next 573 drivers/gpu/drm/selftests/test-drm_mm.c struct drm_mm_node *nodes, *node, *next; next 683 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_for_each_node_safe(node, next, &mm) next 692 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_for_each_node_safe(node, next, &mm) next 880 drivers/gpu/drm/selftests/test-drm_mm.c struct drm_mm_node *nodes, *node, *next; next 947 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_for_each_node_safe(node, next, &mm) next 956 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_for_each_node_safe(node, next, &mm) next 1041 drivers/gpu/drm/selftests/test-drm_mm.c struct drm_mm_node *nodes, *node, *next; next 1073 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_for_each_node_safe(node, next, &mm) next 1082 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_for_each_node_safe(node, next, &mm) next 1093 drivers/gpu/drm/selftests/test-drm_mm.c struct drm_mm_node *node, *next; next 1125 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_for_each_node_safe(node, next, &mm) { next 1156 drivers/gpu/drm/selftests/test-drm_mm.c struct drm_mm_node *next = list_next_entry(hole, node_list); next 1164 drivers/gpu/drm/selftests/test-drm_mm.c if (next->allocated) next 1167 drivers/gpu/drm/selftests/test-drm_mm.c next->start, next->size, next->color); next 1261 drivers/gpu/drm/selftests/test-drm_mm.c e->link.next = NULL; next 1266 drivers/gpu/drm/selftests/test-drm_mm.c e->link.next = &e->link; next 1272 drivers/gpu/drm/selftests/test-drm_mm.c if (!e->link.next) { next 1408 drivers/gpu/drm/selftests/test-drm_mm.c struct drm_mm_node *node, *next; next 1501 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_for_each_node_safe(node, next, &mm) next 1521 drivers/gpu/drm/selftests/test-drm_mm.c struct drm_mm_node *node, *next; next 1598 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_for_each_node_safe(node, next, &mm) next 1621 drivers/gpu/drm/selftests/test-drm_mm.c struct drm_mm_node *nodes, *node, *next; next 1707 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_for_each_node_safe(node, next, &mm) next 1715 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_for_each_node_safe(node, next, &mm) next 1735 drivers/gpu/drm/selftests/test-drm_mm.c struct drm_mm_node *nodes, *node, *next; next 1807 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_for_each_node_safe(node, next, &mm) next 1815 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_for_each_node_safe(node, next, &mm) next 2160 drivers/gpu/drm/selftests/test-drm_mm.c struct drm_mm_node *node, *next; next 2242 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_for_each_node_safe(node, next, &mm) next 2263 drivers/gpu/drm/selftests/test-drm_mm.c struct drm_mm_node *node, *next; next 2343 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_for_each_node_safe(node, next, &mm) next 318 drivers/gpu/drm/sis/sis_mm.c struct sis_memblock *entry, *next; next 336 drivers/gpu/drm/sis/sis_mm.c list_for_each_entry_safe(entry, next, &file_priv->obj_list, next 997 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c struct dma_page *d_page, *next; next 1009 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c list_for_each_entry_safe(d_page, next, &ttm_dma->pages_list, next 1039 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c list_for_each_entry_safe(d_page, next, &ttm_dma->pages_list, next 182 drivers/gpu/drm/udl/udl_main.c node = udl->urbs.list.next; /* have reserved one with sem */ next 283 drivers/gpu/drm/udl/udl_main.c entry = udl->urbs.list.next; next 168 drivers/gpu/drm/vboxvideo/vbva_base.c u32 next; next 176 drivers/gpu/drm/vboxvideo/vbva_base.c next = (vbva_ctx->vbva->record_free_index + 1) % VBVA_MAX_RECORDS; next 179 drivers/gpu/drm/vboxvideo/vbva_base.c if (next == vbva_ctx->vbva->record_first_index) next 183 drivers/gpu/drm/vboxvideo/vbva_base.c if (next == vbva_ctx->vbva->record_first_index) next 188 drivers/gpu/drm/vboxvideo/vbva_base.c vbva_ctx->vbva->record_free_index = next; next 490 drivers/gpu/drm/vc4/vc4_gem.c struct vc4_exec_info *next; next 493 drivers/gpu/drm/vc4/vc4_gem.c next = vc4_first_bin_job(vc4); next 500 drivers/gpu/drm/vc4/vc4_gem.c if (next && next->perfmon == exec->perfmon) next 117 drivers/gpu/drm/vc4/vc4_irq.c struct vc4_exec_info *next, *exec = vc4_first_bin_job(vc4); next 123 drivers/gpu/drm/vc4/vc4_irq.c next = vc4_first_bin_job(vc4); next 129 drivers/gpu/drm/vc4/vc4_irq.c if (next && next->perfmon == exec->perfmon) next 56 drivers/gpu/drm/via/via_dmablit.c uint32_t next; next 74 drivers/gpu/drm/via/via_dmablit.c dma_addr_t next = vsg->chain_start; next 83 drivers/gpu/drm/via/via_dmablit.c dma_unmap_single(&pdev->dev, next, sizeof(*desc_ptr), DMA_TO_DEVICE); next 85 drivers/gpu/drm/via/via_dmablit.c next = (dma_addr_t) desc_ptr->next; next 114 drivers/gpu/drm/via/via_dmablit.c dma_addr_t next = 0 | VIA_DMA_DPR_EC; next 141 drivers/gpu/drm/via/via_dmablit.c desc_ptr->next = (uint32_t) next; next 142 drivers/gpu/drm/via/via_dmablit.c next = dma_map_single(&pdev->dev, desc_ptr, sizeof(*desc_ptr), next 161 drivers/gpu/drm/via/via_dmablit.c vsg->chain_start = next; next 214 drivers/gpu/drm/via/via_mm.c struct via_memblock *entry, *next; next 231 drivers/gpu/drm/via/via_mm.c list_for_each_entry_safe(entry, next, &file_priv->obj_list, next 479 drivers/gpu/drm/vmwgfx/ttm_object.c list = tfile->ref_list.next; next 353 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c struct vmw_ctx_bindinfo *entry, *next; next 356 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c list_for_each_entry_safe(entry, next, &cbs->list, ctx_list) next 395 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c struct vmw_ctx_bindinfo *entry, *next; next 398 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c list_for_each_entry_safe(entry, next, head, res_list) next 447 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c struct vmw_ctx_bindinfo *entry, *next; next 449 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c list_for_each_entry_safe(entry, next, &from->list, ctx_list) { next 1238 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c struct vmw_ctx_bindinfo *entry, *next; next 1240 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c list_for_each_entry_safe(entry, next, &cbs->list, ctx_list) next 381 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c struct vmw_cmdbuf_header *entry, *next; next 385 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c list_for_each_entry_safe(entry, next, &ctx->hw_submitted, list) { next 516 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c struct vmw_cmdbuf_header *entry, *next; next 529 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c list_for_each_entry_safe(entry, next, &man->error, list) { next 125 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c struct vmw_cmdbuf_res *entry, *next; next 127 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c list_for_each_entry_safe(entry, next, list, head) { next 162 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c struct vmw_cmdbuf_res *entry, *next; next 165 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c list_for_each_entry_safe(entry, next, list, head) { next 320 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c struct vmw_cmdbuf_res *entry, *next; next 322 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c list_for_each_entry_safe(entry, next, &man->list, head) next 324 drivers/gpu/drm/vmwgfx/vmwgfx_drv.h bool (*next)(struct vmw_piter *); next 953 drivers/gpu/drm/vmwgfx/vmwgfx_drv.h return viter->next(viter); next 108 drivers/gpu/drm/vmwgfx/vmwgfx_ldu.c entry = list_entry(lds->active.next, typeof(*entry), active); next 47 drivers/gpu/drm/vmwgfx/vmwgfx_marker.c struct vmw_marker *marker, *next; next 50 drivers/gpu/drm/vmwgfx/vmwgfx_marker.c list_for_each_entry_safe(marker, next, &queue->head, head) { next 76 drivers/gpu/drm/vmwgfx/vmwgfx_marker.c struct vmw_marker *marker, *next; next 90 drivers/gpu/drm/vmwgfx/vmwgfx_marker.c list_for_each_entry_safe(marker, next, &queue->head, head) { next 716 drivers/gpu/drm/vmwgfx/vmwgfx_resource.c struct vmw_resource *res, *next; next 723 drivers/gpu/drm/vmwgfx/vmwgfx_resource.c list_for_each_entry_safe(res, next, &vbo->res_list, mob_head) { next 550 drivers/gpu/drm/vmwgfx/vmwgfx_shader.c struct vmw_dx_shader *entry, *next; next 554 drivers/gpu/drm/vmwgfx/vmwgfx_shader.c list_for_each_entry_safe(entry, next, list, cotable_head) { next 436 drivers/gpu/drm/vmwgfx/vmwgfx_so.c struct vmw_view *entry, *next; next 440 drivers/gpu/drm/vmwgfx/vmwgfx_so.c list_for_each_entry_safe(entry, next, list, cotable_head) next 456 drivers/gpu/drm/vmwgfx/vmwgfx_so.c struct vmw_view *entry, *next; next 460 drivers/gpu/drm/vmwgfx/vmwgfx_so.c list_for_each_entry_safe(entry, next, list, srf_head) next 333 drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c viter->next = &__vmw_piter_non_sg_next; next 337 drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c viter->next = &__vmw_piter_non_sg_next; next 343 drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c viter->next = &__vmw_piter_sg_next; next 149 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c struct page *entry, *next; next 151 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c list_for_each_entry_safe(entry, next, &ctx->page_list, lru) { next 60 drivers/gpu/host1x/intr.c struct host1x_waitlist *waiter, *next, *prev; next 62 drivers/gpu/host1x/intr.c list_for_each_entry_safe(waiter, next, head, list) { next 140 drivers/gpu/host1x/intr.c struct host1x_waitlist *waiter, *next; next 142 drivers/gpu/host1x/intr.c list_for_each_entry_safe(waiter, next, head, list) { next 316 drivers/gpu/host1x/intr.c struct host1x_waitlist *waiter, *next; next 318 drivers/gpu/host1x/intr.c list_for_each_entry_safe(waiter, next, next 1541 drivers/gpu/ipu-v3/ipu-image-convert.c run = list_entry(chan->done_q.next, next 1184 drivers/greybus/interface.c struct gb_bundle *next; next 1197 drivers/greybus/interface.c list_for_each_entry_safe(bundle, next, &intf->bundles, links) next 54 drivers/greybus/manifest.c struct manifest_desc *next; next 56 drivers/greybus/manifest.c list_for_each_entry_safe(descriptor, next, &intf->manifest_descs, links) next 79 drivers/greybus/manifest.c struct manifest_desc *next; next 81 drivers/greybus/manifest.c list_for_each_entry_safe(descriptor, next, &intf->manifest_descs, links) next 232 drivers/greybus/manifest.c struct manifest_desc *desc, *next, *tmp; next 240 drivers/greybus/manifest.c list_for_each_entry_safe(desc, next, &intf->manifest_descs, links) { next 286 drivers/greybus/manifest.c list_for_each_entry_safe(desc, next, &list, links) { next 347 drivers/hid/hid-bigbenff.c bigben->report = list_entry(report_list->next, next 985 drivers/hid/hid-core.c hid->report_enum[type].report_list.next, next 1191 drivers/hid/hid-core.c __u8 *next; next 1245 drivers/hid/hid-core.c while ((next = fetch_item(start, end, &item)) != NULL) { next 1246 drivers/hid/hid-core.c start = next; next 198 drivers/hid/hid-cougar.c struct hid_input *next, *hidinput = NULL; next 234 drivers/hid/hid-cougar.c list_for_each_entry_safe(hidinput, next, &hdev->inputs, list) { next 630 drivers/hid/hid-debug.c list = report_enum->report_list.next; next 644 drivers/hid/hid-debug.c list = list->next; next 78 drivers/hid/hid-gaff.c hidinput = list_entry(hid->inputs.next, struct hid_input, list); next 86 drivers/hid/hid-gaff.c report_ptr = report_ptr->next; next 137 drivers/hid/hid-holtekff.c hidinput = list_entry(hid->inputs.next, struct hid_input, list); next 145 drivers/hid/hid-holtekff.c report = list_entry(report_list->next, struct hid_report, list); next 1860 drivers/hid/hid-input.c struct hid_input *next, *hidinput = NULL; next 1924 drivers/hid/hid-input.c list_for_each_entry_safe(hidinput, next, &hid->inputs, list) { next 1961 drivers/hid/hid-input.c struct hid_input *hidinput, *next; next 1965 drivers/hid/hid-input.c list_for_each_entry_safe(hidinput, next, &hid->inputs, list) { next 61 drivers/hid/hid-lg2ff.c hidinput = list_entry(hid->inputs.next, struct hid_input, list); next 53 drivers/hid/hid-lg3ff.c struct hid_report *report = list_entry(report_list->next, struct hid_report, list); next 91 drivers/hid/hid-lg3ff.c struct hid_report *report = list_entry(report_list->next, struct hid_report, list); next 130 drivers/hid/hid-lg3ff.c hidinput = list_entry(hid->inputs.next, struct hid_input, list); next 1259 drivers/hid/hid-lg4ff.c struct hid_report *report = list_entry(report_list->next, struct hid_report, list); next 1273 drivers/hid/hid-lg4ff.c hidinput = list_entry(hid->inputs.next, struct hid_input, list); next 61 drivers/hid/hid-lgff.c struct hid_report *report = list_entry(report_list->next, struct hid_report, list); next 103 drivers/hid/hid-lgff.c struct hid_report *report = list_entry(report_list->next, struct hid_report, list); next 128 drivers/hid/hid-lgff.c hidinput = list_entry(hid->inputs.next, struct hid_input, list); next 2053 drivers/hid/hid-logitech-hidpp.c struct hid_input *hidinput = list_entry(hid->inputs.next, struct hid_input, list); next 2063 drivers/hid/hid-logitech-hidpp.c struct hid_input *hidinput = list_entry(hid->inputs.next, struct hid_input, list); next 2109 drivers/hid/hid-logitech-hidpp.c hidinput = list_entry(hid->inputs.next, struct hid_input, list); next 88 drivers/hid/hid-mf.c input_ptr = input_ptr->next; next 339 drivers/hid/hid-microsoft.c hidinput = list_entry(hdev->inputs.next, struct hid_input, list); next 103 drivers/hid/hid-pl.c report_ptr = report_ptr->next; next 68 drivers/hid/hid-sjoy.c report_ptr = report_ptr->next; next 942 drivers/hid/hid-sony.c struct hid_input *hidinput = list_entry(sc->hdev->inputs.next, next 1828 drivers/hid/hid-sony.c struct hid_report *report = list_entry(report_list->next, next 2281 drivers/hid/hid-sony.c hidinput = list_entry(sc->hdev->inputs.next, struct hid_input, list); next 2430 drivers/hid/hid-sony.c if (sc->list_node.next) { next 112 drivers/hid/hid-steelseries.c struct hid_report *report = list_entry(report_list->next, struct hid_report, list); next 136 drivers/hid/hid-tmff.c hidinput = list_entry(hid->inputs.next, struct hid_input, list); next 65 drivers/hid/hid-zpff.c hidinput = list_entry(hid->inputs.next, struct hid_input, list); next 105 drivers/hid/intel-ish-hid/ishtp/client-buffers.c rb = list_entry(cl->free_rb_list.list.next, struct ishtp_cl_rb, next 115 drivers/hid/intel-ish-hid/ishtp/client-buffers.c rb = list_entry(cl->in_process_list.list.next, next 138 drivers/hid/intel-ish-hid/ishtp/client-buffers.c tx_buf = list_entry(cl->tx_free_list.list.next, next 150 drivers/hid/intel-ish-hid/ishtp/client-buffers.c tx_buf = list_entry(cl->tx_list.list.next, next 44 drivers/hid/intel-ish-hid/ishtp/client.c struct ishtp_cl_rb *next; next 48 drivers/hid/intel-ish-hid/ishtp/client.c list_for_each_entry_safe(rb, next, &cl->dev->read_list.list, list) next 481 drivers/hid/intel-ish-hid/ishtp/client.c rb = list_entry(cl->free_rb_list.list.next, struct ishtp_cl_rb, list); next 676 drivers/hid/intel-ish-hid/ishtp/client.c cl_msg = list_entry(cl->tx_list.list.next, struct ishtp_cl_tx_ring, next 754 drivers/hid/intel-ish-hid/ishtp/client.c cl_msg = list_entry(cl->tx_list.list.next, struct ishtp_cl_tx_ring, next 896 drivers/hid/intel-ish-hid/ishtp/client.c new_rb = list_entry(cl->free_rb_list.list.next, next 1016 drivers/hid/intel-ish-hid/ishtp/client.c new_rb = list_entry(cl->free_rb_list.list.next, next 1229 drivers/hid/usbhid/hid-pidff.c struct hid_input *hidinput = list_entry(hid->inputs.next, next 79 drivers/hid/usbhid/hiddev.c list = report_enum->report_list.next; next 89 drivers/hid/usbhid/hiddev.c list = report->list.next; next 1475 drivers/hid/wacom_sys.c int group, next; next 1481 drivers/hid/wacom_sys.c next = cur->id; next 1484 drivers/hid/wacom_sys.c next_led = wacom_led_find(wacom, group, ++next); next 167 drivers/hsi/clients/cmt_speech.c entry = list_entry(head->next, struct char_queue, list); next 1323 drivers/hsi/clients/cmt_speech.c struct list_head *cursor, *next; next 1326 drivers/hsi/clients/cmt_speech.c list_for_each_safe(cursor, next, head) { next 74 drivers/hv/ring_buffer.c u32 next = ring_info->ring_buffer->write_index; next 76 drivers/hv/ring_buffer.c return next; next 250 drivers/hwmon/ibmpex.c struct ibmpex_bmc_data *p, *next; next 252 drivers/hwmon/ibmpex.c list_for_each_entry_safe(p, next, &driver_data.bmc_data, list) next 582 drivers/hwmon/ibmpex.c struct ibmpex_bmc_data *p, *next; next 585 drivers/hwmon/ibmpex.c list_for_each_entry_safe(p, next, &driver_data.bmc_data, list) next 48 drivers/hwmon/pmbus/pmbus_core.c struct pmbus_sensor *next; next 407 drivers/hwmon/pmbus/pmbus_core.c for (sensor = data->sensors; sensor; sensor = sensor->next) { next 593 drivers/hwmon/pmbus/pmbus_core.c for (sensor = data->sensors; sensor; sensor = sensor->next) { next 1103 drivers/hwmon/pmbus/pmbus_core.c sensor->next = data->sensors; next 699 drivers/hwtracing/coresight/coresight.c struct coresight_node *nd, *next; next 701 drivers/hwtracing/coresight/coresight.c list_for_each_entry_safe(nd, next, path, link) { next 302 drivers/hwtracing/intel_th/msu.c return win->entry.next == &win->msc->win_list; next 506 drivers/hwtracing/stm/policy.c next: next 526 drivers/hwtracing/stm/policy.c goto next; next 64 drivers/i2c/busses/scx200_acb.c struct scx200_acb_iface *next; next 461 drivers/i2c/busses/scx200_acb.c iface->next = scx200_acb_list; next 589 drivers/i2c/busses/scx200_acb.c scx200_acb_list = iface->next; next 1131 drivers/i2c/i2c-core-base.c struct i2c_client *client, *next; next 1151 drivers/i2c/i2c-core-base.c list_for_each_entry_safe(client, next, &adap->userspace_clients, next 1553 drivers/i2c/i2c-core-base.c struct i2c_client *client, *next; next 1574 drivers/i2c/i2c-core-base.c list_for_each_entry_safe(client, next, &adap->userspace_clients, next 775 drivers/iio/adc/ina2xx-adc.c struct timespec64 next, now, delta; next 785 drivers/iio/adc/ina2xx-adc.c ktime_get_ts64(&next); next 798 drivers/iio/adc/ina2xx-adc.c ktime_get_ts64(&next); next 816 drivers/iio/adc/ina2xx-adc.c timespec64_add_ns(&next, 1000 * sampling_us); next 817 drivers/iio/adc/ina2xx-adc.c delta = timespec64_sub(next, now); next 61 drivers/iio/inkern.c struct iio_map_internal *mapi, *next; next 64 drivers/iio/inkern.c list_for_each_entry_safe(mapi, next, &iio_map_list, l) { next 897 drivers/infiniband/core/cm.c work = list_entry(cm_id_priv->work_list.next, struct cm_work, list); next 1753 drivers/infiniband/core/cma.c dev_id_priv = list_entry(id_priv->listen_list.next, next 1822 drivers/infiniband/core/cma.c mc = container_of(id_priv->mc_list.next, next 3439 drivers/infiniband/core/cma.c if (bind_list->owners.first->next) next 4652 drivers/infiniband/core/cma.c id_priv = list_entry(cma_dev->id_list.next, next 237 drivers/infiniband/core/counters.c goto next; next 241 drivers/infiniband/core/counters.c next: next 377 drivers/infiniband/core/counters.c goto next; next 381 drivers/infiniband/core/counters.c next: next 359 drivers/infiniband/core/fmr_pool.c struct ib_pool_fmr *fmr, *next; next 368 drivers/infiniband/core/fmr_pool.c list_for_each_entry_safe(fmr, next, &pool->free_list, list) { next 427 drivers/infiniband/core/fmr_pool.c fmr = list_entry(pool->free_list.next, struct ib_pool_fmr, list); next 147 drivers/infiniband/core/iwcm.c work = list_entry(cm_id_priv->work_free_list.next, struct iwcm_work, next 1028 drivers/infiniband/core/iwcm.c work = list_entry(cm_id_priv->work_list.next, next 1037 drivers/infiniband/core/mad.c send_wr->cur_seg = container_of(send_wr->rmpp_list.next, next 1319 drivers/infiniband/core/mad.c next_send_buf = send_buf->next; next 2398 drivers/infiniband/core/mad.c mad_send_wr = list_entry(mad_agent_priv->wait_list.next, next 2442 drivers/infiniband/core/mad.c if (mad_agent_priv->wait_list.next == &mad_send_wr->agent_list) next 2548 drivers/infiniband/core/mad.c mad_list = container_of(qp_info->overflow_list.next, next 2761 drivers/infiniband/core/mad.c local = list_entry(mad_agent_priv->local_list.next, next 2887 drivers/infiniband/core/mad.c mad_send_wr = list_entry(mad_agent_priv->wait_list.next, next 2939 drivers/infiniband/core/mad.c recv_wr.next = NULL; next 3010 drivers/infiniband/core/mad.c mad_list = list_entry(qp_info->recv_queue.list.next, next 397 drivers/infiniband/core/mad_rmpp.c if (seg->list.next == rmpp_list) next 400 drivers/infiniband/core/mad_rmpp.c return container_of(seg->list.next, struct ib_mad_recv_buf, list); next 402 drivers/infiniband/core/multicast.c member = list_entry(group->active_list.next, next 444 drivers/infiniband/core/multicast.c member = list_entry(group->pending_list.next, next 497 drivers/infiniband/core/multicast.c member = list_entry(group->pending_list.next, next 1345 drivers/infiniband/core/nldev.c goto next; next 1371 drivers/infiniband/core/nldev.c next: idx++; next 75 drivers/infiniband/core/rw.c reg->inv_wr.next = ®->reg_wr.wr; next 78 drivers/infiniband/core/rw.c reg->inv_wr.next = NULL; next 146 drivers/infiniband/core/rw.c prev->wr.wr.next = ®->inv_wr; next 148 drivers/infiniband/core/rw.c prev->wr.wr.next = ®->reg_wr.wr; next 151 drivers/infiniband/core/rw.c reg->reg_wr.wr.next = ®->wr.wr; next 176 drivers/infiniband/core/rw.c prev->wr.wr.next = NULL; next 232 drivers/infiniband/core/rw.c rdma_wr->wr.next = i + 1 < ctx->nr_ops ? next 453 drivers/infiniband/core/rw.c ctx->reg->reg_wr.wr.next = &rdma_wr->wr; next 515 drivers/infiniband/core/rw.c if (ctx->reg[0].inv_wr.next) next 534 drivers/infiniband/core/rw.c last_wr->next = chain_wr; next 881 drivers/infiniband/core/sa_query.c if (ib_nl_request_list.next == &query->list) next 985 drivers/infiniband/core/sa_query.c query = list_entry(ib_nl_request_list.next, next 427 drivers/infiniband/core/ucma.c uevent = list_entry(file->event_list.next, struct ucma_event, list); next 776 drivers/infiniband/core/umem_odp.c struct interval_tree_node *node, *next; next 783 drivers/infiniband/core/umem_odp.c node; node = next) { next 787 drivers/infiniband/core/umem_odp.c next = interval_tree_iter_next(node, start, last - 1); next 334 drivers/infiniband/core/user_mad.c recv_buf = container_of(recv_buf->list.next, next 395 drivers/infiniband/core/user_mad.c packet = list_entry(file->recv_list.next, struct ib_umad_packet, list); next 2004 drivers/infiniband/core/uverbs_cmd.c struct ib_send_wr *wr = NULL, *last, *next; next 2080 drivers/infiniband/core/uverbs_cmd.c next = &ud->wr; next 2096 drivers/infiniband/core/uverbs_cmd.c next = &rdma->wr; next 2113 drivers/infiniband/core/uverbs_cmd.c next = &atomic->wr; next 2117 drivers/infiniband/core/uverbs_cmd.c next_size = sizeof(*next); next 2118 drivers/infiniband/core/uverbs_cmd.c next = alloc_wr(next_size, user_wr->num_sge); next 2119 drivers/infiniband/core/uverbs_cmd.c if (!next) { next 2130 drivers/infiniband/core/uverbs_cmd.c next->ex.imm_data = next 2133 drivers/infiniband/core/uverbs_cmd.c next->ex.invalidate_rkey = user_wr->ex.invalidate_rkey; next 2137 drivers/infiniband/core/uverbs_cmd.c wr = next; next 2139 drivers/infiniband/core/uverbs_cmd.c last->next = next; next 2140 drivers/infiniband/core/uverbs_cmd.c last = next; next 2142 drivers/infiniband/core/uverbs_cmd.c next->next = NULL; next 2143 drivers/infiniband/core/uverbs_cmd.c next->wr_id = user_wr->wr_id; next 2144 drivers/infiniband/core/uverbs_cmd.c next->num_sge = user_wr->num_sge; next 2145 drivers/infiniband/core/uverbs_cmd.c next->opcode = user_wr->opcode; next 2146 drivers/infiniband/core/uverbs_cmd.c next->send_flags = user_wr->send_flags; next 2148 drivers/infiniband/core/uverbs_cmd.c if (next->num_sge) { next 2149 drivers/infiniband/core/uverbs_cmd.c next->sg_list = (void *) next + next 2151 drivers/infiniband/core/uverbs_cmd.c if (copy_from_user(next->sg_list, sgls + sg_ind, next 2152 drivers/infiniband/core/uverbs_cmd.c next->num_sge * next 2157 drivers/infiniband/core/uverbs_cmd.c sg_ind += next->num_sge; next 2159 drivers/infiniband/core/uverbs_cmd.c next->sg_list = NULL; next 2165 drivers/infiniband/core/uverbs_cmd.c for (next = wr; next; next = next->next) { next 2167 drivers/infiniband/core/uverbs_cmd.c if (next == bad_wr) next 2181 drivers/infiniband/core/uverbs_cmd.c next = wr->next; next 2183 drivers/infiniband/core/uverbs_cmd.c wr = next; next 2197 drivers/infiniband/core/uverbs_cmd.c struct ib_recv_wr *wr = NULL, *last, *next; next 2237 drivers/infiniband/core/uverbs_cmd.c (U32_MAX - ALIGN(sizeof *next, sizeof (struct ib_sge))) / next 2243 drivers/infiniband/core/uverbs_cmd.c next = kmalloc(ALIGN(sizeof *next, sizeof (struct ib_sge)) + next 2246 drivers/infiniband/core/uverbs_cmd.c if (!next) { next 2252 drivers/infiniband/core/uverbs_cmd.c wr = next; next 2254 drivers/infiniband/core/uverbs_cmd.c last->next = next; next 2255 drivers/infiniband/core/uverbs_cmd.c last = next; next 2257 drivers/infiniband/core/uverbs_cmd.c next->next = NULL; next 2258 drivers/infiniband/core/uverbs_cmd.c next->wr_id = user_wr->wr_id; next 2259 drivers/infiniband/core/uverbs_cmd.c next->num_sge = user_wr->num_sge; next 2261 drivers/infiniband/core/uverbs_cmd.c if (next->num_sge) { next 2262 drivers/infiniband/core/uverbs_cmd.c next->sg_list = (void *) next + next 2263 drivers/infiniband/core/uverbs_cmd.c ALIGN(sizeof *next, sizeof (struct ib_sge)); next 2264 drivers/infiniband/core/uverbs_cmd.c if (copy_from_user(next->sg_list, sgls + sg_ind, next 2265 drivers/infiniband/core/uverbs_cmd.c next->num_sge * next 2270 drivers/infiniband/core/uverbs_cmd.c sg_ind += next->num_sge; next 2272 drivers/infiniband/core/uverbs_cmd.c next->sg_list = NULL; next 2282 drivers/infiniband/core/uverbs_cmd.c next = wr->next; next 2284 drivers/infiniband/core/uverbs_cmd.c wr = next; next 2294 drivers/infiniband/core/uverbs_cmd.c struct ib_recv_wr *wr, *next; next 2320 drivers/infiniband/core/uverbs_cmd.c for (next = wr; next; next = next->next) { next 2322 drivers/infiniband/core/uverbs_cmd.c if (next == bad_wr) next 2332 drivers/infiniband/core/uverbs_cmd.c next = wr->next; next 2334 drivers/infiniband/core/uverbs_cmd.c wr = next; next 2344 drivers/infiniband/core/uverbs_cmd.c struct ib_recv_wr *wr, *next; next 2371 drivers/infiniband/core/uverbs_cmd.c for (next = wr; next; next = next->next) { next 2373 drivers/infiniband/core/uverbs_cmd.c if (next == bad_wr) next 2383 drivers/infiniband/core/uverbs_cmd.c next = wr->next; next 2385 drivers/infiniband/core/uverbs_cmd.c wr = next; next 39 drivers/infiniband/core/uverbs_ioctl.c struct bundle_alloc_head *next; next 122 drivers/infiniband/core/uverbs_ioctl.c buf->next = pbundle->allocated_mem; next 548 drivers/infiniband/core/uverbs_ioctl.c memblock = memblock->next; next 586 drivers/infiniband/core/uverbs_ioctl.c pbundle->alloc_head.next = NULL; next 252 drivers/infiniband/core/uverbs_main.c event = list_entry(ev_queue->event_list.next, struct ib_uverbs_event, list); next 258 drivers/infiniband/core/uverbs_main.c list_del(ev_queue->event_list.next); next 2124 drivers/infiniband/core/verbs.c fmr = list_entry(fmr_list->next, struct ib_fmr, list); next 2248 drivers/infiniband/core/verbs.c qp = list_entry(xrcd->tgt_qp_list.next, struct ib_qp, xrcd_list); next 2637 drivers/infiniband/core/verbs.c .next = NULL, next 1496 drivers/infiniband/hw/bnxt_re/ib_verbs.c wr = wr->next; next 2316 drivers/infiniband/hw/bnxt_re/ib_verbs.c wr = wr->next; next 2413 drivers/infiniband/hw/bnxt_re/ib_verbs.c wr = wr->next; next 2450 drivers/infiniband/hw/bnxt_re/ib_verbs.c wr = wr->next; next 2503 drivers/infiniband/hw/bnxt_re/ib_verbs.c wr = wr->next; next 2943 drivers/infiniband/hw/bnxt_re/ib_verbs.c rwr.next = NULL; next 2956 drivers/infiniband/hw/bnxt_re/ib_verbs.c swr->next = NULL; next 1709 drivers/infiniband/hw/bnxt_re/main.c struct bnxt_re_dev *rdev, *next; next 1721 drivers/infiniband/hw/bnxt_re/main.c list_for_each_entry_safe_reverse(rdev, next, &to_be_deleted, list) { next 667 drivers/infiniband/hw/bnxt_re/qplib_fp.c int i, rc = 0, next; next 677 drivers/infiniband/hw/bnxt_re/qplib_fp.c next = srq->start_idx; next 678 drivers/infiniband/hw/bnxt_re/qplib_fp.c srq->start_idx = srq->swq[next].next_idx; next 696 drivers/infiniband/hw/bnxt_re/qplib_fp.c srqe->wr_id[0] = cpu_to_le32((u32)next); next 697 drivers/infiniband/hw/bnxt_re/qplib_fp.c srq->swq[next].wr_id = wqe->wr_id; next 197 drivers/infiniband/hw/cxgb3/cxio_hal.c entry = list_entry(uctx->qpids.next, struct cxio_qpid_list, next 937 drivers/infiniband/hw/cxgb3/iwch_provider.c char *cp, *next; next 942 drivers/infiniband/hw/cxgb3/iwch_provider.c next = info.fw_version + 1; next 943 drivers/infiniband/hw/cxgb3/iwch_provider.c cp = strsep(&next, "."); next 945 drivers/infiniband/hw/cxgb3/iwch_provider.c cp = strsep(&next, "."); next 947 drivers/infiniband/hw/cxgb3/iwch_provider.c cp = strsep(&next, "."); next 451 drivers/infiniband/hw/cxgb3/iwch_qp.c wr = wr->next; next 516 drivers/infiniband/hw/cxgb3/iwch_qp.c wr = wr->next; next 4278 drivers/infiniband/hw/cxgb4/cm.c tmp = timeout_list.next; next 4280 drivers/infiniband/hw/cxgb4/cm.c tmp->next = NULL; next 4329 drivers/infiniband/hw/cxgb4/cm.c if (!ep->entry.next) { next 635 drivers/infiniband/hw/cxgb4/qp.c if (wr->next->opcode == IB_WR_SEND) next 638 drivers/infiniband/hw/cxgb4/qp.c wcwr->stag_inv = cpu_to_be32(wr->next->ex.invalidate_rkey); next 643 drivers/infiniband/hw/cxgb4/qp.c if (wr->next->send_flags & IB_SEND_INLINE) next 644 drivers/infiniband/hw/cxgb4/qp.c build_immd_cmpl(sq, &wcwr->u_cmpl.immd_src, wr->next); next 647 drivers/infiniband/hw/cxgb4/qp.c &wcwr->u_cmpl.isgl_src, wr->next->sg_list, 1, NULL); next 691 drivers/infiniband/hw/cxgb4/qp.c bool send_signaled = (wr->next->send_flags & IB_SEND_SIGNALED) || next 732 drivers/infiniband/hw/cxgb4/qp.c if (wr->next->opcode == IB_WR_SEND) next 740 drivers/infiniband/hw/cxgb4/qp.c swsqe->wr_id = wr->next->wr_id; next 1029 drivers/infiniband/hw/cxgb4/qp.c wr = wr->next; next 1072 drivers/infiniband/hw/cxgb4/qp.c wr = wr->next; next 1123 drivers/infiniband/hw/cxgb4/qp.c wr && wr->next && !wr->next->next && next 1126 drivers/infiniband/hw/cxgb4/qp.c (wr->next->opcode == IB_WR_SEND || next 1127 drivers/infiniband/hw/cxgb4/qp.c wr->next->opcode == IB_WR_SEND_WITH_INV) && next 1128 drivers/infiniband/hw/cxgb4/qp.c wr->next->sg_list[0].length == T4_WRITE_CMPL_MAX_CQE && next 1129 drivers/infiniband/hw/cxgb4/qp.c wr->next->num_sge == 1 && num_wrs >= 2) { next 1245 drivers/infiniband/hw/cxgb4/qp.c wr = wr->next; next 1327 drivers/infiniband/hw/cxgb4/qp.c wr = wr->next; next 1414 drivers/infiniband/hw/cxgb4/qp.c wr = wr->next; next 116 drivers/infiniband/hw/cxgb4/resource.c entry = list_entry(uctx->cqids.next, struct c4iw_qid_list, next 186 drivers/infiniband/hw/cxgb4/resource.c entry = list_entry(uctx->qpids.next, struct c4iw_qid_list, next 55 drivers/infiniband/hw/hfi1/debugfs.h .next = _##name##_seq_next, \ next 637 drivers/infiniband/hw/hfi1/driver.c goto next; next 640 drivers/infiniband/hw/hfi1/driver.c goto next; next 653 drivers/infiniband/hw/hfi1/driver.c goto next; /* just in case */ next 657 drivers/infiniband/hw/hfi1/driver.c goto next; next 666 drivers/infiniband/hw/hfi1/driver.c goto next; next 675 drivers/infiniband/hw/hfi1/driver.c next: next 1776 drivers/infiniband/hw/hfi1/driver.c goto next; next 1779 drivers/infiniband/hw/hfi1/driver.c goto next; next 1791 drivers/infiniband/hw/hfi1/driver.c goto next; /* just in case */ next 1799 drivers/infiniband/hw/hfi1/driver.c next: next 1458 drivers/infiniband/hw/hfi1/pio.c u32 head, next; next 1518 drivers/infiniband/hw/hfi1/pio.c next = head + 1; next 1519 drivers/infiniband/hw/hfi1/pio.c if (next >= sc->sr_size) next 1520 drivers/infiniband/hw/hfi1/pio.c next = 0; next 1526 drivers/infiniband/hw/hfi1/pio.c sc->sr_head = next; next 123 drivers/infiniband/hw/hfi1/rc.c u8 next = qp->s_tail_ack_queue; next 151 drivers/infiniband/hw/hfi1/rc.c if (++next > rvt_size_atomic(&dev->rdi)) next 152 drivers/infiniband/hw/hfi1/rc.c next = 0; next 160 drivers/infiniband/hw/hfi1/rc.c qp->s_acked_ack_queue = next; next 161 drivers/infiniband/hw/hfi1/rc.c qp->s_tail_ack_queue = next; next 3060 drivers/infiniband/hw/hfi1/rc.c u8 next; next 3064 drivers/infiniband/hw/hfi1/rc.c next = qp->r_head_ack_queue + 1; next 3066 drivers/infiniband/hw/hfi1/rc.c if (next > rvt_size_atomic(ib_to_rvt(qp->ibqp.device))) next 3067 drivers/infiniband/hw/hfi1/rc.c next = 0; next 3069 drivers/infiniband/hw/hfi1/rc.c if (unlikely(next == qp->s_acked_ack_queue)) { next 3070 drivers/infiniband/hw/hfi1/rc.c if (!qp->s_ack_queue[next].sent) next 3072 drivers/infiniband/hw/hfi1/rc.c update_ack_queue(qp, next); next 3112 drivers/infiniband/hw/hfi1/rc.c qp->r_head_ack_queue = next; next 3135 drivers/infiniband/hw/hfi1/rc.c u8 next; next 3140 drivers/infiniband/hw/hfi1/rc.c next = qp->r_head_ack_queue + 1; next 3141 drivers/infiniband/hw/hfi1/rc.c if (next > rvt_size_atomic(ib_to_rvt(qp->ibqp.device))) next 3142 drivers/infiniband/hw/hfi1/rc.c next = 0; next 3144 drivers/infiniband/hw/hfi1/rc.c if (unlikely(next == qp->s_acked_ack_queue)) { next 3145 drivers/infiniband/hw/hfi1/rc.c if (!qp->s_ack_queue[next].sent) next 3147 drivers/infiniband/hw/hfi1/rc.c update_ack_queue(qp, next); next 3183 drivers/infiniband/hw/hfi1/rc.c qp->r_head_ack_queue = next; next 15 drivers/infiniband/hw/hfi1/rc.h unsigned int next; next 17 drivers/infiniband/hw/hfi1/rc.h next = n + 1; next 18 drivers/infiniband/hw/hfi1/rc.h if (next > rvt_size_atomic(ib_to_rvt(qp->ibqp.device))) next 19 drivers/infiniband/hw/hfi1/rc.h next = 0; next 20 drivers/infiniband/hw/hfi1/rc.h qp->s_tail_ack_queue = next; next 21 drivers/infiniband/hw/hfi1/rc.h qp->s_acked_ack_queue = next; next 2233 drivers/infiniband/hw/hfi1/tid_rdma.c u8 next; next 2266 drivers/infiniband/hw/hfi1/tid_rdma.c next = qp->r_head_ack_queue + 1; next 2267 drivers/infiniband/hw/hfi1/tid_rdma.c if (next > rvt_size_atomic(ib_to_rvt(qp->ibqp.device))) next 2268 drivers/infiniband/hw/hfi1/tid_rdma.c next = 0; next 2270 drivers/infiniband/hw/hfi1/tid_rdma.c if (unlikely(next == qp->s_tail_ack_queue)) { next 2271 drivers/infiniband/hw/hfi1/tid_rdma.c if (!qp->s_ack_queue[next].sent) { next 2275 drivers/infiniband/hw/hfi1/tid_rdma.c update_ack_queue(qp, next); next 2302 drivers/infiniband/hw/hfi1/tid_rdma.c qp->r_head_ack_queue = next; next 3657 drivers/infiniband/hw/hfi1/tid_rdma.c u8 next; next 3697 drivers/infiniband/hw/hfi1/tid_rdma.c next = qp->r_head_ack_queue + 1; next 3698 drivers/infiniband/hw/hfi1/tid_rdma.c if (next > rvt_size_atomic(ib_to_rvt(qp->ibqp.device))) next 3699 drivers/infiniband/hw/hfi1/tid_rdma.c next = 0; next 3701 drivers/infiniband/hw/hfi1/tid_rdma.c if (unlikely(next == qp->s_acked_ack_queue)) { next 3702 drivers/infiniband/hw/hfi1/tid_rdma.c if (!qp->s_ack_queue[next].sent) next 3704 drivers/infiniband/hw/hfi1/tid_rdma.c update_ack_queue(qp, next); next 3791 drivers/infiniband/hw/hfi1/tid_rdma.c qp->r_head_ack_queue = next; next 4267 drivers/infiniband/hw/hfi1/tid_rdma.c u32 psn, next; next 4355 drivers/infiniband/hw/hfi1/tid_rdma.c for (next = priv->r_tid_tail + 1; ; next++) { next 4356 drivers/infiniband/hw/hfi1/tid_rdma.c if (next > rvt_size_atomic(&dev->rdi)) next 4357 drivers/infiniband/hw/hfi1/tid_rdma.c next = 0; next 4358 drivers/infiniband/hw/hfi1/tid_rdma.c if (next == priv->r_tid_head) next 4360 drivers/infiniband/hw/hfi1/tid_rdma.c e = &qp->s_ack_queue[next]; next 4364 drivers/infiniband/hw/hfi1/tid_rdma.c priv->r_tid_tail = next; next 4932 drivers/infiniband/hw/hfi1/tid_rdma.c u32 next; next 4937 drivers/infiniband/hw/hfi1/tid_rdma.c next = flow->flow_state.r_next_psn; next 4938 drivers/infiniband/hw/hfi1/tid_rdma.c flow->npkts = delta_psn(lpsn, next - 1); next 5180 drivers/infiniband/hw/hfi1/tid_rdma.c u32 hwords, next; next 5253 drivers/infiniband/hw/hfi1/tid_rdma.c next = qpriv->r_tid_ack + 1; next 5254 drivers/infiniband/hw/hfi1/tid_rdma.c if (next > rvt_size_atomic(&dev->rdi)) next 5255 drivers/infiniband/hw/hfi1/tid_rdma.c next = 0; next 5256 drivers/infiniband/hw/hfi1/tid_rdma.c qpriv->r_tid_ack = next; next 5257 drivers/infiniband/hw/hfi1/tid_rdma.c if (qp->s_ack_queue[next].opcode != TID_OP(WRITE_REQ)) next 5259 drivers/infiniband/hw/hfi1/tid_rdma.c nreq = ack_to_tid_req(&qp->s_ack_queue[next]); next 318 drivers/infiniband/hw/hfi1/user_sdma.c static u8 next; next 328 drivers/infiniband/hw/hfi1/user_sdma.c mapping[hash] = next; next 329 drivers/infiniband/hw/hfi1/user_sdma.c next = (next + 1) & 0x7F; next 122 drivers/infiniband/hw/hns/hns_roce_cmd.c cmd->free_head = context->next; next 153 drivers/infiniband/hw/hns/hns_roce_cmd.c context->next = cmd->free_head; next 242 drivers/infiniband/hw/hns/hns_roce_cmd.c hr_cmd->context[i].next = i + 1; next 245 drivers/infiniband/hw/hns/hns_roce_cmd.c hr_cmd->context[hr_cmd->max_cmds - 1].next = -1; next 597 drivers/infiniband/hw/hns/hns_roce_device.h int next; next 155 drivers/infiniband/hw/hns/hns_roce_hem.h list_entry(hem->chunk_list.next, next 168 drivers/infiniband/hw/hns/hns_roce_hem.h if (iter->chunk->list.next == &iter->hem->chunk_list) { next 173 drivers/infiniband/hw/hns/hns_roce_hem.h iter->chunk = list_entry(iter->chunk->list.next, next 92 drivers/infiniband/hw/hns/hns_roce_hw_v1.c for (nreq = 0; wr; ++nreq, wr = wr->next) { next 365 drivers/infiniband/hw/hns/hns_roce_hw_v1.c for (nreq = 0; wr; ++nreq, wr = wr->next) { next 1011 drivers/infiniband/hw/hns/hns_roce_hw_v1.c send_wr.next = NULL; next 279 drivers/infiniband/hw/hns/hns_roce_hw_v2.c for (nreq = 0; wr; ++nreq, wr = wr->next) { next 643 drivers/infiniband/hw/hns/hns_roce_hw_v2.c for (nreq = 0; wr; ++nreq, wr = wr->next) { next 6249 drivers/infiniband/hw/hns/hns_roce_hw_v2.c for (nreq = 0; wr; ++nreq, wr = wr->next) { next 434 drivers/infiniband/hw/i40iw/i40iw.h u32 *next) next 440 drivers/infiniband/hw/i40iw/i40iw.h resource_num = find_next_zero_bit(resource_array, max_resources, *next); next 449 drivers/infiniband/hw/i40iw/i40iw.h *next = resource_num + 1; next 450 drivers/infiniband/hw/i40iw/i40iw.h if (*next == max_resources) next 451 drivers/infiniband/hw/i40iw/i40iw.h *next = 0; next 328 drivers/infiniband/hw/i40iw/i40iw_ctrl.c entry = head->next; next 331 drivers/infiniband/hw/i40iw/i40iw_ctrl.c entry = (lastentry != head) ? lastentry->next : NULL; next 60 drivers/infiniband/hw/i40iw/i40iw_puda.c buf = (struct i40iw_puda_buf *)list->next; next 837 drivers/infiniband/hw/i40iw/i40iw_puda.c nextbuf = buf->next; next 870 drivers/infiniband/hw/i40iw/i40iw_puda.c buf->next = nextbuf; next 1417 drivers/infiniband/hw/i40iw/i40iw_puda.c tmpbuf = (struct i40iw_puda_buf *)rxlist->next; next 1422 drivers/infiniband/hw/i40iw/i40iw_puda.c tmpbuf = (struct i40iw_puda_buf *)plist->next; next 82 drivers/infiniband/hw/i40iw/i40iw_puda.h struct i40iw_puda_buf *next; /* for alloclist in rsrc struct */ next 369 drivers/infiniband/hw/i40iw/i40iw_utils.c cqp_request = list_entry(cqp->cqp_avail_reqs.next, next 739 drivers/infiniband/hw/i40iw/i40iw_utils.c entry = (void *)list->next; next 2273 drivers/infiniband/hw/i40iw/i40iw_verbs.c ib_wr = ib_wr->next; next 2329 drivers/infiniband/hw/i40iw/i40iw_verbs.c ib_wr = ib_wr->next; next 811 drivers/infiniband/hw/mlx4/alias_GUID.c cb_ctx = list_entry(det->cb_list.next, next 644 drivers/infiniband/hw/mlx4/mad.c wr.wr.next = NULL; next 1326 drivers/infiniband/hw/mlx4/mad.c recv_wr.next = NULL; next 1444 drivers/infiniband/hw/mlx4/mad.c wr.wr.next = NULL; next 3583 drivers/infiniband/hw/mlx4/qp.c for (nreq = 0; wr; ++nreq, wr = wr->next) { next 3821 drivers/infiniband/hw/mlx4/qp.c if (wr->next) next 3881 drivers/infiniband/hw/mlx4/qp.c for (nreq = 0; wr; ++nreq, wr = wr->next) { next 4485 drivers/infiniband/hw/mlx4/qp.c .next = NULL, next 80 drivers/infiniband/hw/mlx4/srq.c struct mlx4_wqe_srq_next_seg *next; next 147 drivers/infiniband/hw/mlx4/srq.c next = get_wqe(srq, i); next 148 drivers/infiniband/hw/mlx4/srq.c next->next_wqe_index = next 151 drivers/infiniband/hw/mlx4/srq.c for (scatter = (void *) (next + 1); next 152 drivers/infiniband/hw/mlx4/srq.c (void *) scatter < (void *) next + desc_size; next 288 drivers/infiniband/hw/mlx4/srq.c struct mlx4_wqe_srq_next_seg *next; next 293 drivers/infiniband/hw/mlx4/srq.c next = get_wqe(srq, srq->tail); next 294 drivers/infiniband/hw/mlx4/srq.c next->next_wqe_index = cpu_to_be16(wqe_index); next 304 drivers/infiniband/hw/mlx4/srq.c struct mlx4_wqe_srq_next_seg *next; next 320 drivers/infiniband/hw/mlx4/srq.c for (nreq = 0; wr; ++nreq, wr = wr->next) { next 335 drivers/infiniband/hw/mlx4/srq.c next = get_wqe(srq, srq->head); next 336 drivers/infiniband/hw/mlx4/srq.c srq->head = be16_to_cpu(next->next_wqe_index); next 337 drivers/infiniband/hw/mlx4/srq.c scat = (struct mlx4_wqe_data_seg *) (next + 1); next 343 drivers/infiniband/hw/mlx5/cq.c tail = qp->sq.w_list[idx].next; next 345 drivers/infiniband/hw/mlx5/cq.c tail = qp->sq.w_list[idx].next; next 411 drivers/infiniband/hw/mlx5/cq.c wq->last_poll = wq->w_list[idx].next; next 574 drivers/infiniband/hw/mlx5/cq.c struct mlx5_ib_wc *soft_wc, *next; next 577 drivers/infiniband/hw/mlx5/cq.c list_for_each_entry_safe(soft_wc, next, &cq->wc_list, list) { next 2476 drivers/infiniband/hw/mlx5/devx.c event = list_entry(ev_queue->event_list.next, next 2486 drivers/infiniband/hw/mlx5/devx.c list_del(ev_queue->event_list.next); next 488 drivers/infiniband/hw/mlx5/gsi.c for (; wr; wr = wr->next) { next 491 drivers/infiniband/hw/mlx5/gsi.c cur_wr.wr.next = NULL; next 259 drivers/infiniband/hw/mlx5/mlx5_ib.h u16 next; next 698 drivers/infiniband/hw/mlx5/odp.c struct ib_umem_odp *next; next 701 drivers/infiniband/hw/mlx5/odp.c next = odp_next(odp); next 702 drivers/infiniband/hw/mlx5/odp.c if (unlikely(!next || ib_umem_start(next) != io_virt)) { next 704 drivers/infiniband/hw/mlx5/odp.c io_virt, next); next 707 drivers/infiniband/hw/mlx5/odp.c odp = next; next 732 drivers/infiniband/hw/mlx5/odp.c struct pf_frame *next; next 896 drivers/infiniband/hw/mlx5/odp.c frame->next = head; next 912 drivers/infiniband/hw/mlx5/odp.c head = frame->next; next 926 drivers/infiniband/hw/mlx5/odp.c head = frame->next; next 4950 drivers/infiniband/hw/mlx5/qp.c qp->sq.w_list[idx].next = qp->sq.cur_post; next 5002 drivers/infiniband/hw/mlx5/qp.c for (nreq = 0; wr; nreq++, wr = wr->next) { next 5382 drivers/infiniband/hw/mlx5/qp.c for (nreq = 0; wr; nreq++, wr = wr->next) { next 6447 drivers/infiniband/hw/mlx5/qp.c .next = NULL, next 136 drivers/infiniband/hw/mlx5/srq.c struct mlx5_wqe_srq_next_seg *next; next 159 drivers/infiniband/hw/mlx5/srq.c next = get_wqe(srq, i); next 160 drivers/infiniband/hw/mlx5/srq.c next->next_wqe_index = next 408 drivers/infiniband/hw/mlx5/srq.c struct mlx5_wqe_srq_next_seg *next; next 413 drivers/infiniband/hw/mlx5/srq.c next = get_wqe(srq, srq->tail); next 414 drivers/infiniband/hw/mlx5/srq.c next->next_wqe_index = cpu_to_be16(wqe_index); next 424 drivers/infiniband/hw/mlx5/srq.c struct mlx5_wqe_srq_next_seg *next; next 441 drivers/infiniband/hw/mlx5/srq.c for (nreq = 0; wr; nreq++, wr = wr->next) { next 456 drivers/infiniband/hw/mlx5/srq.c next = get_wqe(srq, srq->head); next 457 drivers/infiniband/hw/mlx5/srq.c srq->head = be16_to_cpu(next->next_wqe_index); next 458 drivers/infiniband/hw/mlx5/srq.c scat = (struct mlx5_wqe_data_seg *)(next + 1); next 181 drivers/infiniband/hw/mthca/mthca_cmd.c int next; next 424 drivers/infiniband/hw/mthca/mthca_cmd.c dev->cmd.free_head = context->next; next 460 drivers/infiniband/hw/mthca/mthca_cmd.c context->next = dev->cmd.free_head; next 570 drivers/infiniband/hw/mthca/mthca_cmd.c dev->cmd.context[i].next = i + 1; next 573 drivers/infiniband/hw/mthca/mthca_cmd.c dev->cmd.context[dev->cmd.max_cmds - 1].next = -1; next 1728 drivers/infiniband/hw/mthca/mthca_cmd.c enum ib_qp_state next, u32 num, int is_ee, next 1775 drivers/infiniband/hw/mthca/mthca_cmd.c if (op[cur][next] == CMD_ERR2RST_QPEE) { next 1790 drivers/infiniband/hw/mthca/mthca_cmd.c op[cur][next], CMD_TIME_CLASS_C); next 1824 drivers/infiniband/hw/mthca/mthca_cmd.c op_mod, op[cur][next], CMD_TIME_CLASS_C); next 309 drivers/infiniband/hw/mthca/mthca_cmd.h enum ib_qp_state next, u32 num, int is_ee, next 104 drivers/infiniband/hw/mthca/mthca_memfree.h NULL : list_entry(icm->chunk_list.next, next 117 drivers/infiniband/hw/mthca/mthca_memfree.h if (iter->chunk->list.next == &iter->icm->chunk_list) { next 122 drivers/infiniband/hw/mthca/mthca_memfree.h iter->chunk = list_entry(iter->chunk->list.next, next 1169 drivers/infiniband/hw/mthca/mthca_qp.c struct mthca_next_seg *next; next 1217 drivers/infiniband/hw/mthca/mthca_qp.c next = get_recv_wqe(qp, i); next 1218 drivers/infiniband/hw/mthca/mthca_qp.c next->nda_op = cpu_to_be32(((i + 1) & (qp->rq.max - 1)) << next 1220 drivers/infiniband/hw/mthca/mthca_qp.c next->ee_nds = cpu_to_be32(size); next 1222 drivers/infiniband/hw/mthca/mthca_qp.c for (scatter = (void *) (next + 1); next 1223 drivers/infiniband/hw/mthca/mthca_qp.c (void *) scatter < (void *) next + (1 << qp->rq.wqe_shift); next 1229 drivers/infiniband/hw/mthca/mthca_qp.c next = get_send_wqe(qp, i); next 1230 drivers/infiniband/hw/mthca/mthca_qp.c next->nda_op = cpu_to_be32((((i + 1) & (qp->sq.max - 1)) << next 1236 drivers/infiniband/hw/mthca/mthca_qp.c next = get_recv_wqe(qp, i); next 1237 drivers/infiniband/hw/mthca/mthca_qp.c next->nda_op = htonl((((i + 1) % qp->rq.max) << next 1653 drivers/infiniband/hw/mthca/mthca_qp.c for (nreq = 0; wr; ++nreq, wr = wr->next) { next 1849 drivers/infiniband/hw/mthca/mthca_qp.c for (nreq = 0; wr; wr = wr->next) { next 1957 drivers/infiniband/hw/mthca/mthca_qp.c for (nreq = 0; wr; ++nreq, wr = wr->next) { next 2178 drivers/infiniband/hw/mthca/mthca_qp.c for (nreq = 0; wr; ++nreq, wr = wr->next) { next 2234 drivers/infiniband/hw/mthca/mthca_qp.c struct mthca_next_seg *next; next 2246 drivers/infiniband/hw/mthca/mthca_qp.c next = get_send_wqe(qp, index); next 2248 drivers/infiniband/hw/mthca/mthca_qp.c next = get_recv_wqe(qp, index); next 2250 drivers/infiniband/hw/mthca/mthca_qp.c *dbd = !!(next->ee_nds & cpu_to_be32(MTHCA_NEXT_DBD)); next 2251 drivers/infiniband/hw/mthca/mthca_qp.c if (next->ee_nds & cpu_to_be32(0x3f)) next 2252 drivers/infiniband/hw/mthca/mthca_qp.c *new_wqe = (next->nda_op & cpu_to_be32(~0x3f)) | next 2253 drivers/infiniband/hw/mthca/mthca_qp.c (next->ee_nds & cpu_to_be32(0x3f)); next 183 drivers/infiniband/hw/mthca/mthca_srq.c struct mthca_next_seg *next; next 185 drivers/infiniband/hw/mthca/mthca_srq.c next = wqe = get_wqe(srq, i); next 189 drivers/infiniband/hw/mthca/mthca_srq.c next->nda_op = htonl(((i + 1) << srq->wqe_shift) | 1); next 192 drivers/infiniband/hw/mthca/mthca_srq.c next->nda_op = 0; next 502 drivers/infiniband/hw/mthca/mthca_srq.c for (nreq = 0; wr; wr = wr->next) { next 592 drivers/infiniband/hw/mthca/mthca_srq.c for (nreq = 0; wr; ++nreq, wr = wr->next) { next 2200 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c wr = wr->next; next 2267 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c wr = wr->next; next 2334 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c wr = wr->next; next 599 drivers/infiniband/hw/qedr/qedr_roce_cm.c if (wr->next) { next 602 drivers/infiniband/hw/qedr/qedr_roce_cm.c *bad_wr = wr->next; next 660 drivers/infiniband/hw/qedr/qedr_roce_cm.c wr = wr->next; next 3431 drivers/infiniband/hw/qedr/verbs.c wr = wr->next; next 3541 drivers/infiniband/hw/qedr/verbs.c wr = wr->next; next 3644 drivers/infiniband/hw/qedr/verbs.c wr = wr->next; next 46 drivers/infiniband/hw/qib/qib_debugfs.c .next = _##name##_seq_next, \ next 68 drivers/infiniband/hw/qib/qib_diag.c struct qib_diag_client *next; next 85 drivers/infiniband/hw/qib/qib_diag.c client_pool = dc->next; next 91 drivers/infiniband/hw/qib/qib_diag.c dc->next = NULL; next 109 drivers/infiniband/hw/qib/qib_diag.c dd->diag_client = dc->next; next 114 drivers/infiniband/hw/qib/qib_diag.c if (dc == tdc->next) { next 115 drivers/infiniband/hw/qib/qib_diag.c tdc->next = dc->next; next 119 drivers/infiniband/hw/qib/qib_diag.c tdc = tdc->next; next 126 drivers/infiniband/hw/qib/qib_diag.c rdc->next = client_pool; next 202 drivers/infiniband/hw/qib/qib_diag.c client_pool = dc->next; next 530 drivers/infiniband/hw/qib/qib_diag.c dc->next = dd->diag_client; next 690 drivers/infiniband/hw/qib/qib_diag.c struct diag_observer_list_elt *next; next 708 drivers/infiniband/hw/qib/qib_diag.c olp->next = dd->diag_observer_list; next 725 drivers/infiniband/hw/qib/qib_diag.c dd->diag_observer_list = olp->next; next 751 drivers/infiniband/hw/qib/qib_diag.c olp = olp->next; next 1189 drivers/infiniband/hw/qib/qib_init.c .next = NULL, next 253 drivers/infiniband/hw/qib/qib_qsfp.c static int qsfp_cks(struct qib_pportdata *ppd, int first, int next) next 260 drivers/infiniband/hw/qib/qib_qsfp.c while (first < next) { next 1689 drivers/infiniband/hw/qib/qib_rc.c unsigned next; next 1691 drivers/infiniband/hw/qib/qib_rc.c next = n + 1; next 1692 drivers/infiniband/hw/qib/qib_rc.c if (next > QIB_MAX_RDMA_ATOMIC) next 1693 drivers/infiniband/hw/qib/qib_rc.c next = 0; next 1694 drivers/infiniband/hw/qib/qib_rc.c qp->s_tail_ack_queue = next; next 1939 drivers/infiniband/hw/qib/qib_rc.c u8 next; next 1943 drivers/infiniband/hw/qib/qib_rc.c next = qp->r_head_ack_queue + 1; next 1945 drivers/infiniband/hw/qib/qib_rc.c if (next > QIB_MAX_RDMA_ATOMIC) next 1946 drivers/infiniband/hw/qib/qib_rc.c next = 0; next 1948 drivers/infiniband/hw/qib/qib_rc.c if (unlikely(next == qp->s_tail_ack_queue)) { next 1949 drivers/infiniband/hw/qib/qib_rc.c if (!qp->s_ack_queue[next].sent) next 1951 drivers/infiniband/hw/qib/qib_rc.c qib_update_ack_queue(qp, next); next 1994 drivers/infiniband/hw/qib/qib_rc.c qp->r_head_ack_queue = next; next 2011 drivers/infiniband/hw/qib/qib_rc.c u8 next; next 2015 drivers/infiniband/hw/qib/qib_rc.c next = qp->r_head_ack_queue + 1; next 2016 drivers/infiniband/hw/qib/qib_rc.c if (next > QIB_MAX_RDMA_ATOMIC) next 2017 drivers/infiniband/hw/qib/qib_rc.c next = 0; next 2019 drivers/infiniband/hw/qib/qib_rc.c if (unlikely(next == qp->s_tail_ack_queue)) { next 2020 drivers/infiniband/hw/qib/qib_rc.c if (!qp->s_ack_queue[next].sent) next 2022 drivers/infiniband/hw/qib/qib_rc.c qib_update_ack_queue(qp, next); next 2057 drivers/infiniband/hw/qib/qib_rc.c qp->r_head_ack_queue = next; next 342 drivers/infiniband/hw/qib/qib_sdma.c lp = ppd->sdma_activelist.next; next 373 drivers/infiniband/hw/qib/qib_sdma.c lp = ppd->sdma_activelist.next; next 1253 drivers/infiniband/hw/qib/qib_user_sdma.c list_entry(pktlist->next, struct qib_user_sdma_pkt, next 373 drivers/infiniband/hw/qib/qib_verbs.c priv = list_entry(list->next, struct qib_qp_priv, iowait); next 571 drivers/infiniband/hw/qib/qib_verbs.c struct list_head *l = dev->txreq_free.next; next 601 drivers/infiniband/hw/qib/qib_verbs.c struct list_head *l = dev->txreq_free.next; next 643 drivers/infiniband/hw/qib/qib_verbs.c priv = list_entry(dev->txwait.next, struct qib_qp_priv, next 1167 drivers/infiniband/hw/qib/qib_verbs.c priv = list_entry(list->next, struct qib_qp_priv, iowait); next 1629 drivers/infiniband/hw/qib/qib_verbs.c struct list_head *l = dev->txreq_free.next; next 1665 drivers/infiniband/hw/qib/qib_verbs.c struct list_head *l = dev->txreq_free.next; next 806 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c wr = wr->next; next 897 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c wr = wr->next; next 75 drivers/infiniband/sw/rdmavt/cq.c u32 next; next 98 drivers/infiniband/sw/rdmavt/cq.c next = 0; next 100 drivers/infiniband/sw/rdmavt/cq.c next = head + 1; next 103 drivers/infiniband/sw/rdmavt/cq.c if (unlikely(next == tail || cq->cq_full)) { next 137 drivers/infiniband/sw/rdmavt/cq.c RDMA_WRITE_UAPI_ATOMIC(u_wc->head, next); next 140 drivers/infiniband/sw/rdmavt/cq.c k_wc->head = next; next 205 drivers/infiniband/sw/rdmavt/mr.c r = rkt->next; next 214 drivers/infiniband/sw/rdmavt/mr.c rkt->next = (r + 1) & (rkt->max - 1); next 791 drivers/infiniband/sw/rdmavt/qp.c qpp = &q->next) { next 794 drivers/infiniband/sw/rdmavt/qp.c rcu_dereference_protected(qp->next, next 1134 drivers/infiniband/sw/rdmavt/qp.c RCU_INIT_POINTER(qp->next, NULL); next 1429 drivers/infiniband/sw/rdmavt/qp.c qp->next = rdi->qp_dev->qp_table[n]; next 1836 drivers/infiniband/sw/rdmavt/qp.c for (; wr; wr = wr->next) { next 1838 drivers/infiniband/sw/rdmavt/qp.c u32 next; next 1847 drivers/infiniband/sw/rdmavt/qp.c next = wq->head + 1; next 1848 drivers/infiniband/sw/rdmavt/qp.c if (next >= qp->r_rq.size) next 1849 drivers/infiniband/sw/rdmavt/qp.c next = 0; next 1850 drivers/infiniband/sw/rdmavt/qp.c if (next == READ_ONCE(wq->tail)) { next 1877 drivers/infiniband/sw/rdmavt/qp.c smp_store_release(&wq->head, next); next 2004 drivers/infiniband/sw/rdmavt/qp.c u32 next; next 2071 drivers/infiniband/sw/rdmavt/qp.c next = qp->s_head + 1; next 2072 drivers/infiniband/sw/rdmavt/qp.c if (next >= qp->s_size) next 2073 drivers/infiniband/sw/rdmavt/qp.c next = 0; next 2155 drivers/infiniband/sw/rdmavt/qp.c qp->s_head = next; next 2208 drivers/infiniband/sw/rdmavt/qp.c call_send = qp->s_head == READ_ONCE(qp->s_last) && !wr->next; next 2210 drivers/infiniband/sw/rdmavt/qp.c for (; wr; wr = wr->next) { next 2250 drivers/infiniband/sw/rdmavt/qp.c for (; wr; wr = wr->next) { next 2252 drivers/infiniband/sw/rdmavt/qp.c u32 next; next 2262 drivers/infiniband/sw/rdmavt/qp.c next = wq->head + 1; next 2263 drivers/infiniband/sw/rdmavt/qp.c if (next >= srq->rq.size) next 2264 drivers/infiniband/sw/rdmavt/qp.c next = 0; next 2265 drivers/infiniband/sw/rdmavt/qp.c if (next == READ_ONCE(wq->tail)) { next 2280 drivers/infiniband/sw/rdmavt/qp.c smp_store_release(&wq->head, next); next 2768 drivers/infiniband/sw/rdmavt/qp.c qp = rcu_dereference(pqp->next); next 71 drivers/infiniband/sw/rxe/rxe.h u32 crc, void *next, size_t len) next 80 drivers/infiniband/sw/rxe/rxe.h err = crypto_shash_update(shash, next, len); next 83 drivers/infiniband/sw/rxe/rxe.h return crc32_le(crc, next, len); next 315 drivers/infiniband/sw/rxe/rxe_recv.c if (mce->qp_list.next != &mcg->qp_list) next 400 drivers/infiniband/sw/rxe/rxe_verbs.c wr = wr->next; next 708 drivers/infiniband/sw/rxe/rxe_verbs.c wr = wr->next; next 769 drivers/infiniband/sw/rxe/rxe_verbs.c wr = wr->next; next 278 drivers/infiniband/sw/siw/siw_cm.c work = list_entry(cep->work_freelist.next, struct siw_cm_work, next 24 drivers/infiniband/sw/siw/siw_mem.c u32 id, next; next 26 drivers/infiniband/sw/siw/siw_mem.c get_random_bytes(&next, 4); next 27 drivers/infiniband/sw/siw/siw_mem.c next &= 0x00ffffff; next 29 drivers/infiniband/sw/siw/siw_mem.c if (xa_alloc_cyclic(&sdev->mem_xa, &id, m, limit, &next, next 95 drivers/infiniband/sw/siw/siw_mem.c u32 id, next; next 111 drivers/infiniband/sw/siw/siw_mem.c get_random_bytes(&next, 4); next 112 drivers/infiniband/sw/siw/siw_mem.c next &= 0x00ffffff; next 114 drivers/infiniband/sw/siw/siw_mem.c if (xa_alloc_cyclic(&sdev->mem_xa, &id, mem, limit, &next, next 1227 drivers/infiniband/sw/siw/siw_qp_tx.c tmp->next = fifo_list; next 1233 drivers/infiniband/sw/siw/siw_qp_tx.c qp->tx_list.next = NULL; next 1241 drivers/infiniband/sw/siw/siw_qp_tx.c qp->tx_list.next = NULL; next 704 drivers/infiniband/sw/siw/siw_verbs.c wr = wr->next; next 724 drivers/infiniband/sw/siw/siw_verbs.c wr = wr->next; next 924 drivers/infiniband/sw/siw/siw_verbs.c wr = wr->next; next 1068 drivers/infiniband/sw/siw/siw_verbs.c wr = wr->next; next 1789 drivers/infiniband/sw/siw/siw_verbs.c wr = wr->next; next 225 drivers/infiniband/ulp/ipoib/ipoib_cm.c p = list_entry(priv->cm.rx_flush_list.next, typeof(*p), list); next 342 drivers/infiniband/ulp/ipoib/ipoib_cm.c wr->next = NULL; next 942 drivers/infiniband/ulp/ipoib/ipoib_cm.c p = list_entry(priv->cm.passive_ids.next, typeof(*p), list); next 1359 drivers/infiniband/ulp/ipoib/ipoib_cm.c p = list_entry(priv->cm.start_list.next, typeof(*p), list); next 1413 drivers/infiniband/ulp/ipoib/ipoib_cm.c p = list_entry(priv->cm.reap_list.next, typeof(*p), list); next 129 drivers/infiniband/ulp/ipoib/ipoib_fs.c .next = ipoib_mcg_seq_next, next 234 drivers/infiniband/ulp/ipoib/ipoib_fs.c .next = ipoib_path_seq_next, next 230 drivers/infiniband/ulp/ipoib/ipoib_verbs.c priv->rx_wr.next = NULL; next 208 drivers/infiniband/ulp/iser/iscsi_iser.c tx_desc->inv_wr.next = NULL; next 209 drivers/infiniband/ulp/iser/iscsi_iser.c tx_desc->reg_wr.wr.next = NULL; next 376 drivers/infiniband/ulp/iser/iser_memory.c inv_wr->next = next_wr; next 414 drivers/infiniband/ulp/iser/iser_memory.c wr->wr.next = &tx_desc->send_wr; next 461 drivers/infiniband/ulp/iser/iser_memory.c wr->wr.next = &tx_desc->send_wr; next 987 drivers/infiniband/ulp/iser/iser_verbs.c wr.next = NULL; next 1013 drivers/infiniband/ulp/iser/iser_verbs.c wr->next = wr + 1; next 1018 drivers/infiniband/ulp/iser/iser_verbs.c wr->next = NULL; /* mark end of work requests list */ next 1048 drivers/infiniband/ulp/iser/iser_verbs.c wr->next = NULL; next 1055 drivers/infiniband/ulp/iser/iser_verbs.c if (tx_desc->inv_wr.next) next 1057 drivers/infiniband/ulp/iser/iser_verbs.c else if (tx_desc->reg_wr.wr.next) next 815 drivers/infiniband/ulp/isert/ib_isert.c rx_wr->next = rx_wr + 1; next 819 drivers/infiniband/ulp/isert/ib_isert.c rx_wr->next = NULL; /* mark end of work requests list */ next 846 drivers/infiniband/ulp/isert/ib_isert.c rx_wr.next = NULL; next 867 drivers/infiniband/ulp/isert/ib_isert.c send_wr.next = NULL; next 1231 drivers/infiniband/ulp/srp/ib_srp.c .next = NULL, next 1473 drivers/infiniband/ulp/srp/ib_srp.c if (state->fmr.next >= state->fmr.end) { next 1496 drivers/infiniband/ulp/srp/ib_srp.c *state->fmr.next++ = fmr; next 1532 drivers/infiniband/ulp/srp/ib_srp.c if (state->fr.next >= state->fr.end) { next 1573 drivers/infiniband/ulp/srp/ib_srp.c wr.wr.next = NULL; next 1584 drivers/infiniband/ulp/srp/ib_srp.c *state->fr.next++ = desc; next 1651 drivers/infiniband/ulp/srp/ib_srp.c state->fmr.next = req->fmr_list; next 1673 drivers/infiniband/ulp/srp/ib_srp.c state->fr.next = req->fr_list; next 1732 drivers/infiniband/ulp/srp/ib_srp.c state.gen.next = next_mr; next 1951 drivers/infiniband/ulp/srp/ib_srp.c ret = srp_map_idb(ch, req, state.gen.next, state.gen.end, next 2085 drivers/infiniband/ulp/srp/ib_srp.c wr.next = NULL; next 2107 drivers/infiniband/ulp/srp/ib_srp.c wr.next = NULL; next 328 drivers/infiniband/ulp/srp/ib_srp.h struct ib_pool_fmr **next; next 332 drivers/infiniband/ulp/srp/ib_srp.h struct srp_fr_desc **next; next 336 drivers/infiniband/ulp/srp/ib_srp.h void **next; next 834 drivers/infiniband/ulp/srpt/ib_srpt.c wr.next = NULL; next 857 drivers/infiniband/ulp/srpt/ib_srpt.c .next = NULL, next 2863 drivers/infiniband/ulp/srpt/ib_srpt.c send_wr.next = NULL; next 294 drivers/input/gameport/gameport.c struct gameport_event *e, *next; next 299 drivers/input/gameport/gameport.c list_for_each_entry_safe(e, next, &gameport_event_list, node) { next 409 drivers/input/gameport/gameport.c struct gameport_event *event, *next; next 414 drivers/input/gameport/gameport.c list_for_each_entry_safe(event, next, &gameport_event_list, node) { next 1210 drivers/input/input.c .next = input_devices_seq_next, next 1274 drivers/input/input.c .next = input_handlers_seq_next, next 2086 drivers/input/input.c struct input_handle *handle, *next; next 2092 drivers/input/input.c list_for_each_entry_safe(handle, next, &dev->h_list, d_node) next 2321 drivers/input/input.c struct input_handle *handle, *next; next 2325 drivers/input/input.c list_for_each_entry_safe(handle, next, &handler->h_list, h_node) next 926 drivers/input/serio/i8042-x86ia64io.h id = id->next; next 176 drivers/input/serio/serio.c struct serio_event *e, *next; next 181 drivers/input/serio/serio.c list_for_each_entry_safe(e, next, &serio_event_list, node) { next 297 drivers/input/serio/serio.c struct serio_event *event, *next; next 302 drivers/input/serio/serio.c list_for_each_entry_safe(event, next, &serio_event_list, node) { next 630 drivers/input/serio/serio.c s = list_entry(s->child_node.next, next 719 drivers/input/serio/serio.c struct serio *s, *next; next 722 drivers/input/serio/serio.c list_for_each_entry_safe(s, next, &serio->children, child_node) { next 769 drivers/input/touchscreen/cyttsp4_core.c int next; next 771 drivers/input/touchscreen/cyttsp4_core.c for (nbyte = 0, *axis = 0, next = 0; nbyte < size; nbyte++) { next 775 drivers/input/touchscreen/cyttsp4_core.c __func__, *axis, *axis, size, max, xy_data, next, next 776 drivers/input/touchscreen/cyttsp4_core.c xy_data[next], xy_data[next], bofs); next 777 drivers/input/touchscreen/cyttsp4_core.c *axis = (*axis * 256) + (xy_data[next] >> bofs); next 778 drivers/input/touchscreen/cyttsp4_core.c next++; next 786 drivers/input/touchscreen/cyttsp4_core.c __func__, *axis, *axis, size, max, xy_data, next, next 787 drivers/input/touchscreen/cyttsp4_core.c xy_data[next], xy_data[next]); next 546 drivers/input/touchscreen/sur40.c new_buf = list_entry(sur40->buf_list.next, struct sur40_buffer, list); next 220 drivers/interconnect/core.c struct icc_node *next, *prev = NULL; next 225 drivers/interconnect/core.c next = path->reqs[i].node; next 231 drivers/interconnect/core.c if (!prev || next->provider != prev->provider) { next 232 drivers/interconnect/core.c prev = next; next 237 drivers/interconnect/core.c ret = next->provider->set(prev, next); next 241 drivers/interconnect/core.c prev = next; next 3231 drivers/iommu/amd_iommu.c struct iommu_resv_region *entry, *next; next 3233 drivers/iommu/amd_iommu.c list_for_each_entry_safe(entry, next, head, list) next 1418 drivers/iommu/amd_iommu_init.c struct amd_iommu *iommu, *next; next 1420 drivers/iommu/amd_iommu_init.c for_each_iommu_safe(iommu, next) { next 2070 drivers/iommu/amd_iommu_init.c struct unity_map_entry *entry, *next; next 2072 drivers/iommu/amd_iommu_init.c list_for_each_entry_safe(entry, next, &amd_iommu_unity_map, list) { next 431 drivers/iommu/amd_iommu_types.h #define for_each_iommu_safe(iommu, next) \ next 432 drivers/iommu/amd_iommu_types.h list_for_each_entry_safe((iommu), (next), &amd_iommu_list, list) next 2718 drivers/iommu/arm-smmu-v3.c struct iommu_resv_region *entry, *next; next 2720 drivers/iommu/arm-smmu-v3.c list_for_each_entry_safe(entry, next, head, list) next 1549 drivers/iommu/arm-smmu.c struct iommu_resv_region *entry, *next; next 1551 drivers/iommu/arm-smmu.c list_for_each_entry_safe(entry, next, head, list) next 227 drivers/iommu/dma-iommu.c if (window->node.next == &bridge->dma_ranges && next 559 drivers/iommu/dmar.c struct acpi_dmar_header *iter, *next; next 562 drivers/iommu/dmar.c for (iter = start; iter < end; iter = next) { next 563 drivers/iommu/dmar.c next = (void *)iter + iter->length; next 568 drivers/iommu/dmar.c } else if (next > end) { next 797 drivers/iommu/exynos-iommu.c struct sysmmu_drvdata *data, *next; next 805 drivers/iommu/exynos-iommu.c list_for_each_entry_safe(data, next, &domain->clients, domain_node) { next 843 drivers/iommu/exynos-iommu.c struct sysmmu_drvdata *data, *next; next 859 drivers/iommu/exynos-iommu.c list_for_each_entry_safe(data, next, &domain->clients, domain_node) { next 1012 drivers/iommu/intel-iommu.c goto next; next 1033 drivers/iommu/intel-iommu.c next: next 1109 drivers/iommu/intel-iommu.c goto next; next 1132 drivers/iommu/intel-iommu.c next: next 5726 drivers/iommu/intel-iommu.c struct iommu_resv_region *entry, *next; next 5728 drivers/iommu/intel-iommu.c list_for_each_entry_safe(entry, next, head, list) next 384 drivers/iommu/iommu.c struct iommu_resv_region *region, *next; next 391 drivers/iommu/iommu.c list_for_each_entry_safe(region, next, &group_resv_regions, list) { next 733 drivers/iommu/iova.c struct iova *prev = NULL, *next = NULL; next 742 drivers/iommu/iova.c next = alloc_and_init_iova(pfn_hi + 1, iova->pfn_hi); next 743 drivers/iommu/iova.c if (next == NULL) next 754 drivers/iommu/iova.c if (next) { next 755 drivers/iommu/iova.c iova_insert_rbtree(&iovad->rbroot, next, NULL); next 353 drivers/iommu/virtio-iommu.c struct interval_tree_node *node, *next; next 356 drivers/iommu/virtio-iommu.c next = interval_tree_iter_first(&vdomain->mappings, iova, last); next 357 drivers/iommu/virtio-iommu.c while (next) { next 358 drivers/iommu/virtio-iommu.c node = next; next 360 drivers/iommu/virtio-iommu.c next = interval_tree_iter_next(node, iova, last); next 845 drivers/iommu/virtio-iommu.c struct iommu_resv_region *entry, *next; next 847 drivers/iommu/virtio-iommu.c list_for_each_entry_safe(entry, next, head, list) next 13 drivers/isdn/capi/capilib.c struct capilib_msgidqueue *next; next 41 drivers/isdn/capi/capilib.c np->msgidpool[i].next = np->msgidfree; next 51 drivers/isdn/capi/capilib.c np->msgidfree = mq->next; next 53 drivers/isdn/capi/capilib.c mq->next = NULL; next 55 drivers/isdn/capi/capilib.c np->msgidlast->next = mq; next 66 drivers/isdn/capi/capilib.c for (pp = &np->msgidqueue; *pp; pp = &(*pp)->next) { next 69 drivers/isdn/capi/capilib.c *pp = mq->next; next 72 drivers/isdn/capi/capilib.c mq->next = np->msgidfree; next 99 drivers/isdn/capi/kcapi_proc.c .next = controller_next, next 106 drivers/isdn/capi/kcapi_proc.c .next = controller_next, next 181 drivers/isdn/capi/kcapi_proc.c .next = applications_next, next 188 drivers/isdn/capi/kcapi_proc.c .next = applications_next, next 223 drivers/isdn/capi/kcapi_proc.c .next = capi_driver_next, next 914 drivers/isdn/hardware/mISDN/hfcmulti.c struct hfc_multi *hc, *next, *pcmmaster = NULL; next 931 drivers/isdn/hardware/mISDN/hfcmulti.c list_for_each_entry_safe(hc, next, &HFClist, list) { next 942 drivers/isdn/hardware/mISDN/hfcmulti.c list_for_each_entry_safe(hc, next, &HFClist, list) { next 1117 drivers/isdn/hardware/mISDN/hfcmulti.c struct hfc_multi *pos, *next, *plx_last_hc; next 1219 drivers/isdn/hardware/mISDN/hfcmulti.c list_for_each_entry_safe(pos, next, &HFClist, list) { next 5464 drivers/isdn/hardware/mISDN/hfcmulti.c struct hfc_multi *card, *next; next 5467 drivers/isdn/hardware/mISDN/hfcmulti.c list_for_each_entry_safe(card, next, &HFClist, list) next 2122 drivers/isdn/hardware/mISDN/hfcsusb.c struct hfcsusb *next; next 2130 drivers/isdn/hardware/mISDN/hfcsusb.c list_for_each_entry_safe(hw, next, &HFClist, list) next 513 drivers/isdn/mISDN/dsp_cmx.c member = list_entry(conf->mlist.next, struct dsp_conf_member, list); next 672 drivers/isdn/mISDN/dsp_cmx.c member = list_entry(conf->mlist.next, struct dsp_conf_member, next 686 drivers/isdn/mISDN/dsp_cmx.c member = list_entry(conf->mlist.next, struct dsp_conf_member, next 688 drivers/isdn/mISDN/dsp_cmx.c nextm = list_entry(member->list.next, struct dsp_conf_member, next 1121 drivers/isdn/mISDN/dsp_cmx.c member = list_entry(conf->mlist.next, struct dsp_conf_member, next 1460 drivers/isdn/mISDN/dsp_cmx.c other = (list_entry(conf->mlist.next, next 1705 drivers/isdn/mISDN/dsp_cmx.c member = list_entry(conf->mlist.next, next 1272 drivers/isdn/mISDN/l1oip_core.c struct l1oip *hc, *next; next 1274 drivers/isdn/mISDN/l1oip_core.c list_for_each_entry_safe(hc, next, &l1oip_ilist, list) next 283 drivers/isdn/mISDN/socket.c struct mISDNchannel *bchan, *next; next 297 drivers/isdn/mISDN/socket.c list_for_each_entry_safe(bchan, next, next 68 drivers/isdn/mISDN/timerdev.c struct mISDNtimer *timer, *next; next 85 drivers/isdn/mISDN/timerdev.c list_for_each_entry_safe(timer, next, &dev->expired, list) { next 28 drivers/leds/trigger/ledtrig-pattern.c struct led_pattern *next; next 41 drivers/leds/trigger/ledtrig-pattern.c data->curr = data->next; next 45 drivers/leds/trigger/ledtrig-pattern.c if (data->next == data->patterns + data->npatterns - 1) next 46 drivers/leds/trigger/ledtrig-pattern.c data->next = data->patterns; next 48 drivers/leds/trigger/ledtrig-pattern.c data->next++; next 65 drivers/leds/trigger/ledtrig-pattern.c step_brightness = abs(data->next->brightness - data->curr->brightness); next 68 drivers/leds/trigger/ledtrig-pattern.c if (data->next->brightness > data->curr->brightness) next 82 drivers/leds/trigger/ledtrig-pattern.c if (data->curr->brightness == data->next->brightness) { next 88 drivers/leds/trigger/ledtrig-pattern.c if (!data->next->delta_t) { next 138 drivers/leds/trigger/ledtrig-pattern.c data->next = data->patterns + 1; next 765 drivers/lightnvm/pblk-recovery.c goto next; next 770 drivers/lightnvm/pblk-recovery.c goto next; next 775 drivers/lightnvm/pblk-recovery.c goto next; next 786 drivers/lightnvm/pblk-recovery.c next: next 68 drivers/macintosh/adb-iop.c current_req = req->next; next 245 drivers/macintosh/adb-iop.c req->next = NULL; next 251 drivers/macintosh/adb-iop.c last_req->next = req; next 635 drivers/macintosh/adb.c ap = &(*ap)->next; next 636 drivers/macintosh/adb.c req->next = NULL; next 736 drivers/macintosh/adb.c state->completed = req->next; next 178 drivers/macintosh/macio-adb.c req->next = NULL; next 185 drivers/macintosh/macio-adb.c last_req->next = req; next 224 drivers/macintosh/macio-adb.c current_req = req->next; next 244 drivers/macintosh/macio-adb.c current_req = req->next; next 119 drivers/macintosh/smu.c cmd = list_entry(smu->cmd_list.next, struct smu_cmd, link); next 738 drivers/macintosh/smu.c newcmd = list_entry(smu->cmd_i2c_list.next, next 988 drivers/macintosh/smu.c prop->next = NULL; next 514 drivers/macintosh/via-cuda.c req->next = NULL; next 521 drivers/macintosh/via-cuda.c last_req->next = req; next 645 drivers/macintosh/via-cuda.c current_req = req->next; next 699 drivers/macintosh/via-cuda.c current_req = req->next; next 208 drivers/macintosh/via-macii.c req.next = current_req; next 244 drivers/macintosh/via-macii.c req->next = NULL; next 252 drivers/macintosh/via-macii.c last_req->next = req; next 426 drivers/macintosh/via-macii.c current_req = req->next; next 505 drivers/macintosh/via-macii.c current_req = req->next; next 1174 drivers/macintosh/via-pmu.c req->next = NULL; next 1180 drivers/macintosh/via-pmu.c last_req->next = req; next 1384 drivers/macintosh/via-pmu.c next: next 1476 drivers/macintosh/via-pmu.c goto next; next 1517 drivers/macintosh/via-pmu.c current_req = req->next; next 1567 drivers/macintosh/via-pmu.c current_req = req->next; next 2138 drivers/macintosh/via-pmu.c for (list = &all_pmu_pvt; (list = list->next) != &all_pmu_pvt; ) { next 89 drivers/macintosh/windfarm_core.c unsigned long next, delay; next 91 drivers/macintosh/windfarm_core.c next = jiffies; next 99 drivers/macintosh/windfarm_core.c if (time_after_eq(jiffies, next)) { next 114 drivers/macintosh/windfarm_core.c next += HZ; next 117 drivers/macintosh/windfarm_core.c delay = next - jiffies; next 314 drivers/macintosh/windfarm_smu_controls.c fct = list_entry(smu_fans.next, struct smu_fan_control, link); next 467 drivers/macintosh/windfarm_smu_sensors.c ads = list_entry(smu_ads.next, struct smu_ad_sensor, link); next 989 drivers/mailbox/bcm-flexrm-mailbox.c void *next; next 1052 drivers/mailbox/bcm-flexrm-mailbox.c next = flexrm_write_descs(msg, nhcnt, reqid, next 1056 drivers/mailbox/bcm-flexrm-mailbox.c if (IS_ERR(next)) { next 1057 drivers/mailbox/bcm-flexrm-mailbox.c ret = PTR_ERR(next); next 1063 drivers/mailbox/bcm-flexrm-mailbox.c ring->bd_write_offset = (unsigned long)(next - ring->bd_base); next 90 drivers/md/bcache/alloc.c unsigned int next = c->nbuckets * c->sb.bucket_size / 1024; next 101 drivers/md/bcache/alloc.c } while (atomic_cmpxchg(&c->rescale, r, r + next) != r); next 23 drivers/md/bcache/bset.c struct bkey *k, *next; next 25 drivers/md/bcache/bset.c for (k = i->start; k < bset_bkey_last(i); k = next) { next 26 drivers/md/bcache/bset.c next = bkey_next(k); next 36 drivers/md/bcache/bset.c if (next < bset_bkey_last(i) && next 38 drivers/md/bcache/bset.c &START_KEY(next) : next) > 0) next 113 drivers/md/bcache/bset.c struct bkey *k = iter->data->k, *next = bkey_next(k); next 115 drivers/md/bcache/bset.c if (next < iter->data->end && next 117 drivers/md/bcache/bset.c &START_KEY(next) : next) > 0) { next 787 drivers/md/bcache/btree.c if (c->shrink.list.next) next 678 drivers/md/bcache/journal.c unsigned int next = (ja->cur_idx + 1) % ca->sb.njournal_buckets; next 681 drivers/md/bcache/journal.c if (next == ja->discard_idx) next 684 drivers/md/bcache/journal.c ja->cur_idx = next; next 228 drivers/md/bcache/sysfs.c next_io = wb ? div64_s64(dc->writeback_rate.next-local_clock(), next 205 drivers/md/bcache/util.c d->next += div_u64(done * NSEC_PER_SEC, atomic_long_read(&d->rate)); next 213 drivers/md/bcache/util.c if (time_before64(now + NSEC_PER_SEC * 5LLU / 2LLU, d->next)) next 214 drivers/md/bcache/util.c d->next = now + NSEC_PER_SEC * 5LLU / 2LLU; next 216 drivers/md/bcache/util.c if (time_after64(now - NSEC_PER_SEC * 2, d->next)) next 217 drivers/md/bcache/util.c d->next = now - NSEC_PER_SEC * 2; next 219 drivers/md/bcache/util.c return time_after64(d->next, now) next 220 drivers/md/bcache/util.c ? div_u64(d->next - now, NSEC_PER_SEC / HZ) next 439 drivers/md/bcache/util.h uint64_t next; next 450 drivers/md/bcache/util.h d->next = local_clock(); next 93 drivers/md/bcache/writeback.c dc->writeback_rate.next + NSEC_PER_MSEC))) { next 396 drivers/md/bcache/writeback.c struct keybuf_key *next, *keys[MAX_WRITEBACKS_IN_PASS], *w; next 412 drivers/md/bcache/writeback.c next = bch_keybuf_next(&dc->writeback_keys); next 416 drivers/md/bcache/writeback.c next) { next 421 drivers/md/bcache/writeback.c BUG_ON(ptr_stale(dc->disk.c, &next->key, 0)); next 447 drivers/md/bcache/writeback.c &START_KEY(&next->key))) next 450 drivers/md/bcache/writeback.c size += KEY_SIZE(&next->key); next 451 drivers/md/bcache/writeback.c keys[nk++] = next; next 452 drivers/md/bcache/writeback.c } while ((next = bch_keybuf_next(&dc->writeback_keys))); next 585 drivers/md/bcache/writeback.c goto next; next 602 drivers/md/bcache/writeback.c next: next 734 drivers/md/dm-bufio.c list_entry(write_list->next, struct dm_buffer, write_list); next 866 drivers/md/dm-bufio.c b = list_entry(c->reserved_buffers.next, next 1708 drivers/md/dm-bufio.c struct dm_buffer *b = list_entry(c->reserved_buffers.next, next 1748 drivers/md/dm-bufio.c struct dm_buffer *b = list_entry(c->reserved_buffers.next, next 41 drivers/md/dm-cache-policy-smq.c unsigned next:28; next 134 drivers/md/dm-cache-policy-smq.c return to_entry(es, e->next); next 151 drivers/md/dm-cache-policy-smq.c e->next = l->head; next 167 drivers/md/dm-cache-policy-smq.c e->next = INDEXER_NULL; next 171 drivers/md/dm-cache-policy-smq.c tail->next = l->tail = to_index(es, e); next 189 drivers/md/dm-cache-policy-smq.c e->next = to_index(es, old); next 190 drivers/md/dm-cache-policy-smq.c prev->next = old->prev = to_index(es, e); next 200 drivers/md/dm-cache-policy-smq.c struct entry *next = l_next(es, e); next 203 drivers/md/dm-cache-policy-smq.c prev->next = e->next; next 205 drivers/md/dm-cache-policy-smq.c l->head = e->next; next 207 drivers/md/dm-cache-policy-smq.c if (next) next 208 drivers/md/dm-cache-policy-smq.c next->prev = e->prev; next 719 drivers/md/dm-cache-policy-smq.c e->next = INDEXER_NULL; next 82 drivers/md/dm-delay.c struct dm_delay_info *delayed, *next; next 88 drivers/md/dm-delay.c list_for_each_entry_safe(delayed, next, &dc->delayed_bios, list) { next 113 drivers/md/dm-init.c char *next; next 123 drivers/md/dm-init.c next = str_field_delimit(&field[i], ','); next 148 drivers/md/dm-init.c return next; next 192 drivers/md/dm-init.c char *next; next 202 drivers/md/dm-init.c next = str_field_delimit(&field[i], ';'); next 223 drivers/md/dm-init.c return next; next 1245 drivers/md/dm-integrity.c struct rb_node *next; next 1255 drivers/md/dm-integrity.c next = rb_next(&node->node); next 1256 drivers/md/dm-integrity.c if (unlikely(!next)) next 1259 drivers/md/dm-integrity.c next_node = container_of(next, struct journal_node, node); next 1265 drivers/md/dm-integrity.c struct rb_node *next; next 1271 drivers/md/dm-integrity.c next = rb_next(&node->node); next 1272 drivers/md/dm-integrity.c if (unlikely(!next)) next 1275 drivers/md/dm-integrity.c next_node = container_of(next, struct journal_node, node); next 2125 drivers/md/dm-integrity.c struct bio *next = flushes->bi_next; next 2128 drivers/md/dm-integrity.c flushes = next; next 2906 drivers/md/dm-integrity.c ic->reboot_notifier.next = NULL; next 3172 drivers/md/dm-integrity.c pl[i - 1].next = &pl[i]; next 3175 drivers/md/dm-integrity.c pl[i].next = NULL; next 193 drivers/md/dm-io.c dp->context_ptr = pl->next; next 546 drivers/md/dm-ioctl.c old_nl->next = (uint32_t) ((void *) nl - next 550 drivers/md/dm-ioctl.c nl->next = 0; next 592 drivers/md/dm-ioctl.c info->old_vers->next = (uint32_t) ((void *)info->vers - next 597 drivers/md/dm-ioctl.c info->vers->next = 0; next 1198 drivers/md/dm-ioctl.c spec->next = outptr - outbuf; next 1270 drivers/md/dm-ioctl.c static int next_target(struct dm_target_spec *last, uint32_t next, void *end, next 1273 drivers/md/dm-ioctl.c *spec = (struct dm_target_spec *) ((unsigned char *) last + next); next 1288 drivers/md/dm-ioctl.c uint32_t next = param->data_start; next 1299 drivers/md/dm-ioctl.c r = next_target(spec, next, end, &spec, &target_params); next 1314 drivers/md/dm-ioctl.c next = spec->next; next 243 drivers/md/dm-kcopyd.c struct page_list *next; next 246 drivers/md/dm-kcopyd.c next = pl->next; next 251 drivers/md/dm-kcopyd.c pl->next = kc->pages; next 256 drivers/md/dm-kcopyd.c pl = next; next 274 drivers/md/dm-kcopyd.c kc->pages = pl->next; next 277 drivers/md/dm-kcopyd.c pl->next = *pages; next 294 drivers/md/dm-kcopyd.c struct page_list *next; next 297 drivers/md/dm-kcopyd.c next = pl->next; next 299 drivers/md/dm-kcopyd.c pl = next; next 309 drivers/md/dm-kcopyd.c struct page_list *pl = NULL, *next; next 312 drivers/md/dm-kcopyd.c next = alloc_pl(GFP_KERNEL); next 313 drivers/md/dm-kcopyd.c if (!next) { next 318 drivers/md/dm-kcopyd.c next->next = pl; next 319 drivers/md/dm-kcopyd.c pl = next; next 395 drivers/md/dm-kcopyd.c zero_page_list.next = &zero_page_list; next 448 drivers/md/dm-kcopyd.c job = list_entry(jobs->next, struct kcopyd_job, list); next 469 drivers/md/dm-log-writes.c goto next; next 486 drivers/md/dm-log-writes.c goto next; next 495 drivers/md/dm-log-writes.c next: next 68 drivers/md/dm-queue-length.c struct path_info *pi, *next; next 70 drivers/md/dm-queue-length.c list_for_each_entry_safe(pi, next, paths, list) { next 442 drivers/md/dm-region-hash.c struct dm_region *reg, *next; next 483 drivers/md/dm-region-hash.c list_for_each_entry_safe(reg, next, &recovered, list) { next 489 drivers/md/dm-region-hash.c list_for_each_entry_safe(reg, next, &failed_recovered, list) { next 494 drivers/md/dm-region-hash.c list_for_each_entry_safe(reg, next, &clean, list) { next 651 drivers/md/dm-region-hash.c reg = list_entry(rh->quiesced_regions.next, next 34 drivers/md/dm-round-robin.c struct path_info *pi, *next; next 36 drivers/md/dm-round-robin.c list_for_each_entry_safe(pi, next, paths, list) { next 189 drivers/md/dm-round-robin.c pi = list_entry(s->valid_paths.next, struct path_info, list); next 64 drivers/md/dm-service-time.c struct path_info *pi, *next; next 66 drivers/md/dm-service-time.c list_for_each_entry_safe(pi, next, paths, list) { next 1745 drivers/md/dm-snap.c struct rb_node *next; next 1750 drivers/md/dm-snap.c next = rb_first(&s->out_of_order_tree); next 1751 drivers/md/dm-snap.c while (next) { next 1752 drivers/md/dm-snap.c pe = rb_entry(next, struct dm_snap_pending_exception, next 1756 drivers/md/dm-snap.c next = rb_next(next); next 213 drivers/md/dm-stats.c s = container_of(stats->list.next, struct dm_stat, list_entry); next 219 drivers/md/dm-table.c struct list_head *tmp, *next; next 221 drivers/md/dm-table.c list_for_each_safe(tmp, next, devices) { next 2327 drivers/md/dm-thin.c tc = list_entry_rcu(pool->active_thins.next, struct thin_c, list); next 124 drivers/md/dm-uevent.c struct dm_uevent *event, *next; next 126 drivers/md/dm-uevent.c list_for_each_entry_safe(event, next, events, elist) { next 639 drivers/md/dm-writecache.c struct rb_node *next; next 643 drivers/md/dm-writecache.c next = rb_next(&e->rb_node); next 645 drivers/md/dm-writecache.c if (unlikely(!next)) next 646 drivers/md/dm-writecache.c next = rb_first(&wc->freetree); next 647 drivers/md/dm-writecache.c wc->current_free = next ? container_of(next, struct wc_entry, rb_node) : NULL; next 651 drivers/md/dm-writecache.c e = container_of(wc->freelist.next, struct wc_entry, lru); next 688 drivers/md/dm-writecache.c wc->lru.next = LIST_POISON1; next 690 drivers/md/dm-writecache.c wc->freelist.next = LIST_POISON1; next 717 drivers/md/dm-writecache.c e = container_of(wc->lru.next, struct wc_entry, lru); next 728 drivers/md/dm-writecache.c if (unlikely(e->lru.next == &wc->lru)) next 730 drivers/md/dm-writecache.c e2 = container_of(e->lru.next, struct wc_entry, lru); next 1365 drivers/md/dm-writecache.c wb = list_entry(list->next, struct writeback_struct, endio_entry); next 1402 drivers/md/dm-writecache.c c = list_entry(list->next, struct copy_struct, endio_entry); next 1448 drivers/md/dm-writecache.c list.next->prev = list.prev->next = &list; next 1381 drivers/md/dm-zoned-metadata.c goto next; next 1402 drivers/md/dm-zoned-metadata.c goto next; next 1424 drivers/md/dm-zoned-metadata.c next: next 2341 drivers/md/dm-zoned-metadata.c struct dmz_mblock *mblk, *next; next 2381 drivers/md/dm-zoned-metadata.c rbtree_postorder_for_each_entry_safe(mblk, next, root, node) { next 847 drivers/md/dm.c struct list_head *tmp, *next; next 849 drivers/md/dm.c list_for_each_safe(tmp, next, devices) { next 344 drivers/md/md-bitmap.c struct buffer_head *next = bh->b_this_page; next 346 drivers/md/md-bitmap.c bh = next; next 307 drivers/md/md.c _tmp = all_mddevs.next; \ next 316 drivers/md/md.c _tmp = _tmp->next;}) \ next 4461 drivers/md/md.c = list_entry(mddev->disks.next, next 6280 drivers/md/md.c rdev0 = list_entry(pending_raid_disks.next, next 6521 drivers/md/md.c = list_entry(mddev->disks.next, next 7976 drivers/md/md.c tmp = all_mddevs.next; next 7978 drivers/md/md.c tmp = mddev->all_mddevs.next; next 8101 drivers/md/md.c .next = md_seq_next, next 8688 drivers/md/md.c int next = (last_mark+1) % SYNC_MARKS; next 8690 drivers/md/md.c mddev->resync_mark = mark[next]; next 8691 drivers/md/md.c mddev->resync_mark_cnt = mark_cnt[next]; next 8692 drivers/md/md.c mark[next] = jiffies; next 8693 drivers/md/md.c mark_cnt[next] = io_sectors - atomic_read(&mddev->recovery_active); next 8694 drivers/md/md.c last_mark = next; next 9305 drivers/md/md.c .next = NULL, next 9571 drivers/md/md.c node_detected_dev = list_entry(all_detected_devices.next, next 122 drivers/md/persistent-data/dm-space-map-metadata.c unsigned next = brb_next(brb, brb->end); next 128 drivers/md/persistent-data/dm-space-map-metadata.c if (next == brb->begin) next 135 drivers/md/persistent-data/dm-space-map-metadata.c brb->end = next; next 809 drivers/md/raid1.c struct bio *next = bio->bi_next; next 821 drivers/md/raid1.c bio = next; next 909 drivers/md/raid10.c struct bio *next = bio->bi_next; next 921 drivers/md/raid10.c bio = next; next 1094 drivers/md/raid10.c struct bio *next = bio->bi_next; next 1106 drivers/md/raid10.c bio = next; next 4435 drivers/md/raid10.c sector_t next, safe, last; next 4471 drivers/md/raid10.c next = first_dev_address(conf->reshape_progress - 1, next 4480 drivers/md/raid10.c if (next + conf->offset_diff < safe) next 4492 drivers/md/raid10.c next = last_dev_address(conf->reshape_progress, &conf->geo); next 4502 drivers/md/raid10.c if (next > safe + conf->offset_diff) next 517 drivers/md/raid5-cache.c struct stripe_head *sh, *next; next 519 drivers/md/raid5-cache.c list_for_each_entry_safe(sh, next, &io->stripe_list, log_list) { next 531 drivers/md/raid5-cache.c struct r5l_io_unit *io, *next; next 535 drivers/md/raid5-cache.c list_for_each_entry_safe(io, next, &log->running_ios, log_sibling) { next 547 drivers/md/raid5-cache.c struct r5l_io_unit *io, *next; next 551 drivers/md/raid5-cache.c list_for_each_entry_safe(io, next, &log->running_ios, log_sibling) { next 1202 drivers/md/raid5-cache.c struct r5l_io_unit *io, *next; next 1207 drivers/md/raid5-cache.c list_for_each_entry_safe(io, next, &log->finished_ios, log_sibling) { next 1402 drivers/md/raid5-cache.c struct stripe_head *sh, *next; next 1409 drivers/md/raid5-cache.c list_for_each_entry_safe(sh, next, &conf->r5c_full_stripe_list, lru) { next 1416 drivers/md/raid5-cache.c list_for_each_entry_safe(sh, next, next 1960 drivers/md/raid5-cache.c struct stripe_head *sh, *next; next 1962 drivers/md/raid5-cache.c list_for_each_entry_safe(sh, next, cached_stripe_list, lru) { next 1973 drivers/md/raid5-cache.c struct stripe_head *sh, *next; next 1975 drivers/md/raid5-cache.c list_for_each_entry_safe(sh, next, cached_stripe_list, lru) next 2432 drivers/md/raid5-cache.c struct stripe_head *sh, *next; next 2439 drivers/md/raid5-cache.c list_for_each_entry_safe(sh, next, &ctx->cached_list, lru) { next 402 drivers/md/raid5-ppl.c struct stripe_head *sh, *next; next 409 drivers/md/raid5-ppl.c list_for_each_entry_safe(sh, next, &io->stripe_list, log_list) { next 439 drivers/md/raid5.c first = (conf->inactive_list + hash)->next; next 896 drivers/md/raid5.c struct list_head *first, *next = NULL; next 904 drivers/md/raid5.c first = conf->pending_list.next; next 915 drivers/md/raid5.c first = data->sibling.next; next 916 drivers/md/raid5.c next = data->sibling.next; next 927 drivers/md/raid5.c if (next != &conf->pending_list) next 928 drivers/md/raid5.c conf->next_pending_data = list_entry(next, next 2340 drivers/md/raid5.c nsh = list_entry(newstripes.next, struct stripe_head, lru); next 2413 drivers/md/raid5.c nsh = list_entry(newstripes.next, struct stripe_head, lru); next 4608 drivers/md/raid5.c struct stripe_head *sh, *next; next 4612 drivers/md/raid5.c list_for_each_entry_safe(sh, next, &head_sh->batch_list, batch_list) { next 5065 drivers/md/raid5.c struct list_head *l = conf->delayed_list.next; next 5086 drivers/md/raid5.c struct stripe_head *sh = list_entry(head.next, struct stripe_head, lru); next 5363 drivers/md/raid5.c sh = list_entry(handle_list->next, typeof(*sh), lru); next 5368 drivers/md/raid5.c if (conf->hold_list.next == conf->last_hold) next 5371 drivers/md/raid5.c conf->last_hold = conf->hold_list.next; next 5433 drivers/md/raid5.c if (cb->list.next && !list_empty(&cb->list)) { next 5477 drivers/md/raid5.c if (cb->list.next == NULL) { next 5975 drivers/md/raid5.c sh = list_entry(stripes.next, struct stripe_head, lru); next 118 drivers/media/common/saa7146/saa7146_fops.c struct saa7146_buf *buf,*next = NULL; next 127 drivers/media/common/saa7146/saa7146_fops.c buf = list_entry(q->queue.next,struct saa7146_buf,vb.queue); next 130 drivers/media/common/saa7146/saa7146_fops.c next = list_entry(q->queue.next,struct saa7146_buf, vb.queue); next 133 drivers/media/common/saa7146/saa7146_fops.c buf, q->queue.prev, q->queue.next); next 134 drivers/media/common/saa7146/saa7146_fops.c buf->activate(dev,buf,next); next 346 drivers/media/common/saa7146/saa7146_fops.c buf = list_entry(q->stream.next, struct videobuf_buffer, stream); next 312 drivers/media/common/saa7146/saa7146_hlp.c int cur = 0, next = 0; next 328 drivers/media/common/saa7146/saa7146_hlp.c for( cur = 0, next = 1; next < *count; next++) { next 329 drivers/media/common/saa7146/saa7146_hlp.c if( values[cur] != values[next]) next 330 drivers/media/common/saa7146/saa7146_hlp.c values[++cur] = values[next]; next 997 drivers/media/common/saa7146/saa7146_hlp.c void saa7146_set_capture(struct saa7146_dev *dev, struct saa7146_buf *buf, struct saa7146_buf *next) next 1003 drivers/media/common/saa7146/saa7146_hlp.c DEB_CAP("buf:%p, next:%p\n", buf, next); next 135 drivers/media/common/saa7146/saa7146_vbi.c static void saa7146_set_vbi_capture(struct saa7146_dev *dev, struct saa7146_buf *buf, struct saa7146_buf *next) next 207 drivers/media/common/saa7146/saa7146_vbi.c struct saa7146_buf *next) next 212 drivers/media/common/saa7146/saa7146_vbi.c DEB_VBI("dev:%p, buf:%p, next:%p\n", dev, buf, next); next 213 drivers/media/common/saa7146/saa7146_vbi.c saa7146_set_vbi_capture(dev,buf,next); next 1020 drivers/media/common/saa7146/saa7146_video.c struct saa7146_buf *next) next 1025 drivers/media/common/saa7146/saa7146_video.c saa7146_set_capture(dev,buf,next); next 430 drivers/media/common/siano/smscoreapi.c struct list_head *next; next 433 drivers/media/common/siano/smscoreapi.c for (next = g_smscore_registry.next; next 434 drivers/media/common/siano/smscoreapi.c next != &g_smscore_registry; next 435 drivers/media/common/siano/smscoreapi.c next = next->next) { next 436 drivers/media/common/siano/smscoreapi.c entry = (struct smscore_registry_entry_t *) next; next 527 drivers/media/common/siano/smscoreapi.c struct list_head *next, *first; next 535 drivers/media/common/siano/smscoreapi.c for (next = first->next; next 536 drivers/media/common/siano/smscoreapi.c next != first && !rc; next 537 drivers/media/common/siano/smscoreapi.c next = next->next) { next 539 drivers/media/common/siano/smscoreapi.c (struct smscore_device_t *) next; next 565 drivers/media/common/siano/smscoreapi.c struct list_head *next, *first; next 571 drivers/media/common/siano/smscoreapi.c for (next = first->next; next != first;) { next 573 drivers/media/common/siano/smscoreapi.c (struct smscore_device_notifyee_t *) next; next 574 drivers/media/common/siano/smscoreapi.c next = next->next; next 592 drivers/media/common/siano/smscoreapi.c client = (struct smscore_client_t *) coredev->clients.next; next 1213 drivers/media/common/siano/smscoreapi.c cb = (struct smscore_buffer_t *) coredev->buffers.next; next 1639 drivers/media/common/siano/smscoreapi.c cb = (struct smscore_buffer_t *) coredev->buffers.next; next 1765 drivers/media/common/siano/smscoreapi.c (struct smscore_idlist_t *) client->idlist.next; next 2140 drivers/media/common/siano/smscoreapi.c g_smscore_notifyees.next; next 2151 drivers/media/common/siano/smscoreapi.c g_smscore_registry.next; next 1219 drivers/media/common/siano/smsdvb-main.c smsdvb_unregister_client((struct smsdvb_client_t *)g_smsdvb_clients.next); next 499 drivers/media/dvb-core/dmxdev.c list_for_each_entry(feed, &dmxdevfilter->feed.ts, next) next 520 drivers/media/dvb-core/dmxdev.c list_for_each_entry(feed, &filter->feed.ts, next) { next 580 drivers/media/dvb-core/dmxdev.c list_for_each_entry(feed, &dmxdevfilter->feed.ts, next) { next 600 drivers/media/dvb-core/dmxdev.c list_for_each_entry_safe(feed, tmp, &dmxdevfilter->feed.ts, next) { next 601 drivers/media/dvb-core/dmxdev.c list_del(&feed->next); next 774 drivers/media/dvb-core/dmxdev.c list_for_each_entry(feed, &filter->feed.ts, next) { next 891 drivers/media/dvb-core/dmxdev.c list_add(&feed->next, &filter->feed.ts); next 908 drivers/media/dvb-core/dmxdev.c list_for_each_entry_safe(feed, tmp, &filter->feed.ts, next) { next 913 drivers/media/dvb-core/dmxdev.c list_del(&feed->next); next 191 drivers/media/dvb-core/dvb_demux.c } while ((f = f->next) && sec->is_filtering); next 899 drivers/media/dvb-core/dvb_demux.c dvbdmxfilter->next = dvbdmxfeed->filter; next 948 drivers/media/dvb-core/dvb_demux.c } while ((f = f->next)); next 1046 drivers/media/dvb-core/dvb_demux.c dvbdmxfeed->filter = dvbdmxfilter->next; next 1048 drivers/media/dvb-core/dvb_demux.c while (f->next != dvbdmxfilter) next 1049 drivers/media/dvb-core/dvb_demux.c f = f->next; next 1050 drivers/media/dvb-core/dvb_demux.c f->next = f->next->next; next 102 drivers/media/dvb-core/dvb_vb2.c buf = list_entry(ctx->dvb_q.next, next 286 drivers/media/dvb-core/dvb_vb2.c ctx->buf = list_entry(ctx->dvb_q.next, next 1217 drivers/media/dvb-frontends/drx39xyj/drx_driver.h struct drx_version_list *next; next 163 drivers/media/dvb-frontends/rtl2832_sdr.c buf = list_entry(dev->queued_bufs.next, next 410 drivers/media/dvb-frontends/rtl2832_sdr.c buf = list_entry(dev->queued_bufs.next, next 268 drivers/media/i2c/et8ek8/et8ek8_driver.c const struct et8ek8_reg *next; next 277 drivers/media/i2c/et8ek8/et8ek8_driver.c next = regs; next 284 drivers/media/i2c/et8ek8/et8ek8_driver.c while (next->type != ET8EK8_REG_TERM && next 285 drivers/media/i2c/et8ek8/et8ek8_driver.c next->type != ET8EK8_REG_DELAY) { next 290 drivers/media/i2c/et8ek8/et8ek8_driver.c if (WARN(next->type != ET8EK8_REG_8BIT && next 291 drivers/media/i2c/et8ek8/et8ek8_driver.c next->type != ET8EK8_REG_16BIT, next 292 drivers/media/i2c/et8ek8/et8ek8_driver.c "Invalid type = %d", next->type)) { next 300 drivers/media/i2c/et8ek8/et8ek8_driver.c next++; next 316 drivers/media/i2c/et8ek8/et8ek8_driver.c if (next->type == ET8EK8_REG_DELAY) { next 317 drivers/media/i2c/et8ek8/et8ek8_driver.c msleep(next->val); next 322 drivers/media/i2c/et8ek8/et8ek8_driver.c next++; next 323 drivers/media/i2c/et8ek8/et8ek8_driver.c regs = next; next 326 drivers/media/i2c/et8ek8/et8ek8_driver.c } while (next->type != ET8EK8_REG_TERM); next 363 drivers/media/i2c/et8ek8/et8ek8_driver.c struct et8ek8_reglist **next = &meta->reglist[0].ptr; next 365 drivers/media/i2c/et8ek8/et8ek8_driver.c while (*next) { next 366 drivers/media/i2c/et8ek8/et8ek8_driver.c if ((*next)->type == type) next 367 drivers/media/i2c/et8ek8/et8ek8_driver.c return *next; next 369 drivers/media/i2c/et8ek8/et8ek8_driver.c next++; next 571 drivers/media/i2c/imx274.c const struct reg_8 *next; next 579 drivers/media/i2c/imx274.c for (next = table;; next++) { next 580 drivers/media/i2c/imx274.c if ((next->addr != range_start + range_count) || next 581 drivers/media/i2c/imx274.c (next->addr == IMX274_TABLE_END) || next 582 drivers/media/i2c/imx274.c (next->addr == IMX274_TABLE_WAIT_MS) || next 601 drivers/media/i2c/imx274.c if (next->addr == IMX274_TABLE_END) next 604 drivers/media/i2c/imx274.c if (next->addr == IMX274_TABLE_WAIT_MS) { next 605 drivers/media/i2c/imx274.c msleep_range(next->val); next 610 drivers/media/i2c/imx274.c val = next->val; next 613 drivers/media/i2c/imx274.c range_start = next->addr; next 357 drivers/media/i2c/tvp514x.c const struct tvp514x_reg *next = reglist; next 359 drivers/media/i2c/tvp514x.c for (; next->token != TOK_TERM; next++) { next 360 drivers/media/i2c/tvp514x.c if (next->token == TOK_DELAY) { next 361 drivers/media/i2c/tvp514x.c msleep(next->val); next 365 drivers/media/i2c/tvp514x.c if (next->token == TOK_SKIP) next 368 drivers/media/i2c/tvp514x.c err = tvp514x_write_reg(sd, next->reg, (u8) next->val); next 84 drivers/media/mc/mc-device.c int next = id & MEDIA_ENT_ID_FLAG_NEXT; next 89 drivers/media/mc/mc-device.c if (((media_entity_id(entity) == id) && !next) || next 90 drivers/media/mc/mc-device.c ((media_entity_id(entity) > id) && next)) { next 586 drivers/media/mc/mc-device.c struct media_entity_notify *notify, *next; next 621 drivers/media/mc/mc-device.c list_for_each_entry_safe(notify, next, &mdev->entity_notify, list) next 802 drivers/media/mc/mc-device.c struct media_entity *next; next 821 drivers/media/mc/mc-device.c list_for_each_entry_safe(entity, next, &mdev->entities, graph_obj.list) next 249 drivers/media/mc/mc-entity.c graph->stack[graph->top].link = entity->links.next; next 312 drivers/media/mc/mc-entity.c struct media_entity *next; next 318 drivers/media/mc/mc-entity.c link_top(graph) = link_top(graph)->next; next 327 drivers/media/mc/mc-entity.c next = media_entity_other(entity, link); next 330 drivers/media/mc/mc-entity.c if (media_entity_enum_test_and_set(&graph->ent_enum, next)) { next 331 drivers/media/mc/mc-entity.c link_top(graph) = link_top(graph)->next; next 334 drivers/media/mc/mc-entity.c next->name); next 339 drivers/media/mc/mc-entity.c link_top(graph) = link_top(graph)->next; next 340 drivers/media/mc/mc-entity.c stack_push(graph, next); next 342 drivers/media/mc/mc-entity.c next->name); next 2928 drivers/media/pci/bt8xx/bttv-driver.c buf = list_entry(fh->cap.stream.next,struct bttv_buffer,vb.stream); next 3473 drivers/media/pci/bt8xx/bttv-driver.c item = list_entry(btv->capture.next, struct bttv_buffer, vb.queue); next 3481 drivers/media/pci/bt8xx/bttv-driver.c (item->vb.queue.next != &btv->capture)) { next 3482 drivers/media/pci/bt8xx/bttv-driver.c item = list_entry(item->vb.queue.next, struct bttv_buffer, vb.queue); next 3632 drivers/media/pci/bt8xx/bttv-driver.c item = list_entry(btv->capture.next, struct bttv_buffer, vb.queue); next 3638 drivers/media/pci/bt8xx/bttv-driver.c item = list_entry(btv->vcapture.next, struct bttv_buffer, vb.queue); next 3726 drivers/media/pci/bt8xx/bttv-driver.c new = list_entry(btv->vcapture.next, struct bttv_buffer, vb.queue); next 548 drivers/media/pci/bt8xx/bttv-risc.c unsigned long next = btv->main.dma + ((slot+2) << 2); next 552 drivers/media/pci/bt8xx/bttv-risc.c btv->main.cpu[slot+1] = cpu_to_le32(next); next 564 drivers/media/pci/bt8xx/bttv-risc.c risc->jmp[1] = cpu_to_le32(next); next 644 drivers/media/pci/bt8xx/bttv-risc.c if (set->top->vb.queue.next) next 649 drivers/media/pci/bt8xx/bttv-risc.c if (set->top->vb.queue.next) next 651 drivers/media/pci/bt8xx/bttv-risc.c if (set->bottom->vb.queue.next) next 666 drivers/media/pci/bt8xx/bttv-risc.c if (set->top->vb.queue.next) next 677 drivers/media/pci/bt8xx/bttv-risc.c if (set->bottom->vb.queue.next) next 155 drivers/media/pci/cobalt/cobalt-omnitek.c dma_addr_t next = desc->bus; next 164 drivers/media/pci/cobalt/cobalt-omnitek.c WARN_ON(next & 3); next 209 drivers/media/pci/cobalt/cobalt-omnitek.c next += sizeof(struct sg_dma_descriptor); next 210 drivers/media/pci/cobalt/cobalt-omnitek.c d->next_h = (u32)((u64)next >> 32); next 211 drivers/media/pci/cobalt/cobalt-omnitek.c d->next_l = (u32)next | next 253 drivers/media/pci/cobalt/cobalt-omnitek.c next += sizeof(struct sg_dma_descriptor); next 263 drivers/media/pci/cobalt/cobalt-omnitek.c d->next_h = (u32)((u64)next >> 32); next 264 drivers/media/pci/cobalt/cobalt-omnitek.c d->next_l = (u32)next | (to_pci ? WRITE_TO_PCI : 0); next 272 drivers/media/pci/cobalt/cobalt-omnitek.c struct sg_dma_desc_info *next) next 277 drivers/media/pci/cobalt/cobalt-omnitek.c if (next == NULL) { next 281 drivers/media/pci/cobalt/cobalt-omnitek.c d->next_h = (u32)((u64)next->bus >> 32); next 282 drivers/media/pci/cobalt/cobalt-omnitek.c d->next_l = (u32)next->bus | direction | INTERRUPT_ENABLE; next 40 drivers/media/pci/cobalt/cobalt-omnitek.h struct sg_dma_desc_info *next); next 689 drivers/media/pci/cx18/cx18-fileops.c buf = list_entry(s->vb_capture.next, next 1176 drivers/media/pci/cx23885/cx23885-417.c struct cx23885_buffer *buf = list_entry(dmaq->active.next, next 1184 drivers/media/pci/cx23885/cx23885-417.c struct cx23885_buffer *buf = list_entry(dmaq->active.next, next 434 drivers/media/pci/cx23885/cx23885-core.c buf = list_entry(q->active.next, next 1651 drivers/media/pci/cx23885/cx23885-core.c buf = list_entry(q->active.next, struct cx23885_buffer, next 154 drivers/media/pci/cx23885/cx23885-dvb.c struct cx23885_buffer *buf = list_entry(dmaq->active.next, next 221 drivers/media/pci/cx23885/cx23885-vbi.c struct cx23885_buffer *buf = list_entry(dmaq->active.next, next 237 drivers/media/pci/cx23885/cx23885-vbi.c struct cx23885_buffer *buf = list_entry(dmaq->active.next, next 95 drivers/media/pci/cx23885/cx23885-video.c buf = list_entry(q->active.next, next 488 drivers/media/pci/cx23885/cx23885-video.c struct cx23885_buffer *buf = list_entry(dmaq->active.next, next 504 drivers/media/pci/cx23885/cx23885-video.c struct cx23885_buffer *buf = list_entry(dmaq->active.next, next 113 drivers/media/pci/cx25821/cx25821-video.c buf = list_entry(dmaq->active.next, next 266 drivers/media/pci/cx25821/cx25821-video.c struct cx25821_buffer *buf = list_entry(dmaq->active.next, next 284 drivers/media/pci/cx25821/cx25821-video.c struct cx25821_buffer *buf = list_entry(dmaq->active.next, next 733 drivers/media/pci/cx88/cx88-blackbird.c buf = list_entry(dmaq->active.next, struct cx88_buffer, list); next 741 drivers/media/pci/cx88/cx88-blackbird.c struct cx88_buffer *buf = list_entry(dmaq->active.next, next 769 drivers/media/pci/cx88/cx88-blackbird.c struct cx88_buffer *buf = list_entry(dmaq->active.next, next 523 drivers/media/pci/cx88/cx88-core.c buf = list_entry(q->active.next, next 125 drivers/media/pci/cx88/cx88-dvb.c buf = list_entry(dmaq->active.next, struct cx88_buffer, list); next 140 drivers/media/pci/cx88/cx88-dvb.c struct cx88_buffer *buf = list_entry(dmaq->active.next, next 204 drivers/media/pci/cx88/cx88-mpeg.c buf = list_entry(q->active.next, struct cx88_buffer, list); next 277 drivers/media/pci/cx88/cx88-mpeg.c buf = list_entry(q->active.next, struct cx88_buffer, list); next 106 drivers/media/pci/cx88/cx88-vbi.c buf = list_entry(q->active.next, struct cx88_buffer, list); next 198 drivers/media/pci/cx88/cx88-vbi.c struct cx88_buffer *buf = list_entry(dmaq->active.next, next 217 drivers/media/pci/cx88/cx88-vbi.c struct cx88_buffer *buf = list_entry(dmaq->active.next, next 411 drivers/media/pci/cx88/cx88-video.c buf = list_entry(q->active.next, struct cx88_buffer, list); next 533 drivers/media/pci/cx88/cx88-video.c struct cx88_buffer *buf = list_entry(dmaq->active.next, next 551 drivers/media/pci/cx88/cx88-video.c struct cx88_buffer *buf = list_entry(dmaq->active.next, next 906 drivers/media/pci/intel/ipu3/ipu3-cio2.c unsigned int i, j, next = q->bufs_next; next 938 drivers/media/pci/intel/ipu3/ipu3-cio2.c if (bufs_queued <= 1 || fbpt_rp == next) next 940 drivers/media/pci/intel/ipu3/ipu3-cio2.c next = (fbpt_rp + 1) % CIO2_MAX_BUFFERS; next 949 drivers/media/pci/intel/ipu3/ipu3-cio2.c if (!q->bufs[next]) { next 950 drivers/media/pci/intel/ipu3/ipu3-cio2.c q->bufs[next] = b; next 951 drivers/media/pci/intel/ipu3/ipu3-cio2.c entry = &q->fbpt[next * CIO2_MAX_LOPS]; next 954 drivers/media/pci/intel/ipu3/ipu3-cio2.c q->bufs_next = (next + 1) % CIO2_MAX_BUFFERS; next 961 drivers/media/pci/intel/ipu3/ipu3-cio2.c dev_dbg(&cio2->pci_dev->dev, "entry %i was full!\n", next); next 962 drivers/media/pci/intel/ipu3/ipu3-cio2.c next = (next + 1) % CIO2_MAX_BUFFERS; next 217 drivers/media/pci/ivtv/ivtv-irq.c buf = list_entry(s->q_predma.list.next, struct ivtv_buffer, list); next 66 drivers/media/pci/ivtv/ivtv-queue.c buf = list_entry(q->list.next, struct ivtv_buffer, list); next 67 drivers/media/pci/ivtv/ivtv-queue.c list_del_init(q->list.next); next 79 drivers/media/pci/ivtv/ivtv-queue.c struct ivtv_buffer *buf = list_entry(from->list.next, struct ivtv_buffer, list); next 81 drivers/media/pci/ivtv/ivtv-queue.c list_move_tail(from->list.next, &to->list); next 267 drivers/media/pci/saa7134/saa7134-core.c struct saa7134_buf *next = NULL; next 279 drivers/media/pci/saa7134/saa7134-core.c next = list_entry(q->queue.next, struct saa7134_buf, next 282 drivers/media/pci/saa7134/saa7134-core.c buf->activate(dev, buf, next); next 307 drivers/media/pci/saa7134/saa7134-core.c struct saa7134_buf *buf,*next = NULL; next 314 drivers/media/pci/saa7134/saa7134-core.c buf = list_entry(q->queue.next, struct saa7134_buf, entry); next 316 drivers/media/pci/saa7134/saa7134-core.c buf, q->queue.prev, q->queue.next); next 319 drivers/media/pci/saa7134/saa7134-core.c next = list_entry(q->queue.next, struct saa7134_buf, entry); next 321 drivers/media/pci/saa7134/saa7134-core.c buf->activate(dev, buf, next); next 323 drivers/media/pci/saa7134/saa7134-core.c q->queue.prev, q->queue.next); next 1190 drivers/media/pci/saa7134/saa7134-core.c list_for_each_entry(mops, &mops_list, next) next 1342 drivers/media/pci/saa7134/saa7134-core.c list_for_each_entry(mops, &mops_list, next) next 1379 drivers/media/pci/saa7134/saa7134-core.c struct saa7134_buf *buf, *next; next 1384 drivers/media/pci/saa7134/saa7134-core.c next = buf; next 1393 drivers/media/pci/saa7134/saa7134-core.c next = list_entry(q->queue.next, struct saa7134_buf, next 1395 drivers/media/pci/saa7134/saa7134-core.c buf->activate(dev, buf, next); next 1504 drivers/media/pci/saa7134/saa7134-core.c list_add_tail(&ops->next,&mops_list); next 1514 drivers/media/pci/saa7134/saa7134-core.c list_del(&ops->next); next 33 drivers/media/pci/saa7134/saa7134-ts.c struct saa7134_buf *next) next 42 drivers/media/pci/saa7134/saa7134-ts.c if (NULL == next) next 43 drivers/media/pci/saa7134/saa7134-ts.c next = buf; next 45 drivers/media/pci/saa7134/saa7134-ts.c ts_dbg("- [top] buf=%p next=%p\n", buf, next); next 47 drivers/media/pci/saa7134/saa7134-ts.c saa_writel(SAA7134_RS_BA2(5),saa7134_buffer_base(next)); next 50 drivers/media/pci/saa7134/saa7134-ts.c ts_dbg("- [bottom] buf=%p next=%p\n", buf, next); next 51 drivers/media/pci/saa7134/saa7134-ts.c saa_writel(SAA7134_RS_BA1(5),saa7134_buffer_base(next)); next 71 drivers/media/pci/saa7134/saa7134-vbi.c struct saa7134_buf *next) next 822 drivers/media/pci/saa7134/saa7134-video.c struct saa7134_buf *next) next 471 drivers/media/pci/saa7134/saa7134.h struct saa7134_buf *next); next 530 drivers/media/pci/saa7134/saa7134.h struct list_head next; next 83 drivers/media/pci/solo6x10/solo6x10-g723.c for (ss = pstr->substream; ss != NULL; ss = ss->next) { next 351 drivers/media/pci/solo6x10/solo6x10-g723.c ss; ss = ss->next, i++) next 724 drivers/media/pci/solo6x10/solo6x10-v4l2-enc.c solo_enc->vidq_active.next, next 339 drivers/media/pci/solo6x10/solo6x10-v4l2.c solo_dev->vidq_active.next, next 207 drivers/media/pci/tw5864/tw5864-core.c goto next; next 234 drivers/media/pci/tw5864/tw5864-core.c next: next 437 drivers/media/pci/tw5864/tw5864-video.c struct tw5864_buf *buf = list_entry(input->active.next, next 459 drivers/media/pci/tw5864/tw5864-video.c struct tw5864_buf *buf = list_entry(input->active.next, next 403 drivers/media/pci/tw68/tw68-core.c buf = container_of(dev->active.next, struct tw68_buf, list); next 495 drivers/media/pci/tw68/tw68-video.c container_of(dev->active.next, struct tw68_buf, list); next 510 drivers/media/pci/tw68/tw68-video.c container_of(dev->active.next, struct tw68_buf, list); next 987 drivers/media/pci/tw68/tw68-video.c buf = list_entry(dev->active.next, struct tw68_buf, list); next 40 drivers/media/pci/tw686x/tw686x-audio.c struct tw686x_audio_buf *next = NULL; next 54 drivers/media/pci/tw686x/tw686x-audio.c next = list_first_entry(&ac->buf_list, next 56 drivers/media/pci/tw686x/tw686x-audio.c list_move_tail(&next->list, &ac->buf_list); next 58 drivers/media/pci/tw686x/tw686x-audio.c ac->curr_bufs[pb] = next; next 62 drivers/media/pci/tw686x/tw686x-audio.c if (!done || !next) next 74 drivers/media/pci/tw686x/tw686x-audio.c reg_write(dev, reg, next->dma); next 298 drivers/media/pci/tw686x/tw686x-audio.c ss; ss = ss->next, i++) next 1245 drivers/media/platform/am437x/am437x-vpfe.c vpfe->next_frm = list_entry(vpfe->dma_queue.next, next 1993 drivers/media/platform/am437x/am437x-vpfe.c vpfe->next_frm = list_entry(vpfe->dma_queue.next, next 2061 drivers/media/platform/am437x/am437x-vpfe.c vpfe->next_frm = list_entry(vpfe->dma_queue.next, next 191 drivers/media/platform/atmel/atmel-isi.c isi->active = list_entry(isi->video_buffer_list.next, next 330 drivers/media/platform/atmel/atmel-isi.c desc = list_entry(isi->dma_desc_head.next, next 92 drivers/media/platform/davinci/vpbe_display.c layer->next_frm = list_entry(layer->dma_queue.next, next 279 drivers/media/platform/davinci/vpbe_display.c layer->next_frm = layer->cur_frm = list_entry(layer->dma_queue.next, next 340 drivers/media/platform/davinci/vpbe_display.c layer->next_frm = list_entry(layer->dma_queue.next, next 461 drivers/media/platform/davinci/vpfe_capture.c vpfe_dev->next_frm = list_entry(vpfe_dev->dma_queue.next, next 1434 drivers/media/platform/davinci/vpfe_capture.c vpfe_dev->next_frm = list_entry(vpfe_dev->dma_queue.next, next 206 drivers/media/platform/davinci/vpif_capture.c common->cur_frm = common->next_frm = list_entry(common->dma_queue.next, next 297 drivers/media/platform/davinci/vpif_capture.c common->next_frm = list_entry(common->dma_queue.next, next 345 drivers/media/platform/davinci/vpif_capture.c common->next_frm = list_entry(common->dma_queue.next, next 199 drivers/media/platform/davinci/vpif_display.c list_entry(common->dma_queue.next, next 284 drivers/media/platform/davinci/vpif_display.c common->next_frm = list_entry(common->dma_queue.next, next 309 drivers/media/platform/davinci/vpif_display.c common->next_frm = list_entry(common->dma_queue.next, next 1464 drivers/media/platform/exynos4-is/fimc-capture.c buf = list_entry(fimc->vid_cap.active_buf_q.next, next 692 drivers/media/platform/exynos4-is/fimc-core.h buf = list_entry(vid_cap->active_buf_q.next, next 718 drivers/media/platform/exynos4-is/fimc-core.h buf = list_entry(vid_cap->pending_buf_q.next, next 201 drivers/media/platform/exynos4-is/fimc-lite.h struct flite_buffer *buf = list_entry(dev->active_buf_q.next, next 216 drivers/media/platform/exynos4-is/fimc-lite.h struct flite_buffer *buf = list_entry(dev->pending_buf_q.next, next 376 drivers/media/platform/exynos4-is/media-dev.c p = list_entry(fmd->pipelines.next, typeof(*p), list); next 296 drivers/media/platform/fsl-viu.c buf = list_entry(vidq->active.next, struct viu_buf, vb.queue); next 316 drivers/media/platform/fsl-viu.c buf = list_entry(vidq->queued.next, struct viu_buf, vb.queue); next 350 drivers/media/platform/fsl-viu.c buf = list_entry(vidq->active.next, struct viu_buf, vb.queue); next 505 drivers/media/platform/fsl-viu.c dev, &vidq->queued, vidq->queued.next, next 963 drivers/media/platform/fsl-viu.c buf = list_entry(vidq->active.next, struct viu_buf, next 968 drivers/media/platform/fsl-viu.c buf = list_entry(vidq->queued.next, struct viu_buf, next 1074 drivers/media/platform/fsl-viu.c buf = list_entry(vidq->active.next, next 1545 drivers/media/platform/fsl-viu.c struct v4l2_subdev *sdev = list_entry(v4l2_dev->subdevs.next, next 542 drivers/media/platform/omap/omap_vout.c vout->next_frm = list_entry(vout->dma_queue.next, next 1002 drivers/media/platform/omap/omap_vout.c vout->next_frm = vout->cur_frm = list_entry(vout->dma_queue.next, next 1060 drivers/media/platform/omap3isp/isp.c struct isp_reg *next = reg_list; next 1062 drivers/media/platform/omap3isp/isp.c for (; next->reg != ISP_TOK_TERM; next++) next 1063 drivers/media/platform/omap3isp/isp.c next->val = isp_reg_readl(isp, next->mmio_range, next->reg); next 1075 drivers/media/platform/omap3isp/isp.c struct isp_reg *next = reg_list; next 1077 drivers/media/platform/omap3isp/isp.c for (; next->reg != ISP_TOK_TERM; next++) next 1078 drivers/media/platform/omap3isp/isp.c isp_reg_writel(isp, next->val, next->mmio_range, next->reg); next 1001 drivers/media/platform/pxa_camera.c pcdev->active = list_entry(pcdev->capture.next, next 851 drivers/media/platform/rcar-vin/rcar-dma.c buf = list_entry(vin->buf_list.next, struct rvin_buffer, list); next 548 drivers/media/platform/rcar_fdp1.c struct fdp1_field_buffer *next; next 934 drivers/media/platform/rcar_fdp1.c if (job->next) next 935 drivers/media/platform/rcar_fdp1.c fdp1_write(fdp1, job->next->addrs[0], FD1_RPF2_ADDR_Y); next 1233 drivers/media/platform/rcar_fdp1.c job->next = fdp1_peek_queued_field(ctx); next 136 drivers/media/platform/s5p-mfc/s5p_mfc.c b = list_entry(lh->next, struct s5p_mfc_buf, list); next 227 drivers/media/platform/s5p-mfc/s5p_mfc.c dst_buf = list_entry(ctx->dst_queue.next, next 265 drivers/media/platform/s5p-mfc/s5p_mfc.c src_buf = list_entry(ctx->src_queue.next, struct s5p_mfc_buf, list); next 423 drivers/media/platform/s5p-mfc/s5p_mfc.c src_buf = list_entry(ctx->src_queue.next, struct s5p_mfc_buf, next 538 drivers/media/platform/s5p-mfc/s5p_mfc.c src_buf = list_entry(ctx->src_queue.next, next 577 drivers/media/platform/s5p-mfc/s5p_mfc.c src_buf = list_entry(ctx->src_queue.next, next 612 drivers/media/platform/s5p-mfc/s5p_mfc.c mb_entry = list_entry(ctx->dst_queue.next, struct s5p_mfc_buf, next 1114 drivers/media/platform/s5p-mfc/s5p_mfc_enc.c mb_entry = list_entry((&ctx->ref_queue)->next, next 1134 drivers/media/platform/s5p-mfc/s5p_mfc_enc.c dst_mb = list_entry(ctx->dst_queue.next, struct s5p_mfc_buf, list); next 1151 drivers/media/platform/s5p-mfc/s5p_mfc_enc.c dst_mb = list_entry(ctx->dst_queue.next, next 1192 drivers/media/platform/s5p-mfc/s5p_mfc_enc.c src_mb = list_entry(ctx->src_queue.next, struct s5p_mfc_buf, list); next 1198 drivers/media/platform/s5p-mfc/s5p_mfc_enc.c dst_mb = list_entry(ctx->dst_queue.next, struct s5p_mfc_buf, list); next 1255 drivers/media/platform/s5p-mfc/s5p_mfc_enc.c mb_entry = list_entry(ctx->src_queue.next, struct s5p_mfc_buf, next 1267 drivers/media/platform/s5p-mfc/s5p_mfc_enc.c mb_entry = list_entry(ctx->dst_queue.next, struct s5p_mfc_buf, next 1179 drivers/media/platform/s5p-mfc/s5p_mfc_opr_v5.c temp_vb = list_entry(ctx->src_queue.next, struct s5p_mfc_buf, list); next 1216 drivers/media/platform/s5p-mfc/s5p_mfc_opr_v5.c src_mb = list_entry(ctx->src_queue.next, struct s5p_mfc_buf, next 1236 drivers/media/platform/s5p-mfc/s5p_mfc_opr_v5.c dst_mb = list_entry(ctx->dst_queue.next, struct s5p_mfc_buf, list); next 1255 drivers/media/platform/s5p-mfc/s5p_mfc_opr_v5.c temp_vb = list_entry(ctx->src_queue.next, struct s5p_mfc_buf, list); next 1274 drivers/media/platform/s5p-mfc/s5p_mfc_opr_v5.c dst_mb = list_entry(ctx->dst_queue.next, struct s5p_mfc_buf, list); next 1300 drivers/media/platform/s5p-mfc/s5p_mfc_opr_v5.c temp_vb = list_entry(ctx->src_queue.next, struct s5p_mfc_buf, list); next 1812 drivers/media/platform/s5p-mfc/s5p_mfc_opr_v6.c temp_vb = list_entry(ctx->src_queue.next, struct s5p_mfc_buf, list); next 1856 drivers/media/platform/s5p-mfc/s5p_mfc_opr_v6.c src_mb = list_entry(ctx->src_queue.next, struct s5p_mfc_buf, list); next 1874 drivers/media/platform/s5p-mfc/s5p_mfc_opr_v6.c dst_mb = list_entry(ctx->dst_queue.next, struct s5p_mfc_buf, list); next 1894 drivers/media/platform/s5p-mfc/s5p_mfc_opr_v6.c temp_vb = list_entry(ctx->src_queue.next, struct s5p_mfc_buf, list); next 1910 drivers/media/platform/s5p-mfc/s5p_mfc_opr_v6.c dst_mb = list_entry(ctx->dst_queue.next, struct s5p_mfc_buf, list); next 306 drivers/media/platform/sh_vou.c buf = list_entry(vou_dev->buf_list.next, struct sh_vou_buffer, list); next 316 drivers/media/platform/sh_vou.c buf = list_entry(buf->list.next, struct sh_vou_buffer, list); next 1067 drivers/media/platform/sh_vou.c vou_dev->active = list_entry(vou_dev->buf_list.next, next 1074 drivers/media/platform/sh_vou.c struct sh_vou_buffer *new = list_entry(vou_dev->active->list.next, next 1347 drivers/media/platform/sh_vou.c struct v4l2_subdev *sd = list_entry(v4l2_dev->subdevs.next, next 204 drivers/media/platform/sti/delta/delta-v4l2.c struct delta_dts *next; next 212 drivers/media/platform/sti/delta/delta-v4l2.c list_for_each_entry_safe(dts, next, &ctx->dts, list) next 250 drivers/media/platform/stm32/stm32-dcmi.c buf = list_entry(dcmi->buffers.next, struct dcmi_buf, list); next 809 drivers/media/platform/stm32/stm32-dcmi.c buf = list_entry(dcmi->buffers.next, struct dcmi_buf, list); next 181 drivers/media/platform/sunxi/sun4i-csi/sun4i_dma.c unsigned int next; next 184 drivers/media/platform/sunxi/sun4i-csi/sun4i_dma.c next = !(reg & CSI_BUF_CTRL_DBS); next 187 drivers/media/platform/sunxi/sun4i-csi/sun4i_dma.c sun4i_csi_buffer_mark_done(csi, next, sequence); next 190 drivers/media/platform/sunxi/sun4i-csi/sun4i_dma.c return sun4i_csi_buffer_fill_slot(csi, next); next 816 drivers/media/platform/ti-vpe/cal.c buf = list_entry(dma_q->active.next, struct cal_buffer, list); next 1287 drivers/media/platform/ti-vpe/cal.c buf = list_entry(dma_q->active.next, struct cal_buffer, list); next 484 drivers/media/platform/ti-vpe/vpdma.c list->next = list->buf.addr; next 498 drivers/media/platform/ti-vpe/vpdma.c list->next = list->buf.addr; next 510 drivers/media/platform/ti-vpe/vpdma.c list->next = NULL; next 533 drivers/media/platform/ti-vpe/vpdma.c list_size = (list->next - list->buf.addr) >> 4; next 636 drivers/media/platform/ti-vpe/vpdma.c cfd = list->next; next 645 drivers/media/platform/ti-vpe/vpdma.c list->next = cfd + 1; next 665 drivers/media/platform/ti-vpe/vpdma.c cfd = list->next; next 674 drivers/media/platform/ti-vpe/vpdma.c list->next = cfd + 1; next 703 drivers/media/platform/ti-vpe/vpdma.c ctd = list->next; next 712 drivers/media/platform/ti-vpe/vpdma.c list->next = ctd + 1; next 728 drivers/media/platform/ti-vpe/vpdma.c ctd = list->next; next 737 drivers/media/platform/ti-vpe/vpdma.c list->next = ctd + 1; next 836 drivers/media/platform/ti-vpe/vpdma.c dtd = list->next; next 855 drivers/media/platform/ti-vpe/vpdma.c list->next = dtd + 1; next 904 drivers/media/platform/ti-vpe/vpdma.c dtd = list->next; next 926 drivers/media/platform/ti-vpe/vpdma.c list->next = dtd + 1; next 27 drivers/media/platform/ti-vpe/vpdma.h void *next; next 336 drivers/media/platform/via-camera.c return list_entry(cam->buffer_queue.next, struct via_buffer, queue); next 270 drivers/media/platform/vivid/vivid-kthread-cap.c vid_out_buf = list_entry(dev->vid_out_active.next, next 696 drivers/media/platform/vivid/vivid-kthread-cap.c vid_cap_buf = list_entry(dev->vid_cap_active.next, struct vivid_buffer, list); next 702 drivers/media/platform/vivid/vivid-kthread-cap.c vbi_cap_buf = list_entry(dev->vbi_cap_active.next, next 932 drivers/media/platform/vivid/vivid-kthread-cap.c buf = list_entry(dev->vid_cap_active.next, next 947 drivers/media/platform/vivid/vivid-kthread-cap.c buf = list_entry(dev->vbi_cap_active.next, next 61 drivers/media/platform/vivid/vivid-kthread-out.c vid_out_buf = list_entry(dev->vid_out_active.next, next 68 drivers/media/platform/vivid/vivid-kthread-out.c vbi_out_buf = list_entry(dev->vbi_out_active.next, next 277 drivers/media/platform/vivid/vivid-kthread-out.c buf = list_entry(dev->vid_out_active.next, next 292 drivers/media/platform/vivid/vivid-kthread-out.c buf = list_entry(dev->vbi_out_active.next, next 97 drivers/media/platform/vivid/vivid-sdr-cap.c sdr_cap_buf = list_entry(dev->sdr_cap_active.next, next 301 drivers/media/platform/vivid/vivid-sdr-cap.c buf = list_entry(dev->sdr_cap_active.next, next 801 drivers/media/platform/vsp1/vsp1_dl.c struct vsp1_dl_list *next = list_next_entry(dl, chain); next 803 drivers/media/platform/vsp1/vsp1_dl.c dl->header->next_header = next->dma; next 1155 drivers/media/platform/vsp1/vsp1_dl.c struct vsp1_dl_list *dl, *next; next 1160 drivers/media/platform/vsp1/vsp1_dl.c list_for_each_entry_safe(dl, next, &dlm->free, list) { next 67 drivers/media/platform/vsp1/vsp1_drm.c struct vsp1_entity *next, unsigned int next_pad) next 77 drivers/media/platform/vsp1/vsp1_drm.c prev->sink = next; next 108 drivers/media/platform/vsp1/vsp1_drm.c uif->sink = next; next 539 drivers/media/platform/vsp1/vsp1_drm.c struct vsp1_entity *next; next 552 drivers/media/platform/vsp1/vsp1_drm.c list_for_each_entry_safe(entity, next, &pipe->entities, list_pipe) { next 316 drivers/media/platform/vsp1/vsp1_video.c struct vsp1_vb2_buffer *next = NULL; next 334 drivers/media/platform/vsp1/vsp1_video.c next = list_first_entry(&video->irqqueue, next 346 drivers/media/platform/vsp1/vsp1_video.c return next; next 223 drivers/media/usb/airspy/airspy.c buf = list_entry(s->queued_bufs.next, next 448 drivers/media/usb/airspy/airspy.c buf = list_entry(s->queued_bufs.next, next 403 drivers/media/usb/au0828/au0828-video.c *buf = list_entry(dma_q->active.next, struct au0828_buffer, list); next 478 drivers/media/usb/au0828/au0828-video.c *buf = list_entry(dma_q->active.next, struct au0828_buffer, list); next 875 drivers/media/usb/au0828/au0828-video.c buf = list_entry(vidq->active.next, struct au0828_buffer, list); next 905 drivers/media/usb/au0828/au0828-video.c buf = list_entry(vbiq->active.next, struct au0828_buffer, list); next 351 drivers/media/usb/cpia2/cpia2.h struct framebuf *next; next 2262 drivers/media/usb/cpia2/cpia2_core.c cam->buffers[i].next = &cam->buffers[i+1]; next 2269 drivers/media/usb/cpia2/cpia2_core.c cam->buffers[i].next = cam->buffers; next 2276 drivers/media/usb/cpia2/cpia2_core.c cam->workbuff = cam->curbuff->next; next 139 drivers/media/usb/cpia2/cpia2_usb.c cam->workbuff = cam->workbuff->next; next 233 drivers/media/usb/cpia2/cpia2_usb.c for (ptr = cam->workbuff->next; next 235 drivers/media/usb/cpia2/cpia2_usb.c ptr = ptr->next) next 739 drivers/media/usb/cpia2/cpia2_usb.c cam->workbuff = cam->curbuff->next; next 1278 drivers/media/usb/cx231xx/cx231xx-417.c buf = list_entry(dma_q->active.next, next 1334 drivers/media/usb/cx231xx/cx231xx-417.c buf = list_entry(dma_q->active.next, next 88 drivers/media/usb/cx231xx/cx231xx-core.c list_add_tail(&ops->next, &cx231xx_extension_devlist); next 108 drivers/media/usb/cx231xx/cx231xx-core.c list_del(&ops->next); next 119 drivers/media/usb/cx231xx/cx231xx-core.c list_for_each_entry(ops, &cx231xx_extension_devlist, next) { next 133 drivers/media/usb/cx231xx/cx231xx-core.c list_for_each_entry(ops, &cx231xx_extension_devlist, next) { next 614 drivers/media/usb/cx231xx/cx231xx-vbi.c *buf = list_entry(dma_q->active.next, struct cx231xx_buffer, vb.queue); next 244 drivers/media/usb/cx231xx/cx231xx-video.c *buf = list_entry(dma_q->active.next, struct cx231xx_buffer, vb.queue); next 739 drivers/media/usb/cx231xx/cx231xx.h struct list_head next; next 1088 drivers/media/usb/em28xx/em28xx-core.c list_add_tail(&ops->next, &em28xx_extension_devlist); next 1114 drivers/media/usb/em28xx/em28xx-core.c list_del(&ops->next); next 1126 drivers/media/usb/em28xx/em28xx-core.c list_for_each_entry(ops, &em28xx_extension_devlist, next) { next 1141 drivers/media/usb/em28xx/em28xx-core.c list_for_each_entry(ops, &em28xx_extension_devlist, next) { next 1158 drivers/media/usb/em28xx/em28xx-core.c list_for_each_entry(ops, &em28xx_extension_devlist, next) { next 1174 drivers/media/usb/em28xx/em28xx-core.c list_for_each_entry(ops, &em28xx_extension_devlist, next) { next 608 drivers/media/usb/em28xx/em28xx-video.c buf = list_entry(dma_q->active.next, struct em28xx_buffer, list); next 1175 drivers/media/usb/em28xx/em28xx-video.c buf = list_entry(vidq->active.next, struct em28xx_buffer, list); next 1210 drivers/media/usb/em28xx/em28xx-video.c buf = list_entry(vbiq->active.next, struct em28xx_buffer, list); next 787 drivers/media/usb/em28xx/em28xx.h struct list_head next; next 447 drivers/media/usb/hackrf/hackrf.c buffer = list_entry(buffer_list->next, struct hackrf_buffer, list); next 104 drivers/media/usb/hdpvr/hdpvr-video.c for (p = q->next; p != q;) { next 111 drivers/media/usb/hdpvr/hdpvr-video.c tmp = p->next; next 196 drivers/media/usb/hdpvr/hdpvr-video.c buf = list_entry(dev->free_buff_list.next, struct hdpvr_buffer, next 237 drivers/media/usb/hdpvr/hdpvr-video.c buf = list_entry(dev->rec_buff_list.next, struct hdpvr_buffer, next 153 drivers/media/usb/msi2500/msi2500.c buf = list_entry(dev->queued_bufs.next, struct msi2500_frame_buf, list); next 554 drivers/media/usb/msi2500/msi2500.c buf = list_entry(dev->queued_bufs.next, next 382 drivers/media/usb/pvrusb2/pvrusb2-io.c while ((lp = sp->queued_list.next) != &sp->queued_list) { next 549 drivers/media/usb/pvrusb2/pvrusb2-io.c struct list_head *lp = sp->idle_list.next; next 556 drivers/media/usb/pvrusb2/pvrusb2-io.c struct list_head *lp = sp->ready_list.next; next 185 drivers/media/usb/pwc/pwc-if.c buf = list_entry(pdev->queued_bufs.next, struct pwc_frame_buf, list); next 551 drivers/media/usb/pwc/pwc-if.c buf = list_entry(pdev->queued_bufs.next, struct pwc_frame_buf, next 555 drivers/media/usb/s2255/s2255drv.c buf = list_entry(vc->buf_list.next, next 384 drivers/media/usb/stkwebcam/stk-webcam.c list_move_tail(dev->sio_avail.next, next 857 drivers/media/usb/tm6000/tm6000-core.c list_for_each_entry(ops, &tm6000_extension_devlist, next) { next 871 drivers/media/usb/tm6000/tm6000-core.c list_add_tail(&ops->next, &tm6000_extension_devlist); next 891 drivers/media/usb/tm6000/tm6000-core.c list_del(&ops->next); next 902 drivers/media/usb/tm6000/tm6000-core.c list_for_each_entry(ops, &tm6000_extension_devlist, next) { next 916 drivers/media/usb/tm6000/tm6000-core.c list_for_each_entry(ops, &tm6000_extension_devlist, next) { next 91 drivers/media/usb/tm6000/tm6000-video.c *buf = list_entry(dma_q->active.next, next 1435 drivers/media/usb/tm6000/tm6000-video.c buf = list_entry(fh->vb_vidq.stream.next, struct tm6000_buffer, vb.stream); next 272 drivers/media/usb/tm6000/tm6000.h struct list_head next; next 570 drivers/media/usb/ttusb-dec/ttusb_dec.c for (item = dec->filter_info_list.next; item != &dec->filter_info_list; next 571 drivers/media/usb/ttusb-dec/ttusb_dec.c item = item->next) { next 780 drivers/media/usb/ttusb-dec/ttusb_dec.c if ((item = dec->urb_frame_list.next) != &dec->urb_frame_list) { next 1606 drivers/media/usb/ttusb-dec/ttusb_dec.c while ((item = dec->urb_frame_list.next) != &dec->urb_frame_list) { next 1624 drivers/media/usb/ttusb-dec/ttusb_dec.c while ((item = dec->filter_info_list.next) != &dec->filter_info_list) { next 1299 drivers/media/usb/usbvision/usbvision-core.c (*f) = list_entry(usbvision->inqueue.next, next 754 drivers/media/usb/usbvision/usbvision-video.c f = list_entry(usbvision->outqueue.next, next 756 drivers/media/usb/usbvision/usbvision-video.c list_del(usbvision->outqueue.next); next 950 drivers/media/usb/usbvision/usbvision-video.c frame = list_entry(usbvision->outqueue.next, next 952 drivers/media/usb/usbvision/usbvision-video.c list_del(usbvision->outqueue.next); next 858 drivers/media/usb/uvc/uvc_ctrl.c int next) next 873 drivers/media/usb/uvc/uvc_ctrl.c if ((map->id == v4l2_id) && !next) { next 880 drivers/media/usb/uvc/uvc_ctrl.c (map->id > v4l2_id) && next) { next 893 drivers/media/usb/uvc/uvc_ctrl.c int next = v4l2_id & V4L2_CTRL_FLAG_NEXT_CTRL; next 902 drivers/media/usb/uvc/uvc_ctrl.c __uvc_find_control(entity, v4l2_id, mapping, &ctrl, next); next 903 drivers/media/usb/uvc/uvc_ctrl.c if (ctrl && !next) next 907 drivers/media/usb/uvc/uvc_ctrl.c if (ctrl == NULL && !next) next 1512 drivers/media/usb/uvc/uvc_driver.c if (forward->chain.next || forward->chain.prev) { next 1598 drivers/media/usb/uvc/uvc_driver.c if (term->chain.next || term->chain.prev) { next 1657 drivers/media/usb/uvc/uvc_driver.c if (entity->chain.next || entity->chain.prev) { next 1843 drivers/media/usb/uvc/uvc_driver.c if (term->chain.next || term->chain.prev) next 529 drivers/media/usb/zr364xx/zr364xx.c buf = list_entry(dma_q->active.next, next 537 drivers/media/v4l2-core/v4l2-async.c if (!notifier || !notifier->asd_list.next) next 140 drivers/media/v4l2-core/v4l2-compat-ioctl32.c compat_caddr_t next; next 187 drivers/media/v4l2-core/v4l2-compat-ioctl32.c if (put_user(clipcount ? kclips + 1 : NULL, &kclips->next)) next 40 drivers/media/v4l2-core/v4l2-ctrls.c u32 next; next 2234 drivers/media/v4l2-core/v4l2-ctrls.c ref = ref->next; next 2330 drivers/media/v4l2-core/v4l2-ctrls.c new_ref->next = hdl->buckets[bucket]; next 3370 drivers/media/v4l2-core/v4l2-ctrls.c h->next = 0; next 3396 drivers/media/v4l2-core/v4l2-ctrls.c mref->helper->next = i; next 3492 drivers/media/v4l2-core/v4l2-ctrls.c idx = helpers[idx].next; next 3859 drivers/media/v4l2-core/v4l2-ctrls.c tmp_idx = helpers[tmp_idx].next; next 3875 drivers/media/v4l2-core/v4l2-ctrls.c idx = helpers[idx].next; next 3896 drivers/media/v4l2-core/v4l2-ctrls.c idx = helpers[idx].next; next 90 drivers/media/v4l2-core/v4l2-device.c struct v4l2_subdev *sd, *next; next 99 drivers/media/v4l2-core/v4l2-device.c list_for_each_entry_safe(sd, next, &v4l2_dev->subdevs, list) { next 684 drivers/media/v4l2-core/videobuf-core.c buf = list_entry(q->stream.next, struct videobuf_buffer, stream); next 1069 drivers/media/v4l2-core/videobuf-core.c q->read_buf = list_entry(q->stream.next, next 1129 drivers/media/v4l2-core/videobuf-core.c buf = list_entry(q->stream.next, next 1137 drivers/media/v4l2-core/videobuf-core.c q->read_buf = list_entry(q->stream.next, next 911 drivers/message/fusion/mptbase.c mf = list_entry(ioc->FreeQ.next, MPT_FRAME_HDR, next 1130 drivers/message/fusion/mptbase.c mpt_add_chain(void *pAddr, u8 next, u16 length, dma_addr_t dma_addr) next 1136 drivers/message/fusion/mptbase.c pChain->NextChainOffset = next; next 1150 drivers/message/fusion/mptbase.c mpt_add_chain_64bit(void *pAddr, u8 next, u16 length, dma_addr_t dma_addr) next 1159 drivers/message/fusion/mptbase.c pChain->NextChainOffset = next; next 595 drivers/message/fusion/mptbase.h typedef void (*MPT_ADD_CHAIN)(void *pAddr, u8 next, u16 length, next 812 drivers/message/fusion/mptbase.h void *next; next 336 drivers/message/fusion/mptsas.c struct fw_event_work *fw_event, *next; next 357 drivers/message/fusion/mptsas.c list_for_each_entry_safe(fw_event, next, &ioc->fw_event_list, list) { next 557 drivers/message/fusion/mptsas.c struct mptsas_device_info *sas_info, *next; next 566 drivers/message/fusion/mptsas.c list_for_each_entry_safe(sas_info, next, &ioc->sas_device_info_list, next 660 drivers/message/fusion/mptsas.c struct mptsas_device_info *sas_info, *next; next 722 drivers/message/fusion/mptsas.c list_for_each_entry_safe(sas_info, next, &ioc->sas_device_info_list, next 792 drivers/message/fusion/mptsas.c struct mptsas_device_info *sas_info, *next; next 797 drivers/message/fusion/mptsas.c list_for_each_entry_safe(sas_info, next, &ioc->sas_device_info_list, next 812 drivers/message/fusion/mptsas.c struct mptsas_device_info *sas_info, *next; next 815 drivers/message/fusion/mptsas.c list_for_each_entry_safe(sas_info, next, &ioc->sas_device_info_list, next 1171 drivers/message/fusion/mptsas.c target_reset_list = list_entry(head->next, next 1244 drivers/message/fusion/mptsas.c target_reset_list = list_entry(head->next, next 147 drivers/message/fusion/mptscsih.c chainBuf = list_entry(ioc->FreeChainQ.next, MPT_FRAME_HDR, next 1454 drivers/message/fusion/mptscsih.c int next; next 1465 drivers/message/fusion/mptscsih.c next = ioc->ChainToChain[chain_idx]; next 1483 drivers/message/fusion/mptscsih.c chain_idx = next; next 1539 drivers/mfd/ab8500-debugfs.c while ((action = action->next) != NULL) next 102 drivers/mfd/ipaq-micro.c micro->msg = list_entry(micro->queue.next, next 55 drivers/mfd/sta2x11-mfd.c return list_entry(sta2x11_mfd_list.next, next 200 drivers/mfd/twl4030-power.c u8 delay, u8 next) next 214 drivers/mfd/twl4030-power.c err = twl4030_write_script_byte(address++, next); next 116 drivers/misc/altera-stapl/altera.c struct altera_procinfo *next; next 2382 drivers/misc/altera-stapl/altera.c procptr->next = NULL; next 2389 drivers/misc/altera-stapl/altera.c while (tmpptr->next != NULL) next 2390 drivers/misc/altera-stapl/altera.c tmpptr = tmpptr->next; next 2391 drivers/misc/altera-stapl/altera.c tmpptr->next = procptr; next 2484 drivers/misc/altera-stapl/altera.c proc_list = procptr->next; next 191 drivers/misc/genwqe/card_dev.c struct list_head *node, *next; next 196 drivers/misc/genwqe/card_dev.c list_for_each_safe(node, next, &cfile->map_list) { next 229 drivers/misc/genwqe/card_dev.c struct list_head *node, *next; next 233 drivers/misc/genwqe/card_dev.c list_for_each_safe(node, next, &cfile->pin_list) { next 212 drivers/misc/habanalabs/command_submission.c struct hl_cs *next; next 220 drivers/misc/habanalabs/command_submission.c next = list_first_entry_or_null( next 224 drivers/misc/habanalabs/command_submission.c if ((next) && (!next->tdr_active)) { next 225 drivers/misc/habanalabs/command_submission.c next->tdr_active = true; next 226 drivers/misc/habanalabs/command_submission.c schedule_delayed_work(&next->work_tdr, next 407 drivers/misc/habanalabs/memory.c struct hl_vm_va_block *prev, *next; next 418 drivers/misc/habanalabs/memory.c next = list_next_entry(va_block, node); next 419 drivers/misc/habanalabs/memory.c if (&next->node != va_list && va_block->end + 1 == next->start) { next 420 drivers/misc/habanalabs/memory.c next->start = va_block->start; next 421 drivers/misc/habanalabs/memory.c next->size = next->end - next->start; next 71 drivers/misc/ibmasm/command.c struct list_head *next; next 76 drivers/misc/ibmasm/command.c next = sp->command_queue.next; next 77 drivers/misc/ibmasm/command.c list_del_init(next); next 78 drivers/misc/ibmasm/command.c cmd = list_entry(next, struct command, queue_node); next 203 drivers/misc/lkdtm/bugs.c test_head.next = redirection; next 228 drivers/misc/lkdtm/bugs.c item.node.next = redirection; next 1021 drivers/misc/mei/bus.c struct mei_cl_device *cldev, *next; next 1024 drivers/misc/mei/bus.c list_for_each_entry_safe(cldev, next, &bus->device_list, bus_list) next 304 drivers/misc/mei/client.c struct mei_me_client *me_cl, *next; next 307 drivers/misc/mei/client.c list_for_each_entry_safe(me_cl, next, &dev->me_clients, list) next 393 drivers/misc/mei/client.c struct mei_cl_cb *cb, *next; next 395 drivers/misc/mei/client.c list_for_each_entry_safe(cb, next, head, list) { next 413 drivers/misc/mei/client.c struct mei_cl_cb *cb, *next; next 415 drivers/misc/mei/client.c list_for_each_entry_safe(cb, next, head, list) { next 429 drivers/misc/mei/client.c struct mei_cl_cb *cb, *next; next 431 drivers/misc/mei/client.c list_for_each_entry_safe(cb, next, head, list) next 861 drivers/misc/mei/hbm.c struct mei_cl_cb *cb, *next; next 864 drivers/misc/mei/hbm.c list_for_each_entry_safe(cb, next, &dev->ctrl_rd_list, list) { next 31 drivers/misc/mei/interrupt.c struct mei_cl_cb *cb, *next; next 34 drivers/misc/mei/interrupt.c list_for_each_entry_safe(cb, next, cmpl_list, list) { next 372 drivers/misc/mei/interrupt.c struct mei_cl_cb *cb, *next; next 390 drivers/misc/mei/interrupt.c list_for_each_entry_safe(cb, next, &dev->write_waiting_list, list) { next 401 drivers/misc/mei/interrupt.c list_for_each_entry_safe(cb, next, &dev->ctrl_wr_list, list) { next 445 drivers/misc/mei/interrupt.c list_for_each_entry_safe(cb, next, &dev->write_list, list) { next 135 drivers/misc/mic/vop/vop_debugfs.c desc->flags, desc->next); next 64 drivers/misc/sgi-gru/grufile.c struct list_head *entry, *next; next 73 drivers/misc/sgi-gru/grufile.c list_for_each_safe(entry, next, &vdata->vd_head) { next 118 drivers/misc/sgi-gru/grukservices.c void *next __gru_cacheline_aligned__;/* CL 1 */ next 554 drivers/misc/sgi-gru/grukservices.c mq->next = &mq->data; next 839 drivers/misc/sgi-gru/grukservices.c struct message_header *mhdr = mq->next; next 840 drivers/misc/sgi-gru/grukservices.c void *next, *pnext; next 848 drivers/misc/sgi-gru/grukservices.c pnext = mq->next; next 849 drivers/misc/sgi-gru/grukservices.c next = pnext + GRU_CACHE_LINE_BYTES * lines; next 850 drivers/misc/sgi-gru/grukservices.c if (next == mq->limit) { next 851 drivers/misc/sgi-gru/grukservices.c next = mq->start; next 853 drivers/misc/sgi-gru/grukservices.c } else if (pnext < mq->start2 && next >= mq->start2) { next 859 drivers/misc/sgi-gru/grukservices.c mq->next = next; next 871 drivers/misc/sgi-gru/grukservices.c struct message_header *mhdr = mq->next; next 877 drivers/misc/sgi-gru/grukservices.c mhdr = mq->next; next 233 drivers/misc/sgi-gru/gruprocfs.c .next = seq_next, next 240 drivers/misc/sgi-gru/gruprocfs.c .next = seq_next, next 245 drivers/misc/sgi-xp/xpc.h struct xpc_fifo_entry_uv *next; next 268 drivers/misc/sgi-xp/xpc.h struct xpc_fifo_entry_uv next; /* FOR XPC INTERNAL USE ONLY */ next 290 drivers/misc/sgi-xp/xpc.h struct xpc_fifo_entry_uv next; next 976 drivers/misc/sgi-xp/xpc_uv.c head->first = first->next; next 983 drivers/misc/sgi-xp/xpc_uv.c first->next = NULL; next 995 drivers/misc/sgi-xp/xpc_uv.c last->next = NULL; next 998 drivers/misc/sgi-xp/xpc_uv.c head->last->next = last; next 1106 drivers/misc/sgi-xp/xpc_uv.c &msg_slot->next); next 1352 drivers/misc/sgi-xp/xpc_uv.c msg_slot = container_of(entry, struct xpc_send_msg_slot_uv, next); next 1361 drivers/misc/sgi-xp/xpc_uv.c xpc_put_fifo_entry_uv(&ch->sn.uv.msg_slot_free_list, &msg_slot->next); next 1462 drivers/misc/sgi-xp/xpc_uv.c xpc_put_fifo_entry_uv(&ch_uv->recv_msg_list, &msg_slot->hdr.u.next); next 1657 drivers/misc/sgi-xp/xpc_uv.c hdr.u.next); next 525 drivers/misc/vmw_vmci/vmci_context.c list_item = context->datagram_queue.next; next 551 drivers/misc/vmw_vmci/vmci_context.c list_item = context->datagram_queue.next; next 1936 drivers/mmc/core/mmc_test.c unsigned int next; next 1947 drivers/mmc/core/mmc_test.c next = rnd_next; next 1951 drivers/mmc/core/mmc_test.c rnd_next = next; next 1959 drivers/mmc/core/mmc_test.c next = rnd_next; next 1963 drivers/mmc/core/mmc_test.c rnd_next = next; next 311 drivers/mmc/core/sdio_cis.c this->next = NULL; next 315 drivers/mmc/core/sdio_cis.c prev = &this->next; next 362 drivers/mmc/core/sdio_cis.c tuple = tuple->next; next 403 drivers/mmc/core/sdio_cis.c tuple = tuple->next; next 1057 drivers/mmc/core/sdio_uart.c for (tpl = func->tuples; tpl; tpl = tpl->next) { next 1574 drivers/mmc/host/atmel-mci.c slot = list_entry(host->queue.next, next 1802 drivers/mmc/host/dw_mmc.c slot = list_entry(host->queue.next, next 465 drivers/mmc/host/mmci.c int mmci_prep_data(struct mmci_host *host, struct mmc_data *data, bool next) next 472 drivers/mmc/host/mmci.c err = host->ops->prep_data(host, data, next); next 474 drivers/mmc/host/mmci.c if (next && !err) next 853 drivers/mmc/host/mmci.c bool next) next 861 drivers/mmc/host/mmci.c if (next) next 888 drivers/mmc/host/mmci.c struct mmci_dmae_next *next = &dmae->next_data; next 893 drivers/mmc/host/mmci.c WARN_ON(!data->host_cookie && (next->desc || next->chan)); next 895 drivers/mmc/host/mmci.c dmae->desc_current = next->desc; next 896 drivers/mmc/host/mmci.c dmae->cur = next->chan; next 897 drivers/mmc/host/mmci.c next->desc = NULL; next 898 drivers/mmc/host/mmci.c next->chan = NULL; next 913 drivers/mmc/host/mmci.c struct mmci_dmae_next *next = &dmae->next_data; next 921 drivers/mmc/host/mmci.c if (dmae->desc_current == next->desc) next 924 drivers/mmc/host/mmci.c if (dmae->cur == next->chan) { next 929 drivers/mmc/host/mmci.c next->desc = NULL; next 930 drivers/mmc/host/mmci.c next->chan = NULL; next 357 drivers/mmc/host/mmci.h bool next); next 435 drivers/mmc/host/mmci.h bool next); next 77 drivers/mmc/host/mmci_stm32_sdmmc.c struct mmc_data *data, bool next) next 80 drivers/mmc/host/mmci_stm32_sdmmc.c if (!next && data->host_cookie == host->next_cookie) next 311 drivers/mmc/host/mtk-sd.c u32 next; next 329 drivers/mmc/host/mtk-sd.c u32 next; next 1637 drivers/mmc/host/mtk-sd.c gpd->next = lower_32_bits(dma_addr); next 1649 drivers/mmc/host/mtk-sd.c bd[i].next = lower_32_bits(dma_addr); next 1200 drivers/mmc/host/omap_hsmmc.c struct omap_hsmmc_next *next, next 1205 drivers/mmc/host/omap_hsmmc.c if (!next && data->host_cookie && next 1214 drivers/mmc/host/omap_hsmmc.c if (next || data->host_cookie != host->next_data.cookie) { next 1227 drivers/mmc/host/omap_hsmmc.c if (next) { next 1228 drivers/mmc/host/omap_hsmmc.c next->dma_len = dma_len; next 1229 drivers/mmc/host/omap_hsmmc.c data->host_cookie = ++next->cookie < 0 ? 1 : next->cookie; next 486 drivers/mmc/host/usdhi6rol0.c struct scatterlist *next = sg_next(host->sg); next 490 drivers/mmc/host/usdhi6rol0.c if (!next) next 492 drivers/mmc/host/usdhi6rol0.c host->sg = next; next 494 drivers/mmc/host/usdhi6rol0.c if (WARN(next && sg_dma_len(next) % data->blksz, next 496 drivers/mmc/host/usdhi6rol0.c sg_dma_len(next), data->blksz)) next 477 drivers/mtd/devices/block2mtd.c struct list_head *pos, *next; next 480 drivers/mtd/devices/block2mtd.c list_for_each_safe(pos, next, &blkmtd_device_list) { next 220 drivers/mtd/devices/ms02-nv.c mp->next = root_ms02nv_mtd; next 251 drivers/mtd/devices/ms02-nv.c root_ms02nv_mtd = mp->next; next 91 drivers/mtd/devices/ms02-nv.h struct mtd_info *next; next 62 drivers/mtd/devices/slram.c struct slram_mtd_list *next; next 135 drivers/mtd/devices/slram.c curmtd = &(*curmtd)->next; next 144 drivers/mtd/devices/slram.c (*curmtd)->next = NULL; next 204 drivers/mtd/devices/slram.c nextitem = slram_mtdlist->next; next 524 drivers/mtd/mtd_blkdevs.c struct mtd_blktrans_dev *dev, *next; next 527 drivers/mtd/mtd_blkdevs.c list_for_each_entry_safe(dev, next, &tr->devs, list) next 556 drivers/mtd/mtd_blkdevs.c if (!blktrans_notifier.list.next) next 588 drivers/mtd/mtd_blkdevs.c struct mtd_blktrans_dev *dev, *next; next 595 drivers/mtd/mtd_blkdevs.c list_for_each_entry_safe(dev, next, &tr->devs, list) next 609 drivers/mtd/mtd_blkdevs.c if (blktrans_notifier.list.next) next 639 drivers/mtd/mtdpart.c struct mtd_part *child, *next; next 642 drivers/mtd/mtdpart.c list_for_each_entry_safe(child, next, &mtd_partitions, list) { next 668 drivers/mtd/mtdpart.c struct mtd_part *slave, *next; next 672 drivers/mtd/mtdpart.c list_for_each_entry_safe(slave, next, &mtd_partitions, list) next 685 drivers/mtd/mtdpart.c struct mtd_part *slave, *next; next 689 drivers/mtd/mtdpart.c list_for_each_entry_safe(slave, next, &mtd_partitions, list) next 60 drivers/mtd/parsers/cmdlinepart.c struct cmdline_mtd_partition *next; next 267 drivers/mtd/parsers/cmdlinepart.c this_mtd->next = partitions; next 316 drivers/mtd/parsers/cmdlinepart.c for (part = partitions; part; part = part->next) { next 33 drivers/mtd/parsers/redboot.c struct fis_list *next; next 218 drivers/mtd/parsers/redboot.c prev = &(*prev)->next; next 219 drivers/mtd/parsers/redboot.c new_fl->next = *prev; next 230 drivers/mtd/parsers/redboot.c for (tmp_fl = fl; tmp_fl->next; tmp_fl = tmp_fl->next) { next 231 drivers/mtd/parsers/redboot.c if (tmp_fl->img->flash_base + tmp_fl->img->size + master->erasesize <= tmp_fl->next->img->flash_base) { next 278 drivers/mtd/parsers/redboot.c if(fl->next && fl->img->flash_base + fl->img->size + master->erasesize <= fl->next->img->flash_base) { next 281 drivers/mtd/parsers/redboot.c parts[i].size = fl->next->img->flash_base - parts[i].offset; next 286 drivers/mtd/parsers/redboot.c fl = fl->next; next 294 drivers/mtd/parsers/redboot.c fl = fl->next; next 1111 drivers/mtd/spi-nor/spi-nor.c struct spi_nor_erase_command *cmd, *next; next 1113 drivers/mtd/spi-nor/spi-nor.c list_for_each_entry_safe(cmd, next, erase_list, list) { next 1201 drivers/mtd/spi-nor/spi-nor.c struct spi_nor_erase_command *cmd, *next; next 1208 drivers/mtd/spi-nor/spi-nor.c list_for_each_entry_safe(cmd, next, &erase_list, list) { next 825 drivers/mtd/ubi/attach.c aeb = list_entry(ai->free.next, struct ubi_ainf_peb, u.list); next 657 drivers/mtd/ubi/block.c struct ubiblock *next; next 661 drivers/mtd/ubi/block.c list_for_each_entry_safe(dev, next, &ubiblock_devices, list) { next 454 drivers/mtd/ubi/debug.c .next = eraseblk_count_seq_next, next 209 drivers/mtd/ubi/wl.c wrk = list_entry(ubi->works.next, struct ubi_work, list); next 1679 drivers/mtd/ubi/wl.c wrk = list_entry(ubi->works.next, struct ubi_work, list); next 133 drivers/net/appletalk/ipddp.c for(rt = ipddp_route_list; rt != NULL; rt = rt->next) next 204 drivers/net/appletalk/ipddp.c rt->next = NULL; next 217 drivers/net/appletalk/ipddp.c rt->next = ipddp_route_list; next 241 drivers/net/appletalk/ipddp.c *r = tmp->next; next 246 drivers/net/appletalk/ipddp.c r = &tmp->next; next 260 drivers/net/appletalk/ipddp.c for(f = ipddp_route_list; f != NULL; f = f->next) next 333 drivers/net/appletalk/ipddp.c p = ipddp_route_list->next; next 21 drivers/net/appletalk/ipddp.h struct ipddp_route *next; next 269 drivers/net/appletalk/ltpc.c struct xmitQel *next; next 287 drivers/net/appletalk/ltpc.c qel->next = NULL; next 291 drivers/net/appletalk/ltpc.c xmQtl->next = qel; next 311 drivers/net/appletalk/ltpc.c xmQhd = qel->next; next 83 drivers/net/bonding/bond_alb.c entry->next = TLB_NULL_INDEX; next 106 drivers/net/bonding/bond_alb.c u32 next_index = tx_hash_table[index].next; next 208 drivers/net/bonding/bond_alb.c hash_table[hash_index].next = next_index; next 808 drivers/net/bonding/bond_alb.c u32 next; next 811 drivers/net/bonding/bond_alb.c next = bond_info->rx_hashtbl[ip_src_hash].src_first; next 812 drivers/net/bonding/bond_alb.c bond_info->rx_hashtbl[ip_dst_hash].src_next = next; next 813 drivers/net/bonding/bond_alb.c if (next != RLB_NULL_INDEX) next 814 drivers/net/bonding/bond_alb.c bond_info->rx_hashtbl[next].src_prev = ip_dst_hash; next 3442 drivers/net/bonding/bond_main.c struct net_device *ldev, *next, *now, *dev_stack[MAX_NEST_DEV + 1]; next 3450 drivers/net/bonding/bond_main.c next = NULL; next 3456 drivers/net/bonding/bond_main.c next = ldev; next 3465 drivers/net/bonding/bond_main.c if (!next) { next 3468 drivers/net/bonding/bond_main.c next = dev_stack[--cur]; next 3472 drivers/net/bonding/bond_main.c now = next; next 257 drivers/net/bonding/bond_procfs.c .next = bond_info_seq_next, next 23 drivers/net/dsa/bcm_sf2_cfp.c struct list_head next; next 547 drivers/net/dsa/bcm_sf2_cfp.c list_for_each_entry(rule, &priv->cfp.rules_list, next) { next 565 drivers/net/dsa/bcm_sf2_cfp.c list_for_each_entry(rule, &priv->cfp.rules_list, next) { next 909 drivers/net/dsa/bcm_sf2_cfp.c list_add_tail(&rule->next, &priv->cfp.rules_list); next 989 drivers/net/dsa/bcm_sf2_cfp.c list_del(&rule->next); next 1164 drivers/net/dsa/bcm_sf2_cfp.c list_for_each_entry_safe_reverse(rule, n, &priv->cfp.rules_list, next) next 1186 drivers/net/dsa/bcm_sf2_cfp.c list_for_each_entry(rule, &priv->cfp.rules_list, next) { next 281 drivers/net/ethernet/3com/3c515.c u32 next; next 294 drivers/net/ethernet/3com/3c515.c u32 next; next 339 drivers/net/ethernet/3com/3c515.c next:8; /* The media type to try next. */ next 732 drivers/net/ethernet/3com/3c515.c dev->if_port = media_tbl[dev->if_port].next; next 823 drivers/net/ethernet/3com/3c515.c vp->rx_ring[i].next = next 826 drivers/net/ethernet/3com/3c515.c vp->rx_ring[i].next = 0; next 837 drivers/net/ethernet/3com/3c515.c vp->rx_ring[i - 1].next = next 923 drivers/net/ethernet/3com/3c515.c media_tbl[dev->if_port].next; next 1029 drivers/net/ethernet/3com/3c515.c vp->tx_ring[entry].next = 0; next 1041 drivers/net/ethernet/3com/3c515.c prev_entry->next = isa_virt_to_bus(&vp->tx_ring[entry]); next 1572 drivers/net/ethernet/3com/3c515.c vp = list_entry(root_corkscrew_dev.next, next 544 drivers/net/ethernet/3com/3c59x.c __le32 next; /* Last entry points to 0. */ next 564 drivers/net/ethernet/3com/3c59x.c __le32 next; /* Last entry points to 0. */ next 724 drivers/net/ethernet/3com/3c59x.c next:8; /* The media type to try next. */ next 1582 drivers/net/ethernet/3com/3c59x.c dev->if_port = media_tbl[dev->if_port].next; next 1736 drivers/net/ethernet/3com/3c59x.c vp->rx_ring[i].next = cpu_to_le32(vp->rx_ring_dma + sizeof(struct boom_rx_desc) * (i+1)); next 1759 drivers/net/ethernet/3com/3c59x.c vp->rx_ring[i-1].next = cpu_to_le32(vp->rx_ring_dma); next 1839 drivers/net/ethernet/3com/3c59x.c dev->if_port = media_tbl[dev->if_port].next; next 2145 drivers/net/ethernet/3com/3c59x.c vp->tx_ring[entry].next = 0; next 2214 drivers/net/ethernet/3com/3c59x.c prev_entry->next = cpu_to_le32(vp->tx_ring_dma + entry * sizeof(struct boom_tx_desc)); next 20 drivers/net/ethernet/8390/8390.h unsigned char next; /* pointer to next packet. */ next 1381 drivers/net/ethernet/8390/axnet_cs.c rx_frame.next); next 1415 drivers/net/ethernet/8390/axnet_cs.c rx_frame.status, rx_frame.next, next 1422 drivers/net/ethernet/8390/axnet_cs.c next_frame = rx_frame.next; next 703 drivers/net/ethernet/8390/lib8390.c if (rx_frame.next != next_frame && next 704 drivers/net/ethernet/8390/lib8390.c rx_frame.next != next_frame + 1 && next 705 drivers/net/ethernet/8390/lib8390.c rx_frame.next != next_frame - num_rx_pages && next 706 drivers/net/ethernet/8390/lib8390.c rx_frame.next != next_frame + 1 - num_rx_pages) { next 717 drivers/net/ethernet/8390/lib8390.c rx_frame.next); next 745 drivers/net/ethernet/8390/lib8390.c rx_frame.status, rx_frame.next, next 752 drivers/net/ethernet/8390/lib8390.c next_frame = rx_frame.next; next 202 drivers/net/ethernet/8390/stnic.c hdr->next = buf[0] >> 8; next 211 drivers/net/ethernet/8390/stnic.c ring_page, hdr->status, hdr->next, hdr->count); next 350 drivers/net/ethernet/agere/et131x.c struct tcb *next; /* Next entry in ring */ next 1764 drivers/net/ethernet/agere/et131x.c tcb->next = tcb + 1; next 1770 drivers/net/ethernet/agere/et131x.c tcb->next = NULL; next 2020 drivers/net/ethernet/agere/et131x.c rfd = list_entry(rx_ring->recv_list.next, next 2142 drivers/net/ethernet/agere/et131x.c struct fbr_desc *next; next 2149 drivers/net/ethernet/agere/et131x.c next = (struct fbr_desc *)(fbr->ring_virtaddr) + next 2156 drivers/net/ethernet/agere/et131x.c next->addr_hi = fbr->bus_high[buff_index]; next 2157 drivers/net/ethernet/agere/et131x.c next->addr_lo = fbr->bus_low[buff_index]; next 2158 drivers/net/ethernet/agere/et131x.c next->word2 = buff_index; next 2248 drivers/net/ethernet/agere/et131x.c element = rx_local->recv_list.next; next 2563 drivers/net/ethernet/agere/et131x.c tx_ring->send_tail->next = tcb; next 2569 drivers/net/ethernet/agere/et131x.c WARN_ON(tcb->next != NULL); next 2608 drivers/net/ethernet/agere/et131x.c tx_ring->tcb_qhead = tcb->next; next 2616 drivers/net/ethernet/agere/et131x.c tcb->next = NULL; next 2624 drivers/net/ethernet/agere/et131x.c tx_ring->tcb_qtail->next = tcb; next 2684 drivers/net/ethernet/agere/et131x.c tx_ring->tcb_qtail->next = tcb; next 2708 drivers/net/ethernet/agere/et131x.c struct tcb *next = tcb->next; next 2710 drivers/net/ethernet/agere/et131x.c tx_ring->send_head = next; next 2712 drivers/net/ethernet/agere/et131x.c if (next == NULL) next 2761 drivers/net/ethernet/agere/et131x.c tx_ring->send_head = tcb->next; next 2762 drivers/net/ethernet/agere/et131x.c if (tcb->next == NULL) next 2776 drivers/net/ethernet/agere/et131x.c tx_ring->send_head = tcb->next; next 2777 drivers/net/ethernet/agere/et131x.c if (tcb->next == NULL) next 683 drivers/net/ethernet/alteon/acenic.h struct net_device *next; next 320 drivers/net/ethernet/altera/altera_sgdma.c csrwr32(lower_32_bits(ndesc_phys), desc, sgdma_descroffs(next)); next 427 drivers/net/ethernet/altera/altera_sgdma.c entry = list_entry((list)->next, type, member); \ next 436 drivers/net/ethernet/altera/altera_sgdma.c entry = list_entry((list)->next, type, member); \ next 15 drivers/net/ethernet/altera/altera_sgdmahw.h u32 next; next 769 drivers/net/ethernet/amd/amd8111e.h struct net_device *next; next 251 drivers/net/ethernet/amd/declance.c struct net_device *next; next 1048 drivers/net/ethernet/amd/declance.c dev = lp->next; next 1266 drivers/net/ethernet/amd/declance.c lp->next = root_lance_dev; next 1309 drivers/net/ethernet/amd/declance.c root_lance_dev = lp->next; next 298 drivers/net/ethernet/amd/pcnet32.c struct net_device *next; next 1991 drivers/net/ethernet/amd/pcnet32.c lp->next = pcnet32_dev; next 3036 drivers/net/ethernet/amd/pcnet32.c next_dev = lp->next; next 818 drivers/net/ethernet/apple/bmac.c unsigned int counter, cur = curval, next = nxtval; next 822 drivers/net/ethernet/apple/bmac.c next = ((next & 0x00FF) << 8) | (next >> 8); next 832 drivers/net/ethernet/apple/bmac.c if ((next & 0x0001) == 0) low_data_set = 0; next 835 drivers/net/ethernet/apple/bmac.c next = next >> 1; next 535 drivers/net/ethernet/apple/mace.c int fill, next, len; next 540 drivers/net/ethernet/apple/mace.c next = fill + 1; next 541 drivers/net/ethernet/apple/mace.c if (next >= N_TX_RING) next 542 drivers/net/ethernet/apple/mace.c next = 0; next 543 drivers/net/ethernet/apple/mace.c if (next == mp->tx_empty) { next 562 drivers/net/ethernet/apple/mace.c np = mp->tx_cmds + NCMDS_TX * next; next 567 drivers/net/ethernet/apple/mace.c mp->tx_fill = next; next 575 drivers/net/ethernet/apple/mace.c if (++next >= N_TX_RING) next 576 drivers/net/ethernet/apple/mace.c next = 0; next 577 drivers/net/ethernet/apple/mace.c if (next == mp->tx_empty) next 880 drivers/net/ethernet/apple/mace.c int i, nb, stat, next; next 892 drivers/net/ethernet/apple/mace.c next = i + 1; next 893 drivers/net/ethernet/apple/mace.c if (next >= N_RX_RING) next 894 drivers/net/ethernet/apple/mace.c next = 0; next 895 drivers/net/ethernet/apple/mace.c np = mp->rx_cmds + next; next 896 drivers/net/ethernet/apple/mace.c if (next != mp->rx_fill && next 949 drivers/net/ethernet/apple/mace.c next = i + 1; next 950 drivers/net/ethernet/apple/mace.c if (next >= N_RX_RING) next 951 drivers/net/ethernet/apple/mace.c next = 0; next 952 drivers/net/ethernet/apple/mace.c if (next == mp->rx_empty) next 975 drivers/net/ethernet/apple/mace.c i = next; next 308 drivers/net/ethernet/aquantia/atlantic/aq_ring.c next_ = buff_->next, next 330 drivers/net/ethernet/aquantia/atlantic/aq_ring.c next_ = buff_->next, next 390 drivers/net/ethernet/aquantia/atlantic/aq_ring.c next_ = buff_->next, next 43 drivers/net/ethernet/aquantia/atlantic/aq_ring.h u16 next; next 705 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c buff->next = 0U; next 709 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c buff->next = aq_ring_next_dx(ring, next 759 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c buff->next = 0U; next 769 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c buff->next = rxd_wb->next_desc_ptr; next 773 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c buff->next = next 243 drivers/net/ethernet/atheros/ag71xx.c u32 next; next 997 drivers/net/ethernet/atheros/ag71xx.c desc->next = (u32)(ring->descs_dma + next 1070 drivers/net/ethernet/atheros/ag71xx.c desc->next = (u32)(ring->descs_dma + next 1074 drivers/net/ethernet/atheros/ag71xx.c desc, desc->next); next 1440 drivers/net/ethernet/atheros/ag71xx.c struct sk_buff *next, *skb; next 1483 drivers/net/ethernet/atheros/ag71xx.c goto next; next 1498 drivers/net/ethernet/atheros/ag71xx.c next: next 1507 drivers/net/ethernet/atheros/ag71xx.c list_for_each_entry_safe(skb, next, &rx_list, list) next 1737 drivers/net/ethernet/atheros/ag71xx.c ag->stop_desc->next = (u32)ag->stop_desc_dma; next 78 drivers/net/ethernet/atheros/alx/main.c u16 cur, next, count = 0; next 80 drivers/net/ethernet/atheros/alx/main.c next = cur = rxq->write_idx; next 81 drivers/net/ethernet/atheros/alx/main.c if (++next == alx->rx_ringsz) next 82 drivers/net/ethernet/atheros/alx/main.c next = 0; next 85 drivers/net/ethernet/atheros/alx/main.c while (!cur_buf->skb && next != rxq->read_idx) { next 125 drivers/net/ethernet/atheros/alx/main.c cur = next; next 126 drivers/net/ethernet/atheros/alx/main.c if (++next == alx->rx_ringsz) next 127 drivers/net/ethernet/atheros/alx/main.c next = 0; next 1853 drivers/net/ethernet/atheros/atlx/atl1.c goto next; next 1878 drivers/net/ethernet/atheros/atlx/atl1.c next: next 290 drivers/net/ethernet/aurora/nb8800.c unsigned int next; next 299 drivers/net/ethernet/aurora/nb8800.c next = (last + 1) % RX_DESC_COUNT; next 301 drivers/net/ethernet/aurora/nb8800.c rxd = &priv->rx_descs[next]; next 311 drivers/net/ethernet/aurora/nb8800.c nb8800_receive(dev, next, len); next 320 drivers/net/ethernet/aurora/nb8800.c last = next; next 339 drivers/net/ethernet/aurora/nb8800.c if (priv->rx_descs[next].report) next 396 drivers/net/ethernet/aurora/nb8800.c unsigned int next; next 423 drivers/net/ethernet/aurora/nb8800.c next = priv->tx_next; next 424 drivers/net/ethernet/aurora/nb8800.c txb = &priv->tx_bufs[next]; next 425 drivers/net/ethernet/aurora/nb8800.c txd = &priv->tx_descs[next]; next 428 drivers/net/ethernet/aurora/nb8800.c next = (next + 1) % TX_DESC_COUNT; next 442 drivers/net/ethernet/aurora/nb8800.c desc->n_addr = priv->tx_bufs[next].dma_desc; next 461 drivers/net/ethernet/aurora/nb8800.c priv->tx_next = next; next 767 drivers/net/ethernet/broadcom/bcmsysport.c goto next; next 786 drivers/net/ethernet/broadcom/bcmsysport.c goto next; next 794 drivers/net/ethernet/broadcom/bcmsysport.c goto next; next 804 drivers/net/ethernet/broadcom/bcmsysport.c goto next; next 836 drivers/net/ethernet/broadcom/bcmsysport.c next: next 534 drivers/net/ethernet/broadcom/bnx2x/bnx2x_init.h u64 next; next 921 drivers/net/ethernet/broadcom/bnx2x/bnx2x_init_ops.h t2[i].next = (u64)(t2_mapping + next 610 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c u8 *next = base; next 622 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c memcpy(next, &pos->u, size); next 625 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c counter, next); next 626 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c next += stride + size; next 780 drivers/net/ethernet/broadcom/bnxt/bnxt.c u16 next, max = rxr->rx_agg_bmap_size; next 782 drivers/net/ethernet/broadcom/bnxt/bnxt.c next = find_next_zero_bit(rxr->rx_agg_bmap, max, idx); next 783 drivers/net/ethernet/broadcom/bnxt/bnxt.c if (next >= max) next 784 drivers/net/ethernet/broadcom/bnxt/bnxt.c next = find_first_zero_bit(rxr->rx_agg_bmap, max); next 785 drivers/net/ethernet/broadcom/bnxt/bnxt.c return next; next 657 drivers/net/ethernet/broadcom/cnic.c u32 next) next 661 drivers/net/ethernet/broadcom/cnic.c id_tbl->next = next; next 699 drivers/net/ethernet/broadcom/cnic.c id = find_next_zero_bit(id_tbl->table, id_tbl->max, id_tbl->next); next 702 drivers/net/ethernet/broadcom/cnic.c if (id_tbl->next != 0) { next 703 drivers/net/ethernet/broadcom/cnic.c id = find_first_zero_bit(id_tbl->table, id_tbl->next); next 704 drivers/net/ethernet/broadcom/cnic.c if (id >= id_tbl->next) next 711 drivers/net/ethernet/broadcom/cnic.c id_tbl->next = (id + 1) & (id_tbl->max - 1); next 1009 drivers/net/ethernet/broadcom/cnic.c struct bnx2x_bd_chain_next *next = next 1015 drivers/net/ethernet/broadcom/cnic.c next->addr_hi = (u64) info->dma.pg_map_arr[j] >> 32; next 1016 drivers/net/ethernet/broadcom/cnic.c next->addr_lo = info->dma.pg_map_arr[j] & 0xffffffff; next 5757 drivers/net/ethernet/broadcom/cnic.c udev = list_entry(cnic_udev_list.next, struct cnic_uio_dev, next 144 drivers/net/ethernet/broadcom/cnic.h u32 next; next 1793 drivers/net/ethernet/broadcom/genet/bcmgenet.c goto next; next 1822 drivers/net/ethernet/broadcom/genet/bcmgenet.c goto next; next 1843 drivers/net/ethernet/broadcom/genet/bcmgenet.c goto next; next 1880 drivers/net/ethernet/broadcom/genet/bcmgenet.c next: next 7903 drivers/net/ethernet/broadcom/tg3.c segs = segs->next; next 7904 drivers/net/ethernet/broadcom/tg3.c nskb->next = NULL; next 686 drivers/net/ethernet/brocade/bna/bnad.c goto next; next 719 drivers/net/ethernet/brocade/bna/bnad.c next: next 1045 drivers/net/ethernet/cavium/liquidio/octeon_device.c while (dispatch->next != dispatch) { next 1046 drivers/net/ethernet/cavium/liquidio/octeon_device.c temp = dispatch->next; next 615 drivers/net/ethernet/cavium/liquidio/octeon_network.h if (root->prev == root && root->next == root) next 618 drivers/net/ethernet/cavium/liquidio/octeon_network.h node = root->next; next 97 drivers/net/ethernet/cavium/thunder/nicvf_queues.c struct pgcache *pgcache, *next; next 166 drivers/net/ethernet/cavium/thunder/nicvf_queues.c next = &rbdr->pgcache[rbdr->pgidx]; next 167 drivers/net/ethernet/cavium/thunder/nicvf_queues.c page = next->page; next 595 drivers/net/ethernet/cavium/thunder/nicvf_queues.c goto next; next 599 drivers/net/ethernet/cavium/thunder/nicvf_queues.c goto next; next 617 drivers/net/ethernet/cavium/thunder/nicvf_queues.c next: next 83 drivers/net/ethernet/chelsio/cxgb3/cxgb3_defs.h if ((void *)e->next >= (void *)t->tid_tab && next 84 drivers/net/ethernet/chelsio/cxgb3/cxgb3_defs.h (void *)e->next < (void *)&t->atid_tab[t->natids]) next 102 drivers/net/ethernet/chelsio/cxgb3/cxgb3_defs.h if ((void *)e->next >= (void *)t->tid_tab && next 103 drivers/net/ethernet/chelsio/cxgb3/cxgb3_defs.h (void *)e->next < (void *)&t->atid_tab[t->natids]) next 508 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c p->next = t->afree; next 527 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c p->next = t->sfree; next 663 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c t->afree = p->next; next 685 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c t->sfree = p->next; next 1187 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c t->stid_tab[nstids - 1].next = &t->stid_tab[nstids]; next 1192 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c t->atid_tab[natids - 1].next = &t->atid_tab[natids]; next 144 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.h union listen_entry *next; next 149 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.h union active_open_entry *next; next 243 drivers/net/ethernet/chelsio/cxgb3/l2t.c for (p = &d->l2tab[hash].first; *p; p = &(*p)->next) next 245 drivers/net/ethernet/chelsio/cxgb3/l2t.c *p = e->next; next 334 drivers/net/ethernet/chelsio/cxgb3/l2t.c for (e = d->l2tab[hash].first; e; e = e->next) next 347 drivers/net/ethernet/chelsio/cxgb3/l2t.c e->next = d->l2tab[hash].first; next 409 drivers/net/ethernet/chelsio/cxgb3/l2t.c for (e = d->l2tab[hash].first; e; e = e->next) next 66 drivers/net/ethernet/chelsio/cxgb3/l2t.h struct l2t_entry *next; /* next l2t_entry on chain */ next 90 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c .next = seq_tab_next, next 1072 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c .next = devlog_next, next 1216 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c .next = mboxlog_next, next 1920 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c .next = mps_tcam_next, next 2552 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c .next = dcb_info_next, next 3038 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c .next = sge_queue_next, next 1224 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c t->afree = p->next; next 1241 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c p->next = t->afree; next 1498 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c t->atid_tab[natids - 1].next = &t->atid_tab[natids]; next 216 drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c const struct cxgb4_next_header *next; next 229 drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c next = is_ipv6 ? cxgb4_ipv6_jumps : cxgb4_ipv4_jumps; next 232 drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c for (i = 0; next[i].jump; i++) { next 233 drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c if (next[i].offoff != cls->knode.sel->offoff || next 234 drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c next[i].shift != cls->knode.sel->offshift || next 235 drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c next[i].mask != cls->knode.sel->offmask || next 236 drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c next[i].offset != cls->knode.sel->off) next 248 drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c if (next[i].match_off == off && next 249 drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c next[i].match_val == val && next 250 drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c next[i].match_mask == mask) { next 270 drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c link->match_field = next[i].jump; next 89 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h union aopen_entry *next; next 287 drivers/net/ethernet/chelsio/cxgb4/l2t.c for (p = &d->l2tab[e->hash].first; *p; p = &(*p)->next) next 289 drivers/net/ethernet/chelsio/cxgb4/l2t.c *p = e->next; next 290 drivers/net/ethernet/chelsio/cxgb4/l2t.c e->next = NULL; next 329 drivers/net/ethernet/chelsio/cxgb4/l2t.c for (p = &d->l2tab[e->hash].first; *p; p = &(*p)->next) next 331 drivers/net/ethernet/chelsio/cxgb4/l2t.c *p = e->next; next 332 drivers/net/ethernet/chelsio/cxgb4/l2t.c e->next = NULL; next 444 drivers/net/ethernet/chelsio/cxgb4/l2t.c for (e = d->l2tab[hash].first; e; e = e->next) next 468 drivers/net/ethernet/chelsio/cxgb4/l2t.c e->next = d->l2tab[hash].first; next 545 drivers/net/ethernet/chelsio/cxgb4/l2t.c for (e = d->l2tab[hash].first; e; e = e->next) next 734 drivers/net/ethernet/chelsio/cxgb4/l2t.c .next = l2t_seq_next, next 80 drivers/net/ethernet/chelsio/cxgb4/l2t.h struct l2t_entry *next; /* next l2t_entry on chain */ next 2023 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c .next = mboxlog_next, next 2177 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c .next = sge_queue_next, next 2323 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c .next = sge_qstats_next, next 2421 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c .next = interfaces_next, next 135 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c pool->next, count, 0); next 137 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c pool->next = 0; next 142 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c pool->next = i + count; next 143 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c if (pool->next >= ppm->pool_index_max) next 144 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c pool->next = 0; next 150 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c pool->next); next 165 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c ppm->next, count, 0); next 167 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c ppm->next = 0; next 174 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c ppm->next = i + count; next 175 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c if (ppm->max_index_in_edram && (ppm->next >= ppm->max_index_in_edram)) next 176 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c ppm->next = 0; next 177 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c else if (ppm->next >= ppm->bmap_index_max) next 178 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c ppm->next = 0; next 183 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c __func__, i, count, i + ppm->pool_rsvd, ppm->next, next 207 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c if (i < pool->next) next 208 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c pool->next = i; next 212 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c __func__, cpu, i, pool->next); next 219 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c if (i < ppm->next) next 220 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c ppm->next = i; next 223 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c pr_debug("%s: idx %d, next %u.\n", __func__, i, ppm->next); next 271 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c nr_pages, npods, ppm->next, caller_data); next 479 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c ppm->next = 0; next 123 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.h unsigned int next; /* next possible free index */ next 145 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.h unsigned int next; next 844 drivers/net/ethernet/cisco/enic/enic_main.c while (!buf->os_buf && (buf->next != wq->to_clean)) { next 849 drivers/net/ethernet/cisco/enic/enic_main.c wq->to_use = buf->next; next 50 drivers/net/ethernet/cisco/enic/vnic_rq.c buf->next = rq->bufs[0]; next 53 drivers/net/ethernet/cisco/enic/vnic_rq.c buf->next = rq->bufs[i + 1]; next 55 drivers/net/ethernet/cisco/enic/vnic_rq.c buf->next = buf + 1; next 196 drivers/net/ethernet/cisco/enic/vnic_rq.c buf = buf->next; next 69 drivers/net/ethernet/cisco/enic/vnic_rq.h struct vnic_rq_buf *next; next 132 drivers/net/ethernet/cisco/enic/vnic_rq.h buf = buf->next; next 183 drivers/net/ethernet/cisco/enic/vnic_rq.h rq->to_clean = buf->next; next 50 drivers/net/ethernet/cisco/enic/vnic_wq.c buf->next = wq->bufs[0]; next 51 drivers/net/ethernet/cisco/enic/vnic_wq.c buf->next->prev = buf; next 54 drivers/net/ethernet/cisco/enic/vnic_wq.c buf->next = wq->bufs[i + 1]; next 55 drivers/net/ethernet/cisco/enic/vnic_wq.c buf->next->prev = buf; next 57 drivers/net/ethernet/cisco/enic/vnic_wq.c buf->next = buf + 1; next 58 drivers/net/ethernet/cisco/enic/vnic_wq.c buf->next->prev = buf; next 206 drivers/net/ethernet/cisco/enic/vnic_wq.c buf = wq->to_clean = buf->next; next 54 drivers/net/ethernet/cisco/enic/vnic_wq.h struct vnic_wq_buf *next; next 148 drivers/net/ethernet/cisco/enic/vnic_wq.h buf = buf->next; next 169 drivers/net/ethernet/cisco/enic/vnic_wq.h wq->to_clean = buf->next; next 557 drivers/net/ethernet/dec/tulip/de2104x.c goto next; next 562 drivers/net/ethernet/dec/tulip/de2104x.c goto next; next 591 drivers/net/ethernet/dec/tulip/de2104x.c next: next 762 drivers/net/ethernet/dec/tulip/de4x5.c __le32 next; next 1189 drivers/net/ethernet/dec/tulip/de4x5.c lp->rx_ring[i].next = 0; next 1207 drivers/net/ethernet/dec/tulip/de4x5.c lp->rx_ring[i].next = 0; next 389 drivers/net/ethernet/dec/tulip/tulip.h struct mediainfo *next; next 341 drivers/net/ethernet/dlink/dl2k.c int cid, next; next 374 drivers/net/ethernet/dlink/dl2k.c next = psib[i++]; next 375 drivers/net/ethernet/dlink/dl2k.c if ((cid == 0 && next == 0) || (cid == 0xff && next == 0xff)) { next 392 drivers/net/ethernet/dlink/dl2k.c j = (next - i > 255) ? 255 : next - i; next 404 drivers/net/ethernet/dlink/dl2k.c i = next; next 76 drivers/net/ethernet/ec_bhf.c __le32 next; next 379 drivers/net/ethernet/ec_bhf.c u32 next; next 382 drivers/net/ethernet/ec_bhf.c next = (u8 *)(desc + 1) - priv->rx_buf.buf; next 384 drivers/net/ethernet/ec_bhf.c next = 0; next 385 drivers/net/ethernet/ec_bhf.c next |= RXHDR_NEXT_VALID; next 386 drivers/net/ethernet/ec_bhf.c desc->header.next = cpu_to_le32(next); next 1320 drivers/net/ethernet/fealnx.c struct fealnx_desc *next; next 1330 drivers/net/ethernet/fealnx.c next = np->cur_tx_copy->next_desc_logical; next 1331 drivers/net/ethernet/fealnx.c next->skbuff = skb; next 1332 drivers/net/ethernet/fealnx.c next->control = TXIC | TXLD | CRCEnable | PADEnable; next 1333 drivers/net/ethernet/fealnx.c next->control |= (skb->len << PKTSShift); /* pkt size */ next 1334 drivers/net/ethernet/fealnx.c next->control |= ((skb->len - BPT) << TBSShift); /* buf size */ next 1338 drivers/net/ethernet/fealnx.c next->buffer = pci_map_single(ep->pci_dev, skb->data + BPT, next 1341 drivers/net/ethernet/fealnx.c next->status = TXOWN; next 1344 drivers/net/ethernet/fealnx.c np->cur_tx_copy = next->next_desc_logical; next 1493 drivers/net/ethernet/fealnx.c struct fealnx_desc *next; next 1495 drivers/net/ethernet/fealnx.c next = np->cur_tx->next_desc_logical; next 1496 drivers/net/ethernet/fealnx.c tx_status = next->status; next 1497 drivers/net/ethernet/fealnx.c tx_control = next->control; next 218 drivers/net/ethernet/freescale/fman/fman_mac.h hash_entry = ETH_HASH_ENTRY_OBJ(addr_lst->next); next 2188 drivers/net/ethernet/freescale/gianfar.c struct txbd8 *bdp, *next = NULL; next 2232 drivers/net/ethernet/freescale/gianfar.c next = next_txbd(bdp, base, tx_ring_size); next 2233 drivers/net/ethernet/freescale/gianfar.c buflen = be16_to_cpu(next->length) + next 2251 drivers/net/ethernet/freescale/gianfar.c bdp = next; next 196 drivers/net/ethernet/freescale/ucc_geth.c struct list_head *node = lh->next; next 280 drivers/net/ethernet/hisilicon/hisi_femac.c goto next; next 287 drivers/net/ethernet/hisilicon/hisi_femac.c next: next 547 drivers/net/ethernet/hisilicon/hix5hd2_gmac.c goto next; next 554 drivers/net/ethernet/hisilicon/hix5hd2_gmac.c next: next 125 drivers/net/ethernet/hisilicon/hns/hns_dsaf_ppe.c ppe_cb->next = NULL; next 74 drivers/net/ethernet/hisilicon/hns/hns_dsaf_ppe.h struct hns_ppe_cb *next; /* pointer to next ppe device */ next 205 drivers/net/ethernet/hisilicon/hns3/hnae3.h struct hnae3_ring_chain_node *next; next 438 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c ring = ring->next; next 2726 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c ring->tail_skb->next = new_skb; next 3206 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c cur_chain->next = NULL; next 3208 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c while (tx_ring->next) { next 3209 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c tx_ring = tx_ring->next; next 3216 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c cur_chain->next = chain; next 3231 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c cur_chain->next = NULL; next 3238 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c rx_ring = rx_ring->next; next 3246 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c cur_chain->next = chain; next 3255 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c rx_ring = rx_ring->next; next 3261 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c cur_chain = head->next; next 3263 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c chain = cur_chain->next; next 3267 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c head->next = NULL; next 3278 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c chain = head->next; next 3281 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c chain_tmp = chain->next; next 3290 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c ring->next = group->ring; next 405 drivers/net/ethernet/hisilicon/hns3/hns3_enet.h struct hns3_enet_ring *next; next 636 drivers/net/ethernet/hisilicon/hns3/hns3_enet.h for (pos = (head).ring; pos; pos = pos->next) next 4508 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c for (node = ring_chain; node; node = node->next) { next 113 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_mbx.c chain = head->next; next 116 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_mbx.c chain_tmp = chain->next; next 176 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_mbx.c cur_chain->next = new_chain; next 1008 drivers/net/ethernet/hisilicon/hns3/hns3vf/hclgevf_main.c for (node = ring_chain; node; node = node->next) { next 1031 drivers/net/ethernet/hisilicon/hns3/hns3vf/hclgevf_main.c !node->next) { next 1174 drivers/net/ethernet/hp/hp100.c lp->rxring[i].next = ringptr; next 1183 drivers/net/ethernet/hp/hp100.c lp->txring[i].next = ringptr; next 1359 drivers/net/ethernet/hp/hp100.c lp->rxrtail = ringptr->next; next 1492 drivers/net/ethernet/hp/hp100.c if (lp->txrtail->next == lp->txrhead) { next 1544 drivers/net/ethernet/hp/hp100.c lp->txrtail = ringptr->next; next 1617 drivers/net/ethernet/hp/hp100.c lp->txrhead = lp->txrhead->next; next 1945 drivers/net/ethernet/hp/hp100.c lp->rxrhead = lp->rxrhead->next; next 1956 drivers/net/ethernet/hp/hp100.c lp->rxrtail = lp->rxrtail->next; next 544 drivers/net/ethernet/hp/hp100.h struct hp100_ring *next; next 307 drivers/net/ethernet/huawei/hinic/hinic_rx.c curr_skb->next = skb; next 212 drivers/net/ethernet/i825xx/82596.c struct i596_tbd *next; next 1080 drivers/net/ethernet/i825xx/82596.c tbd->next = I596_NULL; next 446 drivers/net/ethernet/i825xx/ether1.c int i, status, addr, next, next2; next 479 drivers/net/ethernet/i825xx/ether1.c next = addr + RFD_SIZE + RBD_SIZE + ETH_FRAME_LEN + 10; next 480 drivers/net/ethernet/i825xx/ether1.c next2 = next + RFD_SIZE + RBD_SIZE + ETH_FRAME_LEN + 10; next 483 drivers/net/ethernet/i825xx/ether1.c next = RX_AREA_START; next 492 drivers/net/ethernet/i825xx/ether1.c init_rfd.rfd_link = next; next 493 drivers/net/ethernet/i825xx/ether1.c init_rbd.rbd_link = next + RFD_SIZE; next 498 drivers/net/ethernet/i825xx/ether1.c addr = next; next 178 drivers/net/ethernet/i825xx/lib82596.c u32 next; next 1002 drivers/net/ethernet/i825xx/lib82596.c tbd->next = I596_NULL; next 598 drivers/net/ethernet/i825xx/sun3_82586.c p->xmit_buffs[i]->next = 0xffff; next 650 drivers/net/ethernet/i825xx/sun3_82586.c rfd[i].next = make16(rfd + (i+1) % (p->num_recv_buffs+rfdadd) ); next 665 drivers/net/ethernet/i825xx/sun3_82586.c rbd[i].next = make16((rbd + (i+1) % p->num_recv_buffs)); next 808 drivers/net/ethernet/i825xx/sun3_82586.c rbd = (struct rbd_struct *) make32(rbd->next); next 826 drivers/net/ethernet/i825xx/sun3_82586.c p->rfd_top = (struct rfd_struct *) make32(p->rfd_top->next); /* step to next RFD */ next 866 drivers/net/ethernet/i825xx/sun3_82586.c rfds = (struct rfd_struct *) make32(rfds->next); next 134 drivers/net/ethernet/i825xx/sun3_82586.h unsigned short next; /* linkoffset to next RFD */ next 165 drivers/net/ethernet/i825xx/sun3_82586.h unsigned short next; /* pointeroffset to next RBD */ next 310 drivers/net/ethernet/i825xx/sun3_82586.h unsigned short next; /* pointeroffset to next TBD */ next 1779 drivers/net/ethernet/ibm/emac/core.c goto next; next 1785 drivers/net/ethernet/ibm/emac/core.c goto next; next 1810 drivers/net/ethernet/ibm/emac/core.c next: next 1850 drivers/net/ethernet/ibm/emac/core.c goto next; next 2224 drivers/net/ethernet/ibm/ibmvnic.c union sub_crq *next; next 2238 drivers/net/ethernet/ibm/ibmvnic.c next = ibmvnic_next_scrq(adapter, adapter->rx_scrq[scrq_num]); next 2240 drivers/net/ethernet/ibm/ibmvnic.c (struct ibmvnic_rx_buff *)be64_to_cpu(next-> next 2243 drivers/net/ethernet/ibm/ibmvnic.c if (next->rx_comp.rc) { next 2245 drivers/net/ethernet/ibm/ibmvnic.c be16_to_cpu(next->rx_comp.rc)); next 2247 drivers/net/ethernet/ibm/ibmvnic.c next->rx_comp.first = 0; next 2253 drivers/net/ethernet/ibm/ibmvnic.c next->rx_comp.first = 0; next 2258 drivers/net/ethernet/ibm/ibmvnic.c length = be32_to_cpu(next->rx_comp.len); next 2259 drivers/net/ethernet/ibm/ibmvnic.c offset = be16_to_cpu(next->rx_comp.off_frame_data); next 2260 drivers/net/ethernet/ibm/ibmvnic.c flags = next->rx_comp.flags; next 2271 drivers/net/ethernet/ibm/ibmvnic.c ntohs(next->rx_comp.vlan_tci)); next 2274 drivers/net/ethernet/ibm/ibmvnic.c next->rx_comp.first = 0; next 2907 drivers/net/ethernet/ibm/ibmvnic.c union sub_crq *next; next 2916 drivers/net/ethernet/ibm/ibmvnic.c next = ibmvnic_next_scrq(adapter, scrq); next 2917 drivers/net/ethernet/ibm/ibmvnic.c for (i = 0; i < next->tx_comp.num_comps; i++) { next 2918 drivers/net/ethernet/ibm/ibmvnic.c if (next->tx_comp.rcs[i]) { next 2920 drivers/net/ethernet/ibm/ibmvnic.c next->tx_comp.rcs[i]); next 2923 drivers/net/ethernet/ibm/ibmvnic.c index = be32_to_cpu(next->tx_comp.correlators[i]); next 2953 drivers/net/ethernet/ibm/ibmvnic.c next->tx_comp.first = 0; next 420 drivers/net/ethernet/intel/e100.c struct rx *next, *prev; next 499 drivers/net/ethernet/intel/e100.c struct cb *next, *prev; next 862 drivers/net/ethernet/intel/e100.c nic->cb_to_use = cb->next; next 895 drivers/net/ethernet/intel/e100.c nic->cb_to_send = nic->cb_to_send->next; next 1823 drivers/net/ethernet/intel/e100.c cb = nic->cb_to_clean = cb->next) { next 1867 drivers/net/ethernet/intel/e100.c nic->cb_to_clean = nic->cb_to_clean->next; next 1894 drivers/net/ethernet/intel/e100.c cb->next = (i + 1 < count) ? cb + 1 : nic->cbs; next 2062 drivers/net/ethernet/intel/e100.c for (rx = nic->rx_to_clean; rx->skb; rx = nic->rx_to_clean = rx->next) { next 2083 drivers/net/ethernet/intel/e100.c for (rx = nic->rx_to_use; !rx->skb; rx = nic->rx_to_use = rx->next) { next 2165 drivers/net/ethernet/intel/e100.c rx->next = (i + 1 < count) ? rx + 1 : nic->rxs; next 101 drivers/net/ethernet/intel/fm10k/fm10k_debugfs.c .next = fm10k_dbg_desc_seq_next, next 108 drivers/net/ethernet/intel/fm10k/fm10k_debugfs.c .next = fm10k_dbg_desc_seq_next, next 2010 drivers/net/ethernet/intel/i40e/i40e_main.c struct i40e_new_mac_filter *i40e_next_filter(struct i40e_new_mac_filter *next) next 2012 drivers/net/ethernet/intel/i40e/i40e_main.c hlist_for_each_entry_continue(next, hlist) { next 2013 drivers/net/ethernet/intel/i40e/i40e_main.c if (!is_broadcast_ether_addr(next->f->macaddr)) next 2014 drivers/net/ethernet/intel/i40e/i40e_main.c return next; next 4168 drivers/net/ethernet/intel/i40e/i40e_main.c tx_ring->next = q_vector->tx.ring; next 4177 drivers/net/ethernet/intel/i40e/i40e_main.c xdp_ring->next = q_vector->tx.ring; next 4183 drivers/net/ethernet/intel/i40e/i40e_main.c rx_ring->next = q_vector->rx.ring; next 4649 drivers/net/ethernet/intel/i40e/i40e_main.c u32 next; next 4665 drivers/net/ethernet/intel/i40e/i40e_main.c next = (val & I40E_QINT_TQCTL_NEXTQ_INDX_MASK) next 4677 drivers/net/ethernet/intel/i40e/i40e_main.c qp = next; next 353 drivers/net/ethernet/intel/i40e/i40e_txrx.h struct i40e_ring *next; /* pointer to next ring in q_vector */ next 469 drivers/net/ethernet/intel/i40e/i40e_txrx.h for (pos = (head).ring; pos != NULL; pos = pos->next) next 309 drivers/net/ethernet/intel/iavf/iavf_main.c rx_ring->next = q_vector->rx.ring; next 335 drivers/net/ethernet/intel/iavf/iavf_main.c tx_ring->next = q_vector->tx.ring; next 327 drivers/net/ethernet/intel/iavf/iavf_txrx.h struct iavf_ring *next; /* pointer to next ring in q_vector */ next 426 drivers/net/ethernet/intel/iavf/iavf_txrx.h for (pos = (head).ring; pos != NULL; pos = pos->next) next 397 drivers/net/ethernet/intel/ice/ice_lib.c int next; next 400 drivers/net/ethernet/intel/ice/ice_lib.c next = curr + 1; next 407 drivers/net/ethernet/intel/ice/ice_lib.c next = ICE_NO_VSI; next 409 drivers/net/ethernet/intel/ice/ice_lib.c next = i; next 411 drivers/net/ethernet/intel/ice/ice_lib.c return next; next 1380 drivers/net/ethernet/intel/ice/ice_lib.c tx_ring->next = q_vector->tx.ring; next 1396 drivers/net/ethernet/intel/ice/ice_lib.c rx_ring->next = q_vector->rx.ring; next 167 drivers/net/ethernet/intel/ice/ice_txrx.h struct ice_ring *next; /* pointer to next ring in q_vector */ next 231 drivers/net/ethernet/intel/ice/ice_txrx.h for (pos = (head).ring; pos; pos = pos->next) next 204 drivers/net/ethernet/intel/igb/igb_main.c .next = NULL, next 309 drivers/net/ethernet/intel/ixgbe/ixgbe.h struct ixgbe_ring *next; /* pointer to next ring in q_vector */ next 438 drivers/net/ethernet/intel/ixgbe/ixgbe.h for (pos = (head).ring; pos != NULL; pos = pos->next) next 809 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c ring->next = head->ring; next 143 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c .next = NULL, next 91 drivers/net/ethernet/intel/ixgbevf/ixgbevf.h struct ixgbevf_ring *next; next 226 drivers/net/ethernet/intel/ixgbevf/ixgbevf.h for (pos = (head).ring; pos != NULL; pos = pos->next) next 2676 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c ring->next = head->ring; next 375 drivers/net/ethernet/korina.c goto next; next 392 drivers/net/ethernet/korina.c goto next; next 424 drivers/net/ethernet/korina.c next: next 1623 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c u16 pcifunc, int next, bool eol) next 1638 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c aq_req.mce.next = next; next 1746 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c next_mce = hlist_entry(mce->node.next, next 1346 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c u16 index, next, end; next 1356 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c next = find_next_bit(map, end, index); next 1357 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c if (*max_area < (next - index)) { next 1358 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c *max_area = next - index; next 1362 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c if (next < end) { next 1363 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c start = next + 1; next 1375 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c u16 index, next; next 1386 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c next = find_next_bit(map, end, index); next 1387 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c if (next <= end) { next 1388 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c fcnt += next - index; next 1389 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c start = next + 1; next 829 drivers/net/ethernet/marvell/octeontx2/af/rvu_struct.h uint64_t next : 16; next 843 drivers/net/ethernet/marvell/octeontx2/af/rvu_struct.h uint64_t next : 16; next 922 drivers/net/ethernet/marvell/skge.c e->next = ring->start; next 925 drivers/net/ethernet/marvell/skge.c e->next = e + 1; next 999 drivers/net/ethernet/marvell/skge.c } while ((e = e->next) != ring->start); next 1026 drivers/net/ethernet/marvell/skge.c } while ((e = e->next) != ring->start); next 2793 drivers/net/ethernet/marvell/skge.c e = e->next; next 2820 drivers/net/ethernet/marvell/skge.c skge->tx_ring.to_use = e->next; next 2837 drivers/net/ethernet/marvell/skge.c e = e->next; next 2873 drivers/net/ethernet/marvell/skge.c for (e = skge->tx_ring.to_clean; e != skge->tx_ring.to_use; e = e->next) { next 3155 drivers/net/ethernet/marvell/skge.c for (e = ring->to_clean; e != ring->to_use; e = e->next) { next 3205 drivers/net/ethernet/marvell/skge.c for (e = ring->to_clean; prefetch(e->next), work_done < budget; e = e->next) { next 3702 drivers/net/ethernet/marvell/skge.c for (e = skge->tx_ring.to_clean; e != skge->tx_ring.to_use; e = e->next) { next 3710 drivers/net/ethernet/marvell/skge.c for (e = skge->rx_ring.to_clean; ; e = e->next) { next 2395 drivers/net/ethernet/marvell/skge.h struct skge_element *next; next 1549 drivers/net/ethernet/mediatek/mtk_eth_soc.c int next = (i + 1) % MTK_DMA_SIZE; next 1550 drivers/net/ethernet/mediatek/mtk_eth_soc.c u32 next_ptr = ring->phys + next * sz; next 292 drivers/net/ethernet/mellanox/mlx4/alloc.c struct mlx4_zone_entry *next = list_first_entry(&entry->list, next 293 drivers/net/ethernet/mellanox/mlx4/alloc.c typeof(*next), next 296 drivers/net/ethernet/mellanox/mlx4/alloc.c if (next->priority == entry->priority) next 297 drivers/net/ethernet/mellanox/mlx4/alloc.c list_add_tail(&next->prio_list, &entry->prio_list); next 130 drivers/net/ethernet/mellanox/mlx4/cmd.c int next; next 352 drivers/net/ethernet/mellanox/mlx4/cmd.c cmd->free_head = context->next; next 399 drivers/net/ethernet/mellanox/mlx4/cmd.c context->next = cmd->free_head; next 692 drivers/net/ethernet/mellanox/mlx4/cmd.c cmd->free_head = context->next; next 766 drivers/net/ethernet/mellanox/mlx4/cmd.c context->next = cmd->free_head; next 2640 drivers/net/ethernet/mellanox/mlx4/cmd.c priv->cmd.context[i].next = i + 1; next 2648 drivers/net/ethernet/mellanox/mlx4/cmd.c priv->cmd.context[priv->cmd.max_cmds - 1].next = -1; next 160 drivers/net/ethernet/mellanox/mlx4/en_netdev.c struct list_head next; next 312 drivers/net/ethernet/mellanox/mlx4/en_netdev.c list_add_tail(&filter->next, &priv->filters); next 325 drivers/net/ethernet/mellanox/mlx4/en_netdev.c list_del(&filter->next); next 425 drivers/net/ethernet/mellanox/mlx4/en_netdev.c list_for_each_entry_safe(filter, tmp, &priv->filters, next) { next 426 drivers/net/ethernet/mellanox/mlx4/en_netdev.c list_move(&filter->next, &del_list); next 431 drivers/net/ethernet/mellanox/mlx4/en_netdev.c list_for_each_entry_safe(filter, tmp, &del_list, next) { next 444 drivers/net/ethernet/mellanox/mlx4/en_netdev.c list_for_each_entry_safe(filter, tmp, &priv->filters, next) { next 453 drivers/net/ethernet/mellanox/mlx4/en_netdev.c list_move(&filter->next, &del_list); next 461 drivers/net/ethernet/mellanox/mlx4/en_netdev.c if (last_filter && (&last_filter->next != priv->filters.next)) next 462 drivers/net/ethernet/mellanox/mlx4/en_netdev.c list_move(&priv->filters, &last_filter->next); next 466 drivers/net/ethernet/mellanox/mlx4/en_netdev.c list_for_each_entry_safe(filter, tmp, &del_list, next) next 135 drivers/net/ethernet/mellanox/mlx4/en_port.c static unsigned long en_stats_adder(__be64 *start, __be64 *next, int num) next 140 drivers/net/ethernet/mellanox/mlx4/en_port.c int offset = next - start; next 719 drivers/net/ethernet/mellanox/mlx4/en_rx.c goto next; next 723 drivers/net/ethernet/mellanox/mlx4/en_rx.c goto next; next 750 drivers/net/ethernet/mellanox/mlx4/en_rx.c goto next; next 757 drivers/net/ethernet/mellanox/mlx4/en_rx.c goto next; next 802 drivers/net/ethernet/mellanox/mlx4/en_rx.c goto next; next 815 drivers/net/ethernet/mellanox/mlx4/en_rx.c goto next; next 824 drivers/net/ethernet/mellanox/mlx4/en_rx.c goto next; next 896 drivers/net/ethernet/mellanox/mlx4/en_rx.c next: next 101 drivers/net/ethernet/mellanox/mlx4/icm.h NULL : list_entry(icm->chunk_list.next, next 114 drivers/net/ethernet/mellanox/mlx4/icm.h if (iter->chunk->list.next == &iter->icm->chunk_list) { next 119 drivers/net/ethernet/mellanox/mlx4/icm.h iter->chunk = list_entry(iter->chunk->list.next, next 184 drivers/net/ethernet/mellanox/mlx4/pd.c uar = list_entry(priv->bf_list.next, struct mlx4_uar, bf_list); next 186 drivers/net/ethernet/mellanox/mlx5/core/cmd.c struct mlx5_cmd_mailbox *next = msg->next; next 190 drivers/net/ethernet/mellanox/mlx5/core/cmd.c for (i = 0; i < n && next; i++) { next 191 drivers/net/ethernet/mellanox/mlx5/core/cmd.c calc_block_sig(next->buf); next 192 drivers/net/ethernet/mellanox/mlx5/core/cmd.c next = next->next; next 229 drivers/net/ethernet/mellanox/mlx5/core/cmd.c struct mlx5_cmd_mailbox *next = ent->out->next; next 239 drivers/net/ethernet/mellanox/mlx5/core/cmd.c for (i = 0; i < n && next; i++) { next 240 drivers/net/ethernet/mellanox/mlx5/core/cmd.c err = verify_block_sig(next->buf); next 244 drivers/net/ethernet/mellanox/mlx5/core/cmd.c next = next->next; next 774 drivers/net/ethernet/mellanox/mlx5/core/cmd.c struct mlx5_cmd_mailbox *next = msg->next; next 806 drivers/net/ethernet/mellanox/mlx5/core/cmd.c for (i = 0; i < n && next; i++) { next 809 drivers/net/ethernet/mellanox/mlx5/core/cmd.c dump_buf(next->buf, dump_len, 1, offset); next 813 drivers/net/ethernet/mellanox/mlx5/core/cmd.c dump_buf(next->buf, sizeof(struct mlx5_cmd_prot_block), 0, offset); next 816 drivers/net/ethernet/mellanox/mlx5/core/cmd.c next = next->next; next 897 drivers/net/ethernet/mellanox/mlx5/core/cmd.c if (ent->in->next) next 898 drivers/net/ethernet/mellanox/mlx5/core/cmd.c lay->in_ptr = cpu_to_be64(ent->in->next->dma); next 900 drivers/net/ethernet/mellanox/mlx5/core/cmd.c if (ent->out->next) next 901 drivers/net/ethernet/mellanox/mlx5/core/cmd.c lay->out_ptr = cpu_to_be64(ent->out->next->dma); next 1120 drivers/net/ethernet/mellanox/mlx5/core/cmd.c struct mlx5_cmd_mailbox *next; next 1131 drivers/net/ethernet/mellanox/mlx5/core/cmd.c next = to->next; next 1133 drivers/net/ethernet/mellanox/mlx5/core/cmd.c if (!next) { next 1139 drivers/net/ethernet/mellanox/mlx5/core/cmd.c block = next->buf; next 1144 drivers/net/ethernet/mellanox/mlx5/core/cmd.c next = next->next; next 1153 drivers/net/ethernet/mellanox/mlx5/core/cmd.c struct mlx5_cmd_mailbox *next; next 1164 drivers/net/ethernet/mellanox/mlx5/core/cmd.c next = from->next; next 1166 drivers/net/ethernet/mellanox/mlx5/core/cmd.c if (!next) { next 1172 drivers/net/ethernet/mellanox/mlx5/core/cmd.c block = next->buf; next 1177 drivers/net/ethernet/mellanox/mlx5/core/cmd.c next = next->next; next 1199 drivers/net/ethernet/mellanox/mlx5/core/cmd.c mailbox->next = NULL; next 1238 drivers/net/ethernet/mellanox/mlx5/core/cmd.c tmp->next = head; next 1239 drivers/net/ethernet/mellanox/mlx5/core/cmd.c block->next = cpu_to_be64(tmp->next ? tmp->next->dma : 0); next 1244 drivers/net/ethernet/mellanox/mlx5/core/cmd.c msg->next = head; next 1249 drivers/net/ethernet/mellanox/mlx5/core/cmd.c tmp = head->next; next 1261 drivers/net/ethernet/mellanox/mlx5/core/cmd.c struct mlx5_cmd_mailbox *head = msg->next; next 1262 drivers/net/ethernet/mellanox/mlx5/core/cmd.c struct mlx5_cmd_mailbox *next; next 1265 drivers/net/ethernet/mellanox/mlx5/core/cmd.c next = head->next; next 1267 drivers/net/ethernet/mellanox/mlx5/core/cmd.c head = next; next 1650 drivers/net/ethernet/mellanox/mlx5/core/cmd.c msg = list_entry(ch->head.next, typeof(*msg), list); next 206 drivers/net/ethernet/mellanox/mlx5/core/en.h struct mlx5_wqe_srq_next_seg next; next 868 drivers/net/ethernet/mellanox/mlx5/core/en_main.c &wqe->next.next_wqe_index); next 551 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c struct mlx5e_neigh_hash_entry *next = NULL; next 555 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c for (next = nhe ? next 563 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c next; next 564 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c next = list_next_or_null_rcu(&rpriv->neigh_update.neigh_list, next 565 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c &next->neigh_list, next 568 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c if (mlx5e_rep_neigh_entry_hold(next)) next 576 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c return next; next 1377 drivers/net/ethernet/mellanox/mlx5/core/en_rx.c mlx5_wq_ll_pop(wq, cqe->wqe_id, &wqe->next.next_wqe_index); next 1438 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c struct mlx5e_encap_entry *next = NULL; next 1444 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c for (next = e ? next 1452 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c next; next 1453 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c next = list_next_or_null_rcu(&nhe->encap_list, next 1454 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c &next->encap_list, next 1457 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c if (mlx5e_encap_take(next)) next 1465 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c if (!next) next 1466 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c return next; next 1469 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c wait_for_completion(&next->res_ready); next 1471 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c if (!(next->flags & MLX5_ENCAP_ENTRY_VALID)) { next 1472 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c e = next; next 1476 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c return next; next 130 drivers/net/ethernet/mellanox/mlx5/core/fs_counters.c struct list_head *next = mlx5_fc_counters_lookup_next(dev, counter->id); next 132 drivers/net/ethernet/mellanox/mlx5/core/fs_counters.c list_add_tail(&counter->list, next); next 678 drivers/net/ethernet/mellanox/mlx5/core/health.c unsigned long next; next 680 drivers/net/ethernet/mellanox/mlx5/core/health.c get_random_bytes(&next, sizeof(next)); next 681 drivers/net/ethernet/mellanox/mlx5/core/health.c next %= HZ; next 682 drivers/net/ethernet/mellanox/mlx5/core/health.c next += jiffies + MLX5_HEALTH_POLL_INTERVAL; next 684 drivers/net/ethernet/mellanox/mlx5/core/health.c return next; next 167 drivers/net/ethernet/mellanox/mlx5/core/pagealloc.c fp = list_entry(dev->priv.free_list.next, struct fw_page, list); next 316 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_icm_pool.c struct mlx5dr_icm_chunk *chunk, *next; next 322 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_icm_pool.c list_for_each_entry_safe(chunk, next, &bucket->free_list, chunk_list) next 328 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_icm_pool.c list_for_each_entry_safe(chunk, next, &bucket->used_list, chunk_list) next 558 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_icm_pool.c struct mlx5dr_icm_mr *icm_mr, *next; next 563 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_icm_pool.c list_for_each_entry_safe(icm_mr, next, &pool->icm_mr_list, mr_list) next 234 drivers/net/ethernet/mellanox/mlx5/core/uar.c up = list_entry(head->next, struct mlx5_uars_page, list); next 28 drivers/net/ethernet/mellanox/mlxfw/mlxfw_mfa2_tlv_multi.c void *next; next 37 drivers/net/ethernet/mellanox/mlxfw/mlxfw_mfa2_tlv_multi.c next = (void *) tlv + tlv_len; next 38 drivers/net/ethernet/mellanox/mlxfw/mlxfw_mfa2_tlv_multi.c return mlxfw_mfa2_tlv_get(mfa2_file, next); next 90 drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c struct mlxsw_afa_set *next; /* Pointer to the next set. */ next 336 drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c block->first_set->next = block->cur_set; next 357 drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c next_set = set->next; next 389 drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c prev_set->next = set; next 417 drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c if (WARN_ON(!block->first_set->next)) next 419 drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c return block->first_set->next->kvdl_index; next 653 drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c block->cur_set->next = set; next 276 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c if (region->list.next != &group->region_list && next 539 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c if (pos->next != &vgroup->vregion_list) { /* not last */ next 540 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c vregion2 = list_entry(pos->next, typeof(*vregion2), next 943 drivers/net/ethernet/micrel/ksz884x.c u32 next; next 1008 drivers/net/ethernet/micrel/ksz884x.c int next; next 1628 drivers/net/ethernet/micrel/ksz884x.c *desc = &info->ring[info->next]; next 1629 drivers/net/ethernet/micrel/ksz884x.c info->next++; next 1630 drivers/net/ethernet/micrel/ksz884x.c info->next &= info->mask; next 3838 drivers/net/ethernet/micrel/ksz884x.c previous->phw->next = cpu_to_le32(phys); next 3840 drivers/net/ethernet/micrel/ksz884x.c previous->phw->next = cpu_to_le32(desc_info->ring_phys); next 3845 drivers/net/ethernet/micrel/ksz884x.c desc_info->last = desc_info->next = 0; next 3869 drivers/net/ethernet/micrel/ksz884x.c info->last = info->next = 0; next 5032 drivers/net/ethernet/micrel/ksz884x.c int next; next 5041 drivers/net/ethernet/micrel/ksz884x.c next = info->next; next 5044 drivers/net/ethernet/micrel/ksz884x.c desc = &info->ring[next]; next 5058 drivers/net/ethernet/micrel/ksz884x.c next++; next 5059 drivers/net/ethernet/micrel/ksz884x.c next &= info->mask; next 5061 drivers/net/ethernet/micrel/ksz884x.c info->next = next; next 5068 drivers/net/ethernet/micrel/ksz884x.c int next; next 5077 drivers/net/ethernet/micrel/ksz884x.c next = info->next; next 5080 drivers/net/ethernet/micrel/ksz884x.c desc = &info->ring[next]; next 5103 drivers/net/ethernet/micrel/ksz884x.c next++; next 5104 drivers/net/ethernet/micrel/ksz884x.c next &= info->mask; next 5106 drivers/net/ethernet/micrel/ksz884x.c info->next = next; next 5113 drivers/net/ethernet/micrel/ksz884x.c int next; next 5122 drivers/net/ethernet/micrel/ksz884x.c next = info->next; next 5125 drivers/net/ethernet/micrel/ksz884x.c desc = &info->ring[next]; next 5162 drivers/net/ethernet/micrel/ksz884x.c next++; next 5163 drivers/net/ethernet/micrel/ksz884x.c next &= info->mask; next 5165 drivers/net/ethernet/micrel/ksz884x.c info->next = next; next 2906 drivers/net/ethernet/myricom/myri10ge/myri10ge.c segs = segs->next; next 2907 drivers/net/ethernet/myricom/myri10ge/myri10ge.c curr->next = NULL; next 2913 drivers/net/ethernet/myricom/myri10ge/myri10ge.c segs = segs->next; next 2914 drivers/net/ethernet/myricom/myri10ge/myri10ge.c curr->next = NULL; next 4059 drivers/net/ethernet/myricom/myri10ge/myri10ge.c .next = NULL, next 770 drivers/net/ethernet/neterion/s2io.c int next = (j + 1) % blk_cnt; next 772 drivers/net/ethernet/neterion/s2io.c tmp_v_addr_next = ring->rx_blocks[next].block_virt_addr; next 774 drivers/net/ethernet/neterion/s2io.c tmp_p_addr_next = ring->rx_blocks[next].block_dma_addr; next 8489 drivers/net/ethernet/neterion/s2io.c lro->last_frag->next = skb; next 658 drivers/net/ethernet/neterion/vxge/vxge-main.c struct list_head *entry, *next; next 659 drivers/net/ethernet/neterion/vxge/vxge-main.c list_for_each_safe(entry, next, &vpath->mac_addr_list) { next 1068 drivers/net/ethernet/neterion/vxge/vxge-main.c struct list_head *entry, *next; next 1075 drivers/net/ethernet/neterion/vxge/vxge-main.c list_for_each_safe(entry, next, &vpath->mac_addr_list) { next 1131 drivers/net/ethernet/neterion/vxge/vxge-main.c struct list_head *entry, *next; next 1195 drivers/net/ethernet/neterion/vxge/vxge-main.c list_for_each_safe(entry, next, list_head) { next 1237 drivers/net/ethernet/neterion/vxge/vxge-main.c list_for_each_safe(entry, next, list_head) { next 1446 drivers/net/ethernet/neterion/vxge/vxge-main.c struct list_head *entry, *next; next 1451 drivers/net/ethernet/neterion/vxge/vxge-main.c list_for_each_safe(entry, next, &vpath->mac_addr_list) { next 2909 drivers/net/ethernet/neterion/vxge/vxge-main.c struct list_head *entry, *next; next 2913 drivers/net/ethernet/neterion/vxge/vxge-main.c list_for_each_safe(entry, next, &vpath->mac_addr_list) { next 2091 drivers/net/ethernet/neterion/vxge/vxge-traffic.h ((struct __vxge_hw_vpath_handle *)(vpath)->vpath_handles.next) next 22 drivers/net/ethernet/netronome/nfp/bpf/jit.c #define nfp_for_each_insn_walk2(nfp_prog, pos, next) \ next 24 drivers/net/ethernet/netronome/nfp/bpf/jit.c next = list_next_entry(pos, l); \ next 26 drivers/net/ethernet/netronome/nfp/bpf/jit.c &(nfp_prog)->insns != &next->l; \ next 28 drivers/net/ethernet/netronome/nfp/bpf/jit.c next = nfp_meta_next(pos)) next 30 drivers/net/ethernet/netronome/nfp/bpf/jit.c #define nfp_for_each_insn_walk3(nfp_prog, pos, next, next2) \ next 32 drivers/net/ethernet/netronome/nfp/bpf/jit.c next = list_next_entry(pos, l), \ next 33 drivers/net/ethernet/netronome/nfp/bpf/jit.c next2 = list_next_entry(next, l); \ next 35 drivers/net/ethernet/netronome/nfp/bpf/jit.c &(nfp_prog)->insns != &next->l && \ next 38 drivers/net/ethernet/netronome/nfp/bpf/jit.c next = nfp_meta_next(pos), \ next 39 drivers/net/ethernet/netronome/nfp/bpf/jit.c next2 = nfp_meta_next(next)) next 3962 drivers/net/ethernet/netronome/nfp/bpf/jit.c struct bpf_insn insn, next; next 3965 drivers/net/ethernet/netronome/nfp/bpf/jit.c next = meta2->insn; next 3973 drivers/net/ethernet/netronome/nfp/bpf/jit.c if (next.code != (BPF_ALU64 | BPF_AND | BPF_K)) next 3978 drivers/net/ethernet/netronome/nfp/bpf/jit.c if (exp_mask[BPF_SIZE(insn.code)] != next.imm) next 3981 drivers/net/ethernet/netronome/nfp/bpf/jit.c if (next.src_reg || next.dst_reg) next 97 drivers/net/ethernet/netronome/nfp/nfp_net_main.c struct nfp_net *nn, *next; next 99 drivers/net/ethernet/netronome/nfp/nfp_net_main.c list_for_each_entry_safe(nn, next, &pf->vnics, vnic_list) next 550 drivers/net/ethernet/netronome/nfp/nfp_net_main.c struct nfp_net *nn, *next; next 588 drivers/net/ethernet/netronome/nfp/nfp_net_main.c list_for_each_entry_safe(nn, next, &pf->vnics, vnic_list) { next 764 drivers/net/ethernet/netronome/nfp/nfp_net_main.c struct nfp_net *nn, *next; next 767 drivers/net/ethernet/netronome/nfp/nfp_net_main.c list_for_each_entry_safe(nn, next, &pf->vnics, vnic_list) { next 148 drivers/net/ethernet/ni/nixge.c nixge_hw_dma_bd_set_addr((bd), next, (addr)) next 752 drivers/net/ethernet/pasemi/pasemi_mac.c goto next; next 775 drivers/net/ethernet/pasemi/pasemi_mac.c next: next 319 drivers/net/ethernet/pensando/ionic/ionic_dev.c cur->next = cq->info; next 322 drivers/net/ethernet/pensando/ionic/ionic_dev.c cur->next = cur + 1; next 360 drivers/net/ethernet/pensando/ionic/ionic_dev.c cq->tail = cq->tail->next; next 405 drivers/net/ethernet/pensando/ionic/ionic_dev.c cur->next = q->info; next 407 drivers/net/ethernet/pensando/ionic/ionic_dev.c cur->next = cur + 1; next 448 drivers/net/ethernet/pensando/ionic/ionic_dev.c q->head = q->head->next; next 489 drivers/net/ethernet/pensando/ionic/ionic_dev.c q->tail = desc_info->next; next 141 drivers/net/ethernet/pensando/ionic/ionic_dev.h struct ionic_cq_info *next; next 157 drivers/net/ethernet/pensando/ionic/ionic_dev.h struct ionic_desc_info *next; next 183 drivers/net/ethernet/pensando/ionic/ionic_main.c adminq->tail = adminq->tail->next; next 176 drivers/net/ethernet/pensando/ionic/ionic_txrx.c q->tail = desc_info->next; next 194 drivers/net/ethernet/pensando/ionic/ionic_txrx.c rxcq->tail = rxcq->tail->next; next 290 drivers/net/ethernet/pensando/ionic/ionic_txrx.c for (cur = q->tail; cur != q->head; cur = cur->next) { next 433 drivers/net/ethernet/pensando/ionic/ionic_txrx.c q->tail = desc_info->next; next 443 drivers/net/ethernet/pensando/ionic/ionic_txrx.c cq->tail = cq->tail->next; next 707 drivers/net/ethernet/pensando/ionic/ionic_txrx.c rewind = rewind->next; next 687 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c cur = list_entry(head->next, nx_mac_list_t, list); next 721 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c cur = list_entry(head->next, nx_mac_list_t, list); next 1217 drivers/net/ethernet/qlogic/netxen/netxen_nic_init.c next: next 1226 drivers/net/ethernet/qlogic/netxen/netxen_nic_init.c goto next; next 1232 drivers/net/ethernet/qlogic/netxen/netxen_nic_init.c goto next; next 1822 drivers/net/ethernet/qlogic/netxen/netxen_nic_init.c buffer = list_entry(head->next, struct netxen_rx_buffer, list); next 1882 drivers/net/ethernet/qlogic/netxen/netxen_nic_init.c buffer = list_entry(head->next, struct netxen_rx_buffer, list); next 109 drivers/net/ethernet/qlogic/qed/qed_cxt.c u64 next; next 971 drivers/net/ethernet/qlogic/qed/qed_cxt.c entries[j].next = cpu_to_be64(val); next 978 drivers/net/ethernet/qlogic/qed/qed_cxt.c entries[j].next = cpu_to_be64(val); next 81 drivers/net/ethernet/qlogic/qede/qede_rdma.c event_node = list_entry(head->next, struct qede_rdma_event_work, next 299 drivers/net/ethernet/qlogic/qla3xxx.c lrg_buf_cb->next = NULL; next 304 drivers/net/ethernet/qlogic/qla3xxx.c qdev->lrg_buf_free_tail->next = lrg_buf_cb; next 356 drivers/net/ethernet/qlogic/qla3xxx.c qdev->lrg_buf_free_head = lrg_buf_cb->next; next 1836 drivers/net/ethernet/qlogic/qla3xxx.c lrg_buf_cb = lrg_buf_cb->next; next 999 drivers/net/ethernet/qlogic/qla3xxx.h struct ql_rcv_buf_cb *next; next 3901 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c cmd = list_entry(head->next, struct qlcnic_cmd_args, list); next 4125 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c cmd = list_entry(head->next, struct qlcnic_cmd_args, list); next 984 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c u32 cur, prev, next; next 988 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c next = state; next 990 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c if ((next < QLC_83XX_IDC_DEV_COLD) || next 991 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c (next > QLC_83XX_IDC_DEV_QUISCENT)) { next 1000 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c if ((next != QLC_83XX_IDC_DEV_COLD) && next 1001 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c (next != QLC_83XX_IDC_DEV_READY)) { next 1004 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c __func__, cur, prev, next); next 1009 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c if (next == QLC_83XX_IDC_DEV_INIT) { next 1015 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c __func__, cur, prev, next); next 626 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c cur = list_entry(head->next, struct qlcnic_mac_vlan_list, list); next 1279 drivers/net/ethernet/qlogic/qlcnic/qlcnic_init.c next: next 1289 drivers/net/ethernet/qlogic/qlcnic/qlcnic_init.c goto next; next 1295 drivers/net/ethernet/qlogic/qlcnic/qlcnic_init.c goto next; next 860 drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c buffer = list_entry(head->next, struct qlcnic_rx_buffer, list); next 1451 drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c buffer = list_entry(head->next, struct qlcnic_rx_buffer, list); next 1532 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c entry = list_entry(head->next, struct qlcnic_async_cmd, next 2103 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c cur = list_entry(head->next, struct qlcnic_mac_vlan_list, list); next 1222 drivers/net/ethernet/sfc/efx.c struct efx_nic *other, *next; next 1230 drivers/net/ethernet/sfc/efx.c list_for_each_entry_safe(other, next, &efx_unassociated_list, next 1267 drivers/net/ethernet/sfc/efx.c struct efx_nic *other, *next; next 1272 drivers/net/ethernet/sfc/efx.c list_for_each_entry_safe(other, next, &efx->secondary_list, node) { next 1162 drivers/net/ethernet/sfc/falcon/efx.c struct ef4_nic *other, *next; next 1170 drivers/net/ethernet/sfc/falcon/efx.c list_for_each_entry_safe(other, next, &ef4_unassociated_list, next 1207 drivers/net/ethernet/sfc/falcon/efx.c struct ef4_nic *other, *next; next 1212 drivers/net/ethernet/sfc/falcon/efx.c list_for_each_entry_safe(other, next, &efx->secondary_list, node) { next 97 drivers/net/ethernet/sfc/falcon/mtd.c struct ef4_mtd_partition *parts, *part, *next; next 107 drivers/net/ethernet/sfc/falcon/mtd.c list_for_each_entry_safe(part, next, &efx->mtd_list, node) next 1010 drivers/net/ethernet/sfc/mcdi.c if (mcdi->async_list.next == &async->list && next 1132 drivers/net/ethernet/sfc/mcdi.c struct efx_mcdi_async_param *async, *next; next 1159 drivers/net/ethernet/sfc/mcdi.c list_for_each_entry_safe(async, next, &mcdi->async_list, list) { next 100 drivers/net/ethernet/sfc/mtd.c struct efx_mtd_partition *parts, *part, *next; next 110 drivers/net/ethernet/sfc/mtd.c list_for_each_entry_safe(part, next, &efx->mtd_list, node) next 1156 drivers/net/ethernet/sfc/ptp.c struct list_head *next; next 1164 drivers/net/ethernet/sfc/ptp.c list_for_each_safe(cursor, next, &ptp->evt_list) { next 1185 drivers/net/ethernet/sfc/ptp.c struct list_head *next; next 1201 drivers/net/ethernet/sfc/ptp.c list_for_each_safe(cursor, next, &ptp->evt_list) { next 1353 drivers/net/ethernet/sfc/ptp.c struct list_head *next; next 1369 drivers/net/ethernet/sfc/ptp.c list_for_each_safe(cursor, next, &efx->ptp_data->evt_list) { next 465 drivers/net/ethernet/sfc/tx.c struct sk_buff *segments, *next; next 475 drivers/net/ethernet/sfc/tx.c next = skb->next; next 476 drivers/net/ethernet/sfc/tx.c skb->next = NULL; next 479 drivers/net/ethernet/sfc/tx.c skb = next; next 572 drivers/net/ethernet/sgi/ioc3-eth.c goto next; next 603 drivers/net/ethernet/sgi/ioc3-eth.c next: next 792 drivers/net/ethernet/silan/sc92031.c goto next; next 813 drivers/net/ethernet/silan/sc92031.c next: next 1245 drivers/net/ethernet/sis/sis190.c struct sis190_phy *cur, *next; next 1247 drivers/net/ethernet/sis/sis190.c list_for_each_entry_safe(cur, next, first_phy, list) { next 148 drivers/net/ethernet/sis/sis900.c struct mii_phy * next; next 631 drivers/net/ethernet/sis/sis900.c mii_phy = mii_phy->next; next 641 drivers/net/ethernet/sis/sis900.c mii_phy->next = sis_priv->mii; next 732 drivers/net/ethernet/sis/sis900.c for (phy=sis_priv->first_mii; phy; phy=phy->next) { next 2483 drivers/net/ethernet/sis/sis900.c sis_priv->first_mii = phy->next; next 237 drivers/net/ethernet/smsc/epic100.c u32 next; next 244 drivers/net/ethernet/smsc/epic100.c u32 next; next 906 drivers/net/ethernet/smsc/epic100.c ep->rx_ring[i].next = ep->rx_ring_dma + next 911 drivers/net/ethernet/smsc/epic100.c ep->rx_ring[i-1].next = ep->rx_ring_dma; next 931 drivers/net/ethernet/smsc/epic100.c ep->tx_ring[i].next = ep->tx_ring_dma + next 934 drivers/net/ethernet/smsc/epic100.c ep->tx_ring[i-1].next = ep->tx_ring_dma; next 658 drivers/net/ethernet/socionext/netsec.c goto next; next 666 drivers/net/ethernet/socionext/netsec.c next: next 1005 drivers/net/ethernet/socionext/netsec.c goto next; next 1031 drivers/net/ethernet/socionext/netsec.c next: next 640 drivers/net/ethernet/sun/cassini.c entry = cp->rx_spare_list.next; next 2329 drivers/net/ethernet/sun/cassini.c goto next; next 2353 drivers/net/ethernet/sun/cassini.c next: next 3274 drivers/net/ethernet/sun/cassini.c goto next; next 3283 drivers/net/ethernet/sun/cassini.c goto next; next 3294 drivers/net/ethernet/sun/cassini.c goto next; next 3301 drivers/net/ethernet/sun/cassini.c goto next; next 3306 drivers/net/ethernet/sun/cassini.c goto next; next 3325 drivers/net/ethernet/sun/cassini.c goto next; next 3330 drivers/net/ethernet/sun/cassini.c next: next 3521 drivers/net/ethernet/sun/niu.c struct page *next = (struct page *) page->mapping; next 3531 drivers/net/ethernet/sun/niu.c page = next; next 6445 drivers/net/ethernet/sun/niu.c struct page *next = next 6450 drivers/net/ethernet/sun/niu.c page = next; next 821 drivers/net/ethernet/sun/sunbmac.c goto next; next 879 drivers/net/ethernet/sun/sunbmac.c next: next 812 drivers/net/ethernet/sun/sungem.c goto next; next 870 drivers/net/ethernet/sun/sungem.c next: next 2025 drivers/net/ethernet/sun/sunhme.c goto next; next 2088 drivers/net/ethernet/sun/sunhme.c next: next 2160 drivers/net/ethernet/sun/sunhme.c goto next; next 2178 drivers/net/ethernet/sun/sunhme.c next: next 2578 drivers/net/ethernet/sun/sunhme.c qp->next = qfe_sbus_list; next 2594 drivers/net/ethernet/sun/sunhme.c for (qp = qfe_sbus_list; qp != NULL; qp = qp->next) { next 2623 drivers/net/ethernet/sun/sunhme.c for (qp = qfe_sbus_list; qp != NULL; qp = qp->next) { next 2646 drivers/net/ethernet/sun/sunhme.c for (qp = qfe_pci_list; qp != NULL; qp = qp->next) { next 2660 drivers/net/ethernet/sun/sunhme.c qp->next = qfe_pci_list; next 3256 drivers/net/ethernet/sun/sunhme.c struct quattro *next = qfe->next; next 3260 drivers/net/ethernet/sun/sunhme.c qfe_pci_list = next; next 3354 drivers/net/ethernet/sun/sunhme.c struct quattro *next = qfe->next; next 3358 drivers/net/ethernet/sun/sunhme.c qfe_sbus_list = next; next 489 drivers/net/ethernet/sun/sunhme.h struct quattro *next; next 484 drivers/net/ethernet/sun/sunqe.c goto next; next 501 drivers/net/ethernet/sun/sunqe.c next: next 983 drivers/net/ethernet/sun/sunqe.c struct sunqec *next = root_qec_dev->next_module; next 991 drivers/net/ethernet/sun/sunqe.c root_qec_dev = next; next 1015 drivers/net/ethernet/sun/sunvnet_common.c BUG_ON(port->tx_bufs[txi].skb->next); next 1017 drivers/net/ethernet/sun/sunvnet_common.c port->tx_bufs[txi].skb->next = skb; next 1034 drivers/net/ethernet/sun/sunvnet_common.c struct sk_buff *next; next 1037 drivers/net/ethernet/sun/sunvnet_common.c next = skb->next; next 1038 drivers/net/ethernet/sun/sunvnet_common.c skb->next = NULL; next 1040 drivers/net/ethernet/sun/sunvnet_common.c skb = next; next 1288 drivers/net/ethernet/sun/sunvnet_common.c segs = segs->next; next 1289 drivers/net/ethernet/sun/sunvnet_common.c curr->next = NULL; next 1570 drivers/net/ethernet/sun/sunvnet_common.c for (m = vp->mcast_list; m; m = m->next) { next 1597 drivers/net/ethernet/sun/sunvnet_common.c m->next = vp->mcast_list; next 1618 drivers/net/ethernet/sun/sunvnet_common.c for (m = vp->mcast_list; m; m = m->next) { next 1644 drivers/net/ethernet/sun/sunvnet_common.c pp = &m->next; next 1657 drivers/net/ethernet/sun/sunvnet_common.c *pp = m->next; next 1792 drivers/net/ethernet/sun/sunvnet_common.c port = list_entry(vp->port_list.next, struct vnet_port, list); next 112 drivers/net/ethernet/sun/sunvnet_common.h struct vnet_mcast_entry *next; next 181 drivers/net/ethernet/ti/cpmac.c struct cpmac_desc *next; next 245 drivers/net/ethernet/ti/cpmac.c dump = dump->next; next 437 drivers/net/ethernet/ti/cpmac.c restart = desc->next; next 445 drivers/net/ethernet/ti/cpmac.c desc = desc->next; next 693 drivers/net/ethernet/ti/cpmac.c desc->hw_next = desc->next->mapping; next 694 drivers/net/ethernet/ti/cpmac.c desc = desc->next; next 962 drivers/net/ethernet/ti/cpmac.c desc->next = &priv->rx_head[(i + 1) % priv->ring_size]; next 963 drivers/net/ethernet/ti/cpmac.c desc->next->prev = desc; next 964 drivers/net/ethernet/ti/cpmac.c desc->hw_next = (u32)desc->next->mapping; next 60 drivers/net/ethernet/ti/cpts.c struct list_head *this, *next; next 64 drivers/net/ethernet/ti/cpts.c list_for_each_safe(this, next, &cpts->events) { next 194 drivers/net/ethernet/ti/cpts.c struct list_head *this, *next; next 201 drivers/net/ethernet/ti/cpts.c list_for_each_safe(this, next, &cpts->events) { next 371 drivers/net/ethernet/ti/cpts.c struct list_head *this, *next; next 382 drivers/net/ethernet/ti/cpts.c list_for_each_safe(this, next, &cpts->events) { next 475 drivers/net/ethernet/ti/netcp_core.c struct netcp_hook_list *next; next 487 drivers/net/ethernet/ti/netcp_core.c list_for_each_entry(next, &netcp_priv->txhook_list_head, list) { next 488 drivers/net/ethernet/ti/netcp_core.c if (next->order > order) next 491 drivers/net/ethernet/ti/netcp_core.c __list_add(&entry->list, next->list.prev, &next->list); next 501 drivers/net/ethernet/ti/netcp_core.c struct netcp_hook_list *next, *n; next 505 drivers/net/ethernet/ti/netcp_core.c list_for_each_entry_safe(next, n, &netcp_priv->txhook_list_head, list) { next 506 drivers/net/ethernet/ti/netcp_core.c if ((next->order == order) && next 507 drivers/net/ethernet/ti/netcp_core.c (next->hook_rtn == hook_rtn) && next 508 drivers/net/ethernet/ti/netcp_core.c (next->hook_data == hook_data)) { next 509 drivers/net/ethernet/ti/netcp_core.c list_del(&next->list); next 511 drivers/net/ethernet/ti/netcp_core.c devm_kfree(netcp_priv->dev, next); next 524 drivers/net/ethernet/ti/netcp_core.c struct netcp_hook_list *next; next 536 drivers/net/ethernet/ti/netcp_core.c list_for_each_entry(next, &netcp_priv->rxhook_list_head, list) { next 537 drivers/net/ethernet/ti/netcp_core.c if (next->order > order) next 540 drivers/net/ethernet/ti/netcp_core.c __list_add(&entry->list, next->list.prev, &next->list); next 550 drivers/net/ethernet/ti/netcp_core.c struct netcp_hook_list *next, *n; next 554 drivers/net/ethernet/ti/netcp_core.c list_for_each_entry_safe(next, n, &netcp_priv->rxhook_list_head, list) { next 555 drivers/net/ethernet/ti/netcp_core.c if ((next->order == order) && next 556 drivers/net/ethernet/ti/netcp_core.c (next->hook_rtn == hook_rtn) && next 557 drivers/net/ethernet/ti/netcp_core.c (next->hook_data == hook_data)) { next 558 drivers/net/ethernet/ti/netcp_core.c list_del(&next->list); next 560 drivers/net/ethernet/ti/netcp_core.c devm_kfree(netcp_priv->dev, next); next 225 drivers/net/ethernet/toshiba/ps3_gelic_net.c for (descr = start_descr; start_descr != descr->next; descr++) { next 227 drivers/net/ethernet/toshiba/ps3_gelic_net.c descr->next_descr_addr = cpu_to_be32(descr->next->bus_addr); next 289 drivers/net/ethernet/toshiba/ps3_gelic_net.c for (descr = descr_in; descr && descr->bus_addr; descr = descr->next) { next 329 drivers/net/ethernet/toshiba/ps3_gelic_net.c descr->next = descr + 1; next 333 drivers/net/ethernet/toshiba/ps3_gelic_net.c (descr - 1)->next = start_descr; next 339 drivers/net/ethernet/toshiba/ps3_gelic_net.c descr->next_descr_addr = cpu_to_be32(descr->next->bus_addr); next 438 drivers/net/ethernet/toshiba/ps3_gelic_net.c descr = descr->next; next 461 drivers/net/ethernet/toshiba/ps3_gelic_net.c descr = descr->next; next 547 drivers/net/ethernet/toshiba/ps3_gelic_net.c tx_chain->tail = tx_chain->tail->next) { next 681 drivers/net/ethernet/toshiba/ps3_gelic_net.c if (card->tx_chain.tail != card->tx_chain.head->next && next 799 drivers/net/ethernet/toshiba/ps3_gelic_net.c card->tx_chain.head = descr->next; next 1060 drivers/net/ethernet/toshiba/ps3_gelic_net.c chain->head = descr->next; next 246 drivers/net/ethernet/toshiba/ps3_gelic_net.h struct gelic_descr *next; next 1593 drivers/net/ethernet/toshiba/ps3_gelic_wireless.c target = list_entry(wl->network_free_list.next, next 283 drivers/net/ethernet/toshiba/spider_net.c descr = descr->next; next 329 drivers/net/ethernet/toshiba/spider_net.c descr->next = descr + 1; next 335 drivers/net/ethernet/toshiba/spider_net.c (descr-1)->next = chain->ring; next 364 drivers/net/ethernet/toshiba/spider_net.c descr = descr->next; next 501 drivers/net/ethernet/toshiba/spider_net.c chain->head = chain->head->next; next 523 drivers/net/ethernet/toshiba/spider_net.c descr = descr->next; next 532 drivers/net/ethernet/toshiba/spider_net.c chain->head = chain->head->next; next 663 drivers/net/ethernet/toshiba/spider_net.c if (descr->next == chain->tail->prev) { next 669 drivers/net/ethernet/toshiba/spider_net.c chain->head = descr->next; next 715 drivers/net/ethernet/toshiba/spider_net.c descr = descr->next; next 727 drivers/net/ethernet/toshiba/spider_net.c descr = descr->next; next 810 drivers/net/ethernet/toshiba/spider_net.c chain->tail = descr->next; next 859 drivers/net/ethernet/toshiba/spider_net.c descr = descr->next; next 1034 drivers/net/ethernet/toshiba/spider_net.c descr = descr->next; next 1060 drivers/net/ethernet/toshiba/spider_net.c descr = descr->next; next 1095 drivers/net/ethernet/toshiba/spider_net.c descr = descr->next; next 1116 drivers/net/ethernet/toshiba/spider_net.c descr = descr->next; next 1155 drivers/net/ethernet/toshiba/spider_net.c chain->tail = descr->next; next 388 drivers/net/ethernet/toshiba/spider_net.h struct spider_net_descr *next; next 1260 drivers/net/ethernet/via/via-rhine.c dma_addr_t next; next 1264 drivers/net/ethernet/via/via-rhine.c next = rp->rx_ring_dma; next 1270 drivers/net/ethernet/via/via-rhine.c next += sizeof(struct rx_desc); next 1271 drivers/net/ethernet/via/via-rhine.c rp->rx_ring[i].next_desc = cpu_to_le32(next); next 1318 drivers/net/ethernet/via/via-rhine.c dma_addr_t next; next 1322 drivers/net/ethernet/via/via-rhine.c next = rp->tx_ring_dma; next 1327 drivers/net/ethernet/via/via-rhine.c next += sizeof(struct tx_desc); next 1328 drivers/net/ethernet/via/via-rhine.c rp->tx_ring[i].next_desc = cpu_to_le32(next); next 321 drivers/net/ethernet/xilinx/ll_temac.h u32 next; /* Physical address of next buffer descriptor */ next 353 drivers/net/ethernet/xilinx/ll_temac_main.c lp->tx_bd_v[i].next = cpu_to_be32(lp->tx_bd_p next 358 drivers/net/ethernet/xilinx/ll_temac_main.c lp->rx_bd_v[i].next = cpu_to_be32(lp->rx_bd_p next 364 drivers/net/ethernet/xilinx/xilinx_axienet.h u32 next; /* Physical address of next buffer descriptor */ next 219 drivers/net/ethernet/xilinx/xilinx_axienet_main.c lp->tx_bd_v[i].next = lp->tx_bd_p + next 225 drivers/net/ethernet/xilinx/xilinx_axienet_main.c lp->rx_bd_v[i].next = lp->rx_bd_p + next 195 drivers/net/ethernet/xscale/ixp4xx_eth.c u32 next; /* pointer to next buffer, unused */ next 602 drivers/net/ethernet/xscale/ixp4xx_eth.c phys, desc->next, desc->buf_len, desc->pkt_len, next 627 drivers/net/ethernet/xscale/ixp4xx_eth.c BUG_ON(tab[n_desc].next); next 120 drivers/net/fddi/skfp/smttimer.c struct smt_timer *next ; next 141 drivers/net/fddi/skfp/smttimer.c next = smc->t.st_queue ; next 144 drivers/net/fddi/skfp/smttimer.c for ( tm = next ; tm ; tm = next) { next 145 drivers/net/fddi/skfp/smttimer.c next = tm->tm_next ; next 64 drivers/net/geneve.c struct list_head next; /* geneve's per namespace list */ next 1228 drivers/net/geneve.c list_for_each_entry(geneve, &gn->geneve_list, next) { next 1317 drivers/net/geneve.c list_add(&geneve->next, &gn->geneve_list); next 1660 drivers/net/geneve.c list_del(&geneve->next); next 1832 drivers/net/geneve.c struct geneve_dev *geneve, *next; next 1841 drivers/net/geneve.c list_for_each_entry_safe(geneve, next, &gn->geneve_list, next) { next 418 drivers/net/hamradio/bpqether.c .next = bpq_seq_next, next 577 drivers/net/hamradio/bpqether.c bpq = list_entry(bpq_devices.next, struct bpqdev, bpq_list); next 213 drivers/net/hamradio/dmascc.c struct scc_info *next; next 290 drivers/net/hamradio/dmascc.c first = info->next; next 591 drivers/net/hamradio/dmascc.c info->next = first; next 2083 drivers/net/hamradio/scc.c .next = scc_net_seq_next, next 138 drivers/net/hamradio/yam.c struct yam_mcs *next; next 385 drivers/net/hamradio/yam.c p = p->next; next 395 drivers/net/hamradio/yam.c p->next = yam_data; next 410 drivers/net/hamradio/yam.c p = p->next; next 825 drivers/net/hamradio/yam.c .next = yam_seq_next, next 1174 drivers/net/hamradio/yam.c yam_data = yam_data->next; next 848 drivers/net/hyperv/hyperv_net.h u32 next; /* next entry for writing */ next 1038 drivers/net/hyperv/netvsc.c while (mrc->first != mrc->next) { next 1070 drivers/net/hyperv/netvsc.c if (mrc->next >= mrc->first) next 1071 drivers/net/hyperv/netvsc.c *filled = mrc->next - mrc->first; next 1073 drivers/net/hyperv/netvsc.c *filled = (count - mrc->first) + mrc->next; next 1101 drivers/net/hyperv/netvsc.c rcd = mrc->slots + mrc->next; next 1105 drivers/net/hyperv/netvsc.c if (++mrc->next == nvdev->recv_completion_cnt) next 1106 drivers/net/hyperv/netvsc.c mrc->next = 0; next 1003 drivers/net/hyperv/rndis_filter.c if (nvchan->mrc.first != nvchan->mrc.next) next 630 drivers/net/ipvlan/ipvlan_main.c struct ipvl_addr *addr, *next; next 633 drivers/net/ipvlan/ipvlan_main.c list_for_each_entry_safe(addr, next, &ipvlan->addrs, anode) { next 694 drivers/net/ipvlan/ipvlan_main.c struct ipvl_dev *ipvlan, *next; next 727 drivers/net/ipvlan/ipvlan_main.c list_for_each_entry_safe(ipvlan, next, &port->ipvlans, pnode) next 67 drivers/net/macsec.c sc = rcu_dereference_bh(sc->next)) next 71 drivers/net/macsec.c sc = rtnl_dereference(sc->next)) next 167 drivers/net/macsec.c struct macsec_rx_sc __rcu *next; next 1381 drivers/net/macsec.c rx_scp = &rx_sc->next, rx_sc = rtnl_dereference(*rx_scp)) { next 1385 drivers/net/macsec.c rcu_assign_pointer(*rx_scp, rx_sc->next); next 1421 drivers/net/macsec.c rcu_assign_pointer(rx_sc->next, secy->rx_sc); next 2617 drivers/net/macsec.c goto next; next 2620 drivers/net/macsec.c goto next; next 2625 drivers/net/macsec.c next: next 3107 drivers/net/macsec.c rcu_assign_pointer(macsec->secy.rx_sc, rx_sc->next); next 1668 drivers/net/macvlan.c struct macvlan_dev *vlan, *next; next 1716 drivers/net/macvlan.c list_for_each_entry_safe(vlan, next, &port->vlans, list) next 592 drivers/net/phy/dp83640.c struct list_head *this, *next; next 595 drivers/net/phy/dp83640.c list_for_each_safe(this, next, &dp83640->rxts) { next 1017 drivers/net/phy/dp83640.c struct list_head *this, *next; next 1021 drivers/net/phy/dp83640.c list_for_each_safe(this, next, &phyter_clocks) { next 1193 drivers/net/phy/dp83640.c struct list_head *this, *next; next 1211 drivers/net/phy/dp83640.c list_for_each_safe(this, next, &clock->phylist) { next 1436 drivers/net/phy/dp83640.c struct list_head *this, *next; next 1455 drivers/net/phy/dp83640.c list_for_each_safe(this, next, &dp83640->rxts) { next 31 drivers/net/phy/mdio-mux.c struct mdio_mux_child_bus *next; next 174 drivers/net/phy/mdio-mux.c cb->next = pb->children; next 201 drivers/net/phy/mdio-mux.c cb = cb->next; next 1639 drivers/net/ppp/ppp_generic.c list = list->next; next 1749 drivers/net/ppp/ppp_generic.c list = list->next; next 1759 drivers/net/ppp/ppp_generic.c list = list->next; next 2568 drivers/net/ppp/ppp_generic.c p->next = NULL; next 2569 drivers/net/ppp/ppp_generic.c fragpp = &p->next; next 173 drivers/net/ppp/pppoe.c ret = ret->next; next 190 drivers/net/ppp/pppoe.c ret = ret->next; next 193 drivers/net/ppp/pppoe.c po->next = pn->hash_table[hash]; next 211 drivers/net/ppp/pppoe.c *src = ret->next; next 215 drivers/net/ppp/pppoe.c src = &ret->next; next 216 drivers/net/ppp/pppoe.c ret = ret->next; next 288 drivers/net/ppp/pppoe.c po = po->next; next 661 drivers/net/ppp/pppoe.c po->next = NULL; next 1042 drivers/net/ppp/pppoe.c po = po->next; next 1071 drivers/net/ppp/pppoe.c if (po->next) next 1072 drivers/net/ppp/pppoe.c po = po->next; next 1097 drivers/net/ppp/pppoe.c .next = pppoe_seq_next, next 136 drivers/net/slip/slhc.c ts[i].next = &(ts[i - 1]); next 138 drivers/net/slip/slhc.c ts[0].next = &(ts[comp->tslot_limit]); next 232 drivers/net/slip/slhc.c struct cstate *cs = lcs->next; next 307 drivers/net/slip/slhc.c cs = cs->next; next 334 drivers/net/slip/slhc.c lcs->next = cs->next; next 335 drivers/net/slip/slhc.c cs->next = ocs->next; next 336 drivers/net/slip/slhc.c ocs->next = cs; next 112 drivers/net/tap.c struct list_head next; next 179 drivers/net/tap.c list_add_tail(&q->next, &tap->queue_list); next 236 drivers/net/tap.c list_del_init(&q->next); next 303 drivers/net/tap.c list_for_each_entry_safe(q, tmp, &tap->queue_list, next) { next 304 drivers/net/tap.c list_del_init(&q->next); next 356 drivers/net/tap.c struct sk_buff *nskb = segs->next; next 358 drivers/net/tap.c segs->next = NULL; next 397 drivers/net/tap.c list_for_each_entry_rcu(tap_major, &major_list, next) { next 1303 drivers/net/tap.c list_for_each_entry(q, &tap->queue_list, next) next 1330 drivers/net/tap.c list_add_tail_rcu(&tap_major->next, &major_list); next 1370 drivers/net/tap.c list_for_each_entry_safe(tap_major, tmp, &major_list, next) { next 1373 drivers/net/tap.c list_del_rcu(&tap_major->next); next 175 drivers/net/tun.c struct list_head next; next 638 drivers/net/tun.c list_add_tail(&tfile->next, &tun->disabled); next 647 drivers/net/tun.c list_del_init(&tfile->next); next 761 drivers/net/tun.c list_for_each_entry(tfile, &tun->disabled, next) { next 777 drivers/net/tun.c list_for_each_entry_safe(tfile, tmp, &tun->disabled, next) { next 1211 drivers/net/tun.c list_for_each_entry(tfile, &tun->disabled, next) { next 3450 drivers/net/tun.c INIT_LIST_HEAD(&tfile->next); next 3631 drivers/net/tun.c list_for_each_entry(tfile, &tun->disabled, next) next 219 drivers/net/usb/cdc_eem.c goto next; next 277 drivers/net/usb/cdc_eem.c goto next; next 323 drivers/net/usb/cdc_eem.c next: next 1695 drivers/net/usb/r8152.c cursor = tp->tx_free.next; next 1725 drivers/net/usb/r8152.c segs = segs->next; next 1726 drivers/net/usb/r8152.c nskb->next = NULL; next 2044 drivers/net/usb/r8152.c struct list_head *cursor, *next, rx_queue; next 2074 drivers/net/usb/r8152.c list_for_each_safe(cursor, next, &rx_queue) { next 2176 drivers/net/usb/r8152.c list_add_tail(&agg->list, next); next 962 drivers/net/virtio_net.c curr_skb->next = nskb; next 177 drivers/net/vxlan.c return list_entry_rcu(fdb->remotes.next, struct vxlan_rdst, list); next 1366 drivers/net/vxlan.c list_for_each_entry(vxlan, &vn->vxlan_list, next) { next 3053 drivers/net/vxlan.c INIT_LIST_HEAD(&vxlan->next); next 3476 drivers/net/vxlan.c list_for_each_entry(tmp, &vn->vxlan_list, next) { next 3640 drivers/net/vxlan.c list_add(&vxlan->next, &vn->vxlan_list); next 4037 drivers/net/vxlan.c list_del(&vxlan->next); next 4229 drivers/net/vxlan.c struct vxlan_dev *vxlan, *next; next 4232 drivers/net/vxlan.c list_for_each_entry_safe(vxlan, next, &vn->vxlan_list, next) { next 4414 drivers/net/vxlan.c struct vxlan_dev *vxlan, *next; next 4422 drivers/net/vxlan.c list_for_each_entry_safe(vxlan, next, &vn->vxlan_list, next) { next 170 drivers/net/wan/hd64572.h unsigned long next; /* pointer to next block descriptor */ next 214 drivers/net/wan/hdlc.c proto = proto->next; next 321 drivers/net/wan/hdlc.c proto->next = first_proto; next 335 drivers/net/wan/hdlc.c p = &((*p)->next); next 337 drivers/net/wan/hdlc.c *p = proto->next; next 119 drivers/net/wan/hdlc_fr.c struct pvc_device *next; /* Sorted in ascending DLCI order */ next 184 drivers/net/wan/hdlc_fr.c pvc = pvc->next; next 201 drivers/net/wan/hdlc_fr.c pvc_p = &(*pvc_p)->next; next 213 drivers/net/wan/hdlc_fr.c pvc->next = *pvc_p; /* Put it in the chain */ next 255 drivers/net/wan/hdlc_fr.c *pvc_p = pvc->next; next 259 drivers/net/wan/hdlc_fr.c pvc_p = &(*pvc_p)->next; next 551 drivers/net/wan/hdlc_fr.c pvc = pvc->next; next 581 drivers/net/wan/hdlc_fr.c pvc = pvc->next; next 592 drivers/net/wan/hdlc_fr.c pvc = pvc->next; next 756 drivers/net/wan/hdlc_fr.c pvc = pvc->next; next 785 drivers/net/wan/hdlc_fr.c pvc = pvc->next; next 856 drivers/net/wan/hdlc_fr.c pvc = pvc->next; next 1035 drivers/net/wan/hdlc_fr.c pvc = pvc->next; next 1152 drivers/net/wan/hdlc_fr.c struct pvc_device *next = pvc->next; next 1161 drivers/net/wan/hdlc_fr.c pvc = next; next 290 drivers/net/wan/ixp4xx_hss.c u32 next; /* pointer to next buffer, unused */ next 574 drivers/net/wan/ixp4xx_hss.c phys, desc->next, desc->buf_len, desc->pkt_len, next 594 drivers/net/wan/ixp4xx_hss.c BUG_ON(tab[n_desc].next); next 1259 drivers/net/wimax/i2400m/fw.c const void *itr, *next, *top; next 1264 drivers/net/wimax/i2400m/fw.c headers++, itr = next) { next 1281 drivers/net/wimax/i2400m/fw.c next = itr + size; next 1283 drivers/net/wimax/i2400m/fw.c next = itr + header_len; next 628 drivers/net/wireless/ath/ar5523/ar5523.c data = (struct ar5523_rx_data *) ar->rx_data_free.next; next 675 drivers/net/wireless/ath/ar5523/ar5523.c data = (struct ar5523_rx_data *) ar->rx_data_used.next; next 695 drivers/net/wireless/ath/ar5523/ar5523.c data = (struct ar5523_rx_data *) ar->rx_data_free.next; next 808 drivers/net/wireless/ath/ar5523/ar5523.c ar->tx_queue_pending.next; next 522 drivers/net/wireless/ath/ath10k/htt_rx.c prev_frag_buf->next = NULL; next 530 drivers/net/wireless/ath/ath10k/htt_rx.c prev_frag_buf->next = frag_buf; next 541 drivers/net/wireless/ath/ath10k/htt_rx.c prev_frag_buf->next = NULL; next 598 drivers/net/wireless/ath/ath10k/htt_rx.c prev_frag_buf->next = NULL; next 606 drivers/net/wireless/ath/ath10k/htt_rx.c prev_frag_buf->next = frag_buf; next 617 drivers/net/wireless/ath/ath10k/htt_rx.c prev_frag_buf->next = NULL; next 1579 drivers/net/wireless/ath/ath10k/mac.c const u8 *next; next 1594 drivers/net/wireless/ath/ath10k/mac.c next = ie + len; next 1596 drivers/net/wireless/ath/ath10k/mac.c if (WARN_ON(next > end)) next 1599 drivers/net/wireless/ath/ath10k/mac.c memmove(ie, next, end - next); next 7284 drivers/net/wireless/ath/ath10k/mac.c goto next; next 7319 drivers/net/wireless/ath/ath10k/mac.c next: next 1582 drivers/net/wireless/ath/ath5k/base.c goto next; next 1595 drivers/net/wireless/ath/ath5k/base.c next: next 166 drivers/net/wireless/ath/ath5k/debug.c .next = reg_next, next 824 drivers/net/wireless/ath/ath6kl/htc_pipe.c target->pipe.htc_packet_pool = (struct htc_packet *) packet->list.next; next 828 drivers/net/wireless/ath/ath6kl/htc_pipe.c packet->list.next = NULL; next 841 drivers/net/wireless/ath/ath6kl/htc_pipe.c packet->list.next = NULL; next 844 drivers/net/wireless/ath/ath6kl/htc_pipe.c packet->list.next = lh; next 330 drivers/net/wireless/ath/ath6kl/main.c le32_to_cpu(debug_buf.next)); next 343 drivers/net/wireless/ath/ath6kl/target.h __le32 next; next 160 drivers/net/wireless/ath/ath9k/htc_drv_beacon.c goto next; next 170 drivers/net/wireless/ath/ath9k/htc_drv_beacon.c goto next; next 184 drivers/net/wireless/ath/ath9k/htc_drv_beacon.c next: next 737 drivers/net/wireless/ath/ath9k/recv.c tbf = list_entry(bf->list.next, struct ath_rxbuf, list); next 2593 drivers/net/wireless/ath/ath9k/xmit.c bf = list_entry(bf_held->list.next, struct ath_buf, next 572 drivers/net/wireless/ath/carl9170/tx.c goto next; next 581 drivers/net/wireless/ath/carl9170/tx.c next: next 59 drivers/net/wireless/ath/wcn36xx/dxe.c struct wcn36xx_dxe_ctl *ctl = ch->head_blk_ctl, *next; next 63 drivers/net/wireless/ath/wcn36xx/dxe.c next = ctl->next; next 65 drivers/net/wireless/ath/wcn36xx/dxe.c ctl = next; next 86 drivers/net/wireless/ath/wcn36xx/dxe.c prev_ctl->next = cur_ctl; next 87 drivers/net/wireless/ath/wcn36xx/dxe.c cur_ctl->next = ch->head_blk_ctl; next 89 drivers/net/wireless/ath/wcn36xx/dxe.c prev_ctl->next = cur_ctl; next 219 drivers/net/wireless/ath/wcn36xx/dxe.c cur_ctl = cur_ctl->next; next 255 drivers/net/wireless/ath/wcn36xx/dxe.c cur = cur->next; next 310 drivers/net/wireless/ath/wcn36xx/dxe.c cur_ctl = cur_ctl->next; next 324 drivers/net/wireless/ath/wcn36xx/dxe.c cur = cur->next; next 389 drivers/net/wireless/ath/wcn36xx/dxe.c ctl = ctl->next; next 579 drivers/net/wireless/ath/wcn36xx/dxe.c ctl = ctl->next; next 692 drivers/net/wireless/ath/wcn36xx/dxe.c ctl_skb = ctl_bd->next; next 753 drivers/net/wireless/ath/wcn36xx/dxe.c ch->head_blk_ctl = ctl_skb->next; next 420 drivers/net/wireless/ath/wcn36xx/dxe.h struct wcn36xx_dxe_ctl *next; next 2327 drivers/net/wireless/ath/wil6210/cfg80211.c ret = vif->probe_client_pending.next; next 3379 drivers/net/wireless/ath/wil6210/wmi.c ret = wil->pending_wmi_ev.next; next 1071 drivers/net/wireless/atmel/atmel.c goto next; next 1079 drivers/net/wireless/atmel/atmel.c goto next; next 1137 drivers/net/wireless/atmel/atmel.c goto next; next 1145 drivers/net/wireless/atmel/atmel.c next: next 453 drivers/net/wireless/broadcom/b43/pio.c pack = list_entry(q->packets_list.next, next 121 drivers/net/wireless/broadcom/b43/sdio.c tuple = tuple->next; next 462 drivers/net/wireless/broadcom/b43legacy/pio.c packet = list_entry(queue->txfree.next, struct b43legacy_pio_txpacket, next 4200 drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c goto next; next 4206 drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c goto next; next 4213 drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c goto next; next 4231 drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c next: next 52 drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c search = search->next; next 400 drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c search = search->next; next 453 drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c search = search->next; next 470 drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c prev->next = search->next; next 472 drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c flow->tdls_entry = search->next; next 491 drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c tdls_entry->next = NULL; next 498 drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c while (search->next) { next 499 drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c search = search->next; next 503 drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c search->next = tdls_entry; next 35 drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.h struct brcmf_flowring_tdls_entry *next; next 1257 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c struct sk_buff *cur, *next; next 1259 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c skb_queue_walk_safe(&bus->glom, cur, next) { next 1675 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c pfirst->len, pfirst->next, next 382 drivers/net/wireless/broadcom/brcm80211/brcmfmac/usb.c req = list_entry(q->next, struct brcmf_usbreq, list); next 383 drivers/net/wireless/broadcom/brcm80211/brcmfmac/usb.c list_del_init(q->next); next 428 drivers/net/wireless/broadcom/brcm80211/brcmfmac/usb.c req = list_entry(q->next, struct brcmf_usbreq, list); next 431 drivers/net/wireless/broadcom/brcm80211/brcmfmac/usb.c list_del(q->next); next 440 drivers/net/wireless/broadcom/brcm80211/brcmfmac/usb.c struct brcmf_usbreq *req, *next; next 442 drivers/net/wireless/broadcom/brcm80211/brcmfmac/usb.c list_for_each_entry_safe(req, next, q, list) { next 153 drivers/net/wireless/broadcom/brcm80211/brcmsmac/aiutils.h struct gpioh_item *next; next 929 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c struct sk_buff *p, *next; next 978 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c skb_queue_walk_safe(&dma_frames, p, next) { next 289 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c struct brcms_timer *t, *next; next 319 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c for (t = wl->timers; t; t = next) { next 320 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c next = t->next; next 1501 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c t->next = wl->timers; next 1567 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c wl->timers = wl->timers->next; next 1578 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c if (tmp->next == t) { next 1579 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c tmp->next = t->next; next 1586 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c tmp = tmp->next; next 44 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.h struct brcms_timer *next; /* for freeing on unload */ next 7710 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c struct sk_buff *next = NULL; next 7733 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c skb_queue_walk_safe(&recv_frames, p, next) { next 564 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_cmn.c pi->next = pi->sh->phy_head; next 590 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_cmn.c pi->sh->phy_head = pi->next; next 591 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_cmn.c else if (pi->sh->phy_head->next == pi) next 592 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_cmn.c pi->sh->phy_head->next = NULL; next 542 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_int.h struct brcms_phy *next; next 38 drivers/net/wireless/broadcom/brcm80211/brcmutil/utils.c WARN_ON(skb->next); next 111 drivers/net/wireless/broadcom/brcm80211/brcmutil/utils.c struct sk_buff *p, *next; next 114 drivers/net/wireless/broadcom/brcm80211/brcmutil/utils.c skb_queue_walk_safe(q, p, next) { next 145 drivers/net/wireless/broadcom/brcm80211/brcmutil/utils.c struct sk_buff *p, *next; next 148 drivers/net/wireless/broadcom/brcm80211/brcmutil/utils.c skb_queue_walk_safe(q, p, next) { next 305 drivers/net/wireless/broadcom/brcm80211/brcmutil/utils.c for (p = p0; p; p = p->next) next 3024 drivers/net/wireless/cisco/airo.c tmp_net = list_entry(ai->network_free_list.next, next 3026 drivers/net/wireless/cisco/airo.c list_del(ai->network_free_list.next); next 4039 drivers/net/wireless/cisco/airo.c u16 next; next 4043 drivers/net/wireless/cisco/airo.c next = IN4500(ai, AUXDATA); next 4046 drivers/net/wireless/cisco/airo.c return next; next 4056 drivers/net/wireless/cisco/airo.c u16 next; next 4064 drivers/net/wireless/cisco/airo.c next = aux_setup(ai, page, offset, &len); next 4078 drivers/net/wireless/cisco/airo.c next = aux_setup(ai, next, 4, &len); next 5696 drivers/net/wireless/cisco/airo.c ai = list_entry(airo_devices.next, struct airo_info, dev_list); next 760 drivers/net/wireless/intel/ipw2x00/ipw2100.c element = priv->msg_free_list.next; next 2681 drivers/net/wireless/intel/ipw2x00/ipw2100.c i = (rxq->next + 1) % rxq->entries; next 2760 drivers/net/wireless/intel/ipw2x00/ipw2100.c rxq->next = (i ? i : rxq->entries) - 1; next 2763 drivers/net/wireless/intel/ipw2x00/ipw2100.c IPW_MEM_HOST_SHARED_RX_WRITE_INDEX, rxq->next); next 2819 drivers/net/wireless/intel/ipw2x00/ipw2100.c element = priv->fw_pend_list.next; next 2854 drivers/net/wireless/intel/ipw2x00/ipw2100.c if (w != txq->next) next 3002 drivers/net/wireless/intel/ipw2x00/ipw2100.c int next = txq->next; next 3016 drivers/net/wireless/intel/ipw2x00/ipw2100.c element = priv->msg_pend_list.next; next 3023 drivers/net/wireless/intel/ipw2x00/ipw2100.c &txq->drv[txq->next], next 3024 drivers/net/wireless/intel/ipw2x00/ipw2100.c (u32) (txq->nic + txq->next * next 3027 drivers/net/wireless/intel/ipw2x00/ipw2100.c packet->index = txq->next; next 3029 drivers/net/wireless/intel/ipw2x00/ipw2100.c tbd = &txq->drv[txq->next]; next 3042 drivers/net/wireless/intel/ipw2x00/ipw2100.c txq->next++; next 3043 drivers/net/wireless/intel/ipw2x00/ipw2100.c txq->next %= txq->entries; next 3051 drivers/net/wireless/intel/ipw2x00/ipw2100.c if (txq->next != next) { next 3057 drivers/net/wireless/intel/ipw2x00/ipw2100.c txq->next); next 3071 drivers/net/wireless/intel/ipw2x00/ipw2100.c int next = txq->next; next 3083 drivers/net/wireless/intel/ipw2x00/ipw2100.c element = priv->tx_pend_list.next; next 3103 drivers/net/wireless/intel/ipw2x00/ipw2100.c tbd = &txq->drv[txq->next]; next 3105 drivers/net/wireless/intel/ipw2x00/ipw2100.c packet->index = txq->next; next 3142 drivers/net/wireless/intel/ipw2x00/ipw2100.c txq->next++; next 3143 drivers/net/wireless/intel/ipw2x00/ipw2100.c txq->next %= txq->entries; next 3154 drivers/net/wireless/intel/ipw2x00/ipw2100.c tbd = &txq->drv[txq->next]; next 3181 drivers/net/wireless/intel/ipw2x00/ipw2100.c txq->next, tbd->host_addr, next 3189 drivers/net/wireless/intel/ipw2x00/ipw2100.c txq->next++; next 3190 drivers/net/wireless/intel/ipw2x00/ipw2100.c txq->next %= txq->entries; next 3200 drivers/net/wireless/intel/ipw2x00/ipw2100.c if (txq->next != next) { next 3205 drivers/net/wireless/intel/ipw2x00/ipw2100.c txq->next); next 3401 drivers/net/wireless/intel/ipw2x00/ipw2100.c element = priv->tx_free_list.next; next 4398 drivers/net/wireless/intel/ipw2x00/ipw2100.c write_register(priv->net_dev, w, q->next); next 4510 drivers/net/wireless/intel/ipw2x00/ipw2100.c priv->tx_queue.next = 0; next 4632 drivers/net/wireless/intel/ipw2x00/ipw2100.c priv->rx_queue.next = priv->rx_queue.entries - 1; next 5816 drivers/net/wireless/intel/ipw2x00/ipw2100.c element = priv->tx_pend_list.next; next 186 drivers/net/wireless/intel/ipw2x00/ipw2100.h u32 next; next 5149 drivers/net/wireless/intel/ipw2x00/ipw2200.c element = rxq->rx_free.next; next 5187 drivers/net/wireless/intel/ipw2x00/ipw2200.c element = rxq->rx_used.next; next 7598 drivers/net/wireless/intel/ipw2x00/ipw2200.c element = priv->ieee->network_free_list.next; next 1603 drivers/net/wireless/intel/ipw2x00/libipw_rx.c target = list_entry(ieee->network_free_list.next, next 1605 drivers/net/wireless/intel/ipw2x00/libipw_rx.c list_del(ieee->network_free_list.next); next 255 drivers/net/wireless/intel/iwlegacy/3945-mac.c element = il->free_frames.next; next 284 drivers/net/wireless/intel/iwlegacy/3945-mac.c element = il->free_frames.next; next 943 drivers/net/wireless/intel/iwlegacy/3945-mac.c element = rxq->rx_free.next; next 1042 drivers/net/wireless/intel/iwlegacy/3945-mac.c element = rxq->rx_used.next; next 263 drivers/net/wireless/intel/iwlegacy/4965-mac.c element = rxq->rx_free.next; next 364 drivers/net/wireless/intel/iwlegacy/4965-mac.c element = rxq->rx_used.next; next 3694 drivers/net/wireless/intel/iwlegacy/4965-mac.c element = il->free_frames.next; next 3723 drivers/net/wireless/intel/iwlegacy/4965-mac.c element = il->free_frames.next; next 1987 drivers/net/wireless/intel/iwlwifi/fw/dbg.c list_entry(list->next, typeof(*entry), list); next 822 drivers/net/wireless/intel/iwlwifi/mvm/tx.c struct sk_buff *tmp, *next; next 835 drivers/net/wireless/intel/iwlwifi/mvm/tx.c next = skb_gso_segment(skb, netdev_flags); next 837 drivers/net/wireless/intel/iwlwifi/mvm/tx.c if (WARN_ON_ONCE(IS_ERR(next))) next 839 drivers/net/wireless/intel/iwlwifi/mvm/tx.c else if (next) next 842 drivers/net/wireless/intel/iwlwifi/mvm/tx.c while (next) { next 843 drivers/net/wireless/intel/iwlwifi/mvm/tx.c tmp = next; next 844 drivers/net/wireless/intel/iwlwifi/mvm/tx.c next = tmp->next; next 876 drivers/net/wireless/intel/iwlwifi/mvm/tx.c tmp->next = NULL; next 374 drivers/net/wireless/intersil/hostap/hostap_ap.c .next = ap_control_proc_next, next 404 drivers/net/wireless/intersil/hostap/hostap_ap.c for (ptr = mac_restrictions->mac_list.next; next 405 drivers/net/wireless/intersil/hostap/hostap_ap.c ptr != &mac_restrictions->mac_list; ptr = ptr->next) { next 455 drivers/net/wireless/intersil/hostap/hostap_ap.c for (ptr = mac_restrictions->mac_list.next, n = ptr->next; next 457 drivers/net/wireless/intersil/hostap/hostap_ap.c ptr = n, n = ptr->next) { next 504 drivers/net/wireless/intersil/hostap/hostap_ap.c for (ptr = ap->sta_list.next, n = ptr->next; ptr != &ap->sta_list; next 505 drivers/net/wireless/intersil/hostap/hostap_ap.c ptr = n, n = ptr->next) { next 576 drivers/net/wireless/intersil/hostap/hostap_ap.c .next = prism2_ap_proc_next, next 1091 drivers/net/wireless/intersil/hostap/hostap_ap.c entry = entry->next; next 1126 drivers/net/wireless/intersil/hostap/hostap_ap.c entry->next = ap->add_sta_proc_entries; next 1975 drivers/net/wireless/intersil/hostap/hostap_ap.c entry = entry->next; next 2335 drivers/net/wireless/intersil/hostap/hostap_ap.c for (ptr = ap->sta_list.next; ptr != NULL && ptr != &ap->sta_list; next 2336 drivers/net/wireless/intersil/hostap/hostap_ap.c ptr = ptr->next) { next 2387 drivers/net/wireless/intersil/hostap/hostap_ap.c for (ptr = ap->sta_list.next; ptr != NULL && ptr != &ap->sta_list; next 2388 drivers/net/wireless/intersil/hostap/hostap_ap.c ptr = ptr->next) { next 3263 drivers/net/wireless/intersil/hostap/hostap_ap.c entry->next = local->ap->wds_oper_entries; next 153 drivers/net/wireless/intersil/hostap/hostap_ap.h struct add_sta_proc_data *next; next 161 drivers/net/wireless/intersil/hostap/hostap_ap.h struct wds_oper_data *next; next 219 drivers/net/wireless/intersil/hostap/hostap_download.c .next = prism2_download_aux_dump_proc_next, next 656 drivers/net/wireless/intersil/hostap/hostap_hw.c entry = list_entry(local->cmd_queue.next, next 702 drivers/net/wireless/intersil/hostap/hostap_hw.c entry = list_entry(local->cmd_queue.next, next 2183 drivers/net/wireless/intersil/hostap/hostap_hw.c cb = cb->next; next 3007 drivers/net/wireless/intersil/hostap/hostap_hw.c entry = list_entry(local->set_tim_list.next, next 3311 drivers/net/wireless/intersil/hostap/hostap_hw.c tx_cb = tx_cb->next; next 250 drivers/net/wireless/intersil/hostap/hostap_main.c entry->next = local->tx_callback; next 267 drivers/net/wireless/intersil/hostap/hostap_main.c cb = cb->next; next 271 drivers/net/wireless/intersil/hostap/hostap_main.c local->tx_callback = cb->next; next 273 drivers/net/wireless/intersil/hostap/hostap_main.c prev->next = cb->next; next 119 drivers/net/wireless/intersil/hostap/hostap_proc.c .next = prism2_wds_proc_next, next 172 drivers/net/wireless/intersil/hostap/hostap_proc.c .next = prism2_bss_list_proc_next, next 349 drivers/net/wireless/intersil/hostap/hostap_proc.c .next = prism2_scan_results_proc_next, next 531 drivers/net/wireless/intersil/hostap/hostap_wlan.h struct hostap_tx_callback_info *next; next 79 drivers/net/wireless/intersil/orinoco/hermes_dld.c char next[0]; /* next PDR starts here */ next 169 drivers/net/wireless/intersil/orinoco/hermes_dld.c pdr = (struct pdr *) pdr->next; next 459 drivers/net/wireless/intersil/orinoco/orinoco_usb.c list_entry(upriv->req_pending.next, struct request_context, next 106 drivers/net/wireless/intersil/p54/main.c u8 *next; next 114 drivers/net/wireless/intersil/p54/main.c next = tim + 2 + dtim_len; next 119 drivers/net/wireless/intersil/p54/main.c memmove(tim, next, skb_tail_pointer(skb) - next); next 433 drivers/net/wireless/intersil/p54/p54spi.c entry = list_entry(priv->tx_pending.next, next 1795 drivers/net/wireless/intersil/prism54/isl_ioctl.c struct list_head *ptr, *next; next 1805 drivers/net/wireless/intersil/prism54/isl_ioctl.c for (ptr = acl->mac_list.next, next = ptr->next; next 1806 drivers/net/wireless/intersil/prism54/isl_ioctl.c ptr != &acl->mac_list; ptr = next, next = ptr->next) { next 98 drivers/net/wireless/marvell/libertas/firmware.c next: next 107 drivers/net/wireless/marvell/libertas/firmware.c goto next; next 187 drivers/net/wireless/marvell/libertas/firmware.c goto next; next 192 drivers/net/wireless/marvell/libertas/firmware.c goto next; next 216 drivers/net/wireless/marvell/libertas/firmware.c next: next 104 drivers/net/wireless/marvell/libertas/if_sdio.c struct if_sdio_packet *next; next 409 drivers/net/wireless/marvell/libertas/if_sdio.c card->packets = packet->next; next 937 drivers/net/wireless/marvell/libertas/if_sdio.c packet->next = NULL; next 956 drivers/net/wireless/marvell/libertas/if_sdio.c while (cur->next) next 957 drivers/net/wireless/marvell/libertas/if_sdio.c cur = cur->next; next 958 drivers/net/wireless/marvell/libertas/if_sdio.c cur->next = packet; next 1240 drivers/net/wireless/marvell/libertas/if_sdio.c card->packets = card->packets->next; next 1284 drivers/net/wireless/marvell/libertas/if_sdio.c card->packets = card->packets->next; next 79 drivers/net/wireless/marvell/libertas/if_spi.c struct list_head *cursor, *next; next 82 drivers/net/wireless/marvell/libertas/if_spi.c list_for_each_safe(cursor, next, &card->cmd_packet_list) { next 87 drivers/net/wireless/marvell/libertas/if_spi.c list_for_each_safe(cursor, next, &card->data_packet_list) { next 898 drivers/net/wireless/marvell/libertas/if_spi.c cmd_packet_list.next); next 912 drivers/net/wireless/marvell/libertas/if_spi.c data_packet_list.next); next 5219 drivers/net/wireless/marvell/mwl8k.c return list_entry(priv->vif_list.next, struct mwl8k_vif, list); next 303 drivers/net/wireless/mediatek/mt76/dma.c skb->prev = skb->next = NULL; next 473 drivers/net/wireless/mediatek/mt76/mt76x02_dfs.c goto next; next 498 drivers/net/wireless/mediatek/mt76/mt76x02_dfs.c goto next; next 507 drivers/net/wireless/mediatek/mt76/mt76x02_dfs.c next: next 49 drivers/net/wireless/mediatek/mt76/mt76x02_usb_core.c if (!iter->next) { next 782 drivers/net/wireless/mediatek/mt76/usb.c skb->prev = skb->next = NULL; next 1936 drivers/net/wireless/realtek/rtlwifi/base.c struct rtl_bssid_entry *entry, *next; next 1938 drivers/net/wireless/realtek/rtlwifi/base.c list_for_each_entry_safe(entry, next, &rtlpriv->scan_list.list, list) { next 1967 drivers/net/wireless/realtek/rtlwifi/base.c struct rtl_bssid_entry *entry, *next; next 1972 drivers/net/wireless/realtek/rtlwifi/base.c list_for_each_entry_safe(entry, next, &rtlpriv->scan_list.list, list) { next 118 drivers/net/wireless/realtek/rtlwifi/pci.h u8 next; next 122 drivers/net/wireless/rsi/rsi_91x_core.c skb = skb->next; next 3916 drivers/net/wireless/ti/wlcore/main.c const u8 *next, *end = skb->data + skb->len; next 3922 drivers/net/wireless/ti/wlcore/main.c next = ie + len; next 3923 drivers/net/wireless/ti/wlcore/main.c memmove(ie, next, end - next); next 3932 drivers/net/wireless/ti/wlcore/main.c const u8 *next, *end = skb->data + skb->len; next 3939 drivers/net/wireless/ti/wlcore/main.c next = ie + len; next 3940 drivers/net/wireless/ti/wlcore/main.c memmove(ie, next, end - next); next 625 drivers/net/wireless/ti/wlcore/tx.c goto next; next 634 drivers/net/wireless/ti/wlcore/tx.c next: next 296 drivers/net/wireless/wl3501_cs.c u16 next, blk_cnt = 0, zero = 0; next 308 drivers/net/wireless/wl3501_cs.c wl3501_get_from_wla(this, this->tx_buffer_head, &next, next 309 drivers/net/wireless/wl3501_cs.c sizeof(next)); next 313 drivers/net/wireless/wl3501_cs.c this->tx_buffer_head = next; next 316 drivers/net/wireless/wl3501_cs.c if (!next && full_len) { next 339 drivers/net/wireless/wl3501_cs.c u16 next; next 342 drivers/net/wireless/wl3501_cs.c wl3501_get_from_wla(this, ptr, &next, sizeof(next)); next 344 drivers/net/wireless/wl3501_cs.c ptr = next; next 468 drivers/net/wireless/wl3501_cs.c u16 bf, sig_bf, next, tmplen, pktlen; next 511 drivers/net/wireless/wl3501_cs.c wl3501_get_from_wla(this, bf, &next, sizeof(next)); next 512 drivers/net/wireless/wl3501_cs.c bf = next; next 528 drivers/net/wireless/wl3501_cs.c wl3501_get_from_wla(this, bf, &next, sizeof(next)); next 529 drivers/net/wireless/wl3501_cs.c bf = next; next 541 drivers/net/wireless/wl3501_cs.c wl3501_get_from_wla(this, bf, &next, sizeof(next)); next 542 drivers/net/wireless/wl3501_cs.c bf = next; next 1179 drivers/net/wireless/wl3501_cs.c u16 ptr, next; next 1208 drivers/net/wireless/wl3501_cs.c next = ptr + WL3501_BLKSZ; next 1209 drivers/net/wireless/wl3501_cs.c while ((next - this->tx_buffer_head) < this->tx_buffer_size) { next 1211 drivers/net/wireless/wl3501_cs.c wl3501_set_to_wla(this, ptr, &next, sizeof(next)); next 1212 drivers/net/wireless/wl3501_cs.c ptr = next; next 1213 drivers/net/wireless/wl3501_cs.c next = ptr + WL3501_BLKSZ; next 1216 drivers/net/wireless/wl3501_cs.c next = 0; next 1217 drivers/net/wireless/wl3501_cs.c wl3501_set_to_wla(this, ptr, &next, sizeof(next)); next 310 drivers/net/xen-netback/rx.c pkt->frag_iter = frag_iter->next; next 809 drivers/net/xen-netfront.c goto next; next 822 drivers/net/xen-netfront.c goto next; next 832 drivers/net/xen-netfront.c next: next 1339 drivers/nfc/pn533/pn533.c u8 *next, nfcid3[NFC_NFCID3_MAXSIZE]; next 1369 drivers/nfc/pn533/pn533.c next = skb_put(skb, 1); /* Next */ next 1370 drivers/nfc/pn533/pn533.c *next = 0; next 1374 drivers/nfc/pn533/pn533.c *next |= 1; next 1378 drivers/nfc/pn533/pn533.c *next |= 2; next 1381 drivers/nfc/pn533/pn533.c *next |= 4; /* We have some Gi */ next 1807 drivers/nfc/pn533/pn533.c u8 *next, *arg, nfcid3[NFC_NFCID3_MAXSIZE]; next 1842 drivers/nfc/pn533/pn533.c next = skb_put(skb, 1); /* Next */ next 1843 drivers/nfc/pn533/pn533.c *next = 0; next 1847 drivers/nfc/pn533/pn533.c *next |= 1; next 1855 drivers/nfc/pn533/pn533.c *next |= 2; next 1859 drivers/nfc/pn533/pn533.c *next |= 4; /* We have some Gi */ next 1861 drivers/nfc/pn533/pn533.c *next = 0; next 106 drivers/nvdimm/badrange.c struct badrange_entry *bre, *next; next 118 drivers/nvdimm/badrange.c list_for_each_entry_safe(bre, next, badrange_list, list) { next 804 drivers/nvdimm/btt.c struct arena_info *arena, *next; next 806 drivers/nvdimm/btt.c list_for_each_entry_safe(arena, next, &btt->arena_list, list) { next 416 drivers/nvdimm/bus.c struct badrange_entry *bre, *next; next 418 drivers/nvdimm/bus.c list_for_each_entry_safe(bre, next, badrange_list, list) { next 565 drivers/nvdimm/namespace_devs.c struct resource *next, struct resource *exist, next 635 drivers/nvdimm/namespace_devs.c struct resource *next = res->sibling, *new_res = NULL; next 651 drivers/nvdimm/namespace_devs.c space_valid(nd_region, ndd, label_id, NULL, next, exist, next 659 drivers/nvdimm/namespace_devs.c if (!loc && next) { next 661 drivers/nvdimm/namespace_devs.c valid.end = min(mapping_end, next->start - 1); next 662 drivers/nvdimm/namespace_devs.c space_valid(nd_region, ndd, label_id, res, next, exist, next 670 drivers/nvdimm/namespace_devs.c if (!loc && !next) { next 673 drivers/nvdimm/namespace_devs.c space_valid(nd_region, ndd, label_id, res, next, exist, next 694 drivers/nvdimm/namespace_devs.c if (strcmp(next->name, label_id->id) == 0) { next 696 drivers/nvdimm/namespace_devs.c rc = adjust_resource(next, next->start next 697 drivers/nvdimm/namespace_devs.c - allocate, resource_size(next) next 699 drivers/nvdimm/namespace_devs.c new_res = next; next 777 drivers/nvdimm/namespace_devs.c struct resource *next = res->sibling; next 780 drivers/nvdimm/namespace_devs.c if (!next || strcmp(res->name, label_id->id) != 0 next 781 drivers/nvdimm/namespace_devs.c || strcmp(next->name, label_id->id) != 0 next 782 drivers/nvdimm/namespace_devs.c || end != next->start) next 784 drivers/nvdimm/namespace_devs.c end += resource_size(next); next 785 drivers/nvdimm/namespace_devs.c nvdimm_free_dpa(ndd, next); next 99 drivers/nvdimm/nd.h #define for_each_dpa_resource_safe(ndd, res, next) \ next 100 drivers/nvdimm/nd.h for (res = (ndd)->dpa.child, next = res ? res->sibling : NULL; \ next 101 drivers/nvdimm/nd.h res; res = next, next = next ? next->sibling : NULL) next 175 drivers/nvdimm/nd.h static const unsigned next[] = { 0, 2, 3, 1 }; next 177 drivers/nvdimm/nd.h return next[seq & 3]; next 3631 drivers/nvme/host/core.c struct nvme_ns *ns, *next; next 3635 drivers/nvme/host/core.c list_for_each_entry_safe(ns, next, &ctrl->namespaces, list) { next 3641 drivers/nvme/host/core.c list_for_each_entry_safe(ns, next, &rm_list, list) next 3764 drivers/nvme/host/core.c struct nvme_ns *ns, *next; next 3790 drivers/nvme/host/core.c list_for_each_entry_safe(ns, next, &ns_list, list) next 343 drivers/nvme/host/multipath.c struct bio *bio, *next; next 346 drivers/nvme/host/multipath.c next = bio_list_get(&head->requeue_list); next 349 drivers/nvme/host/multipath.c while ((bio = next) != NULL) { next 350 drivers/nvme/host/multipath.c next = bio->bi_next; next 1137 drivers/nvme/host/rdma.c .next = NULL, next 1353 drivers/nvme/host/rdma.c wr.next = NULL; next 1361 drivers/nvme/host/rdma.c first->next = ≀ next 1386 drivers/nvme/host/rdma.c wr.next = NULL; next 1135 drivers/nvme/target/fc.c struct nvmet_fc_tgt_assoc *assoc, *next; next 1139 drivers/nvme/target/fc.c list_for_each_entry_safe(assoc, next, next 1155 drivers/nvme/target/fc.c struct nvmet_fc_tgtport *tgtport, *next; next 1163 drivers/nvme/target/fc.c list_for_each_entry_safe(tgtport, next, &nvmet_fc_target_list, next 704 drivers/nvme/target/loop.c struct nvme_loop_ctrl *ctrl, *next; next 710 drivers/nvme/target/loop.c list_for_each_entry_safe(ctrl, next, &nvme_loop_ctrl_list, list) next 482 drivers/nvme/target/rdma.c rsp = list_entry(queue->rsp_wr_wait_list.next, next 454 drivers/nvme/target/tcp.c node = node->next; next 261 drivers/of/base.c for (pp = np->properties; pp; pp = pp->next) { next 731 drivers/of/base.c struct device_node *next; next 736 drivers/of/base.c next = prev ? prev->sibling : node->child; next 737 drivers/of/base.c for (; next; next = next->sibling) next 738 drivers/of/base.c if (of_node_get(next)) next 741 drivers/of/base.c return next; next 759 drivers/of/base.c struct device_node *next; next 763 drivers/of/base.c next = __of_get_next_child(node, prev); next 765 drivers/of/base.c return next; next 780 drivers/of/base.c struct device_node *next; next 787 drivers/of/base.c next = prev ? prev->sibling : node->child; next 788 drivers/of/base.c for (; next; next = next->sibling) { next 789 drivers/of/base.c if (!__of_device_is_available(next)) next 791 drivers/of/base.c if (of_node_get(next)) next 796 drivers/of/base.c return next; next 810 drivers/of/base.c struct device_node *next = NULL; next 819 drivers/of/base.c next = prev->sibling; next 821 drivers/of/base.c next = node->child; next 824 drivers/of/base.c for (; next; next = next->sibling) { next 825 drivers/of/base.c if (!(of_node_name_eq(next, "cpu") || next 826 drivers/of/base.c __of_node_is_type(next, "cpu"))) next 828 drivers/of/base.c if (of_node_get(next)) next 833 drivers/of/base.c return next; next 1093 drivers/of/base.c for (pp = np->properties; pp; pp = pp->next) { next 1818 drivers/of/base.c struct property **next; next 1820 drivers/of/base.c prop->next = NULL; next 1821 drivers/of/base.c next = &np->properties; next 1822 drivers/of/base.c while (*next) { next 1823 drivers/of/base.c if (strcmp(prop->name, (*next)->name) == 0) next 1827 drivers/of/base.c next = &(*next)->next; next 1829 drivers/of/base.c *next = prop; next 1861 drivers/of/base.c struct property **next; next 1863 drivers/of/base.c for (next = &np->properties; *next; next = &(*next)->next) { next 1864 drivers/of/base.c if (*next == prop) next 1867 drivers/of/base.c if (*next == NULL) next 1871 drivers/of/base.c *next = prop->next; next 1872 drivers/of/base.c prop->next = np->deadprops; next 1914 drivers/of/base.c struct property **next, *oldprop; next 1916 drivers/of/base.c for (next = &np->properties; *next; next = &(*next)->next) { next 1917 drivers/of/base.c if (of_prop_cmp((*next)->name, newprop->name) == 0) next 1920 drivers/of/base.c *oldpropp = oldprop = *next; next 1924 drivers/of/base.c newprop->next = oldprop->next; next 1925 drivers/of/base.c *next = newprop; next 1926 drivers/of/base.c oldprop->next = np->deadprops; next 1930 drivers/of/base.c newprop->next = NULL; next 1931 drivers/of/base.c *next = newprop; next 310 drivers/of/dynamic.c struct property *prop, *next; next 312 drivers/of/dynamic.c for (prop = prop_list; prop != NULL; prop = next) { next 313 drivers/of/dynamic.c next = prop->next; next 584 drivers/of/dynamic.c for (propp = &ce->np->deadprops; *propp; propp = &(*propp)->next) { next 586 drivers/of/dynamic.c *propp = ce->prop->next; next 587 drivers/of/dynamic.c ce->prop->next = NULL; next 612 drivers/of/dynamic.c for (propp = &ce->np->deadprops; *propp; propp = &(*propp)->next) { next 614 drivers/of/dynamic.c *propp = ce->prop->next; next 615 drivers/of/dynamic.c ce->prop->next = NULL; next 168 drivers/of/fdt.c pprev = &pp->next; next 196 drivers/of/fdt.c pprev = &pp->next; next 256 drivers/of/fdt.c struct device_node *child, *next; next 270 drivers/of/fdt.c next = child->sibling; next 274 drivers/of/fdt.c child = next; next 216 drivers/of/of_reserved_mem.c struct reserved_mem *this, *next; next 219 drivers/of/of_reserved_mem.c next = &reserved_mem[i + 1]; next 220 drivers/of/of_reserved_mem.c if (!(this->base && next->base)) next 222 drivers/of/of_reserved_mem.c if (this->base + this->size > next->base) { next 226 drivers/of/of_reserved_mem.c next_end = next->base + next->size; next 229 drivers/of/of_reserved_mem.c next->name, &next->base, &next_end); next 354 drivers/of/overlay.c new_prop->next = target->np->deadprops; next 851 drivers/of/overlay.c if (ovcs->cset.entries.next) next 119 drivers/of/pdt.c tail->next = of_pdt_build_one_prop(node, NULL, NULL, NULL, 0); next 120 drivers/of/pdt.c tail = tail->next; next 122 drivers/of/pdt.c tail->next = of_pdt_build_one_prop(node, tail->name, next 124 drivers/of/pdt.c tail = tail->next; next 1126 drivers/of/unittest.c save_next = prop->next; next 1149 drivers/of/unittest.c struct device_node *next, *dup, *child; next 1182 drivers/of/unittest.c next = child->sibling; next 1184 drivers/of/unittest.c child = next; next 1248 drivers/of/unittest.c struct device_node *next = np->sibling; next 1252 drivers/of/unittest.c np = next; next 252 drivers/parisc/ccio-dma.c struct ioc *next; /* Linked list of discovered iocs */ next 1091 drivers/parisc/ccio-dma.c ioc = ioc->next; next 1105 drivers/parisc/ccio-dma.c ioc = ioc->next; next 1130 drivers/parisc/ccio-dma.c ioc = ioc->next; next 1532 drivers/parisc/ccio-dma.c ioc_p = &(*ioc_p)->next; next 28 drivers/parisc/hppb.c struct hppb_card *next; next 33 drivers/parisc/hppb.c .next = NULL, next 52 drivers/parisc/hppb.c while(card->next) { next 53 drivers/parisc/hppb.c card = card->next; next 57 drivers/parisc/hppb.c card->next = kzalloc(sizeof(struct hppb_card), GFP_KERNEL); next 58 drivers/parisc/hppb.c if(!card->next) { next 62 drivers/parisc/hppb.c card = card->next; next 1682 drivers/parisc/sba_iommu.c sba_dev->next = sba_list; next 40 drivers/parport/daisy.c struct daisydev *next; next 65 drivers/parport/daisy.c for (p = &topology; *p && (*p)->devnum<devnum; p = &(*p)->next) next 67 drivers/parport/daisy.c newdev->next = *p; next 184 drivers/parport/daisy.c p = &dev->next; next 187 drivers/parport/daisy.c *p = dev->next; next 222 drivers/parport/daisy.c p = p->next; next 468 drivers/parport/parport_ip32.c dma_addr_t next; next 510 drivers/parport/parport_ip32.c (unsigned int)parport_ip32_dma.next, next 514 drivers/parport/parport_ip32.c ctxval |= parport_ip32_dma.next & next 519 drivers/parport/parport_ip32.c parport_ip32_dma.next += count; next 594 drivers/parport/parport_ip32.c parport_ip32_dma.next = parport_ip32_dma.buf; next 603 drivers/parport/parport_ip32.c (parport_ip32_dma.next & (MACEPAR_CONTEXT_DATA_BOUND - 1)); next 3348 drivers/parport/parport_pc.c priv = list_entry(ports_list.next, next 52 drivers/parport/procfs.c for (dev = port->devices; dev ; dev = dev->next) { next 501 drivers/parport/share.c for (l = all_ports.next, num = 0; l != &all_ports; l = l->next, num++) { next 800 drivers/parport/share.c tmp->next = port->physport->devices; next 944 drivers/parport/share.c par_dev->next = port->physport->devices; next 1020 drivers/parport/share.c if (dev->next) next 1021 drivers/parport/share.c dev->next->prev = dev->prev; next 1023 drivers/parport/share.c dev->prev->next = dev->next; next 1025 drivers/parport/share.c port->devices = dev->next; next 1374 drivers/parport/share.c for (pd = port->devices; !port->cad && pd; pd = pd->next) { next 383 drivers/pci/bus.c struct list_head *next; next 388 drivers/pci/bus.c next = top->devices.next; next 390 drivers/pci/bus.c if (next == &bus->devices) { next 394 drivers/pci/bus.c next = bus->self->bus_list.next; next 398 drivers/pci/bus.c dev = list_entry(next, struct pci_dev, bus_list); next 401 drivers/pci/bus.c next = dev->subordinate->devices.next; next 404 drivers/pci/bus.c next = dev->bus_list.next; next 2708 drivers/pci/controller/pci-tegra.c .next = tegra_pcie_ports_seq_next, next 817 drivers/pci/controller/pcie-rcar.c desc = list_entry(pdev->dev.msi_list.next, struct msi_desc, list); next 145 drivers/pci/hotplug/acpiphp_glue.c struct acpiphp_slot *slot, *next; next 152 drivers/pci/hotplug/acpiphp_glue.c list_for_each_entry_safe(slot, next, &bridge->slots, node) { next 226 drivers/pci/hotplug/cpqphp.h struct pci_func *next; next 248 drivers/pci/hotplug/cpqphp.h struct slot *next; next 267 drivers/pci/hotplug/cpqphp.h struct pci_resource *next; next 278 drivers/pci/hotplug/cpqphp.h struct controller *next; next 466 drivers/pci/hotplug/cpqphp.h node->next = *head; next 274 drivers/pci/hotplug/cpqphp_core.c next_slot = old_slot->next; next 665 drivers/pci/hotplug/cpqphp_core.c slot->next = ctrl->slot; next 1165 drivers/pci/hotplug/cpqphp_core.c ctrl->next = NULL; next 1167 drivers/pci/hotplug/cpqphp_core.c ctrl->next = cpqhp_ctrl_list; next 1247 drivers/pci/hotplug/cpqphp_core.c struct pci_func *next; next 1278 drivers/pci/hotplug/cpqphp_core.c res = res->next; next 1285 drivers/pci/hotplug/cpqphp_core.c res = res->next; next 1292 drivers/pci/hotplug/cpqphp_core.c res = res->next; next 1299 drivers/pci/hotplug/cpqphp_core.c res = res->next; next 1306 drivers/pci/hotplug/cpqphp_core.c ctrl = ctrl->next; next 1311 drivers/pci/hotplug/cpqphp_core.c next = cpqhp_slot_list[loop]; next 1312 drivers/pci/hotplug/cpqphp_core.c while (next != NULL) { next 1313 drivers/pci/hotplug/cpqphp_core.c res = next->io_head; next 1316 drivers/pci/hotplug/cpqphp_core.c res = res->next; next 1320 drivers/pci/hotplug/cpqphp_core.c res = next->mem_head; next 1323 drivers/pci/hotplug/cpqphp_core.c res = res->next; next 1327 drivers/pci/hotplug/cpqphp_core.c res = next->p_mem_head; next 1330 drivers/pci/hotplug/cpqphp_core.c res = res->next; next 1334 drivers/pci/hotplug/cpqphp_core.c res = next->bus_head; next 1337 drivers/pci/hotplug/cpqphp_core.c res = res->next; next 1341 drivers/pci/hotplug/cpqphp_core.c TempSlot = next; next 1342 drivers/pci/hotplug/cpqphp_core.c next = next->next; next 121 drivers/pci/hotplug/cpqphp_ctrl.c slot = slot->next; next 307 drivers/pci/hotplug/cpqphp_ctrl.c if (!((*head)->next)) next 314 drivers/pci/hotplug/cpqphp_ctrl.c if (((*head)->next) && next 315 drivers/pci/hotplug/cpqphp_ctrl.c ((*head)->length > (*head)->next->length)) { next 318 drivers/pci/hotplug/cpqphp_ctrl.c *head = (*head)->next; next 319 drivers/pci/hotplug/cpqphp_ctrl.c current_res->next = (*head)->next; next 320 drivers/pci/hotplug/cpqphp_ctrl.c (*head)->next = current_res; next 325 drivers/pci/hotplug/cpqphp_ctrl.c while (current_res->next && current_res->next->next) { next 326 drivers/pci/hotplug/cpqphp_ctrl.c if (current_res->next->length > current_res->next->next->length) { next 328 drivers/pci/hotplug/cpqphp_ctrl.c next_res = current_res->next; next 329 drivers/pci/hotplug/cpqphp_ctrl.c current_res->next = current_res->next->next; next 330 drivers/pci/hotplug/cpqphp_ctrl.c current_res = current_res->next; next 331 drivers/pci/hotplug/cpqphp_ctrl.c next_res->next = current_res->next; next 332 drivers/pci/hotplug/cpqphp_ctrl.c current_res->next = next_res; next 334 drivers/pci/hotplug/cpqphp_ctrl.c current_res = current_res->next; next 355 drivers/pci/hotplug/cpqphp_ctrl.c if (!((*head)->next)) next 362 drivers/pci/hotplug/cpqphp_ctrl.c if (((*head)->next) && next 363 drivers/pci/hotplug/cpqphp_ctrl.c ((*head)->length < (*head)->next->length)) { next 366 drivers/pci/hotplug/cpqphp_ctrl.c *head = (*head)->next; next 367 drivers/pci/hotplug/cpqphp_ctrl.c current_res->next = (*head)->next; next 368 drivers/pci/hotplug/cpqphp_ctrl.c (*head)->next = current_res; next 373 drivers/pci/hotplug/cpqphp_ctrl.c while (current_res->next && current_res->next->next) { next 374 drivers/pci/hotplug/cpqphp_ctrl.c if (current_res->next->length < current_res->next->next->length) { next 376 drivers/pci/hotplug/cpqphp_ctrl.c next_res = current_res->next; next 377 drivers/pci/hotplug/cpqphp_ctrl.c current_res->next = current_res->next->next; next 378 drivers/pci/hotplug/cpqphp_ctrl.c current_res = current_res->next; next 379 drivers/pci/hotplug/cpqphp_ctrl.c next_res->next = current_res->next; next 380 drivers/pci/hotplug/cpqphp_ctrl.c current_res->next = next_res; next 382 drivers/pci/hotplug/cpqphp_ctrl.c current_res = current_res->next; next 446 drivers/pci/hotplug/cpqphp_ctrl.c split_node->next = node; next 454 drivers/pci/hotplug/cpqphp_ctrl.c *head = node->next; next 457 drivers/pci/hotplug/cpqphp_ctrl.c while (prevnode->next != node) next 458 drivers/pci/hotplug/cpqphp_ctrl.c prevnode = prevnode->next; next 460 drivers/pci/hotplug/cpqphp_ctrl.c prevnode->next = node->next; next 462 drivers/pci/hotplug/cpqphp_ctrl.c node->next = NULL; next 487 drivers/pci/hotplug/cpqphp_ctrl.c while (node->next) { next 489 drivers/pci/hotplug/cpqphp_ctrl.c node = node->next; next 542 drivers/pci/hotplug/cpqphp_ctrl.c for (node = *head; node; node = node->next) { next 567 drivers/pci/hotplug/cpqphp_ctrl.c split_node->next = node->next; next 568 drivers/pci/hotplug/cpqphp_ctrl.c node->next = split_node; next 586 drivers/pci/hotplug/cpqphp_ctrl.c split_node->next = node->next; next 587 drivers/pci/hotplug/cpqphp_ctrl.c node->next = split_node; next 598 drivers/pci/hotplug/cpqphp_ctrl.c *head = node->next; next 601 drivers/pci/hotplug/cpqphp_ctrl.c while (prevnode->next != node) next 602 drivers/pci/hotplug/cpqphp_ctrl.c prevnode = prevnode->next; next 604 drivers/pci/hotplug/cpqphp_ctrl.c prevnode->next = node->next; next 606 drivers/pci/hotplug/cpqphp_ctrl.c node->next = NULL; next 636 drivers/pci/hotplug/cpqphp_ctrl.c for (max = *head; max; max = max->next) { next 663 drivers/pci/hotplug/cpqphp_ctrl.c split_node->next = max->next; next 664 drivers/pci/hotplug/cpqphp_ctrl.c max->next = split_node; next 681 drivers/pci/hotplug/cpqphp_ctrl.c split_node->next = max->next; next 682 drivers/pci/hotplug/cpqphp_ctrl.c max->next = split_node; next 692 drivers/pci/hotplug/cpqphp_ctrl.c *head = max->next; next 694 drivers/pci/hotplug/cpqphp_ctrl.c while (temp && temp->next != max) next 695 drivers/pci/hotplug/cpqphp_ctrl.c temp = temp->next; next 698 drivers/pci/hotplug/cpqphp_ctrl.c temp->next = max->next; next 701 drivers/pci/hotplug/cpqphp_ctrl.c max->next = NULL; next 733 drivers/pci/hotplug/cpqphp_ctrl.c for (node = *head; node; node = node->next) { next 760 drivers/pci/hotplug/cpqphp_ctrl.c split_node->next = node->next; next 761 drivers/pci/hotplug/cpqphp_ctrl.c node->next = split_node; next 780 drivers/pci/hotplug/cpqphp_ctrl.c split_node->next = node->next; next 781 drivers/pci/hotplug/cpqphp_ctrl.c node->next = split_node; next 788 drivers/pci/hotplug/cpqphp_ctrl.c *head = node->next; next 791 drivers/pci/hotplug/cpqphp_ctrl.c while (prevnode->next != node) next 792 drivers/pci/hotplug/cpqphp_ctrl.c prevnode = prevnode->next; next 794 drivers/pci/hotplug/cpqphp_ctrl.c prevnode->next = node->next; next 796 drivers/pci/hotplug/cpqphp_ctrl.c node->next = NULL; next 824 drivers/pci/hotplug/cpqphp_ctrl.c dbg("*head->next = %p\n", (*head)->next); next 826 drivers/pci/hotplug/cpqphp_ctrl.c if (!(*head)->next) next 830 drivers/pci/hotplug/cpqphp_ctrl.c dbg("*head->next->base = 0x%x\n", (*head)->next->base); next 835 drivers/pci/hotplug/cpqphp_ctrl.c if (((*head)->next) && next 836 drivers/pci/hotplug/cpqphp_ctrl.c ((*head)->base > (*head)->next->base)) { next 838 drivers/pci/hotplug/cpqphp_ctrl.c (*head) = (*head)->next; next 839 drivers/pci/hotplug/cpqphp_ctrl.c node1->next = (*head)->next; next 840 drivers/pci/hotplug/cpqphp_ctrl.c (*head)->next = node1; next 846 drivers/pci/hotplug/cpqphp_ctrl.c while (node1->next && node1->next->next) { next 847 drivers/pci/hotplug/cpqphp_ctrl.c if (node1->next->base > node1->next->next->base) { next 849 drivers/pci/hotplug/cpqphp_ctrl.c node2 = node1->next; next 850 drivers/pci/hotplug/cpqphp_ctrl.c node1->next = node1->next->next; next 851 drivers/pci/hotplug/cpqphp_ctrl.c node1 = node1->next; next 852 drivers/pci/hotplug/cpqphp_ctrl.c node2->next = node1->next; next 853 drivers/pci/hotplug/cpqphp_ctrl.c node1->next = node2; next 855 drivers/pci/hotplug/cpqphp_ctrl.c node1 = node1->next; next 861 drivers/pci/hotplug/cpqphp_ctrl.c while (node1 && node1->next) { next 862 drivers/pci/hotplug/cpqphp_ctrl.c if ((node1->base + node1->length) == node1->next->base) { next 865 drivers/pci/hotplug/cpqphp_ctrl.c node1->length += node1->next->length; next 866 drivers/pci/hotplug/cpqphp_ctrl.c node2 = node1->next; next 867 drivers/pci/hotplug/cpqphp_ctrl.c node1->next = node1->next->next; next 870 drivers/pci/hotplug/cpqphp_ctrl.c node1 = node1->next; next 957 drivers/pci/hotplug/cpqphp_ctrl.c struct pci_func *next; next 963 drivers/pci/hotplug/cpqphp_ctrl.c new_slot->next = NULL; next 969 drivers/pci/hotplug/cpqphp_ctrl.c next = cpqhp_slot_list[busnumber]; next 970 drivers/pci/hotplug/cpqphp_ctrl.c while (next->next != NULL) next 971 drivers/pci/hotplug/cpqphp_ctrl.c next = next->next; next 972 drivers/pci/hotplug/cpqphp_ctrl.c next->next = new_slot; next 986 drivers/pci/hotplug/cpqphp_ctrl.c struct pci_func *next; next 991 drivers/pci/hotplug/cpqphp_ctrl.c next = cpqhp_slot_list[old_slot->bus]; next 992 drivers/pci/hotplug/cpqphp_ctrl.c if (next == NULL) next 995 drivers/pci/hotplug/cpqphp_ctrl.c if (next == old_slot) { next 996 drivers/pci/hotplug/cpqphp_ctrl.c cpqhp_slot_list[old_slot->bus] = old_slot->next; next 1002 drivers/pci/hotplug/cpqphp_ctrl.c while ((next->next != old_slot) && (next->next != NULL)) next 1003 drivers/pci/hotplug/cpqphp_ctrl.c next = next->next; next 1005 drivers/pci/hotplug/cpqphp_ctrl.c if (next->next == old_slot) { next 1006 drivers/pci/hotplug/cpqphp_ctrl.c next->next = old_slot->next; next 1025 drivers/pci/hotplug/cpqphp_ctrl.c struct pci_func *next; next 1031 drivers/pci/hotplug/cpqphp_ctrl.c next = cpqhp_slot_list[tempBus]; next 1033 drivers/pci/hotplug/cpqphp_ctrl.c while (!slot_remove(next)) next 1034 drivers/pci/hotplug/cpqphp_ctrl.c next = cpqhp_slot_list[tempBus]; next 1037 drivers/pci/hotplug/cpqphp_ctrl.c next = cpqhp_slot_list[bridge->bus]; next 1039 drivers/pci/hotplug/cpqphp_ctrl.c if (next == NULL) next 1042 drivers/pci/hotplug/cpqphp_ctrl.c if (next == bridge) { next 1043 drivers/pci/hotplug/cpqphp_ctrl.c cpqhp_slot_list[bridge->bus] = bridge->next; next 1047 drivers/pci/hotplug/cpqphp_ctrl.c while ((next->next != bridge) && (next->next != NULL)) next 1048 drivers/pci/hotplug/cpqphp_ctrl.c next = next->next; next 1050 drivers/pci/hotplug/cpqphp_ctrl.c if (next->next != bridge) next 1052 drivers/pci/hotplug/cpqphp_ctrl.c next->next = bridge->next; next 1080 drivers/pci/hotplug/cpqphp_ctrl.c while (func->next != NULL) { next 1081 drivers/pci/hotplug/cpqphp_ctrl.c func = func->next; next 1130 drivers/pci/hotplug/cpqphp_ctrl.c for (slot = ctrl->slot; slot; slot = slot->next) { next 1743 drivers/pci/hotplug/cpqphp_ctrl.c for (ctrl = cpqhp_ctrl_list; ctrl; ctrl = ctrl->next) next 2414 drivers/pci/hotplug/cpqphp_ctrl.c io_node->length, io_node->next); next 2417 drivers/pci/hotplug/cpqphp_ctrl.c mem_node->length, mem_node->next); next 2420 drivers/pci/hotplug/cpqphp_ctrl.c p_mem_node->length, p_mem_node->next); next 2468 drivers/pci/hotplug/cpqphp_ctrl.c bus_node->next = NULL; next 2473 drivers/pci/hotplug/cpqphp_ctrl.c io_node->next = NULL; next 2486 drivers/pci/hotplug/cpqphp_ctrl.c mem_node->next = NULL; next 2496 drivers/pci/hotplug/cpqphp_ctrl.c p_mem_node->next = NULL; next 2566 drivers/pci/hotplug/cpqphp_ctrl.c hold_bus_node->next = func->bus_head; next 2608 drivers/pci/hotplug/cpqphp_ctrl.c hold_IO_node->next = func->io_head; next 2625 drivers/pci/hotplug/cpqphp_ctrl.c hold_IO_node->next = func->io_head; next 2630 drivers/pci/hotplug/cpqphp_ctrl.c hold_IO_node->next = func->io_head; next 2658 drivers/pci/hotplug/cpqphp_ctrl.c hold_mem_node->next = func->mem_head; next 2677 drivers/pci/hotplug/cpqphp_ctrl.c hold_mem_node->next = func->mem_head; next 2682 drivers/pci/hotplug/cpqphp_ctrl.c hold_mem_node->next = func->mem_head; next 2711 drivers/pci/hotplug/cpqphp_ctrl.c hold_p_mem_node->next = func->p_mem_head; next 2728 drivers/pci/hotplug/cpqphp_ctrl.c hold_p_mem_node->next = func->p_mem_head; next 2733 drivers/pci/hotplug/cpqphp_ctrl.c hold_p_mem_node->next = func->p_mem_head; next 2785 drivers/pci/hotplug/cpqphp_ctrl.c io_node->base, io_node->length, io_node->next); next 2790 drivers/pci/hotplug/cpqphp_ctrl.c io_node->next = func->io_head; next 2804 drivers/pci/hotplug/cpqphp_ctrl.c p_mem_node->next = func->p_mem_head; next 2820 drivers/pci/hotplug/cpqphp_ctrl.c mem_node->next = func->mem_head; next 61 drivers/pci/hotplug/cpqphp_nvram.c u8 next; next 72 drivers/pci/hotplug/cpqphp_nvram.c u8 next; next 307 drivers/pci/hotplug/cpqphp_nvram.c resNode = resNode->next; next 332 drivers/pci/hotplug/cpqphp_nvram.c resNode = resNode->next; next 357 drivers/pci/hotplug/cpqphp_nvram.c resNode = resNode->next; next 382 drivers/pci/hotplug/cpqphp_nvram.c resNode = resNode->next; next 388 drivers/pci/hotplug/cpqphp_nvram.c ctrl = ctrl->next; next 455 drivers/pci/hotplug/cpqphp_nvram.c p_byte = &(p_EV_header->next); next 457 drivers/pci/hotplug/cpqphp_nvram.c p_ev_ctrl = (struct ev_hrt_ctrl *) &(p_EV_header->next); next 533 drivers/pci/hotplug/cpqphp_nvram.c mem_node->next = ctrl->mem_head; next 561 drivers/pci/hotplug/cpqphp_nvram.c p_mem_node->next = ctrl->p_mem_head; next 589 drivers/pci/hotplug/cpqphp_nvram.c io_node->next = ctrl->io_head; next 615 drivers/pci/hotplug/cpqphp_nvram.c bus_node->next = ctrl->bus_head; next 558 drivers/pci/hotplug/cpqphp_pci.c struct pci_func *next; next 577 drivers/pci/hotplug/cpqphp_pci.c next = cpqhp_slot_list[sub_bus]; next 579 drivers/pci/hotplug/cpqphp_pci.c while (next != NULL) { next 580 drivers/pci/hotplug/cpqphp_pci.c rc = cpqhp_save_base_addr_length(ctrl, next); next 584 drivers/pci/hotplug/cpqphp_pci.c next = next->next; next 738 drivers/pci/hotplug/cpqphp_pci.c bus_node->next = func->bus_head; next 753 drivers/pci/hotplug/cpqphp_pci.c io_node->next = func->io_head; next 769 drivers/pci/hotplug/cpqphp_pci.c mem_node->next = func->mem_head; next 785 drivers/pci/hotplug/cpqphp_pci.c p_mem_node->next = func->p_mem_head; next 818 drivers/pci/hotplug/cpqphp_pci.c io_node->next = func->io_head; next 835 drivers/pci/hotplug/cpqphp_pci.c p_mem_node->next = func->p_mem_head; next 852 drivers/pci/hotplug/cpqphp_pci.c mem_node->next = func->mem_head; next 889 drivers/pci/hotplug/cpqphp_pci.c io_node->next = func->io_head; next 906 drivers/pci/hotplug/cpqphp_pci.c p_mem_node->next = func->p_mem_head; next 923 drivers/pci/hotplug/cpqphp_pci.c mem_node->next = func->mem_head; next 954 drivers/pci/hotplug/cpqphp_pci.c struct pci_func *next; next 981 drivers/pci/hotplug/cpqphp_pci.c next = cpqhp_slot_list[sub_bus]; next 983 drivers/pci/hotplug/cpqphp_pci.c while (next != NULL) { next 984 drivers/pci/hotplug/cpqphp_pci.c rc = cpqhp_configure_board(ctrl, next); next 988 drivers/pci/hotplug/cpqphp_pci.c next = next->next; next 1035 drivers/pci/hotplug/cpqphp_pci.c struct pci_func *next; next 1079 drivers/pci/hotplug/cpqphp_pci.c next = cpqhp_slot_list[secondary_bus]; next 1081 drivers/pci/hotplug/cpqphp_pci.c while (next != NULL) { next 1082 drivers/pci/hotplug/cpqphp_pci.c rc = cpqhp_valid_replace(ctrl, next); next 1086 drivers/pci/hotplug/cpqphp_pci.c next = next->next; next 1314 drivers/pci/hotplug/cpqphp_pci.c io_node->next = ctrl->io_head; next 1317 drivers/pci/hotplug/cpqphp_pci.c io_node->next = func->io_head; next 1337 drivers/pci/hotplug/cpqphp_pci.c mem_node->next = ctrl->mem_head; next 1340 drivers/pci/hotplug/cpqphp_pci.c mem_node->next = func->mem_head; next 1362 drivers/pci/hotplug/cpqphp_pci.c p_mem_node->next = ctrl->p_mem_head; next 1365 drivers/pci/hotplug/cpqphp_pci.c p_mem_node->next = func->p_mem_head; next 1385 drivers/pci/hotplug/cpqphp_pci.c bus_node->next = ctrl->bus_head; next 1388 drivers/pci/hotplug/cpqphp_pci.c bus_node->next = func->bus_head; next 1431 drivers/pci/hotplug/cpqphp_pci.c t_node = node->next; next 1439 drivers/pci/hotplug/cpqphp_pci.c t_node = node->next; next 1447 drivers/pci/hotplug/cpqphp_pci.c t_node = node->next; next 1455 drivers/pci/hotplug/cpqphp_pci.c t_node = node->next; next 1483 drivers/pci/hotplug/cpqphp_pci.c res = res->next; next 1492 drivers/pci/hotplug/cpqphp_pci.c res = res->next; next 1501 drivers/pci/hotplug/cpqphp_pci.c res = res->next; next 1510 drivers/pci/hotplug/cpqphp_pci.c res = res->next; next 1530 drivers/pci/hotplug/cpqphp_pci.c res = res->next; next 1539 drivers/pci/hotplug/cpqphp_pci.c res = res->next; next 1548 drivers/pci/hotplug/cpqphp_pci.c res = res->next; next 1557 drivers/pci/hotplug/cpqphp_pci.c res = res->next; next 39 drivers/pci/hotplug/cpqphp_sysfs.c res = res->next; next 46 drivers/pci/hotplug/cpqphp_sysfs.c res = res->next; next 53 drivers/pci/hotplug/cpqphp_sysfs.c res = res->next; next 60 drivers/pci/hotplug/cpqphp_sysfs.c res = res->next; next 85 drivers/pci/hotplug/cpqphp_sysfs.c res = res->next; next 92 drivers/pci/hotplug/cpqphp_sysfs.c res = res->next; next 99 drivers/pci/hotplug/cpqphp_sysfs.c res = res->next; next 106 drivers/pci/hotplug/cpqphp_sysfs.c res = res->next; next 108 drivers/pci/hotplug/cpqphp_sysfs.c slot = slot->next; next 214 drivers/pci/hotplug/ibmphp.h struct ebda_rsrc_list *next; next 325 drivers/pci/hotplug/ibmphp.h struct range_node *next; next 356 drivers/pci/hotplug/ibmphp.h struct resource_node *next; next 684 drivers/pci/hotplug/ibmphp.h struct pci_func *next; next 622 drivers/pci/hotplug/ibmphp_core.c func_cur = func_cur->next; next 636 drivers/pci/hotplug/ibmphp_core.c struct slot *slot_cur, *next; next 640 drivers/pci/hotplug/ibmphp_core.c list_for_each_entry_safe(slot_cur, next, &ibmphp_slot_head, next 1036 drivers/pci/hotplug/ibmphp_ebda.c struct bus_info *bus_info, *next; next 1038 drivers/pci/hotplug/ibmphp_ebda.c list_for_each_entry_safe(bus_info, next, &bus_info_head, next 1046 drivers/pci/hotplug/ibmphp_ebda.c struct controller *controller = NULL, *next; next 1049 drivers/pci/hotplug/ibmphp_ebda.c list_for_each_entry_safe(controller, next, &ebda_hpc_head, next 1063 drivers/pci/hotplug/ibmphp_ebda.c struct ebda_pci_rsrc *resource, *next; next 1065 drivers/pci/hotplug/ibmphp_ebda.c list_for_each_entry_safe(resource, next, &ibmphp_ebda_pci_rsrc_head, next 142 drivers/pci/hotplug/ibmphp_pci.c cur_func->next = NULL; next 161 drivers/pci/hotplug/ibmphp_pci.c cur_func->next = newfunc; next 202 drivers/pci/hotplug/ibmphp_pci.c for (prev_func = cur_func; prev_func->next; prev_func = prev_func->next) ; next 203 drivers/pci/hotplug/ibmphp_pci.c prev_func->next = newfunc; next 205 drivers/pci/hotplug/ibmphp_pci.c cur_func->next = newfunc; next 227 drivers/pci/hotplug/ibmphp_pci.c for (prev_func = cur_func; prev_func->next; prev_func = prev_func->next); next 228 drivers/pci/hotplug/ibmphp_pci.c prev_func->next = newfunc; next 274 drivers/pci/hotplug/ibmphp_pci.c for (prev_func = cur_func; prev_func->next; prev_func = prev_func->next); next 275 drivers/pci/hotplug/ibmphp_pci.c prev_func->next = newfunc; next 277 drivers/pci/hotplug/ibmphp_pci.c cur_func->next = newfunc; next 1585 drivers/pci/hotplug/ibmphp_pci.c temp_func = cur_func->next; next 395 drivers/pci/hotplug/ibmphp_res.c range_cur = range_cur->next; next 411 drivers/pci/hotplug/ibmphp_res.c range->next = range_cur; next 416 drivers/pci/hotplug/ibmphp_res.c range->next = NULL; next 417 drivers/pci/hotplug/ibmphp_res.c range_prev->next = range; next 422 drivers/pci/hotplug/ibmphp_res.c range_prev->next = range; next 423 drivers/pci/hotplug/ibmphp_res.c range->next = range_cur; next 430 drivers/pci/hotplug/ibmphp_res.c range_cur = range_cur->next; next 467 drivers/pci/hotplug/ibmphp_res.c if (res->next) next 468 drivers/pci/hotplug/ibmphp_res.c res = res->next; next 481 drivers/pci/hotplug/ibmphp_res.c res = res->next; next 521 drivers/pci/hotplug/ibmphp_res.c range = range->next; next 524 drivers/pci/hotplug/ibmphp_res.c if (res->next) next 525 drivers/pci/hotplug/ibmphp_res.c res = res->next; next 621 drivers/pci/hotplug/ibmphp_res.c range_cur = range_cur->next; next 658 drivers/pci/hotplug/ibmphp_res.c res->next = NULL; next 670 drivers/pci/hotplug/ibmphp_res.c if (res_cur->next) next 671 drivers/pci/hotplug/ibmphp_res.c res_cur = res_cur->next; next 680 drivers/pci/hotplug/ibmphp_res.c res->next = NULL; next 688 drivers/pci/hotplug/ibmphp_res.c res_cur = res_cur->next; next 692 drivers/pci/hotplug/ibmphp_res.c res_prev->next = res; next 693 drivers/pci/hotplug/ibmphp_res.c res->next = NULL; next 711 drivers/pci/hotplug/ibmphp_res.c res_prev->next = res; next 715 drivers/pci/hotplug/ibmphp_res.c res->next = res_cur; next 722 drivers/pci/hotplug/ibmphp_res.c res->next = NULL; next 740 drivers/pci/hotplug/ibmphp_res.c res->next = NULL; next 800 drivers/pci/hotplug/ibmphp_res.c if (res_cur->next) next 801 drivers/pci/hotplug/ibmphp_res.c res_cur = res_cur->next; next 823 drivers/pci/hotplug/ibmphp_res.c if (mem_cur->next) next 824 drivers/pci/hotplug/ibmphp_res.c mem_cur = mem_cur->next; next 835 drivers/pci/hotplug/ibmphp_res.c bus_cur->firstPFMemFromMem = res_cur->next; next 837 drivers/pci/hotplug/ibmphp_res.c res_prev->next = res_cur->next; next 842 drivers/pci/hotplug/ibmphp_res.c if (res_cur->next) next 843 drivers/pci/hotplug/ibmphp_res.c res_cur = res_cur->next; next 858 drivers/pci/hotplug/ibmphp_res.c if (res_cur->next) { next 861 drivers/pci/hotplug/ibmphp_res.c bus_cur->firstIO = res_cur->next; next 864 drivers/pci/hotplug/ibmphp_res.c bus_cur->firstMem = res_cur->next; next 867 drivers/pci/hotplug/ibmphp_res.c bus_cur->firstPFMem = res_cur->next; next 898 drivers/pci/hotplug/ibmphp_res.c if (res_cur->next) { next 900 drivers/pci/hotplug/ibmphp_res.c res_prev->next = res_cur->next; next 902 drivers/pci/hotplug/ibmphp_res.c res_prev->nextRange = res_cur->next; next 904 drivers/pci/hotplug/ibmphp_res.c res_prev->next = NULL; next 907 drivers/pci/hotplug/ibmphp_res.c res_prev->next = NULL; next 938 drivers/pci/hotplug/ibmphp_res.c range = range->next; next 1062 drivers/pci/hotplug/ibmphp_res.c if (!res_cur->next) { next 1181 drivers/pci/hotplug/ibmphp_res.c if (res_cur->next) next 1182 drivers/pci/hotplug/ibmphp_res.c res_cur = res_cur->next; next 1238 drivers/pci/hotplug/ibmphp_res.c range = range->next; next 1304 drivers/pci/hotplug/ibmphp_res.c range = range->next; next 1364 drivers/pci/hotplug/ibmphp_res.c if (res_cur->next) next 1365 drivers/pci/hotplug/ibmphp_res.c res_cur = res_cur->next; next 1377 drivers/pci/hotplug/ibmphp_res.c if (res_cur->next) next 1378 drivers/pci/hotplug/ibmphp_res.c res_cur = res_cur->next; next 1390 drivers/pci/hotplug/ibmphp_res.c if (res_cur->next) next 1391 drivers/pci/hotplug/ibmphp_res.c res_cur = res_cur->next; next 1404 drivers/pci/hotplug/ibmphp_res.c res_cur = res_cur->next; next 1438 drivers/pci/hotplug/ibmphp_res.c range_cur = range_cur->next; next 1452 drivers/pci/hotplug/ibmphp_res.c range_cur = range_cur->next; next 1466 drivers/pci/hotplug/ibmphp_res.c range_cur = range_cur->next; next 1512 drivers/pci/hotplug/ibmphp_res.c if (res_cur->next) next 1513 drivers/pci/hotplug/ibmphp_res.c res_cur = res_cur->next; next 1526 drivers/pci/hotplug/ibmphp_res.c res_cur = res_cur->next; next 1552 drivers/pci/hotplug/ibmphp_res.c struct bus_node *bus_cur = NULL, *next; next 1561 drivers/pci/hotplug/ibmphp_res.c list_for_each_entry_safe(bus_cur, next, &gbuses, bus_list) { next 1568 drivers/pci/hotplug/ibmphp_res.c range_cur = range_cur->next; next 1579 drivers/pci/hotplug/ibmphp_res.c range_cur = range_cur->next; next 1590 drivers/pci/hotplug/ibmphp_res.c range_cur = range_cur->next; next 1600 drivers/pci/hotplug/ibmphp_res.c if (res_cur->next) next 1601 drivers/pci/hotplug/ibmphp_res.c res_cur = res_cur->next; next 1613 drivers/pci/hotplug/ibmphp_res.c if (res_cur->next) next 1614 drivers/pci/hotplug/ibmphp_res.c res_cur = res_cur->next; next 1626 drivers/pci/hotplug/ibmphp_res.c if (res_cur->next) next 1627 drivers/pci/hotplug/ibmphp_res.c res_cur = res_cur->next; next 1640 drivers/pci/hotplug/ibmphp_res.c res_cur = res_cur->next; next 1672 drivers/pci/hotplug/ibmphp_res.c for (pfmem_cur = bus_cur->firstPFMem, pfmem_prev = NULL; pfmem_cur; pfmem_prev = pfmem_cur, pfmem_cur = pfmem_cur->next) { next 1675 drivers/pci/hotplug/ibmphp_res.c pfmem_prev->next = pfmem_cur->next; next 1677 drivers/pci/hotplug/ibmphp_res.c bus_cur->firstPFMem = pfmem_cur->next; next 1680 drivers/pci/hotplug/ibmphp_res.c pfmem_cur->next = NULL; next 1686 drivers/pci/hotplug/ibmphp_res.c pfmem_cur->next = bus_cur->firstPFMemFromMem; next 1719 drivers/pci/hotplug/ibmphp_res.c pfmem->next = bus_cur->firstPFMemFromMem; next 1721 drivers/pci/hotplug/ibmphp_res.c pfmem->next = NULL; next 1778 drivers/pci/hotplug/ibmphp_res.c range = range->next; next 1788 drivers/pci/hotplug/ibmphp_res.c range = range->next; next 1799 drivers/pci/hotplug/ibmphp_res.c range = range->next; next 1812 drivers/pci/hotplug/ibmphp_res.c if (res->next) next 1813 drivers/pci/hotplug/ibmphp_res.c res = res->next; next 1827 drivers/pci/hotplug/ibmphp_res.c if (res->next) next 1828 drivers/pci/hotplug/ibmphp_res.c res = res->next; next 1842 drivers/pci/hotplug/ibmphp_res.c if (res->next) next 1843 drivers/pci/hotplug/ibmphp_res.c res = res->next; next 1858 drivers/pci/hotplug/ibmphp_res.c res = res->next; next 1886 drivers/pci/hotplug/ibmphp_res.c range_cur = range_cur->next; next 197 drivers/pci/hotplug/pnv_php.c struct device_node *child, *next; next 207 drivers/pci/hotplug/pnv_php.c next = child->sibling; next 211 drivers/pci/hotplug/pnv_php.c child = next; next 112 drivers/pci/hotplug/rpadlpar_core.c struct slot *slot, *next; next 114 drivers/pci/hotplug/rpadlpar_core.c list_for_each_entry_safe(slot, next, &rpaphp_slot_head, next 402 drivers/pci/hotplug/rpaphp_core.c struct slot *slot, *next; next 409 drivers/pci/hotplug/rpaphp_core.c list_for_each_entry_safe(slot, next, &rpaphp_slot_head, next 178 drivers/pci/hotplug/s390_pci_hpc.c struct slot *slot, *next; next 180 drivers/pci/hotplug/s390_pci_hpc.c list_for_each_entry_safe(slot, next, &s390_hotplug_slot_list, next 132 drivers/pci/hotplug/shpchp_core.c struct slot *slot, *next; next 134 drivers/pci/hotplug/shpchp_core.c list_for_each_entry_safe(slot, next, &ctrl->slot_list, slot_list) { next 45 drivers/pci/pci-bridge-emul.h u8 next; next 1197 drivers/pci/pci.c hlist_for_each_entry(tmp, &pci_dev->saved_cap_space, next) { next 1511 drivers/pci/pci.c hlist_for_each_entry(tmp, &dev->saved_cap_space, next) next 1522 drivers/pci/pci.c hlist_for_each_entry(tmp, &dev->saved_cap_space, next) { next 3048 drivers/pci/pci.c hlist_add_head(&new_cap->next, &pci_dev->saved_cap_space); next 3125 drivers/pci/pci.c hlist_for_each_entry_safe(tmp, n, &dev->saved_cap_space, next) next 248 drivers/pci/pcie/aspm.c child = list_entry(linkbus->devices.next, struct pci_dev, bus_list); next 397 drivers/pci/proc.c .next = pci_seq_next, next 5128 drivers/pci/quirks.c hlist_add_head(&state->next, &pdev->saved_cap_space); next 167 drivers/pci/search.c n = from ? from->node.next : pci_root_buses.next; next 2074 drivers/pci/setup-bus.c struct pci_dev *next; next 2082 drivers/pci/setup-bus.c next = bridge; next 2084 drivers/pci/setup-bus.c bridge = next; next 2112 drivers/pci/setup-bus.c next = bridge->bus ? bridge->bus->self : NULL; next 2113 drivers/pci/setup-bus.c } while (next); next 150 drivers/pci/switch/switchtec.c stuser = list_entry(stdev->mrpc_queue.next, struct switchtec_user, next 194 drivers/pci/switch/switchtec.c stuser = list_entry(stdev->mrpc_queue.next, struct switchtec_user, next 561 drivers/pci/xen-pcifront.c dev = container_of(bus->devices.next, struct pci_dev, next 50 drivers/pcmcia/rsrc_nonstatic.c struct resource_map *next; next 112 drivers/pcmcia/rsrc_nonstatic.c for (p = map; ; p = p->next) { next 117 drivers/pcmcia/rsrc_nonstatic.c if ((p->next == map) || (p->next->base > base+num-1)) next 126 drivers/pcmcia/rsrc_nonstatic.c q->next = p->next; p->next = q; next 137 drivers/pcmcia/rsrc_nonstatic.c q = p->next; next 144 drivers/pcmcia/rsrc_nonstatic.c p->next = q->next; next 167 drivers/pcmcia/rsrc_nonstatic.c p->next = q->next ; q->next = p; next 460 drivers/pcmcia/rsrc_nonstatic.c ok = inv_probe(m->next, s); next 489 drivers/pcmcia/rsrc_nonstatic.c if (inv_probe(s_data->mem_db.next, s) > 0) next 491 drivers/pcmcia/rsrc_nonstatic.c if (s_data->mem_db_valid.next != &s_data->mem_db_valid) next 498 drivers/pcmcia/rsrc_nonstatic.c for (m = s_data->mem_db.next; m != &s_data->mem_db; m = mm.next) { next 542 drivers/pcmcia/rsrc_nonstatic.c for (m = s_data->mem_db.next; m != &s_data->mem_db; m = mm.next) { next 577 drivers/pcmcia/rsrc_nonstatic.c if (s_data->mem_db_valid.next != &s_data->mem_db_valid) next 612 drivers/pcmcia/rsrc_nonstatic.c for (m = data->map->next; m != data->map; m = m->next) { next 656 drivers/pcmcia/rsrc_nonstatic.c for (m = s_data->io_db.next; m != &s_data->io_db; m = m->next) { next 1022 drivers/pcmcia/rsrc_nonstatic.c data->mem_db.next = &data->mem_db; next 1023 drivers/pcmcia/rsrc_nonstatic.c data->mem_db_valid.next = &data->mem_db_valid; next 1024 drivers/pcmcia/rsrc_nonstatic.c data->io_db.next = &data->io_db; next 1038 drivers/pcmcia/rsrc_nonstatic.c for (p = data->mem_db_valid.next; p != &data->mem_db_valid; p = q) { next 1039 drivers/pcmcia/rsrc_nonstatic.c q = p->next; next 1042 drivers/pcmcia/rsrc_nonstatic.c for (p = data->mem_db.next; p != &data->mem_db; p = q) { next 1043 drivers/pcmcia/rsrc_nonstatic.c q = p->next; next 1046 drivers/pcmcia/rsrc_nonstatic.c for (p = data->io_db.next; p != &data->io_db; p = q) { next 1047 drivers/pcmcia/rsrc_nonstatic.c q = p->next; next 1076 drivers/pcmcia/rsrc_nonstatic.c for (p = data->io_db.next; p != &data->io_db; p = p->next) { next 1132 drivers/pcmcia/rsrc_nonstatic.c for (p = data->mem_db_valid.next; p != &data->mem_db_valid; next 1133 drivers/pcmcia/rsrc_nonstatic.c p = p->next) { next 1142 drivers/pcmcia/rsrc_nonstatic.c for (p = data->mem_db.next; p != &data->mem_db; p = p->next) { next 182 drivers/pcmcia/sa1111_generic.c s->next = dev_get_drvdata(&dev->dev); next 243 drivers/pcmcia/sa1111_generic.c struct sa1111_pcmcia_socket *next, *s = dev_get_drvdata(&dev->dev); next 247 drivers/pcmcia/sa1111_generic.c for (; s; s = next) { next 248 drivers/pcmcia/sa1111_generic.c next = s->next; next 8 drivers/pcmcia/sa1111_generic.h struct sa1111_pcmcia_socket *next; next 178 drivers/perf/qcom_l2_pmu.c struct list_head next; next 797 drivers/perf/qcom_l2_pmu.c list_for_each_entry(cluster, &l2cache_pmu->clusters, next) { next 907 drivers/perf/qcom_l2_pmu.c INIT_LIST_HEAD(&cluster->next); next 908 drivers/perf/qcom_l2_pmu.c list_add(&cluster->next, &l2cache_pmu->clusters); next 140 drivers/perf/xgene_pmu.c struct list_head next; next 1256 drivers/perf/xgene_pmu.c list_for_each_entry(ctx, &xgene_pmu->mcpmus, next) { next 1261 drivers/perf/xgene_pmu.c list_for_each_entry(ctx, &xgene_pmu->mcbpmus, next) { next 1266 drivers/perf/xgene_pmu.c list_for_each_entry(ctx, &xgene_pmu->l3cpmus, next) { next 1271 drivers/perf/xgene_pmu.c list_for_each_entry(ctx, &xgene_pmu->iobpmus, next) { next 1586 drivers/perf/xgene_pmu.c list_add(&ctx->next, &xgene_pmu->l3cpmus); next 1589 drivers/perf/xgene_pmu.c list_add(&ctx->next, &xgene_pmu->iobpmus); next 1592 drivers/perf/xgene_pmu.c list_add(&ctx->next, &xgene_pmu->iobpmus); next 1595 drivers/perf/xgene_pmu.c list_add(&ctx->next, &xgene_pmu->mcbpmus); next 1598 drivers/perf/xgene_pmu.c list_add(&ctx->next, &xgene_pmu->mcpmus); next 1708 drivers/perf/xgene_pmu.c list_add(&ctx->next, &xgene_pmu->l3cpmus); next 1711 drivers/perf/xgene_pmu.c list_add(&ctx->next, &xgene_pmu->iobpmus); next 1714 drivers/perf/xgene_pmu.c list_add(&ctx->next, &xgene_pmu->iobpmus); next 1717 drivers/perf/xgene_pmu.c list_add(&ctx->next, &xgene_pmu->mcbpmus); next 1720 drivers/perf/xgene_pmu.c list_add(&ctx->next, &xgene_pmu->mcpmus); next 1821 drivers/perf/xgene_pmu.c list_for_each_entry(ctx, &xgene_pmu->mcpmus, next) { next 1824 drivers/perf/xgene_pmu.c list_for_each_entry(ctx, &xgene_pmu->mcbpmus, next) { next 1827 drivers/perf/xgene_pmu.c list_for_each_entry(ctx, &xgene_pmu->l3cpmus, next) { next 1830 drivers/perf/xgene_pmu.c list_for_each_entry(ctx, &xgene_pmu->iobpmus, next) { next 1958 drivers/perf/xgene_pmu.c list_for_each_entry(ctx, pmus, next) { next 992 drivers/phy/phy-core.c struct device_node *parent = of_node_get(children), *next; next 998 drivers/phy/phy-core.c next = of_get_parent(parent); next 1000 drivers/phy/phy-core.c parent = next; next 44 drivers/phy/socionext/phy-uniphier-usb2.c struct uniphier_u2phy_priv *next; next 89 drivers/phy/socionext/phy-uniphier-usb2.c priv = priv->next; next 110 drivers/phy/socionext/phy-uniphier-usb2.c struct uniphier_u2phy_priv *priv = NULL, *next = NULL; next 169 drivers/phy/socionext/phy-uniphier-usb2.c priv->next = next; next 170 drivers/phy/socionext/phy-uniphier-usb2.c next = priv; next 123 drivers/pinctrl/mediatek/pinctrl-mtk-common-v2.c pfd->next = pfd->bitpos + c->x_bits > c->sz_reg ? c->x_addrs : 0; next 156 drivers/pinctrl/mediatek/pinctrl-mtk-common-v2.c mtk_rmw(hw, pf->index, pf->offset + pf->next, BIT(nbits_h) - 1, next 169 drivers/pinctrl/mediatek/pinctrl-mtk-common-v2.c h = (mtk_r32(hw, pf->index, pf->offset + pf->next)) next 185 drivers/pinctrl/mediatek/pinctrl-mtk-common-v2.c if (!pf.next) next 204 drivers/pinctrl/mediatek/pinctrl-mtk-common-v2.c if (!pf.next) next 101 drivers/pinctrl/mediatek/pinctrl-mtk-common-v2.h u8 next; next 37 drivers/pinctrl/pinctrl-at91.c struct at91_gpio_chip *next; /* Bank sharing same clock */ next 1702 drivers/pinctrl/pinctrl-at91.c if (!at91_gpio->next) next 1704 drivers/pinctrl/pinctrl-at91.c at91_gpio = at91_gpio->next; next 1779 drivers/pinctrl/pinctrl-at91.c if (prev->next) { next 1780 drivers/pinctrl/pinctrl-at91.c prev = prev->next; next 1782 drivers/pinctrl/pinctrl-at91.c prev->next = at91_gpio; next 341 drivers/platform/mellanox/mlxbf-tmfifo.c idx = virtio16_to_cpu(vdev, desc->next); next 438 drivers/platform/mellanox/mlxbf-tmfifo.c idx = virtio16_to_cpu(vdev, desc->next); next 723 drivers/platform/mellanox/mlxbf-tmfifo.c idx = virtio16_to_cpu(vdev, desc->next); next 605 drivers/platform/x86/dcdbas.c .next = NULL, next 308 drivers/platform/x86/dell_rbu.c ptemp_list = (&packet_data_head.list)->next; next 322 drivers/platform/x86/dell_rbu.c ptemp_list = ptemp_list->next; next 336 drivers/platform/x86/dell_rbu.c ptemp_list = (&packet_data_head.list)->next; next 340 drivers/platform/x86/dell_rbu.c pnext_list = ptemp_list->next; next 469 drivers/platform/x86/intel_menlow.c struct intel_menlow_attribute *pos, *next; next 472 drivers/platform/x86/intel_menlow.c list_for_each_entry_safe(pos, next, &intel_menlow_attr_list, node) { next 813 drivers/platform/x86/wmi.c struct wmi_block *next = NULL; next 815 drivers/platform/x86/wmi.c list_for_each_entry_safe(wblock, next, &wmi_block_list, list) { next 1107 drivers/platform/x86/wmi.c struct wmi_block *wblock, *next; next 1110 drivers/platform/x86/wmi.c list_for_each_entry_safe(wblock, next, &wmi_block_list, list) { next 1147 drivers/platform/x86/wmi.c struct wmi_block *wblock, *next; next 1209 drivers/platform/x86/wmi.c list_for_each_entry_safe(wblock, next, &wmi_block_list, list) { next 124 drivers/pnp/card.c dev_id->next = NULL; next 126 drivers/pnp/card.c while (ptr && ptr->next) next 127 drivers/pnp/card.c ptr = ptr->next; next 129 drivers/pnp/card.c ptr->next = dev_id; next 139 drivers/pnp/card.c struct pnp_id *next; next 143 drivers/pnp/card.c next = id->next; next 145 drivers/pnp/card.c id = next; next 205 drivers/pnp/card.c pos = pos->next; next 342 drivers/pnp/card.c pos = card->devices.next; next 346 drivers/pnp/card.c pos = from->card_list.next; next 352 drivers/pnp/card.c pos = pos->next; next 106 drivers/pnp/core.c struct pnp_id *next; next 110 drivers/pnp/core.c next = id->next; next 112 drivers/pnp/core.c id = next; next 227 drivers/pnp/core.c for (id = dev->id; id; id = id->next) next 39 drivers/pnp/driver.c pos = pos->next; next 302 drivers/pnp/driver.c dev_id->next = NULL; next 304 drivers/pnp/driver.c while (ptr && ptr->next) next 305 drivers/pnp/driver.c ptr = ptr->next; next 307 drivers/pnp/driver.c ptr->next = dev_id; next 449 drivers/pnp/interface.c pos = pos->next; next 34 drivers/pnp/isapnp/compat.c list = from ? from->global_list.next : pnp_cards.next; next 41 drivers/pnp/isapnp/compat.c list = list->next; next 57 drivers/pnp/isapnp/compat.c list = pnp_global.next; next 59 drivers/pnp/isapnp/compat.c list = from->global_list.next; next 67 drivers/pnp/isapnp/compat.c list = list->next; next 72 drivers/pnp/isapnp/compat.c list = card->devices.next; next 74 drivers/pnp/isapnp/compat.c list = from->card_list.next; next 83 drivers/pnp/isapnp/compat.c list = list->next; next 199 drivers/pnp/quirks.c new_option = list_entry(new_option->list.next, next 1204 drivers/pwm/core.c .next = pwm_seq_next, next 958 drivers/rapidio/rio-scan.c if (mport->nnode.next || mport->nnode.prev) next 846 drivers/rapidio/rio.c n = from ? from->global_list.next : rio_devices.next; next 852 drivers/rapidio/rio.c n = n->next; next 1440 drivers/rapidio/rio.c n = from ? from->global_list.next : rio_devices.next; next 1449 drivers/rapidio/rio.c n = n->next; next 364 drivers/remoteproc/qcom_q6v5_mss.c struct qcom_scm_vmperm next; next 373 drivers/remoteproc/qcom_q6v5_mss.c next.vmid = remote_owner ? QCOM_SCM_VMID_MSS_MSA : QCOM_SCM_VMID_HLOS; next 374 drivers/remoteproc/qcom_q6v5_mss.c next.perm = remote_owner ? QCOM_SCM_PERM_RW : QCOM_SCM_PERM_RWX; next 377 drivers/remoteproc/qcom_q6v5_mss.c current_perm, &next, 1); next 793 drivers/rtc/interface.c struct timerqueue_node *next = timerqueue_getnext(&rtc->timerqueue); next 802 drivers/rtc/interface.c while (next) { next 803 drivers/rtc/interface.c if (next->expires >= now) next 805 drivers/rtc/interface.c next = timerqueue_iterate_next(next); next 810 drivers/rtc/interface.c if (!next || ktime_before(timer->node.expires, next->expires)) { next 853 drivers/rtc/interface.c struct timerqueue_node *next = timerqueue_getnext(&rtc->timerqueue); next 858 drivers/rtc/interface.c if (next == &timer->node) { next 862 drivers/rtc/interface.c next = timerqueue_getnext(&rtc->timerqueue); next 863 drivers/rtc/interface.c if (!next) { next 867 drivers/rtc/interface.c alarm.time = rtc_ktime_to_tm(next->expires); next 890 drivers/rtc/interface.c struct timerqueue_node *next; next 901 drivers/rtc/interface.c while ((next = timerqueue_getnext(&rtc->timerqueue))) { next 902 drivers/rtc/interface.c if (next->expires > now) next 906 drivers/rtc/interface.c timer = container_of(next, struct rtc_timer, node); next 925 drivers/rtc/interface.c if (next) { next 930 drivers/rtc/interface.c alarm.time = rtc_ktime_to_tm(next->expires); next 940 drivers/rtc/interface.c timer = container_of(next, struct rtc_timer, node); next 76 drivers/rtc/rtc-88pm80x.c static void rtc_next_alarm_time(struct rtc_time *next, struct rtc_time *now, next 82 drivers/rtc/rtc-88pm80x.c next->tm_year = now->tm_year; next 83 drivers/rtc/rtc-88pm80x.c next->tm_mon = now->tm_mon; next 84 drivers/rtc/rtc-88pm80x.c next->tm_mday = now->tm_mday; next 85 drivers/rtc/rtc-88pm80x.c next->tm_hour = alrm->tm_hour; next 86 drivers/rtc/rtc-88pm80x.c next->tm_min = alrm->tm_min; next 87 drivers/rtc/rtc-88pm80x.c next->tm_sec = alrm->tm_sec; next 90 drivers/rtc/rtc-88pm80x.c next_time = rtc_tm_to_time64(next); next 95 drivers/rtc/rtc-88pm80x.c rtc_time64_to_tm(next_time, next); next 83 drivers/rtc/rtc-88pm860x.c static void rtc_next_alarm_time(struct rtc_time *next, struct rtc_time *now, next 89 drivers/rtc/rtc-88pm860x.c next->tm_year = now->tm_year; next 90 drivers/rtc/rtc-88pm860x.c next->tm_mon = now->tm_mon; next 91 drivers/rtc/rtc-88pm860x.c next->tm_mday = now->tm_mday; next 92 drivers/rtc/rtc-88pm860x.c next->tm_hour = alrm->tm_hour; next 93 drivers/rtc/rtc-88pm860x.c next->tm_min = alrm->tm_min; next 94 drivers/rtc/rtc-88pm860x.c next->tm_sec = alrm->tm_sec; next 97 drivers/rtc/rtc-88pm860x.c rtc_tm_to_time(next, &next_time); next 102 drivers/rtc/rtc-88pm860x.c rtc_time_to_tm(next_time, next); next 1678 drivers/s390/block/dasd.c struct dasd_ccw_req *cqr, *next, *fcqr; next 1830 drivers/s390/block/dasd.c next = NULL; next 1838 drivers/s390/block/dasd.c if (cqr->devlist.next != &device->ccw_queue) { next 1839 drivers/s390/block/dasd.c next = list_entry(cqr->devlist.next, next 1864 drivers/s390/block/dasd.c next = cqr; next 1868 drivers/s390/block/dasd.c if (next && (next->status == DASD_CQR_QUEUED) && next 1870 drivers/s390/block/dasd.c if (device->discipline->start_IO(next) == 0) next 1871 drivers/s390/block/dasd.c expires = next->expires; next 2018 drivers/s390/block/dasd.c cqr = list_entry(device->ccw_queue.next, struct dasd_ccw_req, devlist); next 2087 drivers/s390/block/dasd.c cqr = list_entry(device->ccw_queue.next, struct dasd_ccw_req, devlist); next 2584 drivers/s390/block/dasd.c cqr = list_entry(device->ccw_queue.next, struct dasd_ccw_req, devlist); next 2621 drivers/s390/block/dasd.c list_add(&cqr->devlist, device->ccw_queue.next); next 374 drivers/s390/block/dasd_alias.c if (group->next == device) next 375 drivers/s390/block/dasd_alias.c group->next = NULL; next 682 drivers/s390/block/dasd_alias.c alias_device = group->next; next 694 drivers/s390/block/dasd_alias.c group->next = list_first_entry(&group->aliaslist, next 697 drivers/s390/block/dasd_alias.c group->next = list_first_entry(&alias_device->alias_list, next 236 drivers/s390/block/dasd_diag.c struct dasd_ccw_req *cqr, *next; next 288 drivers/s390/block/dasd_diag.c next = list_entry(device->ccw_queue.next, next 290 drivers/s390/block/dasd_diag.c if (next->status == DASD_CQR_QUEUED) { next 291 drivers/s390/block/dasd_diag.c rc = dasd_start_diag(next); next 293 drivers/s390/block/dasd_diag.c expires = next->expires; next 650 drivers/s390/block/dasd_eckd.h struct dasd_device *next; next 697 drivers/s390/block/dasd_int.h if (left->next != chunk_list) { next 698 drivers/s390/block/dasd_int.h tmp = list_entry(left->next, struct dasd_mchunk, list); next 712 drivers/s390/block/dasd_int.h __list_add(&chunk->list, left, left->next); next 129 drivers/s390/block/dasd_proc.c .next = dasd_devices_next, next 73 drivers/s390/char/con3215.c struct raw3215_req *next; /* pointer to next request */ next 116 drivers/s390/char/con3215.c raw3215_freelist = req->next; next 132 drivers/s390/char/con3215.c req->next = raw3215_freelist; next 925 drivers/s390/char/con3215.c req->next = raw3215_freelist; next 91 drivers/s390/char/monwriter.c struct mon_buf *entry, *next; next 93 drivers/s390/char/monwriter.c list_for_each_entry_safe(entry, next, &monpriv->list, list) next 209 drivers/s390/char/monwriter.c struct mon_buf *entry, *next; next 211 drivers/s390/char/monwriter.c list_for_each_entry_safe(entry, next, &monpriv->list, list) { next 370 drivers/s390/char/raw3270.c rq = list_entry(rp->req_queue.next,struct raw3270_request,list); next 652 drivers/s390/char/raw3270.c rq = list_entry(rp->req_queue.next,struct raw3270_request,list); next 1183 drivers/s390/char/raw3270.c v = list_entry(rp->view_list.next, struct raw3270_view, list); next 257 drivers/s390/char/raw3270.h if (left->next != free_list) { next 258 drivers/s390/char/raw3270.h tmp = list_entry(left->next, struct string, list); next 272 drivers/s390/char/raw3270.h __list_add(&cs->list, left, left->next); next 315 drivers/s390/char/sclp.c req = list_entry(sclp_req_queue.next, struct sclp_req, list); next 145 drivers/s390/char/sclp_con.c list = sclp_con_outqueue.next; next 148 drivers/s390/char/sclp_con.c list = list->next; next 190 drivers/s390/char/sclp_con.c page = sclp_con_pages.next; next 123 drivers/s390/char/sclp_tty.c buffer = list_entry(sclp_tty_outqueue.next, next 195 drivers/s390/char/sclp_tty.c page = sclp_tty_pages.next; next 136 drivers/s390/char/sclp_vt220.c request = list_entry(sclp_vt220_outqueue.next, next 379 drivers/s390/char/sclp_vt220.c list = sclp_vt220_outqueue.next; next 382 drivers/s390/char/sclp_vt220.c list = list->next; next 432 drivers/s390/char/sclp_vt220.c page = (void *) sclp_vt220_empty.next; next 146 drivers/s390/char/tape_core.c req = list_entry(tdev->req_queue.next, struct tape_request, next 877 drivers/s390/char/tape_core.c request = list_entry(device->req_queue.next, struct tape_request, list); next 1156 drivers/s390/char/tape_core.c req = list_entry(device->req_queue.next, next 72 drivers/s390/char/tape_proc.c request = list_entry(device->req_queue.next, next 103 drivers/s390/char/tape_proc.c .next = tape_proc_next, next 465 drivers/s390/char/tty3270.c s = list_entry(tp->rcl_lines.next, struct string, list); next 391 drivers/s390/cio/blacklist.c .next = cio_ignore_proc_seq_next, next 31 drivers/s390/cio/vfio_ccw_cp.c struct list_head next; next 335 drivers/s390/cio/vfio_ccw_cp.c list_add_tail(&chain->next, &cp->ccwchain_list); next 342 drivers/s390/cio/vfio_ccw_cp.c list_del(&chain->next); next 411 drivers/s390/cio/vfio_ccw_cp.c list_for_each_entry(chain, &cp->ccwchain_list, next) { next 494 drivers/s390/cio/vfio_ccw_cp.c list_for_each_entry(iter, &cp->ccwchain_list, next) { next 683 drivers/s390/cio/vfio_ccw_cp.c list_for_each_entry_safe(chain, temp, &cp->ccwchain_list, next) { next 738 drivers/s390/cio/vfio_ccw_cp.c list_for_each_entry(chain, &cp->ccwchain_list, next) { next 751 drivers/s390/cio/vfio_ccw_cp.c list_for_each_entry_continue(chain, &cp->ccwchain_list, next) { next 786 drivers/s390/cio/vfio_ccw_cp.c chain = list_first_entry(&cp->ccwchain_list, struct ccwchain, next); next 822 drivers/s390/cio/vfio_ccw_cp.c list_for_each_entry(chain, &cp->ccwchain_list, next) { next 859 drivers/s390/cio/vfio_ccw_cp.c list_for_each_entry(chain, &cp->ccwchain_list, next) { next 242 drivers/s390/crypto/ap_queue.c ap_msg = list_entry(aq->requestq.next, struct ap_message, list); next 723 drivers/s390/crypto/ap_queue.c struct ap_message *ap_msg, *next; next 725 drivers/s390/crypto/ap_queue.c list_for_each_entry_safe(ap_msg, next, &aq->pendingq, list) { next 731 drivers/s390/crypto/ap_queue.c list_for_each_entry_safe(ap_msg, next, &aq->requestq, list) { next 214 drivers/s390/net/ctcm_main.c *c = ch->next; next 236 drivers/s390/net/ctcm_main.c c = &((*c)->next); next 258 drivers/s390/net/ctcm_main.c ch = ch->next; next 1464 drivers/s390/net/ctcm_main.c c = &(*c)->next; next 1484 drivers/s390/net/ctcm_main.c ch->next = *c; next 132 drivers/s390/net/ctcm_main.h struct channel *next; next 441 drivers/s390/net/lcs.c ipm = list_entry(card->ipm_list.next, next 644 drivers/s390/net/lcs.c int prev, next; next 648 drivers/s390/net/lcs.c next = (index + 1) & (LCS_NUM_BUFFS - 1); next 650 drivers/s390/net/lcs.c if (channel->ccws[next].flags & CCW_FLAG_SUSPEND) { next 689 drivers/s390/net/lcs.c int index, prev, next; next 696 drivers/s390/net/lcs.c next = (index + 1) & (LCS_NUM_BUFFS - 1); next 711 drivers/s390/net/lcs.c channel->ccws[next].flags &= ~CCW_FLAG_PCI; next 223 drivers/s390/net/netiucv.c u16 next; next 630 drivers/s390/net/netiucv.c if (!header->next) next 634 drivers/s390/net/netiucv.c header->next -= offset; next 635 drivers/s390/net/netiucv.c offset += header->next; next 636 drivers/s390/net/netiucv.c header->next -= NETIUCV_HDRLEN; next 637 drivers/s390/net/netiucv.c if (skb_tailroom(pskb) < header->next) { next 639 drivers/s390/net/netiucv.c header->next, skb_tailroom(pskb)); next 642 drivers/s390/net/netiucv.c skb_put(pskb, header->next); next 664 drivers/s390/net/netiucv.c skb_pull(pskb, header->next); next 747 drivers/s390/net/netiucv.c header.next = conn->tx_buff->len + skb->len + NETIUCV_HDRLEN; next 767 drivers/s390/net/netiucv.c header.next = 0; next 1196 drivers/s390/net/netiucv.c header.next = nskb->len + NETIUCV_HDRLEN; next 1198 drivers/s390/net/netiucv.c header.next = 0; next 2149 drivers/s390/net/netiucv.c cp = list_entry(iucv_connection_list.next, next 2608 drivers/s390/net/qeth_core_main.c entry = list_entry(card->qdio.in_buf_pool.entry_list.next, next 1211 drivers/s390/net/qeth_l3_main.c for (im6 = in6_dev->mc_list; im6 != NULL; im6 = im6->next) { next 1537 drivers/s390/scsi/zfcp_erp.c struct list_head *next; next 1550 drivers/s390/scsi/zfcp_erp.c next = adapter->erp_ready_head.next; next 1553 drivers/s390/scsi/zfcp_erp.c if (next != &adapter->erp_ready_head) { next 1554 drivers/s390/scsi/zfcp_erp.c act = list_entry(next, struct zfcp_erp_action, list); next 65 drivers/s390/scsi/zfcp_fc.c unsigned long next = adapter->next_port_scan; next 69 drivers/s390/scsi/zfcp_fc.c if (time_before(now, next)) { next 70 drivers/s390/scsi/zfcp_fc.c delay = next - now; next 55 drivers/sbus/char/bbc_i2c.h struct bbc_i2c_bus *next; next 163 drivers/sbus/char/openprom.c !prop->next || next 164 drivers/sbus/char/openprom.c (len = strlen(prop->next->name)) + 1 > bufsize) next 167 drivers/sbus/char/openprom.c prop = prop->next; next 441 drivers/sbus/char/openprom.c prop = prop->next; next 175 drivers/scsi/BusLogic.c ccb->next = adapter->free_ccbs; next 284 drivers/scsi/BusLogic.c adapter->free_ccbs = ccb->next; next 285 drivers/scsi/BusLogic.c ccb->next = NULL; next 296 drivers/scsi/BusLogic.c adapter->free_ccbs = ccb->next; next 297 drivers/scsi/BusLogic.c ccb->next = NULL; next 320 drivers/scsi/BusLogic.c ccb->next = adapter->free_ccbs; next 2600 drivers/scsi/BusLogic.c ccb->next = NULL; next 2605 drivers/scsi/BusLogic.c adapter->lastccb->next = ccb; next 2746 drivers/scsi/BusLogic.c adapter->firstccb = ccb->next; next 3881 drivers/scsi/BusLogic.c struct blogic_adapter *ha, *next; next 3883 drivers/scsi/BusLogic.c list_for_each_entry_safe(ha, next, &blogic_host_list, host_list) next 855 drivers/scsi/BusLogic.h struct blogic_ccb *next; next 812 drivers/scsi/aacraid/aachba.c scsicmd->list.next = NULL; next 4000 drivers/scsi/aacraid/aachba.c psg->sg[0].next = 0; next 4014 drivers/scsi/aacraid/aachba.c psg->sg[i].next = 0; next 476 drivers/scsi/aacraid/aacraid.h __le32 next; /* reserved for F/W use */ next 484 drivers/scsi/aacraid/aacraid.h u32 next; /* reserved for F/W use */ next 1230 drivers/scsi/aacraid/aacraid.h struct list_head next; // used to link context's into a linked list next 1292 drivers/scsi/aacraid/aacraid.h void *next; /* this is used by the allocator */ next 206 drivers/scsi/aacraid/commctrl.c entry = dev->fib_list.next; next 208 drivers/scsi/aacraid/commctrl.c context = list_entry(entry, struct aac_fib_context, next); next 212 drivers/scsi/aacraid/commctrl.c entry = dev->fib_list.next; next 214 drivers/scsi/aacraid/commctrl.c entry = entry->next; next 217 drivers/scsi/aacraid/commctrl.c list_add_tail(&fibctx->next, &dev->fib_list); next 256 drivers/scsi/aacraid/commctrl.c entry = dev->fib_list.next; next 260 drivers/scsi/aacraid/commctrl.c fibctx = list_entry(entry, struct aac_fib_context, next); next 267 drivers/scsi/aacraid/commctrl.c entry = entry->next; next 292 drivers/scsi/aacraid/commctrl.c entry = fibctx->fib_list.next; next 350 drivers/scsi/aacraid/commctrl.c entry = fibctx->fib_list.next; next 363 drivers/scsi/aacraid/commctrl.c list_del(&fibctx->next); next 397 drivers/scsi/aacraid/commctrl.c entry = dev->fib_list.next; next 401 drivers/scsi/aacraid/commctrl.c fibctx = list_entry(entry, struct aac_fib_context, next); next 407 drivers/scsi/aacraid/commctrl.c entry = entry->next; next 176 drivers/scsi/aacraid/commsup.c fibptr->next = fibptr+1; /* Forward chain the fibs */ next 206 drivers/scsi/aacraid/commsup.c dev->fibs[dev->scsi_host_ptr->can_queue + AAC_NUM_MGT_FIB - 1].next = NULL; next 257 drivers/scsi/aacraid/commsup.c dev->free_fib = fibptr->next; next 299 drivers/scsi/aacraid/commsup.c fibptr->next = fibptr->dev->free_fib; next 1748 drivers/scsi/aacraid/commsup.c entry = aac->fib_list.next; next 1760 drivers/scsi/aacraid/commsup.c struct aac_fib_context *fibctx = list_entry(entry, struct aac_fib_context, next); next 1780 drivers/scsi/aacraid/commsup.c entry = entry->next; next 1824 drivers/scsi/aacraid/commsup.c entry = entry->next; next 2029 drivers/scsi/aacraid/commsup.c entry = dev->fib_list.next; next 2031 drivers/scsi/aacraid/commsup.c entry = entry->next; next 2087 drivers/scsi/aacraid/commsup.c entry = dev->fib_list.next; next 2102 drivers/scsi/aacraid/commsup.c next); next 2120 drivers/scsi/aacraid/commsup.c entry = entry->next; next 2131 drivers/scsi/aacraid/commsup.c entry = entry->next; next 2157 drivers/scsi/aacraid/commsup.c entry = entry->next; next 2187 drivers/scsi/aacraid/commsup.c entry = dev->queues->queue[HostNormCmdQueue].cmdq.next; next 101 drivers/scsi/aacraid/src.c entry = dev->sync_fib_list.next; next 502 drivers/scsi/aha152x.c struct scsi_cmnd *next; /* next sc in queue */ next 554 drivers/scsi/aha152x.c #define SCNEXT(SCpnt) SCDATA(SCpnt)->next next 1110 drivers/scsi/aha152x.c struct scsi_cmnd *next; next 1113 drivers/scsi/aha152x.c next = SCNEXT(ptr); next 1117 drivers/scsi/aha152x.c next = NULL; next 1127 drivers/scsi/aha152x.c ptr = next; next 2250 drivers/scsi/aha152x.c struct scsi_cmnd *next = SCNEXT(ptr); next 2262 drivers/scsi/aha152x.c ptr = next; next 200 drivers/scsi/aic7xxx/aic79xx_core.c u_int prev, u_int next, u_int tid); next 2909 drivers/scsi/aic7xxx/aic79xx_core.c u_int next; next 2961 drivers/scsi/aic7xxx/aic79xx_core.c next = SCB_LIST_NULL; next 2964 drivers/scsi/aic7xxx/aic79xx_core.c next = ahd_inw_scbram(ahd, SCB_NEXT2); next 2967 drivers/scsi/aic7xxx/aic79xx_core.c ahd_outw(ahd, SCB_NEXT2, next); next 8382 drivers/scsi/aic7xxx/aic79xx_core.c u_int next; next 8389 drivers/scsi/aic7xxx/aic79xx_core.c next = *list_head; next 8391 drivers/scsi/aic7xxx/aic79xx_core.c for (scbid = next; !SCBID_IS_NULL(scbid); scbid = next) { next 8407 drivers/scsi/aic7xxx/aic79xx_core.c next = ahd_inw_scbram(ahd, SCB_NEXT); next 8421 drivers/scsi/aic7xxx/aic79xx_core.c ahd_rem_wscb(ahd, scbid, prev, next, tid); next 8424 drivers/scsi/aic7xxx/aic79xx_core.c *list_head = next; next 8482 drivers/scsi/aic7xxx/aic79xx_core.c u_int prev, u_int next, u_int tid) next 8489 drivers/scsi/aic7xxx/aic79xx_core.c ahd_outw(ahd, SCB_NEXT, next); next 8500 drivers/scsi/aic7xxx/aic79xx_core.c if (SCBID_IS_NULL(next) next 8505 drivers/scsi/aic7xxx/aic79xx_core.c return (next); next 472 drivers/scsi/aic7xxx/aic7xxx.h /*31*/ uint8_t next; /* next 593 drivers/scsi/aic7xxx/aic7xxx_core.c q_hscb->next = scb->hscb->tag; next 612 drivers/scsi/aic7xxx/aic7xxx_core.c || scb->hscb->next == SCB_LIST_NULL) next 614 drivers/scsi/aic7xxx/aic7xxx_core.c scb->hscb->tag, scb->hscb->next); next 5804 drivers/scsi/aic7xxx/aic7xxx_core.c prev_scb->hscb->next = scb->hscb->tag; next 5809 drivers/scsi/aic7xxx/aic7xxx_core.c scb->hscb->next = ahc->next_queued_scb->hscb->tag; next 5838 drivers/scsi/aic7xxx/aic7xxx_core.c uint8_t next; next 5949 drivers/scsi/aic7xxx/aic7xxx_core.c next = scb->hscb->next; next 5952 drivers/scsi/aic7xxx/aic7xxx_core.c scb->hscb->next = next; next 5961 drivers/scsi/aic7xxx/aic7xxx_core.c scb->hscb->next = ahc->next_queued_scb->hscb->tag; next 5968 drivers/scsi/aic7xxx/aic7xxx_core.c next = ahc_inb(ahc, WAITING_SCBH); /* Start at head of list. */ next 5971 drivers/scsi/aic7xxx/aic7xxx_core.c while (next != SCB_LIST_NULL) { next 5974 drivers/scsi/aic7xxx/aic7xxx_core.c ahc_outb(ahc, SCBPTR, next); next 5986 drivers/scsi/aic7xxx/aic7xxx_core.c scb_index, next); next 6014 drivers/scsi/aic7xxx/aic7xxx_core.c next = ahc_rem_wscb(ahc, next, prev); next 6017 drivers/scsi/aic7xxx/aic7xxx_core.c prev = next; next 6018 drivers/scsi/aic7xxx/aic7xxx_core.c next = ahc_inb(ahc, SCB_NEXT); next 6023 drivers/scsi/aic7xxx/aic7xxx_core.c prev = next; next 6024 drivers/scsi/aic7xxx/aic7xxx_core.c next = ahc_inb(ahc, SCB_NEXT); next 6141 drivers/scsi/aic7xxx/aic7xxx_core.c u_int next; next 6147 drivers/scsi/aic7xxx/aic7xxx_core.c next = ahc_inb(ahc, DISCONNECTED_SCBH); next 6157 drivers/scsi/aic7xxx/aic7xxx_core.c while (next != SCB_LIST_NULL) { next 6160 drivers/scsi/aic7xxx/aic7xxx_core.c ahc_outb(ahc, SCBPTR, next); next 6170 drivers/scsi/aic7xxx/aic7xxx_core.c if (next == prev) { next 6173 drivers/scsi/aic7xxx/aic7xxx_core.c next, prev); next 6180 drivers/scsi/aic7xxx/aic7xxx_core.c next = next 6181 drivers/scsi/aic7xxx/aic7xxx_core.c ahc_rem_scb_from_disc_list(ahc, prev, next); next 6183 drivers/scsi/aic7xxx/aic7xxx_core.c prev = next; next 6184 drivers/scsi/aic7xxx/aic7xxx_core.c next = ahc_inb(ahc, SCB_NEXT); next 6189 drivers/scsi/aic7xxx/aic7xxx_core.c prev = next; next 6190 drivers/scsi/aic7xxx/aic7xxx_core.c next = ahc_inb(ahc, SCB_NEXT); next 6205 drivers/scsi/aic7xxx/aic7xxx_core.c u_int next; next 6208 drivers/scsi/aic7xxx/aic7xxx_core.c next = ahc_inb(ahc, SCB_NEXT); next 6216 drivers/scsi/aic7xxx/aic7xxx_core.c ahc_outb(ahc, SCB_NEXT, next); next 6218 drivers/scsi/aic7xxx/aic7xxx_core.c ahc_outb(ahc, DISCONNECTED_SCBH, next); next 6220 drivers/scsi/aic7xxx/aic7xxx_core.c return (next); next 6250 drivers/scsi/aic7xxx/aic7xxx_core.c u_int curscb, next; next 6258 drivers/scsi/aic7xxx/aic7xxx_core.c next = ahc_inb(ahc, SCB_NEXT); next 6268 drivers/scsi/aic7xxx/aic7xxx_core.c ahc_outb(ahc, WAITING_SCBH, next); next 6281 drivers/scsi/aic7xxx/aic7xxx_core.c ahc_outb(ahc, SCB_NEXT, next); next 6288 drivers/scsi/aic7xxx/aic7xxx_core.c return next; next 324 drivers/scsi/aic94xx/aic94xx_hwi.c for (i = 0; i < seq->num_escbs; i++, escb = list_entry(escb->list.next, next 1364 drivers/scsi/aic94xx/aic94xx_hwi.c ascb = list_entry(ascb->list.next, struct asd_ascb, list); next 427 drivers/scsi/aic94xx/aic94xx_sds.c u16 next; next 511 drivers/scsi/aic94xx/aic94xx_sds.c __le16 next; next 535 drivers/scsi/aic94xx/aic94xx_sds.c u16 next; next 544 drivers/scsi/aic94xx/aic94xx_sds.c __le16 next; next 726 drivers/scsi/aic94xx/aic94xx_sds.c el = start + le16_to_cpu(el->next); next 3028 drivers/scsi/arcmsr/arcmsr_hba.c ccb = list_entry(head->next, struct CommandControlBlock, list); next 27 drivers/scsi/arm/msgqueue.c msgq->free = mq->next; next 41 drivers/scsi/arm/msgqueue.c mq->next = msgq->free; next 59 drivers/scsi/arm/msgqueue.c msgq->entries[i].next = &msgq->entries[i + 1]; next 61 drivers/scsi/arm/msgqueue.c msgq->entries[NR_MESSAGES - 1].next = NULL; next 85 drivers/scsi/arm/msgqueue.c for (mq = msgq->qe; mq; mq = mq->next) next 102 drivers/scsi/arm/msgqueue.c for (mq = msgq->qe; mq && msgno; mq = mq->next, msgno--); next 131 drivers/scsi/arm/msgqueue.c mq->next = NULL; next 135 drivers/scsi/arm/msgqueue.c mqp = &(*mqp)->next; next 153 drivers/scsi/arm/msgqueue.c mqnext = mq->next; next 20 drivers/scsi/arm/msgqueue.h struct msgqueue_entry *next; next 114 drivers/scsi/arm/queue.c l = queue->free.next; next 191 drivers/scsi/arm/queue.c SCpnt = __queue_remove(queue, queue->head.next); next 116 drivers/scsi/bfa/bfa_cs.h #define bfa_q_first(_q) ((void *)(((struct list_head *) (_q))->next)) next 117 drivers/scsi/bfa/bfa_cs.h #define bfa_q_next(_qe) (((struct list_head *) (_qe))->next) next 983 drivers/scsi/bnx2fc/57xx_hsi_bnx2fc.h struct regpair next; next 152 drivers/scsi/bnx2fc/bnx2fc_fcoe.c struct sk_buff *skb, *next; next 157 drivers/scsi/bnx2fc/bnx2fc_fcoe.c skb_queue_walk_safe(list, skb, next) { next 2764 drivers/scsi/bnx2fc/bnx2fc_fcoe.c struct bnx2fc_hba *hba, *next; next 2781 drivers/scsi/bnx2fc/bnx2fc_fcoe.c list_for_each_entry_safe(hba, next, &to_be_deleted, list) { next 2111 drivers/scsi/bnx2fc/bnx2fc_hwi.c hba->t2_hash_tbl[i].next.lo = addr & 0xffffffff; next 2112 drivers/scsi/bnx2fc/bnx2fc_hwi.c hba->t2_hash_tbl[i].next.hi = addr >> 32; next 438 drivers/scsi/bnx2fc/bnx2fc_io.c cmd_mgr->free_list[index].next; next 493 drivers/scsi/bnx2fc/bnx2fc_io.c cmd_mgr->free_list[index].next; next 618 drivers/scsi/bnx2fc/bnx2fc_tgt.c u32 conn_id, next; next 629 drivers/scsi/bnx2fc/bnx2fc_tgt.c next = hba->next_conn_id; next 639 drivers/scsi/bnx2fc/bnx2fc_tgt.c if (conn_id == next) { next 531 drivers/scsi/bnx2i/bnx2i_init.c hba = list_entry(adapter_list.next, struct bnx2i_hba, link); next 69 drivers/scsi/csiostor/csio_defs.h return ((list->next == list) && (list->prev == list)); next 72 drivers/scsi/csiostor/csio_defs.h #define csio_list_next(elem) (((struct list_head *)(elem))->next) next 938 drivers/scsi/csiostor/csio_lnode.c struct list_head *tmp, *next; next 941 drivers/scsi/csiostor/csio_lnode.c list_for_each_safe(tmp, next, &rnhead->sm.sm_list) { next 122 drivers/scsi/csiostor/csio_scsi.c struct list_head *tmp, *next; next 133 drivers/scsi/csiostor/csio_scsi.c list_for_each_safe(tmp, next, &scm->active_q) { next 1163 drivers/scsi/csiostor/csio_scsi.c struct list_head *tmp, *next; next 1167 drivers/scsi/csiostor/csio_scsi.c list_for_each_safe(tmp, next, q) { next 1236 drivers/scsi/csiostor/csio_scsi.c struct list_head *tmp, *next; next 1246 drivers/scsi/csiostor/csio_scsi.c list_for_each_safe(tmp, next, q) { next 470 drivers/scsi/cxgbi/libcxgbi.c start = idx = pmap->next; next 477 drivers/scsi/cxgbi/libcxgbi.c pmap->next = idx; next 485 drivers/scsi/cxgbi/libcxgbi.c pmap->sport_base + idx, pmap->next); next 494 drivers/scsi/cxgbi/libcxgbi.c pmap->next); next 1142 drivers/scsi/cxgbi/libcxgbi.c struct sk_buff *next; next 1191 drivers/scsi/cxgbi/libcxgbi.c next = skb->next; next 1192 drivers/scsi/cxgbi/libcxgbi.c skb->next = NULL; next 1198 drivers/scsi/cxgbi/libcxgbi.c skb = next; next 451 drivers/scsi/cxgbi/libcxgbi.h unsigned int next; next 711 drivers/scsi/dc395x.c struct DeviceCtlBlk* next = NULL; next 720 drivers/scsi/dc395x.c next = i; next 726 drivers/scsi/dc395x.c if (!next) next 728 drivers/scsi/dc395x.c next = i; next 732 drivers/scsi/dc395x.c return next; next 800 drivers/scsi/dc395x.c start = list_entry(dcb_list_head->next, typeof(*start), list); next 822 drivers/scsi/dc395x.c srb = list_entry(waiting_list_head->next, next 68 drivers/scsi/dpt/dpti_i2o.h struct i2o_device *next; /* Chain */ next 93 drivers/scsi/dpt/dpti_i2o.h struct i2o_controller *next; /* Controller chain */ next 137 drivers/scsi/dpt_i2o.c struct adpt_i2o_post_wait_data *next; next 194 drivers/scsi/dpt_i2o.c adpt_hba *next; next 212 drivers/scsi/dpt_i2o.c for (pHba = hba_chain; pHba; pHba = next) { next 213 drivers/scsi/dpt_i2o.c next = pHba->next; next 239 drivers/scsi/dpt_i2o.c for (pHba = hba_chain; pHba; pHba = pHba->next) { next 250 drivers/scsi/dpt_i2o.c for (pHba = hba_chain; pHba; pHba = next) { next 251 drivers/scsi/dpt_i2o.c next = pHba->next; next 271 drivers/scsi/dpt_i2o.c for (pHba = hba_chain; pHba; pHba = next) { next 272 drivers/scsi/dpt_i2o.c next = pHba->next; next 551 drivers/scsi/dpt_i2o.c for (pHba = hba_chain; pHba; pHba = pHba->next) { next 842 drivers/scsi/dpt_i2o.c pNext = pHba->next; next 853 drivers/scsi/dpt_i2o.c p1 = p1->next; next 973 drivers/scsi/dpt_i2o.c for(p = hba_chain; p->next; p = p->next); next 974 drivers/scsi/dpt_i2o.c p->next = pHba; next 978 drivers/scsi/dpt_i2o.c pHba->next = NULL; next 1038 drivers/scsi/dpt_i2o.c struct i2o_device* next; next 1050 drivers/scsi/dpt_i2o.c for( p1 = hba_chain; p1; p2 = p1,p1=p1->next){ next 1053 drivers/scsi/dpt_i2o.c p2->next = p1->next; next 1055 drivers/scsi/dpt_i2o.c hba_chain = p1->next; next 1090 drivers/scsi/dpt_i2o.c for(d = pHba->devices; d ; d = next){ next 1091 drivers/scsi/dpt_i2o.c next = d->next; next 1169 drivers/scsi/dpt_i2o.c wait_data->next = adpt_post_wait_queue; next 1212 drivers/scsi/dpt_i2o.c for(p1 = adpt_post_wait_queue; p1; p2 = p1, p1 = p1->next) { next 1218 drivers/scsi/dpt_i2o.c p2->next = p1->next; next 1220 drivers/scsi/dpt_i2o.c adpt_post_wait_queue = p1->next; next 1283 drivers/scsi/dpt_i2o.c for(p1 = adpt_post_wait_queue; p1; p1 = p1->next) { next 1295 drivers/scsi/dpt_i2o.c for(p1 = adpt_post_wait_queue; p1; p1 = p1->next) { next 1476 drivers/scsi/dpt_i2o.c d->next = NULL; next 1486 drivers/scsi/dpt_i2o.c for(d = pHba->devices; d ; d = d->next) { next 1512 drivers/scsi/dpt_i2o.c for(d = pHba->devices; d ; d = d->next) { next 1581 drivers/scsi/dpt_i2o.c d->next=pHba->devices; next 1607 drivers/scsi/dpt_i2o.c for (pHba = hba_chain; pHba; pHba = pHba->next) { next 1640 drivers/scsi/dpt_i2o.c for (pHba = hba_chain; pHba; pHba = pHba->next) { next 1963 drivers/scsi/dpt_i2o.c for (pHba = hba_chain; pHba; pHba = pHba->next) { next 2508 drivers/scsi/dpt_i2o.c for (d = pHba->devices; d; d = d->next) { next 2559 drivers/scsi/dpt_i2o.c d->next = NULL; next 2632 drivers/scsi/dpt_i2o.c for (pI2o_dev = pHba->devices; pI2o_dev; pI2o_dev = pI2o_dev->next) { next 3101 drivers/scsi/dpt_i2o.c for(pHba = hba_chain; pHba; pHba = pHba->next) { next 3537 drivers/scsi/dpt_i2o.c adpt_hba *pHba, *next; next 3547 drivers/scsi/dpt_i2o.c for (pHba = hba_chain; pHba; pHba = pHba->next) { next 3555 drivers/scsi/dpt_i2o.c for (pHba = hba_chain; pHba; pHba = next) { next 3556 drivers/scsi/dpt_i2o.c next = pHba->next; next 3564 drivers/scsi/dpt_i2o.c adpt_hba *pHba, *next; next 3566 drivers/scsi/dpt_i2o.c for (pHba = hba_chain; pHba; pHba = next) { next 3567 drivers/scsi/dpt_i2o.c next = pHba->next; next 203 drivers/scsi/dpti.h struct _adpt_hba *next; next 1156 drivers/scsi/esas2r/esas2r.h sgl = a->free_sg_list_head.next; next 1407 drivers/scsi/esas2r/esas2r.h struct list_head *element, *next; next 1409 drivers/scsi/esas2r/esas2r.h list_for_each_safe(element, next, comp_list) { next 437 drivers/scsi/esas2r/esas2r_init.c struct esas2r_mem_desc *memdesc, *next; next 512 drivers/scsi/esas2r/esas2r_init.c list_for_each_entry_safe(memdesc, next, &a->free_sg_list_head, next 518 drivers/scsi/esas2r/esas2r_init.c list_for_each_entry_safe(memdesc, next, &a->vrq_mds_head, next_desc) { next 342 drivers/scsi/esas2r/esas2r_int.c struct list_head *element, *next; next 346 drivers/scsi/esas2r/esas2r_int.c list_for_each_safe(element, next, &a->defer_list) { next 771 drivers/scsi/esas2r/esas2r_io.c struct list_head *next, *element; next 782 drivers/scsi/esas2r/esas2r_io.c list_for_each_safe(element, next, &a->defer_list) { next 815 drivers/scsi/esas2r/esas2r_io.c list_for_each_safe(element, next, &a->active_list) { next 908 drivers/scsi/esas2r/esas2r_main.c struct list_head *element, *next; next 910 drivers/scsi/esas2r/esas2r_main.c list_for_each_safe(element, next, queue) { next 883 drivers/scsi/esp_scsi.c ret = list_entry(head->next, struct esp_cmd_entry, list); next 272 drivers/scsi/fcoe/fcoe_ctlr.c struct fcoe_fcf *next; next 275 drivers/scsi/fcoe/fcoe_ctlr.c list_for_each_entry_safe(fcf, next, &fip->fcfs, list) { next 820 drivers/scsi/fcoe/fcoe_ctlr.c struct fcoe_fcf *next; next 831 drivers/scsi/fcoe/fcoe_ctlr.c list_for_each_entry_safe(fcf, next, &fip->fcfs, list) { next 867 drivers/scsi/fcoe/fcoe_ctlr.c list_for_each_entry_safe(fcf, next, &del_list, list) { next 873 drivers/scsi/fcoe/fcoe_sysfs.c struct fcoe_fcf_device *fcf, *next; next 876 drivers/scsi/fcoe/fcoe_sysfs.c list_for_each_entry_safe(fcf, next, next 228 drivers/scsi/fnic/fnic_fcs.c struct fnic_event *next = NULL; next 237 drivers/scsi/fnic/fnic_fcs.c list_for_each_entry_safe(fevt, next, &fnic->evlist, list) { next 248 drivers/scsi/fnic/fnic_fcs.c struct fnic_event *next = NULL; next 257 drivers/scsi/fnic/fnic_fcs.c list_for_each_entry_safe(fevt, next, &fnic->evlist, list) { next 1315 drivers/scsi/fnic/fnic_fcs.c struct fcoe_vlan *next; next 1324 drivers/scsi/fnic/fnic_fcs.c list_for_each_entry_safe(vlan, next, &fnic->vlans, list) { next 48 drivers/scsi/fnic/vnic_rq.c buf->next = rq->bufs[0]; next 51 drivers/scsi/fnic/vnic_rq.c buf->next = rq->bufs[i + 1]; next 53 drivers/scsi/fnic/vnic_rq.c buf->next = buf + 1; next 179 drivers/scsi/fnic/vnic_rq.c buf = rq->to_clean = buf->next; next 83 drivers/scsi/fnic/vnic_rq.h struct vnic_rq_buf *next; next 143 drivers/scsi/fnic/vnic_rq.h buf = buf->next; next 199 drivers/scsi/fnic/vnic_rq.h rq->to_clean = buf->next; next 68 drivers/scsi/fnic/vnic_wq.c buf->next = wq->bufs[0]; next 71 drivers/scsi/fnic/vnic_wq.c buf->next = wq->bufs[i + 1]; next 73 drivers/scsi/fnic/vnic_wq.c buf->next = buf + 1; next 235 drivers/scsi/fnic/vnic_wq.c buf = wq->to_clean = buf->next; next 70 drivers/scsi/fnic/vnic_wq.h struct vnic_wq_buf *next; next 126 drivers/scsi/fnic/vnic_wq.h buf = buf->next; next 156 drivers/scsi/fnic/vnic_wq.h wq->to_clean = buf->next; next 9638 drivers/scsi/hpsa.c struct hpsa_sas_phy *next; next 9640 drivers/scsi/hpsa.c list_for_each_entry_safe(hpsa_sas_phy, next, next 9665 drivers/scsi/hpsa.c struct hpsa_sas_port *next; next 9670 drivers/scsi/hpsa.c list_for_each_entry_safe(hpsa_sas_port, next, next 704 drivers/scsi/hptiop.c hba->req_list = ret->next; next 712 drivers/scsi/hptiop.c req->next = hba->req_list; next 1226 drivers/scsi/hptiop.c hba->u.mvfrey.internal_req.next = NULL; next 1454 drivers/scsi/hptiop.c hba->reqs[i].next = NULL; next 247 drivers/scsi/hptiop.h struct hptiop_request *next; next 1268 drivers/scsi/ibmvscsi/ibmvfc.c evt = list_entry(vhost->free.next, struct ibmvfc_event, queue); next 447 drivers/scsi/ibmvscsi/ibmvscsi.c pool->next = 0; next 559 drivers/scsi/ibmvscsi/ibmvscsi.c int offset = pool->next; next 564 drivers/scsi/ibmvscsi/ibmvscsi.c pool->next = offset; next 72 drivers/scsi/ibmvscsi/ibmvscsi.h int next; next 666 drivers/scsi/initio.c if ((host->first_avail = scb->next) == NULL) next 668 drivers/scsi/initio.c scb->next = NULL; next 693 drivers/scsi/initio.c cmnd->next = NULL; next 695 drivers/scsi/initio.c host->last_avail->next = cmnd; next 712 drivers/scsi/initio.c scbp->next = NULL; next 714 drivers/scsi/initio.c host->last_pending->next = scbp; next 730 drivers/scsi/initio.c if ((scbp->next = host->first_pending) != NULL) { next 755 drivers/scsi/initio.c first = first->next; next 760 drivers/scsi/initio.c first = first->next; next 777 drivers/scsi/initio.c if ((host->first_pending = tmp->next) == NULL) next 780 drivers/scsi/initio.c prev->next = tmp->next; next 784 drivers/scsi/initio.c tmp->next = NULL; next 788 drivers/scsi/initio.c tmp = tmp->next; next 803 drivers/scsi/initio.c scbp->next = NULL; next 805 drivers/scsi/initio.c host->last_busy->next = scbp; next 820 drivers/scsi/initio.c if ((host->first_busy = tmp->next) == NULL) next 822 drivers/scsi/initio.c tmp->next = NULL; next 847 drivers/scsi/initio.c if ((host->first_busy = tmp->next) == NULL) next 850 drivers/scsi/initio.c prev->next = tmp->next; next 854 drivers/scsi/initio.c tmp->next = NULL; next 862 drivers/scsi/initio.c tmp = tmp->next; next 880 drivers/scsi/initio.c tmp = tmp->next; next 895 drivers/scsi/initio.c scbp->next = NULL; next 897 drivers/scsi/initio.c host->last_done->next = scbp; next 910 drivers/scsi/initio.c if ((host->first_done = tmp->next) == NULL) next 912 drivers/scsi/initio.c tmp->next = NULL; next 947 drivers/scsi/initio.c if ((host->first_pending = tmp->next) == NULL) next 950 drivers/scsi/initio.c prev->next = tmp->next; next 962 drivers/scsi/initio.c tmp = tmp->next; next 977 drivers/scsi/initio.c if ((host->first_busy = tmp->next) == NULL) next 980 drivers/scsi/initio.c prev->next = tmp->next; next 984 drivers/scsi/initio.c tmp->next = NULL; next 996 drivers/scsi/initio.c tmp = tmp->next; next 1250 drivers/scsi/initio.c int next; next 1252 drivers/scsi/initio.c next = host->active->next_state; next 1254 drivers/scsi/initio.c switch (next) { next 1256 drivers/scsi/initio.c next = initio_state_1(host); next 1259 drivers/scsi/initio.c next = initio_state_2(host); next 1262 drivers/scsi/initio.c next = initio_state_3(host); next 1265 drivers/scsi/initio.c next = initio_state_4(host); next 1268 drivers/scsi/initio.c next = initio_state_5(host); next 1271 drivers/scsi/initio.c next = initio_state_6(host); next 1274 drivers/scsi/initio.c next = initio_state_7(host); next 1281 drivers/scsi/initio.c if (next <= 0) next 1282 drivers/scsi/initio.c return next; next 2378 drivers/scsi/initio.c if ((host->first_busy = tmp->next) == NULL) next 2381 drivers/scsi/initio.c prev->next = tmp->next; next 2392 drivers/scsi/initio.c tmp = tmp->next; next 2878 drivers/scsi/initio.c prev->next = tmp; next 2881 drivers/scsi/initio.c prev->next = NULL; next 350 drivers/scsi/initio.h struct scsi_ctrl_blk *next; next 701 drivers/scsi/ipr.c ipr_cmd = list_entry(hrrq->hrrq_free_q.next, next 1461 drivers/scsi/ipr.c res = list_entry(ioa_cfg->free_res_q.next, next 7580 drivers/scsi/ipr.c ipr_cmd->u.res = list_entry(ioa_cfg->used_res_q.next, next 7654 drivers/scsi/ipr.c ipr_cmd->u.res = list_entry(ioa_cfg->used_res_q.next, next 7838 drivers/scsi/ipr.c res = list_entry(ioa_cfg->free_res_q.next, next 8563 drivers/scsi/ipr.c hostrcb = list_entry(ioa_cfg->hostrcb_free_q.next, next 801 drivers/scsi/ips.c item = item->next; next 866 drivers/scsi/ips.c item = item->next; next 1117 drivers/scsi/ips.c scratch->next = NULL; next 2980 drivers/scsi/ips.c item->next = NULL; next 2983 drivers/scsi/ips.c queue->tail->next = item; next 3017 drivers/scsi/ips.c queue->head = item->next; next 3018 drivers/scsi/ips.c item->next = NULL; next 3055 drivers/scsi/ips.c while ((p) && (item != p->next)) next 3056 drivers/scsi/ips.c p = p->next; next 3060 drivers/scsi/ips.c p->next = item->next; next 3062 drivers/scsi/ips.c if (!item->next) next 3065 drivers/scsi/ips.c item->next = NULL; next 978 drivers/scsi/ips.h struct ips_copp_wait_item *next; next 2155 drivers/scsi/isci/request.c enum sci_base_request_states next) next 2163 drivers/scsi/isci/request.c sci_change_state(&ireq->sm, next); next 104 drivers/scsi/isci/task.c task = list_entry(task->list.next, struct sas_task, list)) next 81 drivers/scsi/libfc/fc_disc.c struct fc_disc_port *dp, *next; next 148 drivers/scsi/libfc/fc_disc.c list_for_each_entry_safe(dp, next, &disc_ports, peers) { next 1923 drivers/scsi/libfc/fc_exch.c struct fc_exch *next; next 1927 drivers/scsi/libfc/fc_exch.c list_for_each_entry_safe(ep, next, &pool->ex_list, ex_list) { next 2543 drivers/scsi/libfc/fc_exch.c struct fc_exch_mgr_anchor *ema, *next; next 2546 drivers/scsi/libfc/fc_exch.c list_for_each_entry_safe(ema, next, &lport->ema_list, ema_list) next 1496 drivers/scsi/libiscsi.c conn->task = list_entry(conn->mgmtqueue.next, next 1517 drivers/scsi/libiscsi.c conn->task = list_entry(conn->cmdqueue.next, struct iscsi_task, next 1560 drivers/scsi/libiscsi.c task = list_entry(conn->requeue.next, struct iscsi_task, next 819 drivers/scsi/libsas/sas_ata.c list_del_init(sata_q.next); next 67 drivers/scsi/libsas/sas_discover.c phy = container_of(port->phy_list.next, struct asd_sas_phy, port_phy_el); next 110 drivers/scsi/libsas/sas_event.c sas_phy = container_of(port->phy_list.next, struct asd_sas_phy, next 700 drivers/scsi/libsas/sas_scsi_host.c ssp = list_entry(ha->eh_dev_q.next, typeof(*ssp), eh_list_node); next 613 drivers/scsi/lpfc/lpfc_bsg.c prsp = (struct lpfc_dmabuf *)pcmd->list.next; next 810 drivers/scsi/lpfc/lpfc_bsg.c ed = list_entry(evt->events_to_get.next, typeof(*ed), node); next 817 drivers/scsi/lpfc/lpfc_bsg.c ed = list_entry(evt->events_to_see.next, typeof(*ed), node); next 883 drivers/scsi/lpfc/lpfc_bsg.c struct list_head head, *curr, *next; next 893 drivers/scsi/lpfc/lpfc_bsg.c list_for_each_safe(curr, next, &head) { next 2960 drivers/scsi/lpfc/lpfc_bsg.c struct list_head head, *curr, *next; next 2998 drivers/scsi/lpfc/lpfc_bsg.c list_for_each_safe(curr, next, &head) { next 3042 drivers/scsi/lpfc/lpfc_bsg.c dmp = list_entry(next, struct lpfc_dmabuf, list); next 3056 drivers/scsi/lpfc/lpfc_bsg.c dmp = list_entry(next, struct lpfc_dmabuf, list); next 5622 drivers/scsi/lpfc/lpfc_bsg.c struct lpfc_dmabuf *dmabuf, *next; next 5661 drivers/scsi/lpfc/lpfc_bsg.c list_for_each_entry_safe(dmabuf, next, next 2073 drivers/scsi/lpfc/lpfc_els.c cmdiocb->context2)->list.next, next 2192 drivers/scsi/lpfc/lpfc_hbadisc.c } else if (fcf_pri->list.next == &phba->fcf.fcf_pri_list next 11092 drivers/scsi/lpfc/lpfc_init.c struct lpfc_queue *eq, *next; next 11107 drivers/scsi/lpfc/lpfc_init.c list_for_each_entry_safe(eq, next, &eqlist, _poll_list) { next 11118 drivers/scsi/lpfc/lpfc_init.c struct lpfc_queue *eq, *next; next 11130 drivers/scsi/lpfc/lpfc_init.c list_for_each_entry_safe(eq, next, &phba->poll_list, _poll_list) { next 12531 drivers/scsi/lpfc/lpfc_init.c struct lpfc_dmabuf *dmabuf, *next; next 12602 drivers/scsi/lpfc/lpfc_init.c list_for_each_entry_safe(dmabuf, next, &dma_buffer_list, list) { next 32 drivers/scsi/lpfc/lpfc_scsi.h entry = list_entry((list)->next, type, member); \ next 39 drivers/scsi/lpfc/lpfc_scsi.h list_entry((list)->next, type, member) next 6215 drivers/scsi/lpfc/lpfc_sli.c struct lpfc_dmabuf *dmabuf, *next; next 6218 drivers/scsi/lpfc/lpfc_sli.c list_for_each_entry_safe(dmabuf, next, next 10964 drivers/scsi/lpfc/lpfc_sli.c slp->next, slp->prev, pring->postbufq_cnt); next 11008 drivers/scsi/lpfc/lpfc_sli.c slp->next, slp->prev, pring->postbufq_cnt); next 14396 drivers/scsi/lpfc/lpfc_sli.c struct lpfc_queue *eq, *next; next 14398 drivers/scsi/lpfc/lpfc_sli.c list_for_each_entry_safe(eq, next, &phba->poll_list, _poll_list) next 431 drivers/scsi/megaraid.c scb = list_entry(head->next, scb_t, list); next 433 drivers/scsi/megaraid.c list_del_init(head->next); next 1068 drivers/scsi/megaraid.c struct list_head *pos, *next; next 1071 drivers/scsi/megaraid.c list_for_each_safe(pos, next, &adapter->pending_list) { next 1934 drivers/scsi/megaraid.c struct list_head *pos, *next; next 1945 drivers/scsi/megaraid.c list_for_each_safe(pos, next, &adapter->pending_list) { next 1286 drivers/scsi/megaraid/megaraid_mbox.c scb = list_entry(head->next, scb_t, list); next 1886 drivers/scsi/megaraid/megaraid_mbox.c scb = list_entry(adapter->pend_list.next, scb_t, list); next 3584 drivers/scsi/megaraid/megaraid_mbox.c scb = list_entry(head->next, scb_t, list); next 612 drivers/scsi/megaraid/megaraid_mm.c kioc = list_entry(head->next, uioc_t, list); next 314 drivers/scsi/megaraid/megaraid_sas_base.c cmd = list_entry((&instance->cmd_pool)->next, next 358 drivers/scsi/megaraid/megaraid_sas_base.c list_add(&cmd->list, (&instance->cmd_pool)->next); next 2713 drivers/scsi/megaraid/megaraid_sas_base.c reset_cmd = list_entry((&clist_local)->next, next 3699 drivers/scsi/megaraid/megaraid_sas_base.c cmd = list_entry((&clist_local)->next, next 150 drivers/scsi/mesh.c struct mesh_state *next; next 560 drivers/scsi/mesh.c struct scsi_cmnd *cmd, *prev, *next; next 577 drivers/scsi/mesh.c next = (struct scsi_cmnd *) cmd->host_scribble; next 579 drivers/scsi/mesh.c ms->request_q = next; next 581 drivers/scsi/mesh.c prev->host_scribble = (void *) next; next 582 drivers/scsi/mesh.c if (next == NULL) next 1658 drivers/scsi/mpt3sas/mpt3sas_base.c struct adapter_reply_queue *reply_q, *next; next 1663 drivers/scsi/mpt3sas/mpt3sas_base.c list_for_each_entry_safe(reply_q, next, &ioc->reply_queue_list, list) { next 2816 drivers/scsi/mpt3sas/mpt3sas_base.c struct adapter_reply_queue *reply_q, *next; next 2821 drivers/scsi/mpt3sas/mpt3sas_base.c list_for_each_entry_safe(reply_q, next, &ioc->reply_queue_list, list) { next 3519 drivers/scsi/mpt3sas/mpt3sas_base.c request = list_entry(ioc->internal_free_list.next, next 3572 drivers/scsi/mpt3sas/mpt3sas_base.c request = list_entry(ioc->hpr_free_list.next, next 4096 drivers/scsi/mpt3sas/mpt3sas_scsih.c delayed_event_ack = list_entry(ioc->delayed_event_ack_list.next, next 4106 drivers/scsi/mpt3sas/mpt3sas_scsih.c delayed_sc = list_entry(ioc->delayed_sc_list.next, next 4134 drivers/scsi/mpt3sas/mpt3sas_scsih.c delayed_tr = list_entry(ioc->delayed_tr_volume_list.next, next 4144 drivers/scsi/mpt3sas/mpt3sas_scsih.c delayed_tr = list_entry(ioc->delayed_tr_list.next, next 9627 drivers/scsi/mpt3sas/mpt3sas_scsih.c struct _sas_port *mpt3sas_port, *next; next 9631 drivers/scsi/mpt3sas/mpt3sas_scsih.c list_for_each_entry_safe(mpt3sas_port, next, next 9741 drivers/scsi/mpt3sas/mpt3sas_scsih.c struct _raid_device *raid_device, *next; next 9771 drivers/scsi/mpt3sas/mpt3sas_scsih.c list_for_each_entry_safe(raid_device, next, &ioc->raid_device_list, next 546 drivers/scsi/mpt3sas/mpt3sas_transport.c struct _sas_port *mpt3sas_port, *next; next 552 drivers/scsi/mpt3sas/mpt3sas_transport.c list_for_each_entry_safe(mpt3sas_port, next, &sas_node->sas_port_list, next 607 drivers/scsi/mpt3sas/mpt3sas_transport.c struct _sas_phy *mpt3sas_phy, *next; next 743 drivers/scsi/mpt3sas/mpt3sas_transport.c list_for_each_entry_safe(mpt3sas_phy, next, &mpt3sas_port->phy_list, next 766 drivers/scsi/mpt3sas/mpt3sas_transport.c struct _sas_port *mpt3sas_port, *next; next 778 drivers/scsi/mpt3sas/mpt3sas_transport.c list_for_each_entry_safe(mpt3sas_port, next, &sas_node->sas_port_list, next 70 drivers/scsi/mvsas/mv_sas.c phy = container_of(sha->sas_port[i]->phy_list.next, next 321 drivers/scsi/mvumi.c cmd = list_entry((&mhba->cmd_pool)->next, next 154 drivers/scsi/ncr53c8xx.c struct list_head *elem = head->next; next 198 drivers/scsi/ncr53c8xx.c struct m_link *next; next 202 drivers/scsi/ncr53c8xx.c struct m_vtob *next; next 218 drivers/scsi/ncr53c8xx.c struct m_pool *next; next 239 drivers/scsi/ncr53c8xx.c while (!h[j].next) { next 241 drivers/scsi/ncr53c8xx.c h[j].next = (m_link_s *)mp->getp(mp); next 242 drivers/scsi/ncr53c8xx.c if (h[j].next) next 243 drivers/scsi/ncr53c8xx.c h[j].next->next = NULL; next 249 drivers/scsi/ncr53c8xx.c a = (m_addr_t) h[j].next; next 251 drivers/scsi/ncr53c8xx.c h[j].next = h[j].next->next; next 255 drivers/scsi/ncr53c8xx.c h[j].next = (m_link_s *) (a+s); next 256 drivers/scsi/ncr53c8xx.c h[j].next->next = NULL; next 296 drivers/scsi/ncr53c8xx.c while (q->next && q->next != (m_link_s *) b) { next 297 drivers/scsi/ncr53c8xx.c q = q->next; next 299 drivers/scsi/ncr53c8xx.c if (!q->next) { next 300 drivers/scsi/ncr53c8xx.c ((m_link_s *) a)->next = h[i].next; next 301 drivers/scsi/ncr53c8xx.c h[i].next = (m_link_s *) a; next 304 drivers/scsi/ncr53c8xx.c q->next = q->next->next; next 386 drivers/scsi/ncr53c8xx.c vbp->next = mp->vtob[hc]; next 404 drivers/scsi/ncr53c8xx.c vbpp = &(*vbpp)->next; next 407 drivers/scsi/ncr53c8xx.c *vbpp = (*vbpp)->next; next 418 drivers/scsi/ncr53c8xx.c for (mp = mp0.next; mp && mp->bush != bush; mp = mp->next); next 431 drivers/scsi/ncr53c8xx.c mp->next = mp0.next; next 432 drivers/scsi/ncr53c8xx.c mp0.next = mp; next 439 drivers/scsi/ncr53c8xx.c struct m_pool **pp = &mp0.next; next 442 drivers/scsi/ncr53c8xx.c pp = &(*pp)->next; next 444 drivers/scsi/ncr53c8xx.c *pp = (*pp)->next; next 495 drivers/scsi/ncr53c8xx.c vp = vp->next; next 370 drivers/scsi/pmcraid.c cmd = list_entry(pinstance->free_cmd_pool.next, next 1611 drivers/scsi/pmcraid.c res = list_entry(pinstance->free_res_q.next, next 5545 drivers/scsi/pmcraid.c res = list_entry(pinstance->free_res_q.next, next 208 drivers/scsi/qedi/qedi.h u16 next; next 523 drivers/scsi/qedi/qedi_main.c u16 start_id, u16 next) next 527 drivers/scsi/qedi/qedi_main.c id_tbl->next = next; next 564 drivers/scsi/qedi/qedi_main.c id = find_next_zero_bit(id_tbl->table, id_tbl->max, id_tbl->next); next 567 drivers/scsi/qedi/qedi_main.c if (id_tbl->next != 0) { next 568 drivers/scsi/qedi/qedi_main.c id = find_first_zero_bit(id_tbl->table, id_tbl->next); next 569 drivers/scsi/qedi/qedi_main.c if (id >= id_tbl->next) next 576 drivers/scsi/qedi/qedi_main.c id_tbl->next = (id + 1) & (id_tbl->max - 1); next 1252 drivers/scsi/qla1280.c sp = list_entry(done_q->next, struct srb, list); next 1012 drivers/scsi/qla1280.h struct scsi_qla_host *next; next 3583 drivers/scsi/qla2xxx/qla_os.c BUG_ON(base_vha->list.next == &ha->vp_list); next 1281 drivers/scsi/qla2xxx/qla_sup.c goto next; next 1291 drivers/scsi/qla2xxx/qla_sup.c next: next 6295 drivers/scsi/qla2xxx/qla_target.c tgt->sess_works_list.next, typeof(*prm), next 158 drivers/scsi/qlogicfas.c priv->next = cards; next 216 drivers/scsi/qlogicfas.c for (priv = cards; priv != NULL; priv = priv->next) next 87 drivers/scsi/qlogicfas408.h struct qlogicfas408_priv *next; /* next private struct */ next 690 drivers/scsi/qlogicpti.c while(qlink->next) next 691 drivers/scsi/qlogicpti.c qlink = qlink->next; next 692 drivers/scsi/qlogicpti.c qlink->next = qpti; next 696 drivers/scsi/qlogicpti.c qpti->next = NULL; next 704 drivers/scsi/qlogicpti.c qptichain = qpti->next; next 707 drivers/scsi/qlogicpti.c while(qlink->next != qpti) next 708 drivers/scsi/qlogicpti.c qlink = qlink->next; next 709 drivers/scsi/qlogicpti.c qlink->next = qpti->next; next 711 drivers/scsi/qlogicpti.c qpti->next = NULL; next 1214 drivers/scsi/qlogicpti.c struct scsi_cmnd *next; next 1216 drivers/scsi/qlogicpti.c next = (struct scsi_cmnd *) dq->host_scribble; next 1218 drivers/scsi/qlogicpti.c dq = next; next 359 drivers/scsi/qlogicpti.h struct qlogicpti *next; next 505 drivers/scsi/qlogicpti.h for((qp) = qptichain; (qp); (qp) = (qp)->next) next 98 drivers/scsi/raid_class.c struct raid_component *rc, *next; next 101 drivers/scsi/raid_class.c list_for_each_entry_safe(rc, next, &rd->component_list, node) { next 568 drivers/scsi/scsi.c struct scsi_device *next = NULL; next 572 drivers/scsi/scsi.c while (list->next != &shost->__devices) { next 573 drivers/scsi/scsi.c next = list_entry(list->next, struct scsi_device, siblings); next 575 drivers/scsi/scsi.c if (!scsi_device_get(next)) next 577 drivers/scsi/scsi.c next = NULL; next 578 drivers/scsi/scsi.c list = list->next; next 584 drivers/scsi/scsi.c return next; next 2909 drivers/scsi/scsi_debug.c unsigned long next; next 2915 drivers/scsi/scsi_debug.c next = find_next_zero_bit(map_storep, map_size, index); next 2917 drivers/scsi/scsi_debug.c next = find_next_bit(map_storep, map_size, index); next 2919 drivers/scsi/scsi_debug.c end = min_t(sector_t, sdebug_store_sectors, map_index_to_lba(next)); next 523 drivers/scsi/scsi_devinfo.c char *vendor, *model, *strflags, *next; next 527 drivers/scsi/scsi_devinfo.c next = dev_list; next 528 drivers/scsi/scsi_devinfo.c if (next && next[0] == '"') { next 532 drivers/scsi/scsi_devinfo.c next++; next 542 drivers/scsi/scsi_devinfo.c for (vendor = strsep(&next, ":"); vendor && (vendor[0] != '\0') next 543 drivers/scsi/scsi_devinfo.c && (res == 0); vendor = strsep(&next, ":")) { next 545 drivers/scsi/scsi_devinfo.c model = strsep(&next, ":"); next 547 drivers/scsi/scsi_devinfo.c strflags = strsep(&next, next_check); next 632 drivers/scsi/scsi_devinfo.c if (devinfo_table->scsi_dev_info_list.next == dl->bottom && next 669 drivers/scsi/scsi_devinfo.c dl->bottom = dl->bottom->next; next 671 drivers/scsi/scsi_devinfo.c dl->top = dl->top->next; next 679 drivers/scsi/scsi_devinfo.c dl->bottom = devinfo_table->scsi_dev_info_list.next; next 692 drivers/scsi/scsi_devinfo.c .next = devinfo_seq_next, next 1219 drivers/scsi/scsi_error.c struct scsi_cmnd *scmd, *next; next 1227 drivers/scsi/scsi_error.c list_for_each_entry_safe(scmd, next, work_q, eh_entry) { next 1330 drivers/scsi/scsi_error.c struct scsi_cmnd *scmd, *next; next 1335 drivers/scsi/scsi_error.c scmd = list_entry(cmd_list->next, struct scsi_cmnd, eh_entry); next 1355 drivers/scsi/scsi_error.c list_for_each_entry_safe(scmd, next, cmd_list, eh_entry) next 1406 drivers/scsi/scsi_error.c struct scsi_cmnd *scmd, *stu_scmd, *next; next 1436 drivers/scsi/scsi_error.c list_for_each_entry_safe(scmd, next, next 1471 drivers/scsi/scsi_error.c struct scsi_cmnd *scmd, *bdr_scmd, *next; next 1501 drivers/scsi/scsi_error.c list_for_each_entry_safe(scmd, next, next 1538 drivers/scsi/scsi_error.c struct scsi_cmnd *next, *scmd; next 1553 drivers/scsi/scsi_error.c scmd = list_entry(tmp_list.next, struct scsi_cmnd, eh_entry); next 1567 drivers/scsi/scsi_error.c list_for_each_entry_safe(scmd, next, &tmp_list, eh_entry) { next 1594 drivers/scsi/scsi_error.c struct scsi_cmnd *scmd, *chan_scmd, *next; next 1636 drivers/scsi/scsi_error.c list_for_each_entry_safe(scmd, next, work_q, eh_entry) { next 1666 drivers/scsi/scsi_error.c struct scsi_cmnd *scmd, *next; next 1671 drivers/scsi/scsi_error.c scmd = list_entry(work_q->next, next 1683 drivers/scsi/scsi_error.c list_for_each_entry_safe(scmd, next, work_q, eh_entry) { next 1704 drivers/scsi/scsi_error.c struct scsi_cmnd *scmd, *next; next 1707 drivers/scsi/scsi_error.c list_for_each_entry_safe(scmd, next, work_q, eh_entry) { next 2087 drivers/scsi/scsi_error.c struct scsi_cmnd *scmd, *next; next 2089 drivers/scsi/scsi_error.c list_for_each_entry_safe(scmd, next, done_q, eh_entry) { next 468 drivers/scsi/scsi_lib.c sdev = list_entry(starved_list.next, next 377 drivers/scsi/scsi_proc.c struct device *next = bus_find_next_device(&scsi_bus_type, start); next 380 drivers/scsi/scsi_proc.c return next; next 418 drivers/scsi/scsi_proc.c .next = scsi_seq_next, next 165 drivers/scsi/scsi_scan.c struct async_scan_data *next = list_entry(scanning_hosts.next, next 167 drivers/scsi/scsi_scan.c complete(&next->prev_finished); next 1796 drivers/scsi/scsi_scan.c struct async_scan_data *next = list_entry(scanning_hosts.next, next 1798 drivers/scsi/scsi_scan.c complete(&next->prev_finished); next 1594 drivers/scsi/scsi_transport_fc.c pos = list_entry((head)->next, typeof(*pos), member) next 1039 drivers/scsi/scsi_transport_sas.c struct list_head *ent = port->phy_list.next; next 2332 drivers/scsi/sg.c .next = dev_seq_next, next 2340 drivers/scsi/sg.c .next = dev_seq_next, next 2348 drivers/scsi/sg.c .next = dev_seq_next, next 53 drivers/scsi/sim710.c char *pos = str, *next; next 56 drivers/scsi/sim710.c while(pos != NULL && (next = strchr(pos, ':')) != NULL) { next 57 drivers/scsi/sim710.c int val = (int)simple_strtoul(++next, NULL, 0); next 1747 drivers/scsi/smartpqi/smartpqi_init.c struct pqi_scsi_dev *next; next 1800 drivers/scsi/smartpqi/smartpqi_init.c list_for_each_entry_safe(device, next, &ctrl_info->scsi_device_list, next 1828 drivers/scsi/smartpqi/smartpqi_init.c list_for_each_entry_safe(device, next, &delete_list, next 1857 drivers/scsi/smartpqi/smartpqi_init.c list_for_each_entry_safe(device, next, &add_list, add_list_entry) { next 1957 drivers/scsi/smartpqi/smartpqi_init.c struct pqi_scsi_dev *next; next 2135 drivers/scsi/smartpqi/smartpqi_init.c list_for_each_entry_safe(device, next, &new_device_list_head, next 3882 drivers/scsi/smartpqi/smartpqi_init.c struct pqi_io_request *next; next 3903 drivers/scsi/smartpqi/smartpqi_init.c list_for_each_entry_safe(io_request, next, next 5459 drivers/scsi/smartpqi/smartpqi_init.c struct pqi_io_request *next; next 5470 drivers/scsi/smartpqi/smartpqi_init.c list_for_each_entry_safe(io_request, next, next 5499 drivers/scsi/smartpqi/smartpqi_init.c struct pqi_io_request *next; next 5509 drivers/scsi/smartpqi/smartpqi_init.c list_for_each_entry_safe(io_request, next, next 160 drivers/scsi/smartpqi/smartpqi_sas_transport.c struct pqi_sas_phy *next; next 162 drivers/scsi/smartpqi/smartpqi_sas_transport.c list_for_each_entry_safe(pqi_sas_phy, next, next 187 drivers/scsi/smartpqi/smartpqi_sas_transport.c struct pqi_sas_port *next; next 192 drivers/scsi/smartpqi/smartpqi_sas_transport.c list_for_each_entry_safe(pqi_sas_port, next, next 444 drivers/scsi/snic/snic_debugfs.c .next = snic_trc_seq_next, next 115 drivers/scsi/snic/snic_io.c SNIC_BUG_ON(rqi->list.next == NULL); /* if not added to spl_cmd_list */ next 65 drivers/scsi/snic/vnic_wq.c buf->next = wq->bufs[0]; next 68 drivers/scsi/snic/vnic_wq.c buf->next = wq->bufs[i + 1]; next 70 drivers/scsi/snic/vnic_wq.c buf->next = buf + 1; next 226 drivers/scsi/snic/vnic_wq.c buf = wq->to_clean = buf->next; next 51 drivers/scsi/snic/vnic_wq.h struct vnic_wq_buf *next; next 113 drivers/scsi/snic/vnic_wq.h buf = buf->next; next 143 drivers/scsi/snic/vnic_wq.h wq->to_clean = buf->next; next 1121 drivers/scsi/sym53c8xx_2/sym_hipd.h struct sym_m_link *next; next 1129 drivers/scsi/sym53c8xx_2/sym_hipd.h struct sym_m_vtob *next; next 1160 drivers/scsi/sym53c8xx_2/sym_hipd.h struct sym_m_pool *next; next 64 drivers/scsi/sym53c8xx_2/sym_malloc.c while (!h[j].next) { next 66 drivers/scsi/sym53c8xx_2/sym_malloc.c h[j].next = (m_link_p) M_GET_MEM_CLUSTER(); next 67 drivers/scsi/sym53c8xx_2/sym_malloc.c if (h[j].next) next 68 drivers/scsi/sym53c8xx_2/sym_malloc.c h[j].next->next = NULL; next 74 drivers/scsi/sym53c8xx_2/sym_malloc.c a = h[j].next; next 76 drivers/scsi/sym53c8xx_2/sym_malloc.c h[j].next = h[j].next->next; next 80 drivers/scsi/sym53c8xx_2/sym_malloc.c h[j].next = (m_link_p) (a+s); next 81 drivers/scsi/sym53c8xx_2/sym_malloc.c h[j].next->next = NULL; next 120 drivers/scsi/sym53c8xx_2/sym_malloc.c ((m_link_p) a)->next = h[i].next; next 121 drivers/scsi/sym53c8xx_2/sym_malloc.c h[i].next = (m_link_p) a; next 127 drivers/scsi/sym53c8xx_2/sym_malloc.c while (q->next && q->next != (m_link_p) b) { next 128 drivers/scsi/sym53c8xx_2/sym_malloc.c q = q->next; next 130 drivers/scsi/sym53c8xx_2/sym_malloc.c if (!q->next) { next 131 drivers/scsi/sym53c8xx_2/sym_malloc.c ((m_link_p) a)->next = h[i].next; next 132 drivers/scsi/sym53c8xx_2/sym_malloc.c h[i].next = (m_link_p) a; next 135 drivers/scsi/sym53c8xx_2/sym_malloc.c q->next = q->next->next; next 222 drivers/scsi/sym53c8xx_2/sym_malloc.c vbp->next = mp->vtob[hc]; next 240 drivers/scsi/sym53c8xx_2/sym_malloc.c vbpp = &(*vbpp)->next; next 243 drivers/scsi/sym53c8xx_2/sym_malloc.c *vbpp = (*vbpp)->next; next 255 drivers/scsi/sym53c8xx_2/sym_malloc.c for (mp = mp0.next; next 257 drivers/scsi/sym53c8xx_2/sym_malloc.c mp = mp->next); next 271 drivers/scsi/sym53c8xx_2/sym_malloc.c mp->next = mp0.next; next 272 drivers/scsi/sym53c8xx_2/sym_malloc.c mp0.next = mp; next 282 drivers/scsi/sym53c8xx_2/sym_malloc.c m_pool_p *pp = &mp0.next; next 285 drivers/scsi/sym53c8xx_2/sym_malloc.c pp = &(*pp)->next; next 287 drivers/scsi/sym53c8xx_2/sym_malloc.c *pp = (*pp)->next; next 358 drivers/scsi/sym53c8xx_2/sym_malloc.c vp = vp->next; next 217 drivers/scsi/wd33c93.h struct Scsi_Host *next; next 421 drivers/sh/clk/core.c if (clk->node.next || clk->node.prev) next 24 drivers/sh/intc/virq.c struct intc_virq_list *next; next 28 drivers/sh/intc/virq.c for (entry = head; entry; entry = entry->next) next 93 drivers/sh/intc/virq.c last = &entry->next; next 67 drivers/soc/bcm/brcmstb/pm/pm-mips.c struct brcmstb_mem_transfer *next; next 74 drivers/soc/fsl/dpio/dpio-service.c d = list_entry(dpio_list.next, struct dpaa2_io, node); next 250 drivers/soc/fsl/qbman/bman.c struct bman_pool *next; next 457 drivers/soc/fsl/qbman/qman_test_stash.c hp_cpu->iterator->node.next, next 500 drivers/soc/fsl/qbman/qman_test_stash.c hp_cpu->iterator->node.next, next 266 drivers/spi/spi-bcm63xx.c t = list_entry(t->transfer_list.next, struct spi_transfer, next 317 drivers/spi/spi-bcm63xx.c t = list_entry(t->transfer_list.next, struct spi_transfer, next 214 drivers/spi/spi-mpc52xx-psc.c m = container_of(mps->queue.next, struct spi_message, queue); next 192 drivers/spi/spi-mpc52xx.c ms->transfer = container_of(ms->message->transfers.next, next 287 drivers/spi/spi-mpc52xx.c if (ms->transfer->transfer_list.next == &ms->message->transfers) { next 302 drivers/spi/spi-mpc52xx.c ms->transfer = container_of(ms->transfer->transfer_list.next, next 343 drivers/spi/spi-pic32-sqi.c struct ring_desc *rdesc, *next; next 430 drivers/spi/spi-pic32-sqi.c list_for_each_entry_safe_reverse(rdesc, next, next 781 drivers/spi/spi-pl022.c if (trans->transfer_list.next != &msg->transfers) { next 783 drivers/spi/spi-pl022.c list_entry(trans->transfer_list.next, next 1591 drivers/spi/spi-pl022.c pl022->cur_transfer = list_entry(msg->transfers.next, next 286 drivers/spi/spi-sh.c mesg = list_entry(ss->queue.next, struct spi_message, queue); next 585 drivers/spi/spi-topcliff-pch.c list_for_each_entry_safe(pmsg, tmp, data->queue.next, queue) { next 670 drivers/spi/spi-topcliff-pch.c list_for_each_entry_safe(pmsg, tmp, data->queue.next, queue) { next 1133 drivers/spi/spi-topcliff-pch.c list_for_each_entry_safe(pmsg, tmp, data->queue.next, queue) { next 1155 drivers/spi/spi-topcliff-pch.c data->current_msg = list_entry(data->queue.next, struct spi_message, next 1178 drivers/spi/spi-topcliff-pch.c list_entry(data->current_msg->transfers.next, next 1185 drivers/spi/spi-topcliff-pch.c list_entry(data->cur_trans->transfer_list.next, next 1242 drivers/spi/spi-topcliff-pch.c if ((data->cur_trans->transfer_list.next) == next 256 drivers/spi/spi-txx9.c if (t->transfer_list.next == &m->transfers) next 289 drivers/spi/spi-txx9.c m = container_of(c->queue.next, struct spi_message, queue); next 1484 drivers/spi/spi.c struct spi_message *next; next 1489 drivers/spi/spi.c next = list_first_entry_or_null(&ctlr->queue, struct spi_message, next 1493 drivers/spi/spi.c return next; next 2840 drivers/spi/spi.c if (rxfer->replaced_after->next == &msg->transfers) { next 2857 drivers/spi/spi.c list_move_tail(rxfer->replaced_after->next, next 581 drivers/ssb/sdio.c tuple = tuple->next; next 273 drivers/staging/android/ashmem.c struct ashmem_range *range, *next; next 276 drivers/staging/android/ashmem.c list_for_each_entry_safe(range, next, &asma->unpinned_list, unpinned) next 625 drivers/staging/android/ashmem.c struct ashmem_range *range, *next; next 628 drivers/staging/android/ashmem.c list_for_each_entry_safe(range, next, &asma->unpinned_list, unpinned) { next 693 drivers/staging/android/ashmem.c struct ashmem_range *range, *next; next 697 drivers/staging/android/ashmem.c list_for_each_entry_safe(range, next, &asma->unpinned_list, unpinned) { next 439 drivers/staging/comedi/comedidev.h struct comedi_driver *next; /* Next in list of COMEDI drivers. */ next 960 drivers/staging/comedi/drivers.c for (driv = comedi_drivers; driv; driv = driv->next) { next 975 drivers/staging/comedi/drivers.c for (driv = comedi_drivers; driv; driv = driv->next) { next 1129 drivers/staging/comedi/drivers.c driver->next = comedi_drivers; next 1154 drivers/staging/comedi/drivers.c comedi_drivers = driver->next; next 1156 drivers/staging/comedi/drivers.c for (prev = comedi_drivers; prev->next; prev = prev->next) { next 1157 drivers/staging/comedi/drivers.c if (prev->next == driver) { next 1158 drivers/staging/comedi/drivers.c prev->next = driver->next; next 1530 drivers/staging/comedi/drivers/cb_pcidas64.c devpriv->ai_dma_desc[i].next = next 1545 drivers/staging/comedi/drivers/cb_pcidas64.c devpriv->ao_dma_desc[i].next = next 2955 drivers/staging/comedi/drivers/cb_pcidas64.c next_bits = le32_to_cpu(devpriv->ao_dma_desc[buffer_index].next); next 2957 drivers/staging/comedi/drivers/cb_pcidas64.c devpriv->ao_dma_desc[buffer_index].next = cpu_to_le32(next_bits); next 2962 drivers/staging/comedi/drivers/cb_pcidas64.c next_bits = le32_to_cpu(devpriv->ao_dma_desc[prev_buffer_index].next); next 2964 drivers/staging/comedi/drivers/cb_pcidas64.c devpriv->ao_dma_desc[prev_buffer_index].next = cpu_to_le32(next_bits); next 150 drivers/staging/comedi/drivers/gsc_hpdi.c unsigned int next; next 152 drivers/staging/comedi/drivers/gsc_hpdi.c next = readl(devpriv->plx9080_mmio + PLX_REG_DMAPADR(channel)); next 157 drivers/staging/comedi/drivers/gsc_hpdi.c for (desc = 0; (next < start || next >= start + devpriv->block_size) && next 433 drivers/staging/comedi/drivers/gsc_hpdi.c devpriv->dma_desc[i].next = cpu_to_le32((phys_addr + next 447 drivers/staging/comedi/drivers/gsc_hpdi.c devpriv->dma_desc[i - 1].next = cpu_to_le32(phys_addr | next_bits); next 674 drivers/staging/comedi/drivers/mite.c desc->next = cpu_to_le32(ring->dma_addr + next 687 drivers/staging/comedi/drivers/mite.c desc->next = cpu_to_le32(ring->dma_addr); next 25 drivers/staging/comedi/drivers/mite.h __le32 next; next 46 drivers/staging/comedi/drivers/plx9080.h __le32 next; next 50 drivers/staging/comedi/proc.c for (driv = comedi_drivers; driv; driv = driv->next) { next 529 drivers/staging/exfat/exfat.h struct buf_cache_t *next; next 42 drivers/staging/exfat/exfat_cache.c bp->next = list->next; next 44 drivers/staging/exfat/exfat_cache.c list->next->prev = bp; next 45 drivers/staging/exfat/exfat_cache.c list->next = bp; next 51 drivers/staging/exfat/exfat_cache.c bp->next = list; next 52 drivers/staging/exfat/exfat_cache.c list->prev->next = bp; next 58 drivers/staging/exfat/exfat_cache.c bp->prev->next = bp->next; next 59 drivers/staging/exfat/exfat_cache.c bp->next->prev = bp->prev; next 65 drivers/staging/exfat/exfat_cache.c bp->prev->next = bp->next; next 66 drivers/staging/exfat/exfat_cache.c bp->next->prev = bp->prev; next 138 drivers/staging/exfat/exfat_cache.c p_fs->FAT_cache_lru_list.next = &p_fs->FAT_cache_lru_list; next 147 drivers/staging/exfat/exfat_cache.c p_fs->FAT_cache_array[i].next = NULL; next 152 drivers/staging/exfat/exfat_cache.c p_fs->buf_cache_lru_list.next = &p_fs->buf_cache_lru_list; next 161 drivers/staging/exfat/exfat_cache.c p_fs->buf_cache_array[i].next = NULL; next 495 drivers/staging/exfat/exfat_cache.c bp = p_fs->FAT_cache_lru_list.next; next 507 drivers/staging/exfat/exfat_cache.c bp = bp->next; next 520 drivers/staging/exfat/exfat_cache.c bp = p_fs->FAT_cache_lru_list.next; next 526 drivers/staging/exfat/exfat_cache.c bp = bp->next; next 689 drivers/staging/exfat/exfat_cache.c bp = p_fs->buf_cache_lru_list.next; next 701 drivers/staging/exfat/exfat_cache.c bp = bp->next; next 714 drivers/staging/exfat/exfat_cache.c bp = p_fs->buf_cache_lru_list.next; next 720 drivers/staging/exfat/exfat_cache.c bp = bp->next; next 405 drivers/staging/exfat/exfat_core.c u32 clu, next; next 413 drivers/staging/exfat/exfat_core.c while ((FAT_read(sb, clu, &next) == 0) && next 414 drivers/staging/exfat/exfat_core.c (next != CLUSTER_32(~0))) { next 417 drivers/staging/exfat/exfat_core.c clu = next; next 98 drivers/staging/fwserial/dma_fifo.c struct dma_pending *pending, *next; next 103 drivers/staging/fwserial/dma_fifo.c list_for_each_entry_safe(pending, next, &fifo->pending, link) next 115 drivers/staging/fwserial/dma_fifo.c struct dma_pending *pending, *next; next 120 drivers/staging/fwserial/dma_fifo.c list_for_each_entry_safe(pending, next, &fifo->pending, link) next 221 drivers/staging/fwserial/dma_fifo.c pended->next = fifo->out; next 244 drivers/staging/fwserial/dma_fifo.c struct dma_pending *pending, *next, *tmp; next 264 drivers/staging/fwserial/dma_fifo.c list_for_each_entry_safe(pending, next, &fifo->pending, link) { next 272 drivers/staging/fwserial/dma_fifo.c addr_check(fifo->in, fifo->done, pending->next), next 275 drivers/staging/fwserial/dma_fifo.c pending->next)) next 279 drivers/staging/fwserial/dma_fifo.c fifo->done = pending->next; next 61 drivers/staging/fwserial/dma_fifo.h unsigned int next; next 220 drivers/staging/gdm724x/gdm_mux.c r = list_entry(rx->to_host_list.next, struct mux_rx, next 152 drivers/staging/gdm724x/gdm_usb.c t_sdu = list_entry(tx->free_list.next, struct usb_tx_sdu, list); next 220 drivers/staging/gdm724x/gdm_usb.c r = list_entry(rx->free_list.next, struct usb_rx, free_list); next 395 drivers/staging/gdm724x/gdm_usb.c r = list_entry(rx->to_host_list.next, next 606 drivers/staging/gdm724x/gdm_usb.c t_sdu = list_entry(tx->sdu_list.next, struct usb_tx_sdu, list); next 659 drivers/staging/gdm724x/gdm_usb.c t = list_entry(tx->hci_list.next, struct usb_tx, list); next 88 drivers/staging/greybus/audio_manager.c struct gb_audio_manager_module *module, *next; next 93 drivers/staging/greybus/audio_manager.c list_for_each_entry_safe(module, next, &modules_list, list) { next 89 drivers/staging/greybus/bootrom.c enum next_request_type next, next 92 drivers/staging/greybus/bootrom.c bootrom->next_request = next; next 1162 drivers/staging/isdn/gigaset/bas-gigaset.c struct isow_urbctx_t *done, *next, *ovfl; next 1196 drivers/staging/isdn/gigaset/bas-gigaset.c next = ubc->isooutfree; next 1199 drivers/staging/isdn/gigaset/bas-gigaset.c if (next) { next 1200 drivers/staging/isdn/gigaset/bas-gigaset.c rc = submit_iso_write_urb(next); next 1205 drivers/staging/isdn/gigaset/bas-gigaset.c ubc->isooutfree = next; next 1206 drivers/staging/isdn/gigaset/bas-gigaset.c next = NULL; next 1209 drivers/staging/isdn/gigaset/bas-gigaset.c if (next) { next 1261 drivers/staging/isdn/gigaset/bas-gigaset.c next = ubc->isooutfree; next 1264 drivers/staging/isdn/gigaset/bas-gigaset.c if (next) { next 1266 drivers/staging/isdn/gigaset/bas-gigaset.c rc = submit_iso_write_urb(next); next 1751 drivers/staging/isdn/gigaset/bas-gigaset.c if (cb->next != NULL) { next 1752 drivers/staging/isdn/gigaset/bas-gigaset.c cs->cmdbuf = cb->next; next 2016 drivers/staging/isdn/gigaset/bas-gigaset.c cs->lastcmdbuf->next = cb; next 334 drivers/staging/isdn/gigaset/common.c unsigned next, tail; next 342 drivers/staging/isdn/gigaset/common.c next = (tail + 1) % MAX_EVENTS; next 343 drivers/staging/isdn/gigaset/common.c if (unlikely(next == cs->ev_head)) next 353 drivers/staging/isdn/gigaset/common.c cs->ev_tail = next; next 374 drivers/staging/isdn/gigaset/common.c struct at_state_t *cur, *next; next 376 drivers/staging/isdn/gigaset/common.c list_for_each_entry_safe(cur, next, &cs->temp_at_states, list) { next 837 drivers/staging/isdn/gigaset/common.c cb = cb->next; next 408 drivers/staging/isdn/gigaset/ev-layer.c unsigned next, tail; next 416 drivers/staging/isdn/gigaset/ev-layer.c next = (tail + 1) % MAX_EVENTS; next 417 drivers/staging/isdn/gigaset/ev-layer.c if (unlikely(next == cs->ev_head)) { next 428 drivers/staging/isdn/gigaset/ev-layer.c cs->ev_tail = next; next 780 drivers/staging/isdn/gigaset/ev-layer.c cb->next = NULL; next 1195 drivers/staging/isdn/gigaset/ev-layer.c cb->next = NULL; next 519 drivers/staging/isdn/gigaset/gigaset.h struct cmdbuf_t *next, *prev; next 322 drivers/staging/isdn/gigaset/interface.c cb->next = NULL; next 133 drivers/staging/isdn/gigaset/ser-gigaset.c cs->cmdbuf = cb = cb->next; next 209 drivers/staging/isdn/gigaset/ser-gigaset.c cs->cmdbuf = cb->next; next 250 drivers/staging/isdn/gigaset/ser-gigaset.c cs->lastcmdbuf->next = cb; next 437 drivers/staging/isdn/gigaset/usb-gigaset.c cs->cmdbuf = cb->next; next 501 drivers/staging/isdn/gigaset/usb-gigaset.c cs->lastcmdbuf->next = cb; next 155 drivers/staging/isdn/hysdn/hysdn_defs.h struct HYSDN_CARD *next; /* pointer to next card */ next 95 drivers/staging/isdn/hysdn/hysdn_init.c card->next = NULL; /*end of chain */ next 97 drivers/staging/isdn/hysdn/hysdn_init.c card_last->next = card; /* pointer to next card */ next 129 drivers/staging/isdn/hysdn/hysdn_init.c card_root = card_root->next; next 135 drivers/staging/isdn/hysdn/hysdn_init.c if (tmp->next == card) next 136 drivers/staging/isdn/hysdn/hysdn_init.c tmp->next = card->next; next 138 drivers/staging/isdn/hysdn/hysdn_init.c tmp = tmp->next; next 381 drivers/staging/isdn/hysdn/hysdn_procconf.c card = card->next; /* next entry */ next 407 drivers/staging/isdn/hysdn/hysdn_procconf.c card = card->next; /* point to next card */ next 33 drivers/staging/isdn/hysdn/hysdn_proclog.c struct log_data *next; next 118 drivers/staging/isdn/hysdn/hysdn_proclog.c ib->next = NULL; next 125 drivers/staging/isdn/hysdn/hysdn_proclog.c pd->log_tail->next = ib; /* follows existing messages */ next 129 drivers/staging/isdn/hysdn/hysdn_proclog.c while (pd->log_head->next) { next 131 drivers/staging/isdn/hysdn/hysdn_proclog.c (pd->log_head->next->usage_cnt <= 0)) { next 133 drivers/staging/isdn/hysdn/hysdn_proclog.c pd->log_head = pd->log_head->next; next 188 drivers/staging/isdn/hysdn/hysdn_proclog.c file->private_data = &inf->next; /* next structure */ next 218 drivers/staging/isdn/hysdn/hysdn_proclog.c filep->private_data = &pd->log_tail->next; next 265 drivers/staging/isdn/hysdn/hysdn_proclog.c inf = inf->next; next 272 drivers/staging/isdn/hysdn/hysdn_proclog.c pd->log_head = pd->log_head->next; next 498 drivers/staging/kpc2000/kpc2000/cell_probe.c struct list_head *next; next 500 drivers/staging/kpc2000/kpc2000/cell_probe.c list_for_each_safe(ptr, next, &pcard->uio_devices_list) { next 89 drivers/staging/kpc2000/kpc_dma/dma.c struct kpc_dma_descriptor *next; next 121 drivers/staging/kpc2000/kpc_dma/dma.c next = dma_pool_alloc(eng->desc_pool, GFP_KERNEL | GFP_DMA, &next_handle); next 122 drivers/staging/kpc2000/kpc_dma/dma.c if (!next) next 125 drivers/staging/kpc2000/kpc_dma/dma.c clear_desc(next); next 126 drivers/staging/kpc2000/kpc_dma/dma.c next->MyDMAAddr = next_handle; next 129 drivers/staging/kpc2000/kpc_dma/dma.c cur->Next = next; next 130 drivers/staging/kpc2000/kpc_dma/dma.c cur = next; next 219 drivers/staging/kpc2000/kpc_dma/dma.c struct kpc_dma_descriptor *next = cur->Next; next 224 drivers/staging/kpc2000/kpc_dma/dma.c cur = next; next 201 drivers/staging/media/imx/imx-ic-prpencvf.c struct imx_media_buffer *done, *next; next 219 drivers/staging/media/imx/imx-ic-prpencvf.c next = imx_media_capture_device_next_buf(vdev); next 220 drivers/staging/media/imx/imx-ic-prpencvf.c if (next) { next 221 drivers/staging/media/imx/imx-ic-prpencvf.c phys = vb2_dma_contig_plane_dma_addr(&next->vbuf.vb2_buf, 0); next 222 drivers/staging/media/imx/imx-ic-prpencvf.c priv->active_vb2_buf[priv->ipu_buf_num] = next; next 704 drivers/staging/media/imx/imx-media-capture.c buf = list_entry(priv->ready_q.next, struct imx_media_buffer, next 265 drivers/staging/media/imx/imx-media-csi.c struct imx_media_buffer *done, *next; next 283 drivers/staging/media/imx/imx-media-csi.c next = imx_media_capture_device_next_buf(vdev); next 284 drivers/staging/media/imx/imx-media-csi.c if (next) { next 285 drivers/staging/media/imx/imx-media-csi.c phys = vb2_dma_contig_plane_dma_addr(&next->vbuf.vb2_buf, 0); next 286 drivers/staging/media/imx/imx-media-csi.c priv->active_vb2_buf[priv->ipu_buf_num] = next; next 628 drivers/staging/media/imx/imx7-media-csi.c struct imx_media_buffer *done, *next; next 643 drivers/staging/media/imx/imx7-media-csi.c next = imx_media_capture_device_next_buf(vdev); next 644 drivers/staging/media/imx/imx7-media-csi.c if (next) { next 645 drivers/staging/media/imx/imx7-media-csi.c phys = vb2_dma_contig_plane_dma_addr(&next->vbuf.vb2_buf, 0); next 646 drivers/staging/media/imx/imx7-media-csi.c csi->active_vb2_buf[csi->buf_num] = next; next 98 drivers/staging/media/ipu3/ipu3-css-fw.h struct imgu_fw_binary_xinfo *next __aligned(8); next 154 drivers/staging/media/ipu3/ipu3-css-fw.h u64 next; next 181 drivers/staging/most/dim2/dim2.c list_del(head->next); next 189 drivers/staging/most/dim2/dim2.c list_move_tail(head->next, &hdm_ch->started_list); next 288 drivers/staging/most/dim2/dim2.c list_del(head->next); next 421 drivers/staging/most/dim2/dim2.c list_del(head->next); next 2343 drivers/staging/octeon-usb/octeon-hcd.c struct cvmx_usb_transaction *transaction, *next; next 2346 drivers/staging/octeon-usb/octeon-hcd.c list_for_each_entry_safe(transaction, next, &pipe->transactions, node) { next 99 drivers/staging/octeon/ethernet-tx.c t->next = to_free_list; next 109 drivers/staging/octeon/ethernet-tx.c to_free_list = to_free_list->next; next 445 drivers/staging/octeon/ethernet-tx.c skb->next = to_free_list; next 462 drivers/staging/octeon/ethernet-tx.c t->next = to_free_list; next 473 drivers/staging/octeon/ethernet-tx.c to_free_list = to_free_list->next; next 1358 drivers/staging/qlge/qlge.h struct tx_ring_desc *next; next 178 drivers/staging/rtl8188eu/core/rtw_ap.c plist = phead->next; next 183 drivers/staging/rtl8188eu/core/rtw_ap.c plist = plist->next; next 211 drivers/staging/rtl8188eu/core/rtw_ap.c plist = phead->next; next 216 drivers/staging/rtl8188eu/core/rtw_ap.c plist = plist->next; next 1021 drivers/staging/rtl8188eu/core/rtw_ap.c plist = phead->next; next 1025 drivers/staging/rtl8188eu/core/rtw_ap.c plist = plist->next; next 1081 drivers/staging/rtl8188eu/core/rtw_ap.c plist = phead->next; next 1085 drivers/staging/rtl8188eu/core/rtw_ap.c plist = plist->next; next 1330 drivers/staging/rtl8188eu/core/rtw_ap.c plist = phead->next; next 1336 drivers/staging/rtl8188eu/core/rtw_ap.c plist = plist->next; next 1601 drivers/staging/rtl8188eu/core/rtw_ap.c plist = phead->next; next 1607 drivers/staging/rtl8188eu/core/rtw_ap.c plist = plist->next; next 1730 drivers/staging/rtl8188eu/core/rtw_ap.c plist = phead->next; next 1733 drivers/staging/rtl8188eu/core/rtw_ap.c plist = plist->next; next 26 drivers/staging/rtl8188eu/core/rtw_ioctl_set.c plist = phead->next; next 191 drivers/staging/rtl8188eu/core/rtw_mlme.c plist = phead->next; next 197 drivers/staging/rtl8188eu/core/rtw_mlme.c plist = plist->next; next 215 drivers/staging/rtl8188eu/core/rtw_mlme.c plist = phead->next; next 220 drivers/staging/rtl8188eu/core/rtw_mlme.c plist = plist->next; next 325 drivers/staging/rtl8188eu/core/rtw_mlme.c for (plist = phead->next; plist != phead; plist = plist->next) { next 403 drivers/staging/rtl8188eu/core/rtw_mlme.c plist = phead->next; next 413 drivers/staging/rtl8188eu/core/rtw_mlme.c plist = plist->next; next 696 drivers/staging/rtl8188eu/core/rtw_mlme.c plist = phead->next; next 699 drivers/staging/rtl8188eu/core/rtw_mlme.c ptemp = plist->next; next 1468 drivers/staging/rtl8188eu/core/rtw_mlme.c pmlmepriv->pscanned = phead->next; next 1476 drivers/staging/rtl8188eu/core/rtw_mlme.c pmlmepriv->pscanned = pmlmepriv->pscanned->next; next 1769 drivers/staging/rtl8188eu/core/rtw_mlme_ext.c plist = phead->next; next 1778 drivers/staging/rtl8188eu/core/rtw_mlme_ext.c plist = plist->next; next 5415 drivers/staging/rtl8188eu/core/rtw_mlme_ext.c xmitframe_plist = xmitframe_phead->next; next 5420 drivers/staging/rtl8188eu/core/rtw_mlme_ext.c xmitframe_plist = xmitframe_plist->next; next 181 drivers/staging/rtl8188eu/core/rtw_recv.c plist = phead->next; next 186 drivers/staging/rtl8188eu/core/rtw_recv.c plist = plist->next; next 928 drivers/staging/rtl8188eu/core/rtw_recv.c xmitframe_plist = xmitframe_phead->next; next 933 drivers/staging/rtl8188eu/core/rtw_recv.c xmitframe_plist = xmitframe_plist->next; next 1331 drivers/staging/rtl8188eu/core/rtw_recv.c plist = phead->next; next 1348 drivers/staging/rtl8188eu/core/rtw_recv.c plist = plist->next; next 1379 drivers/staging/rtl8188eu/core/rtw_recv.c plist = plist->next; next 1642 drivers/staging/rtl8188eu/core/rtw_recv.c plist = phead->next; next 1649 drivers/staging/rtl8188eu/core/rtw_recv.c plist = plist->next; next 1673 drivers/staging/rtl8188eu/core/rtw_recv.c plist = phead->next; next 1696 drivers/staging/rtl8188eu/core/rtw_recv.c plist = plist->next; next 142 drivers/staging/rtl8188eu/core/rtw_sta_mgt.c plist = phead->next; next 149 drivers/staging/rtl8188eu/core/rtw_sta_mgt.c plist = plist->next; next 315 drivers/staging/rtl8188eu/core/rtw_sta_mgt.c plist = phead->next; next 320 drivers/staging/rtl8188eu/core/rtw_sta_mgt.c plist = plist->next; next 391 drivers/staging/rtl8188eu/core/rtw_sta_mgt.c plist = phead->next; next 396 drivers/staging/rtl8188eu/core/rtw_sta_mgt.c plist = plist->next; next 427 drivers/staging/rtl8188eu/core/rtw_sta_mgt.c plist = phead->next; next 437 drivers/staging/rtl8188eu/core/rtw_sta_mgt.c plist = plist->next; next 487 drivers/staging/rtl8188eu/core/rtw_sta_mgt.c plist = phead->next; next 490 drivers/staging/rtl8188eu/core/rtw_sta_mgt.c plist = plist->next; next 1335 drivers/staging/rtl8188eu/core/rtw_xmit.c plist = phead->next; next 1340 drivers/staging/rtl8188eu/core/rtw_xmit.c plist = plist->next; next 1364 drivers/staging/rtl8188eu/core/rtw_xmit.c xmitframe_plist = xmitframe_phead->next; next 1369 drivers/staging/rtl8188eu/core/rtw_xmit.c xmitframe_plist = xmitframe_plist->next; next 1404 drivers/staging/rtl8188eu/core/rtw_xmit.c sta_plist = sta_phead->next; next 1422 drivers/staging/rtl8188eu/core/rtw_xmit.c sta_plist = sta_plist->next; next 1746 drivers/staging/rtl8188eu/core/rtw_xmit.c plist = phead->next; next 1751 drivers/staging/rtl8188eu/core/rtw_xmit.c plist = plist->next; next 1818 drivers/staging/rtl8188eu/core/rtw_xmit.c xmitframe_plist = xmitframe_phead->next; next 1823 drivers/staging/rtl8188eu/core/rtw_xmit.c xmitframe_plist = xmitframe_plist->next; next 1899 drivers/staging/rtl8188eu/core/rtw_xmit.c xmitframe_plist = xmitframe_phead->next; next 1904 drivers/staging/rtl8188eu/core/rtw_xmit.c xmitframe_plist = xmitframe_plist->next; next 1946 drivers/staging/rtl8188eu/core/rtw_xmit.c xmitframe_plist = xmitframe_phead->next; next 1951 drivers/staging/rtl8188eu/core/rtw_xmit.c xmitframe_plist = xmitframe_plist->next; next 512 drivers/staging/rtl8188eu/hal/rtl8188eu_xmit.c xmitframe_plist = xmitframe_phead->next; next 516 drivers/staging/rtl8188eu/hal/rtl8188eu_xmit.c xmitframe_plist = xmitframe_plist->next; next 982 drivers/staging/rtl8188eu/os_dep/ioctl_linux.c pmlmepriv->pscanned = phead->next; next 987 drivers/staging/rtl8188eu/os_dep/ioctl_linux.c pmlmepriv->pscanned = pmlmepriv->pscanned->next; next 1239 drivers/staging/rtl8188eu/os_dep/ioctl_linux.c plist = phead->next; next 1253 drivers/staging/rtl8188eu/os_dep/ioctl_linux.c plist = plist->next; next 1323 drivers/staging/rtl8188eu/os_dep/ioctl_linux.c pmlmepriv->pscanned = phead->next; next 1328 drivers/staging/rtl8188eu/os_dep/ioctl_linux.c pmlmepriv->pscanned = pmlmepriv->pscanned->next; next 123 drivers/staging/rtl8188eu/os_dep/xmit_linux.c plist = phead->next; next 129 drivers/staging/rtl8188eu/os_dep/xmit_linux.c plist = plist->next; next 258 drivers/staging/rtl8192e/rtl8192e/rtl_core.h struct tx_ring *next; next 354 drivers/staging/rtl8192e/rtl819x_TSProc.c (*ppTS) = list_entry(pUnusedList->next, next 454 drivers/staging/rtl8192e/rtllib_rx.c while (pList->next != &pTS->RxPendingPktList) { next 456 drivers/staging/rtl8192e/rtllib_rx.c list_entry(pList->next, struct rx_reorder_entry, next 458 drivers/staging/rtl8192e/rtllib_rx.c pList = pList->next; next 460 drivers/staging/rtl8192e/rtllib_rx.c ((struct rx_reorder_entry *)list_entry(pList->next, next 466 drivers/staging/rtl8192e/rtllib_rx.c pReorderEntry->List.next = pList->next; next 467 drivers/staging/rtl8192e/rtllib_rx.c pReorderEntry->List.next->prev = &pReorderEntry->List; next 469 drivers/staging/rtl8192e/rtllib_rx.c pList->next = &pReorderEntry->List; next 646 drivers/staging/rtl8192e/rtllib_rx.c list_entry(ieee->RxReorder_Unused_List.next, next 2672 drivers/staging/rtl8192e/rtllib_rx.c target = list_entry(ieee->network_free_list.next, next 2674 drivers/staging/rtl8192e/rtllib_rx.c list_del(ieee->network_free_list.next); next 40 drivers/staging/rtl8192u/ieee80211/ieee80211_crypt.c for (ptr = ieee->crypt_deinit_list.next, n = ptr->next; next 41 drivers/staging/rtl8192u/ieee80211/ieee80211_crypt.c ptr != &ieee->crypt_deinit_list; ptr = n, n = ptr->next) { next 132 drivers/staging/rtl8192u/ieee80211/ieee80211_crypt.c for (ptr = hcrypt->algs.next; ptr != &hcrypt->algs; ptr = ptr->next) { next 163 drivers/staging/rtl8192u/ieee80211/ieee80211_crypt.c for (ptr = hcrypt->algs.next; ptr != &hcrypt->algs; ptr = ptr->next) { next 224 drivers/staging/rtl8192u/ieee80211/ieee80211_crypt.c for (ptr = hcrypt->algs.next, n = ptr->next; ptr != &hcrypt->algs; next 225 drivers/staging/rtl8192u/ieee80211/ieee80211_crypt.c ptr = n, n = ptr->next) { next 508 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c while (pList->next != &pTS->rx_pending_pkt_list) { next 509 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c if (SN_LESS(pReorderEntry->SeqNum, list_entry(pList->next, struct rx_reorder_entry, List)->SeqNum)) next 510 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c pList = pList->next; next 511 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c else if (SN_EQUAL(pReorderEntry->SeqNum, list_entry(pList->next, struct rx_reorder_entry, List)->SeqNum)) next 516 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c pReorderEntry->List.next = pList->next; next 517 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c pReorderEntry->List.next->prev = &pReorderEntry->List; next 519 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c pList->next = &pReorderEntry->List; next 651 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c pReorderEntry = list_entry(ieee->RxReorder_Unused_List.next, struct rx_reorder_entry, List); next 2439 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c target = list_entry(ieee->network_free_list.next, next 2441 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c list_del(ieee->network_free_list.next); next 372 drivers/staging/rtl8192u/ieee80211/rtl819x_TSProc.c (*ppTS) = list_entry(pUnusedList->next, struct ts_common_info, list); next 459 drivers/staging/rtl8192u/r8192U.h struct buffer *next; next 191 drivers/staging/rtl8712/rtl8712_recv.c plist = phead->next; next 206 drivers/staging/rtl8712/rtl8712_recv.c plist = plist->next; next 232 drivers/staging/rtl8712/rtl8712_recv.c plist = plist->next; next 485 drivers/staging/rtl8712/rtl8712_recv.c plist = phead->next; next 490 drivers/staging/rtl8712/rtl8712_recv.c plist = plist->next; next 513 drivers/staging/rtl8712/rtl8712_recv.c plist = phead->next; next 530 drivers/staging/rtl8712/rtl8712_recv.c plist = plist->next; next 158 drivers/staging/rtl8712/rtl8712_xmit.c xmitframe_plist = xmitframe_phead->next; next 198 drivers/staging/rtl8712/rtl8712_xmit.c sta_plist = sta_phead->next; next 209 drivers/staging/rtl8712/rtl8712_xmit.c sta_plist = sta_plist->next; next 1055 drivers/staging/rtl8712/rtl871x_ioctl_linux.c pmlmepriv->pscanned = phead->next; next 1061 drivers/staging/rtl8712/rtl871x_ioctl_linux.c pmlmepriv->pscanned = pmlmepriv->pscanned->next; next 1207 drivers/staging/rtl8712/rtl871x_ioctl_linux.c plist = phead->next; next 1217 drivers/staging/rtl8712/rtl871x_ioctl_linux.c plist = plist->next; next 1266 drivers/staging/rtl8712/rtl871x_ioctl_linux.c pmlmepriv->pscanned = phead->next; next 1272 drivers/staging/rtl8712/rtl871x_ioctl_linux.c pmlmepriv->pscanned = pmlmepriv->pscanned->next; next 1975 drivers/staging/rtl8712/rtl871x_ioctl_linux.c plist = phead->next; next 2003 drivers/staging/rtl8712/rtl871x_ioctl_linux.c plist = plist->next; next 49 drivers/staging/rtl8712/rtl871x_ioctl_set.c plist = phead->next; next 143 drivers/staging/rtl8712/rtl871x_mlme.c plist = phead->next; next 146 drivers/staging/rtl8712/rtl871x_mlme.c plist = plist->next; next 164 drivers/staging/rtl8712/rtl871x_mlme.c plist = phead->next; next 167 drivers/staging/rtl8712/rtl871x_mlme.c plist = plist->next; next 265 drivers/staging/rtl8712/rtl871x_mlme.c plist = phead->next; next 276 drivers/staging/rtl8712/rtl871x_mlme.c plist = plist->next; next 344 drivers/staging/rtl8712/rtl871x_mlme.c plist = phead->next; next 358 drivers/staging/rtl8712/rtl871x_mlme.c plist = plist->next; next 1080 drivers/staging/rtl8712/rtl871x_mlme.c pmlmepriv->pscanned = phead->next; next 1092 drivers/staging/rtl8712/rtl871x_mlme.c pmlmepriv->pscanned = pmlmepriv->pscanned->next; next 128 drivers/staging/rtl8712/rtl871x_recv.c plist = phead->next; next 131 drivers/staging/rtl8712/rtl871x_recv.c plist = plist->next; next 73 drivers/staging/rtl8712/rtl871x_sta_mgt.c plist = phead->next; next 75 drivers/staging/rtl8712/rtl871x_sta_mgt.c plist = plist->next; next 205 drivers/staging/rtl8712/rtl871x_sta_mgt.c plist = phead->next; next 209 drivers/staging/rtl8712/rtl871x_sta_mgt.c plist = plist->next; next 230 drivers/staging/rtl8712/rtl871x_sta_mgt.c plist = phead->next; next 238 drivers/staging/rtl8712/rtl871x_sta_mgt.c plist = plist->next; next 845 drivers/staging/rtl8712/rtl871x_xmit.c plist = phead->next; next 848 drivers/staging/rtl8712/rtl871x_xmit.c plist = plist->next; next 426 drivers/staging/rtl8723bs/hal/rtl8723bs_xmit.c next: next 461 drivers/staging/rtl8723bs/hal/rtl8723bs_xmit.c goto next; next 468 drivers/staging/rtl8723bs/hal/rtl8723bs_xmit.c goto next; next 69 drivers/staging/rtl8723bs/include/osdep_service_linux.h return list->next; next 518 drivers/staging/unisys/include/iochannel.h struct uiscmdrsp *next; next 46 drivers/staging/unisys/visorhba/visorhba_main.c struct visordisk_info *next; next 304 drivers/staging/unisys/visornic/visornic_main.c skbinlist = skbinlist->next) { next 1296 drivers/staging/unisys/visornic/visornic_main.c curr->next = NULL; next 1301 drivers/staging/unisys/visornic/visornic_main.c prev->next = curr; next 164 drivers/staging/uwb/beacon.c struct uwb_beca_e *bce, *next; next 165 drivers/staging/uwb/beacon.c list_for_each_entry_safe(bce, next, &rc->uwb_beca.list, node) { next 179 drivers/staging/uwb/beacon.c struct uwb_beca_e *bce, *next; next 180 drivers/staging/uwb/beacon.c list_for_each_entry_safe(bce, next, &rc->uwb_beca.list, node) { next 274 drivers/staging/uwb/beacon.c struct uwb_beca_e *bce, *next; next 278 drivers/staging/uwb/beacon.c list_for_each_entry_safe(bce, next, &rc->uwb_beca.list, node) { next 290 drivers/staging/uwb/beacon.c struct uwb_beca_e *bce, *next; next 293 drivers/staging/uwb/beacon.c list_for_each_entry_safe(bce, next, &rc->uwb_beca.list, node) { next 131 drivers/staging/uwb/drp-ie.c int z, i, num_fields = 0, next = 0; next 147 drivers/staging/uwb/drp-ie.c for (i = 0; i < next; i++) { next 156 drivers/staging/uwb/drp-ie.c zones[next].zone_bm = 1 << z; next 157 drivers/staging/uwb/drp-ie.c zones[next].mas_bm = current_bmp; next 158 drivers/staging/uwb/drp-ie.c next++; next 28 drivers/staging/uwb/i1480/dfu/mac.c struct fw_hdr *next; next 36 drivers/staging/uwb/i1480/dfu/mac.c struct fw_hdr *next; next 39 drivers/staging/uwb/i1480/dfu/mac.c next = hdr->next; next 41 drivers/staging/uwb/i1480/dfu/mac.c hdr = next; next 61 drivers/staging/uwb/i1480/dfu/mac.c hdr->next = NULL; next 125 drivers/staging/uwb/i1480/dfu/mac.c prev_hdr = &hdr->next; next 208 drivers/staging/uwb/i1480/dfu/mac.c for (hdr_itr = hdr; hdr_itr != NULL; hdr_itr = hdr_itr->next) { next 265 drivers/staging/vc04_services/bcm2835-camera/mmal-msg.h u32 next; /* next header */ next 130 drivers/staging/vc04_services/interface/vchi/vchi_common.h void *next; next 253 drivers/staging/vc04_services/interface/vchiq_arm/vchiq_arm.c struct bulk_waiter_node *waiter, *next; next 255 drivers/staging/vc04_services/interface/vchiq_arm/vchiq_arm.c list_for_each_entry_safe(waiter, next, next 2048 drivers/staging/vc04_services/interface/vchiq_arm/vchiq_arm.c struct bulk_waiter_node *waiter, *next; next 2050 drivers/staging/vc04_services/interface/vchiq_arm/vchiq_arm.c list_for_each_entry_safe(waiter, next, next 191 drivers/staging/vt6655/desc.h struct vnt_rx_desc *next __aligned(8); next 234 drivers/staging/vt6655/desc.h struct vnt_tx_desc *next __aligned(8); next 549 drivers/staging/vt6655/device_main.c desc->next = &priv->aRD0Ring[(i + 1) % priv->opts.rx_descs0]; next 595 drivers/staging/vt6655/device_main.c desc->next = &priv->aRD1Ring[(i+1) % priv->opts.rx_descs1]; next 662 drivers/staging/vt6655/device_main.c desc->next = &(priv->apTD0Rings[(i+1) % priv->opts.tx_descs[0]]); next 702 drivers/staging/vt6655/device_main.c desc->next = &(priv->apTD1Rings[(i + 1) % priv->opts.tx_descs[1]]); next 756 drivers/staging/vt6655/device_main.c rd = rd->next) { next 900 drivers/staging/vt6655/device_main.c for (desc = priv->apTailTD[idx]; priv->iTDUsed[idx] > 0; desc = desc->next) { next 1176 drivers/staging/vt6655/device_main.c priv->apCurrTD[dma_idx] = head_td->next; next 282 drivers/staging/wlan-ng/hfa384x_usb.c return list_entry(hw->ctlxq.active.next, struct hfa384x_usbctlx, list); next 2835 drivers/staging/wlan-ng/hfa384x_usb.c head = list_entry(hw->ctlxq.pending.next, next 18 drivers/staging/wusbcore/host/whci/asl.c struct whc_qset **next, struct whc_qset **prev) next 24 drivers/staging/wusbcore/host/whci/asl.c n = qset->list_node.next; next 26 drivers/staging/wusbcore/host/whci/asl.c n = n->next; next 31 drivers/staging/wusbcore/host/whci/asl.c *next = container_of(n, struct whc_qset, list_node); next 44 drivers/staging/wusbcore/host/whci/asl.c struct whc_qset *next, *prev; next 49 drivers/staging/wusbcore/host/whci/asl.c qset_get_next_prev(whc, qset, &next, &prev); next 50 drivers/staging/wusbcore/host/whci/asl.c whc_qset_set_link_ptr(&qset->qh.link, next->qset_dma); next 57 drivers/staging/wusbcore/host/whci/asl.c struct whc_qset *prev, *next; next 59 drivers/staging/wusbcore/host/whci/asl.c qset_get_next_prev(whc, qset, &next, &prev); next 72 drivers/staging/wusbcore/host/whci/asl.c whc_qset_set_link_ptr(&prev->qh.link, next->qset_dma); next 1740 drivers/staging/wusbcore/wa-xfer.c struct wa_xfer *xfer, *next; next 1754 drivers/staging/wusbcore/wa-xfer.c list_for_each_entry_safe(xfer, next, &tmp_list, list_node) { next 1771 drivers/staging/wusbcore/wa-xfer.c struct wa_xfer *xfer, *next; next 1786 drivers/staging/wusbcore/wa-xfer.c list_for_each_entry_safe(xfer, next, &tmp_list, list_node) { next 120 drivers/target/iscsi/cxgbit/cxgbit.h struct np_info *next; next 90 drivers/target/iscsi/cxgbit/cxgbit_cm.c p->next = cdev->np_hash_tab[bucket]; next 105 drivers/target/iscsi/cxgbit/cxgbit_cm.c for (p = cdev->np_hash_tab[bucket]; p; p = p->next) { next 122 drivers/target/iscsi/cxgbit/cxgbit_cm.c for (p = *prev; p; prev = &p->next, p = p->next) { next 125 drivers/target/iscsi/cxgbit/cxgbit_cm.c *prev = p->next; next 638 drivers/target/iscsi/iscsi_target_tmr.c goto next; next 646 drivers/target/iscsi/iscsi_target_tmr.c goto next; next 688 drivers/target/iscsi/iscsi_target_tmr.c goto next; next 700 drivers/target/iscsi/iscsi_target_tmr.c next: next 1739 drivers/target/target_core_alua.c struct se_lun *lun, *next; next 1770 drivers/target/target_core_alua.c list_for_each_entry_safe(lun, next, next 703 drivers/target/target_core_pr.c struct se_lun *lun_tmp, *next, *dest_lun; next 726 drivers/target/target_core_pr.c list_for_each_entry_safe(lun_tmp, next, &dev->dev_sep_list, lun_dev_link) { next 2488 drivers/target/target_core_pr.c list_entry(dev->t10_pr.registration_list.next, next 1419 drivers/target/target_core_sbc.c goto next; next 1445 drivers/target/target_core_sbc.c next: next 137 drivers/target/target_core_spc.c bool next = true; next 153 drivers/target/target_core_spc.c if (next) { next 154 drivers/target/target_core_spc.c next = false; next 157 drivers/target/target_core_spc.c next = true; next 259 drivers/target/target_core_tmr.c struct se_cmd *cmd, *next; next 286 drivers/target/target_core_tmr.c list_for_each_entry_safe(cmd, next, &dev->state_list, state_list) { next 316 drivers/target/target_core_tmr.c cmd = list_entry(drain_task_list.next, struct se_cmd, state_list); next 2093 drivers/target/target_core_transport.c cmd = list_entry(dev->delayed_cmd_list.next, next 810 drivers/thermal/thermal_core.c struct thermal_instance *pos, *next; next 814 drivers/thermal/thermal_core.c list_for_each_entry_safe(pos, next, &tz->thermal_instances, tz_node) { next 85 drivers/thunderbolt/cap.c offset = header.basic.next; next 131 drivers/thunderbolt/cap.c offset = header.basic.next; next 171 drivers/thunderbolt/cap.c offset = header.extended_long.next; next 177 drivers/thunderbolt/cap.c offset = header.extended_short.next; next 768 drivers/thunderbolt/switch.c struct tb_port *next; next 782 drivers/thunderbolt/switch.c next = prev->remote; next 784 drivers/thunderbolt/switch.c next = tb_port_at(tb_route(end->sw), prev->sw); next 787 drivers/thunderbolt/switch.c next = prev->remote; next 789 drivers/thunderbolt/switch.c next = tb_upstream_port(prev->sw); next 794 drivers/thunderbolt/switch.c if (next->dual_link_port && next 795 drivers/thunderbolt/switch.c next->link_nr != prev->link_nr) { next 796 drivers/thunderbolt/switch.c next = next->dual_link_port; next 801 drivers/thunderbolt/switch.c return next; next 56 drivers/thunderbolt/tb_regs.h u8 next; next 70 drivers/thunderbolt/tb_regs.h u8 next; next 90 drivers/thunderbolt/tb_regs.h u16 next; next 94 drivers/tty/goldfish.c unsigned long next = next 96 drivers/tty/goldfish.c unsigned long avail = next - addr; next 97 drivers/tty/hvc/hvc_console.c list_for_each_entry(hp, &hvc_structs, next) { next 257 drivers/tty/hvc/hvc_console.c list_del(&(hp->next)); next 803 drivers/tty/hvc/hvc_console.c list_for_each_entry(hp, &hvc_structs, next) { next 969 drivers/tty/hvc/hvc_console.c list_add_tail(&(hp->next), &hvc_structs); next 49 drivers/tty/hvc/hvc_console.h struct list_head next; next 191 drivers/tty/hvc/hvc_iucv.c struct iucv_tty_buffer *ent, *next; next 193 drivers/tty/hvc/hvc_iucv.c list_for_each_entry_safe(ent, next, list, list) { next 955 drivers/tty/hvc/hvc_iucv.c struct iucv_tty_buffer *ent, *next; next 959 drivers/tty/hvc/hvc_iucv.c list_for_each_entry_safe(ent, next, &priv->tty_outqueue, list) next 287 drivers/tty/hvc/hvcs.c struct list_head next; /* list management */ next 641 drivers/tty/hvc/hvcs.c list_for_each_entry(hvcsd, &hvcs_structs, next) { next 694 drivers/tty/hvc/hvcs.c list_del(&(hvcsd->next)); next 805 drivers/tty/hvc/hvcs.c list_add_tail(&(hvcsd->next), &hvcs_structs); next 933 drivers/tty/hvc/hvcs.c list_for_each_entry(hvcsd, &hvcs_structs, next) { next 1071 drivers/tty/hvc/hvcs.c list_for_each_entry(hvcsd, &hvcs_structs, next) { next 240 drivers/tty/n_r3964.c pHeader->next = NULL; next 245 drivers/tty/n_r3964.c pInfo->tx_last->next = pHeader; next 271 drivers/tty/n_r3964.c for (pDump = pHeader; pDump; pDump = pDump->next) next 289 drivers/tty/n_r3964.c pInfo->tx_first = pHeader->next; next 310 drivers/tty/n_r3964.c pHeader->next = NULL; next 315 drivers/tty/n_r3964.c pInfo->rx_last->next = pHeader; next 345 drivers/tty/n_r3964.c pInfo->rx_first = pHeader->next; next 353 drivers/tty/n_r3964.c for (pFind = pInfo->rx_first; pFind; pFind = pFind->next) { next 354 drivers/tty/n_r3964.c if (pFind->next == pHeader) { next 356 drivers/tty/n_r3964.c pFind->next = pHeader->next; next 358 drivers/tty/n_r3964.c if (pFind->next == NULL) { next 535 drivers/tty/n_r3964.c pBlock->next = NULL; next 544 drivers/tty/n_r3964.c for (pClient = pInfo->firstClient; pClient; pClient = pClient->next) { next 747 drivers/tty/n_r3964.c for (pClient = pInfo->firstClient; pClient; pClient = pClient->next) { next 764 drivers/tty/n_r3964.c ppClient = &(*ppClient)->next) { next 770 drivers/tty/n_r3964.c *ppClient = pClient->next; next 803 drivers/tty/n_r3964.c pClient->next = pInfo->firstClient; next 866 drivers/tty/n_r3964.c pMsg->next = NULL; next 871 drivers/tty/n_r3964.c pClient->last_msg->next = pMsg; next 911 drivers/tty/n_r3964.c pClient->first_msg = pMsg->next; next 1027 drivers/tty/n_r3964.c pNext = pClient->next; next 1047 drivers/tty/n_r3964.c pNextHeader = pHeader->next; next 132 drivers/tty/serial/8250/8250_core.c l = l->next; next 158 drivers/tty/serial/8250/8250_core.c i->head = i->head->next; next 29 drivers/tty/serial/8250/8250_hp300.c struct hp300_port *next; /* next port */ next 273 drivers/tty/serial/8250/8250_hp300.c port->next = hp300_ports; next 309 drivers/tty/serial/8250/8250_hp300.c port = port->next; next 70 drivers/tty/serial/ip22zilog.c struct uart_ip22zilog_port *next; next 462 drivers/tty/serial/ip22zilog.c up = up->next; next 484 drivers/tty/serial/ip22zilog.c up = up->next; next 712 drivers/tty/serial/ip22zilog.c up->next->flags |= IP22ZILOG_FLAG_RESET_DONE; next 1096 drivers/tty/serial/ip22zilog.c up[channel].next = &up[channel - 1]; next 1097 drivers/tty/serial/ip22zilog.c up[channel].next = NULL; next 180 drivers/tty/serial/kgdboc.c cons = cons->next; next 1090 drivers/tty/serial/pch_uart.c int next = 1; next 1095 drivers/tty/serial/pch_uart.c while (next) { next 1137 drivers/tty/serial/pch_uart.c next = 0; /* MS ir prioirty is the lowest. So, MS ir next 1147 drivers/tty/serial/pch_uart.c next = 0; next 78 drivers/tty/serial/sunzilog.c struct uart_sunzilog_port *next; next 562 drivers/tty/serial/sunzilog.c up = up->next; next 584 drivers/tty/serial/sunzilog.c up = up->next; next 1107 drivers/tty/serial/sunzilog.c up->next = up + 1; next 1109 drivers/tty/serial/sunzilog.c up->next = NULL; next 1597 drivers/tty/serial/sunzilog.c up = up->next; next 1634 drivers/tty/serial/sunzilog.c up = up->next; next 184 drivers/tty/synclink_gt.c struct cond_wait *next; next 202 drivers/tty/synclink_gt.c __le32 next; /* physical address of next descriptor */ next 212 drivers/tty/synclink_gt.c #define set_desc_next(a,b) (a).next = cpu_to_le32((unsigned int)(b)) next 2316 drivers/tty/synclink_gt.c for (w = info->gpio_wait_q, prev = NULL ; w != NULL ; w = w->next) { next 2321 drivers/tty/synclink_gt.c prev->next = w->next; next 2323 drivers/tty/synclink_gt.c info->gpio_wait_q = w->next; next 3001 drivers/tty/synclink_gt.c w->next = *head; next 3010 drivers/tty/synclink_gt.c for (w = *head, prev = NULL ; w != NULL ; prev = w, w = w->next) { next 3013 drivers/tty/synclink_gt.c prev->next = w->next; next 3015 drivers/tty/synclink_gt.c *head = w->next; next 3025 drivers/tty/synclink_gt.c *head = (*head)->next; next 3364 drivers/tty/synclink_gt.c info->rbufs[i].next = cpu_to_le32(pbufs); next 3366 drivers/tty/synclink_gt.c info->rbufs[i].next = cpu_to_le32(pbufs + ((i+1) * sizeof(struct slgt_desc))); next 3376 drivers/tty/synclink_gt.c info->tbufs[i].next = cpu_to_le32(pbufs + info->rbuf_count * sizeof(struct slgt_desc)); next 3378 drivers/tty/synclink_gt.c info->tbufs[i].next = cpu_to_le32(pbufs + ((info->rbuf_count + i + 1) * sizeof(struct slgt_desc))); next 113 drivers/tty/synclinkmp.c u16 next; /* lower l6 bits of next descriptor addr */ next 3447 drivers/tty/synclinkmp.c info->rx_buf_list[i].next = info->buffer_list_phys; next 3449 drivers/tty/synclinkmp.c info->rx_buf_list[i].next += (i + 1) * sizeof(SCADESC); next 3462 drivers/tty/synclinkmp.c info->tx_buf_list[i].next = info->buffer_list_phys + next 3466 drivers/tty/synclinkmp.c info->tx_buf_list[i].next += (i + 1) * sizeof(SCADESC); next 102 drivers/tty/tty_buffer.c p->next = NULL; next 119 drivers/tty/tty_buffer.c struct tty_buffer *p, *next; next 125 drivers/tty/tty_buffer.c buf->head = p->next; next 131 drivers/tty/tty_buffer.c llist_for_each_entry_safe(p, next, llist, free) next 223 drivers/tty/tty_buffer.c struct tty_buffer *next; next 231 drivers/tty/tty_buffer.c while ((next = smp_load_acquire(&buf->head->next)) != NULL) { next 233 drivers/tty/tty_buffer.c buf->head = next; next 285 drivers/tty/tty_buffer.c smp_store_release(&b->next, n); next 509 drivers/tty/tty_buffer.c struct tty_buffer *next; next 520 drivers/tty/tty_buffer.c next = smp_load_acquire(&head->next); next 526 drivers/tty/tty_buffer.c if (next == NULL) next 528 drivers/tty/tty_buffer.c buf->head = next; next 243 drivers/tty/tty_ldisc.c .next = tty_ldiscs_seq_next, next 76 drivers/tty/tty_ldsem.c struct ldsem_waiter *waiter, *next; next 94 drivers/tty/tty_ldsem.c list_for_each_entry_safe(waiter, next, &sem->read_wait, list) { next 123 drivers/tty/tty_ldsem.c waiter = list_entry(sem->write_wait.next, struct ldsem_waiter, list); next 206 drivers/usb/atm/usbatm.c urb = list_entry(channel->list.next, struct urb, urb_list); next 787 drivers/usb/c67x00/c67x00-sched.c urb = list_entry(ep_data->queue.next, next 894 drivers/usb/c67x00/c67x00-sched.c while (td->td_list.next != &c67x00->td_list) { next 895 drivers/usb/c67x00/c67x00-sched.c td = list_entry(td->td_list.next, struct c67x00_td, td_list); next 1033 drivers/usb/c67x00/c67x00-sched.c tmp = list_entry(td->td_list.next, typeof(*td), td_list); next 1073 drivers/usb/c67x00/c67x00-sched.c if (td->td_list.next == &c67x00->td_list) next 34 drivers/usb/chipidea/otg_fsm.c char *next; next 38 drivers/usb/chipidea/otg_fsm.c next = buf; next 40 drivers/usb/chipidea/otg_fsm.c t = scnprintf(next, size, "%d\n", ci->fsm.a_bus_req); next 42 drivers/usb/chipidea/otg_fsm.c next += t; next 83 drivers/usb/chipidea/otg_fsm.c char *next; next 87 drivers/usb/chipidea/otg_fsm.c next = buf; next 89 drivers/usb/chipidea/otg_fsm.c t = scnprintf(next, size, "%d\n", ci->fsm.a_bus_drop); next 91 drivers/usb/chipidea/otg_fsm.c next += t; next 123 drivers/usb/chipidea/otg_fsm.c char *next; next 127 drivers/usb/chipidea/otg_fsm.c next = buf; next 129 drivers/usb/chipidea/otg_fsm.c t = scnprintf(next, size, "%d\n", ci->fsm.b_bus_req); next 131 drivers/usb/chipidea/otg_fsm.c next += t; next 385 drivers/usb/chipidea/udc.c lastnode->ptr->next = cpu_to_le32(node->dma); next 464 drivers/usb/chipidea/udc.c lastnode->ptr->next = cpu_to_le32(TD_TERMINATE); next 475 drivers/usb/chipidea/udc.c u32 next = firstnode->dma & TD_ADDR_MASK; next 482 drivers/usb/chipidea/udc.c prevlastnode->ptr->next = cpu_to_le32(next); next 496 drivers/usb/chipidea/udc.c hwep->qh.ptr->td.next = cpu_to_le32(firstnode->dma); next 531 drivers/usb/chipidea/udc.c hwep->qh.ptr->td.next = cpu_to_le32(node->dma); next 636 drivers/usb/chipidea/udc.c struct ci_hw_req *hwreq = list_entry(hwep->qh.queue.next, next 1270 drivers/usb/chipidea/udc.c hwep->qh.ptr->td.next |= cpu_to_le32(TD_TERMINATE); /* needed? */ next 22 drivers/usb/chipidea/udc.h __le32 next; next 444 drivers/usb/core/devio.c as = list_entry(ps->async_completed.next, struct async, next 648 drivers/usb/core/devio.c as = list_entry(list->next, struct async, asynclist); next 2760 drivers/usb/core/devio.c ps = list_entry(udev->filelist.next, struct usb_dev_state, list); next 1677 drivers/usb/core/hcd.c urb = list_entry(local_list.next, struct urb, urb_list); next 782 drivers/usb/core/hub.c struct list_head *next; next 788 drivers/usb/core/hub.c next = hub->tt.clear_list.next; next 789 drivers/usb/core/hub.c clear = list_entry(next, struct usb_tt_clear, clear_list); next 949 drivers/usb/core/urb.c victim = list_entry(anchor->urb_list.next, struct urb, next 259 drivers/usb/dwc2/debugfs.c ep->queue.next, ep->queue.prev); next 4684 drivers/usb/dwc2/gadget.c u32 next = DXEPCTL_NEXTEP((epnum + 1) % 15); next 4687 drivers/usb/dwc2/gadget.c dwc2_writel(hsotg, next, DIEPCTL(epnum)); next 4689 drivers/usb/dwc2/gadget.c dwc2_writel(hsotg, next, DOEPCTL(epnum)); next 1626 drivers/usb/dwc2/hcd.c if (!qh_list->next) next 2728 drivers/usb/dwc2/hcd.c qh_ptr = hsotg->periodic_sched_ready.next; next 2745 drivers/usb/dwc2/hcd.c qh_ptr = qh_ptr->next; next 2757 drivers/usb/dwc2/hcd.c qh_ptr = hsotg->non_periodic_sched_inactive.next; next 2779 drivers/usb/dwc2/hcd.c qh_ptr = qh_ptr->next; next 2910 drivers/usb/dwc2/hcd.c qh_ptr = hsotg->periodic_sched_assigned.next; next 2922 drivers/usb/dwc2/hcd.c qh_ptr = qh_ptr->next; next 2928 drivers/usb/dwc2/hcd.c qh_ptr = qh_ptr->next; next 2958 drivers/usb/dwc2/hcd.c qh_ptr = qh_ptr->next; next 3042 drivers/usb/dwc2/hcd.c hsotg->non_periodic_qh_ptr = hsotg->non_periodic_qh_ptr->next; next 3061 drivers/usb/dwc2/hcd.c goto next; next 3065 drivers/usb/dwc2/hcd.c goto next; next 3077 drivers/usb/dwc2/hcd.c next: next 3079 drivers/usb/dwc2/hcd.c hsotg->non_periodic_qh_ptr = hsotg->non_periodic_qh_ptr->next; next 3083 drivers/usb/dwc2/hcd.c hsotg->non_periodic_qh_ptr->next; next 158 drivers/usb/dwc2/hcd_intr.c qh_entry = hsotg->periodic_sched_inactive.next; next 161 drivers/usb/dwc2/hcd_intr.c qh_entry = qh_entry->next; next 1789 drivers/usb/dwc2/hcd_queue.c hsotg->non_periodic_qh_ptr->next; next 3006 drivers/usb/dwc3/gadget.c enum dwc3_link_state next = evtinfo & DWC3_LINK_STATE_MASK; next 3030 drivers/usb/dwc3/gadget.c (next == DWC3_LINK_STATE_RESUME)) { next 3054 drivers/usb/dwc3/gadget.c if (next == DWC3_LINK_STATE_U0) { next 3081 drivers/usb/dwc3/gadget.c switch (next) { next 3098 drivers/usb/dwc3/gadget.c dwc->link_state = next; next 3104 drivers/usb/dwc3/gadget.c enum dwc3_link_state next = evtinfo & DWC3_LINK_STATE_MASK; next 3106 drivers/usb/dwc3/gadget.c if (dwc->link_state != next && next == DWC3_LINK_STATE_U3) next 3109 drivers/usb/dwc3/gadget.c dwc->link_state = next; next 454 drivers/usb/gadget/composite.c void *next = buf + USB_DT_CONFIG_SIZE; next 473 drivers/usb/gadget/composite.c status = usb_descriptor_fillbuf(next, len, next 478 drivers/usb/gadget/composite.c next += status; next 488 drivers/usb/gadget/composite.c status = usb_descriptor_fillbuf(next, len, next 493 drivers/usb/gadget/composite.c next += status; next 496 drivers/usb/gadget/composite.c len = next - buf; next 530 drivers/usb/gadget/composite.c while ((pos = pos->next) != &cdev->configs) { next 1221 drivers/usb/gadget/composite.c int next = cdev->next_string_id; next 1224 drivers/usb/gadget/composite.c if (unlikely(next >= 254)) next 1226 drivers/usb/gadget/composite.c str->id = ++next; next 1229 drivers/usb/gadget/composite.c cdev->next_string_id = next; next 1369 drivers/usb/gadget/composite.c unsigned next = c->next_string_id; next 1370 drivers/usb/gadget/composite.c if (unlikely(n > 254 || (unsigned)next + n > 254)) next 1373 drivers/usb/gadget/composite.c return next + 1; next 439 drivers/usb/gadget/function/f_eem.c goto next; next 494 drivers/usb/gadget/function/f_eem.c goto next; next 515 drivers/usb/gadget/function/f_eem.c next: next 611 drivers/usb/gadget/function/f_hid.c struct f_hidg_req_list *list, *next; next 618 drivers/usb/gadget/function/f_hid.c list_for_each_entry_safe(list, next, &hidg->completed_out_req, list) { next 738 drivers/usb/gadget/function/f_mass_storage.c common->next_buffhd_to_fill = bh->next; next 843 drivers/usb/gadget/function/f_mass_storage.c common->next_buffhd_to_fill = bh->next; next 857 drivers/usb/gadget/function/f_mass_storage.c common->next_buffhd_to_drain = bh->next; next 1471 drivers/usb/gadget/function/f_mass_storage.c common->next_buffhd_to_fill = bh2->next; next 1483 drivers/usb/gadget/function/f_mass_storage.c common->next_buffhd_to_drain = bh->next; next 1536 drivers/usb/gadget/function/f_mass_storage.c common->next_buffhd_to_fill = bh->next; next 1549 drivers/usb/gadget/function/f_mass_storage.c common->next_buffhd_to_fill = bh->next; next 1646 drivers/usb/gadget/function/f_mass_storage.c common->next_buffhd_to_fill = bh->next; next 2620 drivers/usb/gadget/function/f_mass_storage.c bh->next = bh + 1; next 2627 drivers/usb/gadget/function/f_mass_storage.c bh->next = buffhds; next 382 drivers/usb/gadget/function/f_printer.c req = container_of(dev->rx_reqs.next, next 478 drivers/usb/gadget/function/f_printer.c req = container_of(dev->rx_buffers.next, next 594 drivers/usb/gadget/function/f_printer.c req = container_of(dev->tx_reqs.next, struct usb_request, next 859 drivers/usb/gadget/function/f_printer.c req = container_of(dev->rx_buffers.next, struct usb_request, next 866 drivers/usb/gadget/function/f_printer.c req = container_of(dev->rx_buffers.next, struct usb_request, next 873 drivers/usb/gadget/function/f_printer.c req = container_of(dev->tx_reqs_active.next, next 1107 drivers/usb/gadget/function/f_printer.c req = container_of(dev->rx_reqs.next, struct usb_request, list); next 1114 drivers/usb/gadget/function/f_printer.c req = container_of(dev->tx_reqs.next, struct usb_request, list); next 1378 drivers/usb/gadget/function/f_printer.c req = container_of(dev->tx_reqs.next, struct usb_request, next 1388 drivers/usb/gadget/function/f_printer.c req = container_of(dev->rx_reqs.next, next 1395 drivers/usb/gadget/function/f_printer.c req = container_of(dev->rx_buffers.next, next 146 drivers/usb/gadget/function/storage_common.h struct fsg_buffhd *next; next 370 drivers/usb/gadget/function/u_ether.c struct list_head *next; next 372 drivers/usb/gadget/function/u_ether.c next = req->list.next; next 376 drivers/usb/gadget/function/u_ether.c if (next == list) next 379 drivers/usb/gadget/function/u_ether.c req = container_of(next, struct usb_request, list); next 246 drivers/usb/gadget/function/u_serial.c req = list_entry(pool->next, struct usb_request, list); next 319 drivers/usb/gadget/function/u_serial.c req = list_entry(pool->next, struct usb_request, list); next 493 drivers/usb/gadget/function/u_serial.c req = list_entry(head->next, struct usb_request, list); next 472 drivers/usb/gadget/udc/amd5536udc.h u32 next; next 464 drivers/usb/gadget/udc/at91_udc.c req = list_entry(ep->queue.next, struct at91_request, queue); next 1010 drivers/usb/gadget/udc/at91_udc.c req = list_entry(ep->queue.next, next 1297 drivers/usb/gadget/udc/at91_udc.c req = list_entry(ep0->queue.next, struct at91_request, queue); next 471 drivers/usb/gadget/udc/atmel_usba_udc.c req = list_entry(ep->queue.next, struct usba_request, queue); next 501 drivers/usb/gadget/udc/atmel_usba_udc.c req = list_entry(ep->queue.next, next 892 drivers/usb/gadget/udc/atmel_usba_udc.c if (ep->queue.next == &req->queue) { next 1443 drivers/usb/gadget/udc/atmel_usba_udc.c req = list_entry(ep->queue.next, next 1636 drivers/usb/gadget/udc/atmel_usba_udc.c req = list_entry(ep->queue.next, struct usba_request, queue); next 1700 drivers/usb/gadget/udc/atmel_usba_udc.c req = list_entry(ep->queue.next, struct usba_request, queue); next 257 drivers/usb/gadget/udc/atmel_usba_udc.h dma_addr_t next; next 1227 drivers/usb/gadget/udc/bcm63xx_udc.c struct bcm63xx_req *next; next 1229 drivers/usb/gadget/udc/bcm63xx_udc.c next = list_first_entry(&bep->queue, next 1231 drivers/usb/gadget/udc/bcm63xx_udc.c iudma_write(udc, bep->iudma, next); next 2094 drivers/usb/gadget/udc/bcm63xx_udc.c struct bcm63xx_req *next; next 2096 drivers/usb/gadget/udc/bcm63xx_udc.c next = list_first_entry(&bep->queue, next 2098 drivers/usb/gadget/udc/bcm63xx_udc.c iudma_write(udc, iudma, next); next 576 drivers/usb/gadget/udc/bdc/bdc_ep.c req = list_entry(ep->queue.next, struct bdc_req, next 976 drivers/usb/gadget/udc/bdc/bdc_ep.c req = list_entry(ep->queue.next, struct bdc_req, next 1097 drivers/usb/gadget/udc/bdc/bdc_ep.c req = list_entry(ep->queue.next, struct bdc_req, next 334 drivers/usb/gadget/udc/dummy_hcd.c req = list_entry(ep->queue.next, struct dummy_request, queue); next 220 drivers/usb/gadget/udc/fotg210-udc.c req = list_entry(ep->queue.next, next 817 drivers/usb/gadget/udc/fotg210-udc.c req = list_entry(ep->queue.next, next 840 drivers/usb/gadget/udc/fotg210-udc.c struct fotg210_request *req = list_entry(ep->queue.next, next 850 drivers/usb/gadget/udc/fotg210-udc.c struct fotg210_request *req = list_entry(ep->queue.next, next 132 drivers/usb/gadget/udc/fsl_qe_udc.c req = list_entry(ep->queue.next, struct qe_req, queue); next 903 drivers/usb/gadget/udc/fsl_qe_udc.c req = list_entry(ep->queue.next, struct qe_req, queue); next 1171 drivers/usb/gadget/udc/fsl_qe_udc.c ep->tx_req = list_entry(ep->queue.next, struct qe_req, next 218 drivers/usb/gadget/udc/fsl_udc_core.c req = list_entry(ep->queue.next, struct fsl_req, queue); next 957 drivers/usb/gadget/udc/fsl_udc_core.c if (ep->queue.next == &req->queue) { next 962 drivers/usb/gadget/udc/fsl_udc_core.c if (req->queue.next != &ep->queue) { next 965 drivers/usb/gadget/udc/fsl_udc_core.c next_req = list_entry(req->queue.next, struct fsl_req, next 259 drivers/usb/gadget/udc/fusb300_udc.c req = list_entry(ep->queue.next, struct fusb300_request, queue); next 961 drivers/usb/gadget/udc/fusb300_udc.c struct fusb300_request *req = list_entry(ep->queue.next, next 972 drivers/usb/gadget/udc/fusb300_udc.c struct fusb300_request *req = list_entry(ep->queue.next, next 1033 drivers/usb/gadget/udc/fusb300_udc.c req = list_entry(ep->queue.next, next 489 drivers/usb/gadget/udc/goku_udc.c req = list_entry(ep->queue.next, next 524 drivers/usb/gadget/udc/goku_udc.c req = list_entry(ep->queue.next, struct goku_request, queue); next 604 drivers/usb/gadget/udc/goku_udc.c req = list_entry(ep->queue.next, struct goku_request, queue); next 631 drivers/usb/gadget/udc/goku_udc.c req = list_entry(ep->queue.next, struct goku_request, queue); next 649 drivers/usb/gadget/udc/goku_udc.c req = list_entry(ep->queue.next, struct goku_request, queue); next 801 drivers/usb/gadget/udc/goku_udc.c req = list_entry(ep->queue.next, struct goku_request, queue); next 842 drivers/usb/gadget/udc/goku_udc.c if (ep->dma && ep->queue.next == &req->queue && !ep->stopped) { next 870 drivers/usb/gadget/udc/goku_udc.c req = list_entry(ep->queue.next, struct goku_request, next 165 drivers/usb/gadget/udc/gr_udc.c struct gr_dma_desc *next; next 170 drivers/usb/gadget/udc/gr_udc.c next = req->first_desc; next 172 drivers/usb/gadget/udc/gr_udc.c desc = next; next 173 drivers/usb/gadget/udc/gr_udc.c next = desc->next_desc; next 257 drivers/usb/gadget/udc/gr_udc.c struct gr_dma_desc *next; next 259 drivers/usb/gadget/udc/gr_udc.c next = req->first_desc; next 260 drivers/usb/gadget/udc/gr_udc.c if (!next) next 264 drivers/usb/gadget/udc/gr_udc.c desc = next; next 265 drivers/usb/gadget/udc/gr_udc.c next = desc->next_desc; next 457 drivers/usb/gadget/udc/gr_udc.c req->last_desc->next = desc->paddr; next 115 drivers/usb/gadget/udc/gr_udc.h u32 next; next 995 drivers/usb/gadget/udc/lpc32xx_udc.c req = list_entry(ep->queue.next, struct lpc32xx_request, queue); next 1021 drivers/usb/gadget/udc/lpc32xx_udc.c req = list_entry(ep->queue.next, struct lpc32xx_request, queue); next 1441 drivers/usb/gadget/udc/lpc32xx_udc.c req = list_entry(ep->queue.next, struct lpc32xx_request, queue); next 1462 drivers/usb/gadget/udc/lpc32xx_udc.c req = list_entry(ep0->queue.next, struct lpc32xx_request, next 1499 drivers/usb/gadget/udc/lpc32xx_udc.c req = list_entry(ep0->queue.next, struct lpc32xx_request, next 1973 drivers/usb/gadget/udc/lpc32xx_udc.c req = list_entry(ep->queue.next, struct lpc32xx_request, queue); next 2000 drivers/usb/gadget/udc/lpc32xx_udc.c req = list_entry(ep->queue.next, struct lpc32xx_request, queue); next 733 drivers/usb/gadget/udc/m66592-udc.c req = list_entry(ep->queue.next, struct m66592_request, queue); next 907 drivers/usb/gadget/udc/m66592-udc.c req = list_entry(ep->queue.next, struct m66592_request, queue); next 915 drivers/usb/gadget/udc/m66592-udc.c req = list_entry(ep->queue.next, next 938 drivers/usb/gadget/udc/m66592-udc.c req = list_entry(ep->queue.next, struct m66592_request, queue); next 951 drivers/usb/gadget/udc/m66592-udc.c req = list_entry(ep->queue.next, next 1020 drivers/usb/gadget/udc/m66592-udc.c req = list_entry(ep->queue.next, next 1169 drivers/usb/gadget/udc/m66592-udc.c req = list_entry(ep->queue.next, struct m66592_request, queue); next 1318 drivers/usb/gadget/udc/m66592-udc.c req = list_entry(ep->queue.next, struct m66592_request, queue); next 130 drivers/usb/gadget/udc/mv_u3d_core.c curr_trb = list_entry(curr_req->trb_list.next, next 456 drivers/usb/gadget/udc/mv_u3d_core.c req->trb_head = list_entry(req->trb_list.next, next 483 drivers/usb/gadget/udc/mv_u3d_core.c req = list_entry(ep->req_list.next, struct mv_u3d_req, list); next 875 drivers/usb/gadget/udc/mv_u3d_core.c if (ep->queue.next == &req->queue) { next 880 drivers/usb/gadget/udc/mv_u3d_core.c if (req->queue.next != &ep->queue) { next 884 drivers/usb/gadget/udc/mv_u3d_core.c next_req = list_entry(req->queue.next, next 905 drivers/usb/gadget/udc/mv_u3d_core.c curr_req = list_entry(ep->req_list.next, next 1370 drivers/usb/gadget/udc/mv_u3d_core.c req = list_entry(ep->queue.next, struct mv_u3d_req, queue); next 1645 drivers/usb/gadget/udc/mv_u3d_core.c req = list_entry(curr_ep->req_list.next, next 807 drivers/usb/gadget/udc/mv_udc_core.c if (ep->queue.next == &req->queue) { next 812 drivers/usb/gadget/udc/mv_udc_core.c if (req->queue.next != &ep->queue) { next 815 drivers/usb/gadget/udc/mv_udc_core.c next_req = list_entry(req->queue.next, next 1305 drivers/usb/gadget/udc/mv_udc_core.c req = list_entry(ep->queue.next, struct mv_req, queue); next 466 drivers/usb/gadget/udc/net2272.c req = list_entry(ep->queue.next, next 595 drivers/usb/gadget/udc/net2272.c req = list_entry(ep->queue.next, next 617 drivers/usb/gadget/udc/net2272.c req = list_entry(ep->queue.next, struct net2272_request, queue); next 917 drivers/usb/gadget/udc/net2272.c req = list_entry(ep->queue.next, next 953 drivers/usb/gadget/udc/net2272.c if (ep->queue.next == &req->queue) { next 1169 drivers/usb/gadget/udc/net2272.c char *next; next 1177 drivers/usb/gadget/udc/net2272.c next = buf; next 1182 drivers/usb/gadget/udc/net2272.c t = scnprintf(next, size, "%s version %s," next 1193 drivers/usb/gadget/udc/net2272.c next += t; next 1197 drivers/usb/gadget/udc/net2272.c t = scnprintf(next, size, "\ndmareq %02x: %s %s%s%s%s\n", next 1204 drivers/usb/gadget/udc/net2272.c next += t; next 1217 drivers/usb/gadget/udc/net2272.c t = scnprintf(next, size, next 1222 drivers/usb/gadget/udc/net2272.c next += t; next 1234 drivers/usb/gadget/udc/net2272.c t = scnprintf(next, size, next 1248 drivers/usb/gadget/udc/net2272.c next += t; next 1250 drivers/usb/gadget/udc/net2272.c t = scnprintf(next, size, next 1261 drivers/usb/gadget/udc/net2272.c next += t; next 1263 drivers/usb/gadget/udc/net2272.c t = scnprintf(next, size, next 1269 drivers/usb/gadget/udc/net2272.c next += t; next 1273 drivers/usb/gadget/udc/net2272.c t = scnprintf(next, size, next 1278 drivers/usb/gadget/udc/net2272.c next += t; next 1514 drivers/usb/gadget/udc/net2272.c req = list_entry(ep->queue.next, next 1547 drivers/usb/gadget/udc/net2272.c req = list_entry(ep->queue.next, next 1591 drivers/usb/gadget/udc/net2272.c req = list_entry(ep->queue.next, next 1734 drivers/usb/gadget/udc/net2272.c req = list_entry(ep->queue.next, next 1144 drivers/usb/gadget/udc/net2280.c req = list_entry(ep->queue.next, next 1206 drivers/usb/gadget/udc/net2280.c req = list_entry(ep->queue.next, struct net2280_request, queue); next 1233 drivers/usb/gadget/udc/net2280.c req = list_entry(ep->queue.next, next 1282 drivers/usb/gadget/udc/net2280.c if (ep->queue.next == &req->queue) { next 1287 drivers/usb/gadget/udc/net2280.c if (likely(ep->queue.next == &req->queue)) { next 1314 drivers/usb/gadget/udc/net2280.c start_dma(ep, list_entry(ep->queue.next, next 1658 drivers/usb/gadget/udc/net2280.c char *next; next 1666 drivers/usb/gadget/udc/net2280.c next = buf; next 1676 drivers/usb/gadget/udc/net2280.c t = scnprintf(next, size, "%s version " DRIVER_VERSION next 1690 drivers/usb/gadget/udc/net2280.c next += t; next 1705 drivers/usb/gadget/udc/net2280.c t = scnprintf(next, size, next 1711 drivers/usb/gadget/udc/net2280.c next += t; next 1727 drivers/usb/gadget/udc/net2280.c t = scnprintf(next, size, next 1749 drivers/usb/gadget/udc/net2280.c next += t; next 1751 drivers/usb/gadget/udc/net2280.c t = scnprintf(next, size, next 1760 drivers/usb/gadget/udc/net2280.c next += t; next 1765 drivers/usb/gadget/udc/net2280.c t = scnprintf(next, size, next 1774 drivers/usb/gadget/udc/net2280.c next += t; next 1781 drivers/usb/gadget/udc/net2280.c t = scnprintf(next, size, "\nirqs: "); next 1783 drivers/usb/gadget/udc/net2280.c next += t; next 1790 drivers/usb/gadget/udc/net2280.c t = scnprintf(next, size, " %s/%lu", ep->ep.name, ep->irqs); next 1792 drivers/usb/gadget/udc/net2280.c next += t; next 1795 drivers/usb/gadget/udc/net2280.c t = scnprintf(next, size, "\n"); next 1797 drivers/usb/gadget/udc/net2280.c next += t; next 1809 drivers/usb/gadget/udc/net2280.c char *next; next 1815 drivers/usb/gadget/udc/net2280.c next = buf; next 1831 drivers/usb/gadget/udc/net2280.c t = scnprintf(next, size, next 1840 drivers/usb/gadget/udc/net2280.c t = scnprintf(next, size, "ep0 max 64 pio %s\n", next 1845 drivers/usb/gadget/udc/net2280.c next += t; next 1848 drivers/usb/gadget/udc/net2280.c t = scnprintf(next, size, "\t(nothing queued)\n"); next 1852 drivers/usb/gadget/udc/net2280.c next += t; next 1857 drivers/usb/gadget/udc/net2280.c t = scnprintf(next, size, next 1864 drivers/usb/gadget/udc/net2280.c t = scnprintf(next, size, next 1871 drivers/usb/gadget/udc/net2280.c next += t; next 1877 drivers/usb/gadget/udc/net2280.c t = scnprintf(next, size, "\t td %08x " next 1886 drivers/usb/gadget/udc/net2280.c next += t; next 2520 drivers/usb/gadget/udc/net2280.c req = list_entry(ep->queue.next, next 2620 drivers/usb/gadget/udc/net2280.c req = list_entry(ep->queue.next, next 2732 drivers/usb/gadget/udc/net2280.c req = list_entry(ep->queue.next, next 3135 drivers/usb/gadget/udc/net2280.c req = list_entry(ep->queue.next, next 645 drivers/usb/gadget/udc/omap_udc.c req = container_of(ep->queue.next, next 652 drivers/usb/gadget/udc/omap_udc.c req = container_of(ep->queue.next, next 664 drivers/usb/gadget/udc/omap_udc.c req = container_of(ep->queue.next, next 671 drivers/usb/gadget/udc/omap_udc.c req = container_of(ep->queue.next, next 788 drivers/usb/gadget/udc/omap_udc.c req = container_of(ep->queue.next, struct omap_req, queue); next 813 drivers/usb/gadget/udc/omap_udc.c req = container_of(ep->queue.next, struct omap_req, queue); next 1024 drivers/usb/gadget/udc/omap_udc.c if (use_dma && ep->dma_channel && ep->queue.next == &req->queue) { next 1342 drivers/usb/gadget/udc/omap_udc.c req = list_entry(ep->queue.next, struct omap_req, queue); next 1418 drivers/usb/gadget/udc/omap_udc.c req = container_of(ep0->queue.next, struct omap_req, queue); next 1873 drivers/usb/gadget/udc/omap_udc.c req = container_of(ep->queue.next, next 1914 drivers/usb/gadget/udc/omap_udc.c req = container_of(ep->queue.next, next 1948 drivers/usb/gadget/udc/omap_udc.c req = container_of(ep->queue.next, next 1981 drivers/usb/gadget/udc/omap_udc.c req = list_entry(ep->queue.next, struct omap_req, queue); next 224 drivers/usb/gadget/udc/pch_udc.c u32 next; next 1494 drivers/usb/gadget/udc/pch_udc.c req = list_entry(ep->queue.next, struct pch_udc_request, queue); next 1515 drivers/usb/gadget/udc/pch_udc.c dma_addr_t addr = (dma_addr_t)td->next; next 1516 drivers/usb/gadget/udc/pch_udc.c td->next = 0x00; next 1520 drivers/usb/gadget/udc/pch_udc.c addr2 = (dma_addr_t)td->next; next 1569 drivers/usb/gadget/udc/pch_udc.c last->next = dma_addr; next 1574 drivers/usb/gadget/udc/pch_udc.c td->next = req->td_data_phys; next 1660 drivers/usb/gadget/udc/pch_udc.c td_data = phys_to_virt(td_data->next); next 2095 drivers/usb/gadget/udc/pch_udc.c req = list_entry(ep->queue.next, struct pch_udc_request, queue); next 2109 drivers/usb/gadget/udc/pch_udc.c td_data = phys_to_virt(td_data->next); next 2129 drivers/usb/gadget/udc/pch_udc.c req = list_entry(ep->queue.next, struct pch_udc_request, queue); next 2173 drivers/usb/gadget/udc/pch_udc.c req = list_entry(ep->queue.next, struct pch_udc_request, queue); next 2199 drivers/usb/gadget/udc/pch_udc.c addr = (dma_addr_t)td->next; next 2213 drivers/usb/gadget/udc/pch_udc.c req = list_entry(ep->queue.next, struct pch_udc_request, queue); next 2280 drivers/usb/gadget/udc/pch_udc.c req = list_entry(ep->queue.next, struct pch_udc_request, next 2522 drivers/usb/gadget/udc/pch_udc.c ep->td_data->next = ep->td_data_phys; next 955 drivers/usb/gadget/udc/pxa25x_udc.c req = list_entry(ep->queue.next, next 1642 drivers/usb/gadget/udc/pxa25x_udc.c req = list_entry(ep->queue.next, struct pxa25x_request, queue); next 1862 drivers/usb/gadget/udc/pxa25x_udc.c req = list_entry(ep->queue.next, next 776 drivers/usb/gadget/udc/pxa27x_udc.c req = list_entry(ep->queue.next, struct pxa27x_request, queue); next 1942 drivers/usb/gadget/udc/pxa27x_udc.c req = list_entry(ep->queue.next, struct pxa27x_request, queue); next 2038 drivers/usb/gadget/udc/pxa27x_udc.c req = list_entry(ep->queue.next, next 894 drivers/usb/gadget/udc/r8a66597-udc.c return list_entry(ep->queue.next, struct r8a66597_request, queue); next 273 drivers/usb/gadget/udc/s3c-hsudc.c hsreq = list_entry(hsep->queue.next, next 444 drivers/usb/gadget/udc/s3c-hsudc.c hsreq = list_entry(hsep->queue.next, next 481 drivers/usb/gadget/udc/s3c-hsudc.c hsreq = list_entry(hsep->queue.next, next 526 drivers/usb/gadget/udc/s3c-hsudc.c hsreq = list_entry(hsep->queue.next, next 718 drivers/usb/gadget/udc/s3c-hsudc.c hsreq = list_entry(hsep->queue.next, next 731 drivers/usb/gadget/udc/s3c-hsudc.c hsreq = list_entry(hsep->queue.next, next 260 drivers/usb/gadget/udc/s3c2410_udc.c req = list_entry(ep->queue.next, struct s3c2410_request, next 733 drivers/usb/gadget/udc/s3c2410_udc.c req = list_entry(ep->queue.next, struct s3c2410_request, queue); next 802 drivers/usb/gadget/udc/s3c2410_udc.c req = list_entry(ep->queue.next, next 572 drivers/usb/gadget/udc/snps_udc_core.c dma_addr_t addr = (dma_addr_t)td->next; next 579 drivers/usb/gadget/udc/snps_udc_core.c addr_next = (dma_addr_t)td->next; next 619 drivers/usb/gadget/udc/snps_udc_core.c req->td_data->next = req->td_phys; next 810 drivers/usb/gadget/udc/snps_udc_core.c req->td_data->next); next 813 drivers/usb/gadget/udc/snps_udc_core.c td = (struct udc_data_dma *)phys_to_virt(last->next); next 833 drivers/usb/gadget/udc/snps_udc_core.c req->td_data->next = dma_addr; next 852 drivers/usb/gadget/udc/snps_udc_core.c last->next = dma_addr; next 1006 drivers/usb/gadget/udc/snps_udc_core.c td = phys_to_virt(td->next); next 1023 drivers/usb/gadget/udc/snps_udc_core.c td = phys_to_virt(td->next); next 1247 drivers/usb/gadget/udc/snps_udc_core.c req = list_entry(ep->queue.next, next 1274 drivers/usb/gadget/udc/snps_udc_core.c if (ep->queue.next == &req->queue) { next 2125 drivers/usb/gadget/udc/snps_udc_core.c req = list_entry(ep->queue.next, next 2143 drivers/usb/gadget/udc/snps_udc_core.c req = list_entry(ep->queue.next, next 2220 drivers/usb/gadget/udc/snps_udc_core.c req = list_entry(ep->queue.next, next 2341 drivers/usb/gadget/udc/snps_udc_core.c req = list_entry(ep->queue.next, next 2381 drivers/usb/gadget/udc/snps_udc_core.c req = list_entry(ep->queue.next, next 2707 drivers/usb/gadget/udc/snps_udc_core.c req = list_entry(ep->queue.next, next 379 drivers/usb/host/ehci-dbg.c char *next = *nextp; next 398 drivers/usb/host/ehci-dbg.c temp = scnprintf(next, size, next 413 drivers/usb/host/ehci-dbg.c next += temp; next 446 drivers/usb/host/ehci-dbg.c temp = scnprintf(next, size, next 456 drivers/usb/host/ehci-dbg.c next += temp; next 461 drivers/usb/host/ehci-dbg.c temp = scnprintf(next, size, "\n"); next 463 drivers/usb/host/ehci-dbg.c next += temp; next 467 drivers/usb/host/ehci-dbg.c *nextp = next; next 476 drivers/usb/host/ehci-dbg.c char *next; next 481 drivers/usb/host/ehci-dbg.c next = buf->output_buf; next 484 drivers/usb/host/ehci-dbg.c *next = 0; next 493 drivers/usb/host/ehci-dbg.c qh_lines(ehci, qh, &next, &size); next 495 drivers/usb/host/ehci-dbg.c temp = scnprintf(next, size, "\nunlink =\n"); next 497 drivers/usb/host/ehci-dbg.c next += temp; next 502 drivers/usb/host/ehci-dbg.c qh_lines(ehci, qh, &next, &size); next 516 drivers/usb/host/ehci-dbg.c char *next; next 523 drivers/usb/host/ehci-dbg.c next = buf->output_buf; next 526 drivers/usb/host/ehci-dbg.c *next = 0; next 531 drivers/usb/host/ehci-dbg.c temp = scnprintf(next, size, next 534 drivers/usb/host/ehci-dbg.c next += temp; next 537 drivers/usb/host/ehci-dbg.c temp = scnprintf(next, size, next 542 drivers/usb/host/ehci-dbg.c next += temp; next 547 drivers/usb/host/ehci-dbg.c temp = scnprintf(next, size, next 552 drivers/usb/host/ehci-dbg.c next += temp; next 555 drivers/usb/host/ehci-dbg.c temp = scnprintf(next, size, next 560 drivers/usb/host/ehci-dbg.c next += temp; next 562 drivers/usb/host/ehci-dbg.c temp = scnprintf(next, size, next 565 drivers/usb/host/ehci-dbg.c next += temp; next 569 drivers/usb/host/ehci-dbg.c temp = scnprintf(next, size, next 574 drivers/usb/host/ehci-dbg.c next += temp; next 577 drivers/usb/host/ehci-dbg.c temp = scnprintf(next, size, next 585 drivers/usb/host/ehci-dbg.c next += temp; next 590 drivers/usb/host/ehci-dbg.c return next - buf->output_buf; next 628 drivers/usb/host/ehci-dbg.c char *next; next 639 drivers/usb/host/ehci-dbg.c next = buf->output_buf; next 642 drivers/usb/host/ehci-dbg.c temp = scnprintf(next, size, "size = %d\n", ehci->periodic_size); next 644 drivers/usb/host/ehci-dbg.c next += temp; next 657 drivers/usb/host/ehci-dbg.c temp = scnprintf(next, size, "%4d: ", i); next 659 drivers/usb/host/ehci-dbg.c next += temp; next 667 drivers/usb/host/ehci-dbg.c temp = scnprintf(next, size, " qh%d-%04x/%p", next 675 drivers/usb/host/ehci-dbg.c next += temp; next 681 drivers/usb/host/ehci-dbg.c temp = scnprintf(next, size, next 684 drivers/usb/host/ehci-dbg.c next += temp; next 690 drivers/usb/host/ehci-dbg.c temp = output_buf_tds_dir(next, ehci, next 702 drivers/usb/host/ehci-dbg.c temp = scnprintf(next, size, next 709 drivers/usb/host/ehci-dbg.c temp = scnprintf(next, size, next 715 drivers/usb/host/ehci-dbg.c temp = scnprintf(next, size, next 726 drivers/usb/host/ehci-dbg.c next += temp; next 729 drivers/usb/host/ehci-dbg.c temp = scnprintf(next, size, "\n"); next 731 drivers/usb/host/ehci-dbg.c next += temp; next 761 drivers/usb/host/ehci-dbg.c char *next, scratch[80]; next 767 drivers/usb/host/ehci-dbg.c next = buf->output_buf; next 773 drivers/usb/host/ehci-dbg.c size = scnprintf(next, size, next 785 drivers/usb/host/ehci-dbg.c temp = scnprintf(next, size, next 794 drivers/usb/host/ehci-dbg.c next += temp; next 810 drivers/usb/host/ehci-dbg.c temp = scnprintf(next, size, next 815 drivers/usb/host/ehci-dbg.c next += temp; next 819 drivers/usb/host/ehci-dbg.c temp = scnprintf(next, size, next 822 drivers/usb/host/ehci-dbg.c next += temp; next 837 drivers/usb/host/ehci-dbg.c temp = scnprintf(next, size, "structural params 0x%08x\n", i); next 839 drivers/usb/host/ehci-dbg.c next += temp; next 842 drivers/usb/host/ehci-dbg.c temp = scnprintf(next, size, "capability params 0x%08x\n", i); next 844 drivers/usb/host/ehci-dbg.c next += temp; next 849 drivers/usb/host/ehci-dbg.c temp = scnprintf(next, size, fmt, temp, scratch); next 851 drivers/usb/host/ehci-dbg.c next += temp; next 855 drivers/usb/host/ehci-dbg.c temp = scnprintf(next, size, fmt, temp, scratch); next 857 drivers/usb/host/ehci-dbg.c next += temp; next 861 drivers/usb/host/ehci-dbg.c temp = scnprintf(next, size, fmt, temp, scratch); next 863 drivers/usb/host/ehci-dbg.c next += temp; next 865 drivers/usb/host/ehci-dbg.c temp = scnprintf(next, size, "uframe %04x\n", next 868 drivers/usb/host/ehci-dbg.c next += temp; next 874 drivers/usb/host/ehci-dbg.c temp = scnprintf(next, size, fmt, temp, scratch); next 876 drivers/usb/host/ehci-dbg.c next += temp; next 878 drivers/usb/host/ehci-dbg.c temp = scnprintf(next, size, next 883 drivers/usb/host/ehci-dbg.c next += temp; next 888 drivers/usb/host/ehci-dbg.c temp = scnprintf(next, size, "async unlink qh %p\n", next 892 drivers/usb/host/ehci-dbg.c next += temp; next 896 drivers/usb/host/ehci-dbg.c temp = scnprintf(next, size, next 901 drivers/usb/host/ehci-dbg.c next += temp; next 903 drivers/usb/host/ehci-dbg.c temp = scnprintf(next, size, "complete %ld unlink %ld\n", next 906 drivers/usb/host/ehci-dbg.c next += temp; next 117 drivers/usb/host/ehci-q.c qtd = list_entry(qh->qtd_list.next, struct ehci_qtd, qtd_list); next 443 drivers/usb/host/ehci-q.c qh->qtd_list.next == &qtd->qtd_list && next 1043 drivers/usb/host/ehci-q.c qtd = list_entry (qtd_list->next, struct ehci_qtd, next 1118 drivers/usb/host/ehci-q.c qtd = list_entry(qtd_list->next, struct ehci_qtd, qtd_list); next 643 drivers/usb/host/ehci-sched.c ehci->qh_scan_next = list_entry(qh->intr_node.next, next 690 drivers/usb/host/ehci-sched.c else if (ehci->intr_unlink.next == &qh->unlink_node) { next 711 drivers/usb/host/ehci-sched.c else if (ehci->intr_unlink_wait.next == &qh->unlink_node) { next 1486 drivers/usb/host/ehci-sched.c u32 now, base, next, start, period, span, now2; next 1521 drivers/usb/host/ehci-sched.c next = start; next 1536 drivers/usb/host/ehci-sched.c } while (start > next && !done); next 1565 drivers/usb/host/ehci-sched.c next = now + ehci->i_thresh; /* uframe cache */ next 1567 drivers/usb/host/ehci-sched.c next = (now + 2 + 7) & ~0x07; /* full frame cache */ next 1578 drivers/usb/host/ehci-sched.c next = (next - base) & (mod - 1); next 1604 drivers/usb/host/ehci-sched.c if (unlikely(start < next && next 1645 drivers/usb/host/ehci-sched.c start = next + ((start - next) & (period - 1)); next 1784 drivers/usb/host/ehci-sched.c itd = list_entry(iso_sched->td_list.next, next 2187 drivers/usb/host/ehci-sched.c sitd = list_entry(sched->td_list.next, next 42 drivers/usb/host/fhci-mem.c td = list_entry(fhci->empty_tds.next, struct td, node); next 43 drivers/usb/host/fhci-mem.c list_del(fhci->empty_tds.next); next 66 drivers/usb/host/fhci-mem.c ed = list_entry(fhci->empty_eds.next, struct ed, node); next 67 drivers/usb/host/fhci-mem.c list_del(fhci->empty_eds.next); next 75 drivers/usb/host/fhci-q.c td = list_entry(ed->td_list.next, struct td, node); next 87 drivers/usb/host/fhci-q.c td = list_entry(frame->tds_list.next, struct td, frame_lh); next 88 drivers/usb/host/fhci-q.c list_del_init(frame->tds_list.next); next 100 drivers/usb/host/fhci-q.c td = list_entry(frame->tds_list.next, struct td, frame_lh); next 112 drivers/usb/host/fhci-q.c td = list_entry(ed->td_list.next, struct td, node); next 113 drivers/usb/host/fhci-q.c list_del_init(ed->td_list.next); next 117 drivers/usb/host/fhci-q.c ed->td_head = list_entry(ed->td_list.next, struct td, next 132 drivers/usb/host/fhci-q.c td = list_entry(p_list->done_list.next, struct td, node); next 133 drivers/usb/host/fhci-q.c list_del_init(p_list->done_list.next); next 149 drivers/usb/host/fhci-q.c ed->td_head = list_entry(ed->td_list.next, struct td, node); next 174 drivers/usb/host/fhci-q.c ed->td_head = list_entry(ed->td_list.next, struct td, node); next 63 drivers/usb/host/fhci-sched.c if (ed->td_list.next->next != &ed->td_list) { next 65 drivers/usb/host/fhci-sched.c list_entry(ed->td_list.next->next, struct td, next 257 drivers/usb/host/fhci-sched.c struct list_head *node = list->next; next 361 drivers/usb/host/fotg210-hcd.c char *next = *nextp; next 380 drivers/usb/host/fotg210-hcd.c temp = scnprintf(next, size, next 391 drivers/usb/host/fotg210-hcd.c next += temp; next 407 drivers/usb/host/fotg210-hcd.c temp = snprintf(next, size, next 430 drivers/usb/host/fotg210-hcd.c next += temp; next 435 drivers/usb/host/fotg210-hcd.c temp = snprintf(next, size, "\n"); next 440 drivers/usb/host/fotg210-hcd.c next += temp; next 444 drivers/usb/host/fotg210-hcd.c *nextp = next; next 453 drivers/usb/host/fotg210-hcd.c char *next; next 458 drivers/usb/host/fotg210-hcd.c next = buf->output_buf; next 461 drivers/usb/host/fotg210-hcd.c *next = 0; next 470 drivers/usb/host/fotg210-hcd.c qh_lines(fotg210, qh, &next, &size); next 472 drivers/usb/host/fotg210-hcd.c temp = scnprintf(next, size, "\nunlink =\n"); next 474 drivers/usb/host/fotg210-hcd.c next += temp; next 478 drivers/usb/host/fotg210-hcd.c qh_lines(fotg210, qh, &next, &size); next 521 drivers/usb/host/fotg210-hcd.c char *next; next 533 drivers/usb/host/fotg210-hcd.c next = buf->output_buf; next 536 drivers/usb/host/fotg210-hcd.c temp = scnprintf(next, size, "size = %d\n", fotg210->periodic_size); next 538 drivers/usb/host/fotg210-hcd.c next += temp; next 551 drivers/usb/host/fotg210-hcd.c temp = scnprintf(next, size, "%4d: ", i); next 553 drivers/usb/host/fotg210-hcd.c next += temp; next 561 drivers/usb/host/fotg210-hcd.c temp = scnprintf(next, size, " qh%d-%04x/%p", next 569 drivers/usb/host/fotg210-hcd.c next += temp; next 575 drivers/usb/host/fotg210-hcd.c temp = scnprintf(next, size, next 578 drivers/usb/host/fotg210-hcd.c next += temp; next 584 drivers/usb/host/fotg210-hcd.c temp = output_buf_tds_dir(next, next 596 drivers/usb/host/fotg210-hcd.c temp = scnprintf(next, size, next 603 drivers/usb/host/fotg210-hcd.c temp = scnprintf(next, size, next 610 drivers/usb/host/fotg210-hcd.c next += temp; next 613 drivers/usb/host/fotg210-hcd.c temp = scnprintf(next, size, "\n"); next 615 drivers/usb/host/fotg210-hcd.c next += temp; next 645 drivers/usb/host/fotg210-hcd.c char *next, scratch[80]; next 651 drivers/usb/host/fotg210-hcd.c next = buf->output_buf; next 657 drivers/usb/host/fotg210-hcd.c size = scnprintf(next, size, next 670 drivers/usb/host/fotg210-hcd.c temp = scnprintf(next, size, next 679 drivers/usb/host/fotg210-hcd.c next += temp; next 683 drivers/usb/host/fotg210-hcd.c temp = scnprintf(next, size, "structural params 0x%08x\n", i); next 685 drivers/usb/host/fotg210-hcd.c next += temp; next 688 drivers/usb/host/fotg210-hcd.c temp = scnprintf(next, size, "capability params 0x%08x\n", i); next 690 drivers/usb/host/fotg210-hcd.c next += temp; next 695 drivers/usb/host/fotg210-hcd.c temp = scnprintf(next, size, fmt, temp, scratch); next 697 drivers/usb/host/fotg210-hcd.c next += temp; next 701 drivers/usb/host/fotg210-hcd.c temp = scnprintf(next, size, fmt, temp, scratch); next 703 drivers/usb/host/fotg210-hcd.c next += temp; next 707 drivers/usb/host/fotg210-hcd.c temp = scnprintf(next, size, fmt, temp, scratch); next 709 drivers/usb/host/fotg210-hcd.c next += temp; next 711 drivers/usb/host/fotg210-hcd.c temp = scnprintf(next, size, "uframe %04x\n", next 714 drivers/usb/host/fotg210-hcd.c next += temp; next 717 drivers/usb/host/fotg210-hcd.c temp = scnprintf(next, size, "async unlink qh %p\n", next 720 drivers/usb/host/fotg210-hcd.c next += temp; next 724 drivers/usb/host/fotg210-hcd.c temp = scnprintf(next, size, next 729 drivers/usb/host/fotg210-hcd.c next += temp; next 731 drivers/usb/host/fotg210-hcd.c temp = scnprintf(next, size, "complete %ld unlink %ld\n", next 734 drivers/usb/host/fotg210-hcd.c next += temp; next 2075 drivers/usb/host/fotg210-hcd.c qtd = list_entry(qh->qtd_list.next, next 2966 drivers/usb/host/fotg210-hcd.c qtd = list_entry(qtd_list->next, struct fotg210_qtd, next 3035 drivers/usb/host/fotg210-hcd.c qtd = list_entry(qtd_list->next, struct fotg210_qtd, qtd_list); next 3164 drivers/usb/host/fotg210-hcd.c struct fotg210_qh *qh, *next; next 3169 drivers/usb/host/fotg210-hcd.c next = fotg210->async->qh_next.qh; next 3170 drivers/usb/host/fotg210-hcd.c while (next) { next 3171 drivers/usb/host/fotg210-hcd.c qh = next; next 3172 drivers/usb/host/fotg210-hcd.c next = qh->qh_next.qh; next 3595 drivers/usb/host/fotg210-hcd.c fotg210->qh_scan_next = list_entry(qh->intr_node.next, next 4168 drivers/usb/host/fotg210-hcd.c u32 now, next, start, period, span; next 4198 drivers/usb/host/fotg210-hcd.c next = now + fotg210->i_thresh; next 4200 drivers/usb/host/fotg210-hcd.c next = now; next 4206 drivers/usb/host/fotg210-hcd.c excess = (stream->next_uframe - period - next) & (mod - 1); next 4208 drivers/usb/host/fotg210-hcd.c start = next + excess - mod + period * next 4211 drivers/usb/host/fotg210-hcd.c start = next + excess + period; next 4239 drivers/usb/host/fotg210-hcd.c next = start; next 4247 drivers/usb/host/fotg210-hcd.c } while (start > next && !done); next 4383 drivers/usb/host/fotg210-hcd.c itd = list_entry(iso_sched->td_list.next, next 577 drivers/usb/host/imx21-hcd.c td = list_entry(ep_priv->td_list.next, struct td, list); next 242 drivers/usb/host/isp116x-hcd.c urb = container_of(ep->hep->urb_list.next, next 329 drivers/usb/host/isp116x-hcd.c prev = &temp->next; next 331 drivers/usb/host/isp116x-hcd.c *prev = ep->next; next 361 drivers/usb/host/isp116x-hcd.c container_of(ep->hep->urb_list.next, struct urb, urb_list); next 510 drivers/usb/host/isp116x-hcd.c while (last_ep->next) next 511 drivers/usb/host/isp116x-hcd.c last_ep = (last_ep->active = last_ep->next); next 518 drivers/usb/host/isp116x-hcd.c urb = container_of(ep->hep->urb_list.next, next 565 drivers/usb/host/isp116x-hcd.c if ((&isp116x->async)->next != (&isp116x->async)->prev) next 566 drivers/usb/host/isp116x-hcd.c list_move(&isp116x->async, (&isp116x->async)->next); next 803 drivers/usb/host/isp116x-hcd.c prev = &here->next; next 807 drivers/usb/host/isp116x-hcd.c ep->next = here; next 856 drivers/usb/host/isp116x-hcd.c if (ep->hep->urb_list.next == &urb->urb_list) next 321 drivers/usb/host/isp116x.h struct isp116x_ep *next; next 781 drivers/usb/host/isp1362-hcd.c if (isp1362_hcd->async.next != isp1362_hcd->async.prev) { next 783 drivers/usb/host/isp1362-hcd.c list_move(&isp1362_hcd->async, isp1362_hcd->async.next); next 914 drivers/usb/host/isp1362-hcd.c if (epq->active.next == &ep->active) next 1422 drivers/usb/host/isp1362-hcd.c if (ep->hep->urb_list.next == &urb->urb_list) { next 578 drivers/usb/host/isp1362.h container_of(ep->hep->urb_list.next, struct urb, urb_list); \ next 128 drivers/usb/host/max3421-hcd.c struct max3421_hcd *next; next 806 drivers/usb/host/max3421-hcd.c struct urb *urb, *next; next 813 drivers/usb/host/max3421-hcd.c list_for_each_entry_safe(urb, next, &ep->urb_list, urb_list) { next 1900 drivers/usb/host/max3421-hcd.c max3421_hcd->next = max3421_hcd_list; next 1956 drivers/usb/host/max3421-hcd.c for (prev = &max3421_hcd_list; *prev; prev = &(*prev)->next) { next 1973 drivers/usb/host/max3421-hcd.c *prev = max3421_hcd->next; next 23 drivers/usb/host/ohci-dbg.c #define ohci_dbg_sw(ohci, next, size, format, arg...) \ next 25 drivers/usb/host/ohci-dbg.c if (next != NULL) { \ next 27 drivers/usb/host/ohci-dbg.c s_len = scnprintf (*next, *size, format, ## arg ); \ next 28 drivers/usb/host/ohci-dbg.c *size -= s_len; *next += s_len; \ next 34 drivers/usb/host/ohci-dbg.c #define ohci_dbg_nosw(ohci, next, size, format, arg...) \ next 37 drivers/usb/host/ohci-dbg.c s_len = scnprintf(*next, *size, format, ## arg); \ next 38 drivers/usb/host/ohci-dbg.c *size -= s_len; *next += s_len; \ next 46 drivers/usb/host/ohci-dbg.c char **next, next 49 drivers/usb/host/ohci-dbg.c ohci_dbg_sw (ohci, next, size, "%s 0x%08x%s%s%s%s%s%s%s%s%s\n", next 68 drivers/usb/host/ohci-dbg.c char **next, next 72 drivers/usb/host/ohci-dbg.c ohci_dbg_sw (ohci, next, size, "%s %08x\n", label, value); next 101 drivers/usb/host/ohci-dbg.c ohci_dump_status (struct ohci_hcd *controller, char **next, unsigned *size) next 107 drivers/usb/host/ohci-dbg.c ohci_dbg_sw (controller, next, size, next 114 drivers/usb/host/ohci-dbg.c ohci_dbg_sw (controller, next, size, next 129 drivers/usb/host/ohci-dbg.c ohci_dbg_sw (controller, next, size, next 140 drivers/usb/host/ohci-dbg.c next, size); next 143 drivers/usb/host/ohci-dbg.c next, size); next 148 drivers/usb/host/ohci-dbg.c next, size); next 152 drivers/usb/host/ohci-dbg.c next, size); next 155 drivers/usb/host/ohci-dbg.c next, size); next 159 drivers/usb/host/ohci-dbg.c next, size); next 162 drivers/usb/host/ohci-dbg.c next, size); next 165 drivers/usb/host/ohci-dbg.c ohci_readl (controller, ®s->donehead), next, size); next 168 drivers/usb/host/ohci-dbg.c #define dbg_port_sw(hc,num,value,next,size) \ next 169 drivers/usb/host/ohci-dbg.c ohci_dbg_sw (hc, next, size, \ next 194 drivers/usb/host/ohci-dbg.c char **next, next 204 drivers/usb/host/ohci-dbg.c ohci_dbg_sw (controller, next, size, next 215 drivers/usb/host/ohci-dbg.c ohci_dbg_sw (controller, next, size, next 222 drivers/usb/host/ohci-dbg.c ohci_dbg_sw (controller, next, size, next 236 drivers/usb/host/ohci-dbg.c dbg_port_sw (controller, i, temp, next, size); next 492 drivers/usb/host/ohci-dbg.c char *next; next 501 drivers/usb/host/ohci-dbg.c next = buf->page; next 504 drivers/usb/host/ohci-dbg.c temp = scnprintf (next, size, "size = %d\n", NUM_INTS); next 506 drivers/usb/host/ohci-dbg.c next += temp; next 515 drivers/usb/host/ohci-dbg.c temp = scnprintf (next, size, "%2d [%3d]:", i, ohci->load [i]); next 517 drivers/usb/host/ohci-dbg.c next += temp; next 520 drivers/usb/host/ohci-dbg.c temp = scnprintf (next, size, " ed%d/%p", next 523 drivers/usb/host/ohci-dbg.c next += temp; next 539 drivers/usb/host/ohci-dbg.c temp = scnprintf (next, size, next 555 drivers/usb/host/ohci-dbg.c next += temp; next 570 drivers/usb/host/ohci-dbg.c temp = scnprintf (next, size, "\n"); next 572 drivers/usb/host/ohci-dbg.c next += temp; next 588 drivers/usb/host/ohci-dbg.c char *next; next 594 drivers/usb/host/ohci-dbg.c next = buf->page; next 601 drivers/usb/host/ohci-dbg.c ohci_dbg_nosw(ohci, &next, &size, next 611 drivers/usb/host/ohci-dbg.c size -= scnprintf (next, size, next 616 drivers/usb/host/ohci-dbg.c ohci_dump_status(ohci, &next, &size); next 620 drivers/usb/host/ohci-dbg.c ohci_dbg_nosw(ohci, &next, &size, next 625 drivers/usb/host/ohci-dbg.c temp = scnprintf (next, size, next 630 drivers/usb/host/ohci-dbg.c next += temp; next 633 drivers/usb/host/ohci-dbg.c temp = scnprintf (next, size, "fmremaining 0x%08x %sFR=0x%04x\n", next 637 drivers/usb/host/ohci-dbg.c next += temp; next 640 drivers/usb/host/ohci-dbg.c temp = scnprintf (next, size, "periodicstart 0x%04x\n", next 643 drivers/usb/host/ohci-dbg.c next += temp; next 646 drivers/usb/host/ohci-dbg.c temp = scnprintf (next, size, "lsthresh 0x%04x\n", next 649 drivers/usb/host/ohci-dbg.c next += temp; next 651 drivers/usb/host/ohci-dbg.c temp = scnprintf (next, size, "hub poll timer %s\n", next 654 drivers/usb/host/ohci-dbg.c next += temp; next 657 drivers/usb/host/ohci-dbg.c ohci_dump_roothub (ohci, 1, &next, &size); next 256 drivers/usb/host/ohci-hcd.c u16 next = ohci_frame_no(ohci) + 1; next 261 drivers/usb/host/ohci-hcd.c if (unlikely(tick_before(frame, next))) { next 265 drivers/usb/host/ohci-hcd.c frame += (next - frame + ed->interval - 1) & next 281 drivers/usb/host/ohci-hcd.c (u16) (next - frame), next 287 drivers/usb/host/ohci-hcd.c next); next 838 drivers/usb/host/ohci-q.c struct list_head *tmp = td->td_list.next; next 853 drivers/usb/host/ohci-q.c struct td *next; next 855 drivers/usb/host/ohci-q.c next = list_entry (tmp, struct td, td_list); next 856 drivers/usb/host/ohci-q.c tmp = next->td_list.next; next 858 drivers/usb/host/ohci-q.c if (next->urb != urb) next 869 drivers/usb/host/ohci-q.c list_del(&next->td_list); next 871 drivers/usb/host/ohci-q.c ed->hwHeadP = next->hwNextTD | toggle; next 1173 drivers/usb/host/ohci-q.c td = list_entry(ed->td_list.next, struct td, td_list); next 1286 drivers/usb/host/oxu210hp-hcd.c qtd = list_entry(qh->qtd_list.next, next 1980 drivers/usb/host/oxu210hp-hcd.c qtd = list_entry(qtd_list->next, struct ehci_qtd, next 2046 drivers/usb/host/oxu210hp-hcd.c qtd = list_entry(qtd_list->next, struct ehci_qtd, qtd_list); next 2086 drivers/usb/host/oxu210hp-hcd.c struct ehci_qh *next; next 2095 drivers/usb/host/oxu210hp-hcd.c next = qh->reclaim; next 2096 drivers/usb/host/oxu210hp-hcd.c oxu->reclaim = next; next 2116 drivers/usb/host/oxu210hp-hcd.c if (next) { next 2118 drivers/usb/host/oxu210hp-hcd.c start_unlink_async(oxu, next); next 837 drivers/usb/host/r8a66597-hcd.c struct r8a66597_td *td, *next; next 844 drivers/usb/host/r8a66597-hcd.c list_for_each_entry_safe(td, next, list, queue) { next 147 drivers/usb/host/r8a66597.h return list_entry(r8a66597->pipe_queue[pipenum].next, next 312 drivers/usb/host/sl811-hcd.c sl811->next_periodic = ep->next; next 317 drivers/usb/host/sl811-hcd.c ep = container_of(sl811->async.next, next 331 drivers/usb/host/sl811-hcd.c if (ep->schedule.next == &sl811->async) next 334 drivers/usb/host/sl811-hcd.c sl811->next_async = container_of(ep->schedule.next, next 344 drivers/usb/host/sl811-hcd.c urb = container_of(ep->hep->urb_list.next, struct urb, urb_list); next 457 drivers/usb/host/sl811-hcd.c prev = &temp->next; next 459 drivers/usb/host/sl811-hcd.c *prev = ep->next; next 467 drivers/usb/host/sl811-hcd.c sl811->next_periodic = ep->next; next 486 drivers/usb/host/sl811-hcd.c urb = container_of(ep->hep->urb_list.next, struct urb, urb_list); next 700 drivers/usb/host/sl811-hcd.c ->hep->urb_list.next, next 710 drivers/usb/host/sl811-hcd.c ->hep->urb_list.next, next 933 drivers/usb/host/sl811-hcd.c prev = &here->next; next 937 drivers/usb/host/sl811-hcd.c ep->next = here; next 977 drivers/usb/host/sl811-hcd.c if (ep->hep->urb_list.next != &urb->urb_list) { next 1485 drivers/usb/host/sl811-hcd.c ep = ep->next; next 186 drivers/usb/host/sl811.h struct sl811h_ep *next; next 347 drivers/usb/host/u132-hcd.c struct u132_endp *next_endp = list_entry(head->next, next 521 drivers/usb/host/u132-hcd.c struct list_head *next = endp->urb_more.next; next 522 drivers/usb/host/u132-hcd.c struct u132_urbq *urbq = list_entry(next, struct u132_urbq, next 524 drivers/usb/host/u132-hcd.c list_del(next); next 560 drivers/usb/host/u132-hcd.c struct list_head *next = endp->urb_more.next; next 561 drivers/usb/host/u132-hcd.c struct u132_urbq *urbq = list_entry(next, struct u132_urbq, next 563 drivers/usb/host/u132-hcd.c list_del(next); next 2472 drivers/usb/host/u132-hcd.c struct list_head *next = endp->urb_more.next; next 2473 drivers/usb/host/u132-hcd.c struct u132_urbq *urbq = list_entry(next, next 2475 drivers/usb/host/u132-hcd.c list_del(next); next 220 drivers/usb/host/uhci-debug.c struct urb_priv *urbp = list_entry(qh->queue.next, next 222 drivers/usb/host/uhci-debug.c struct uhci_td *td = list_entry(urbp->td_list.next, next 435 drivers/usb/host/uhci-debug.c tmp = tmp->next; next 507 drivers/usb/host/uhci-debug.c tmp = head->next; next 511 drivers/usb/host/uhci-debug.c tmp = tmp->next; next 198 drivers/usb/host/uhci-q.c ntd = list_entry(td->fl_list.next, next 334 drivers/usb/host/uhci-q.c if (qh->queue.next != &urbp->node) { next 360 drivers/usb/host/uhci-q.c td = list_entry(urbp->td_list.next, struct uhci_td, list); next 383 drivers/usb/host/uhci-q.c urbp = list_entry(qh->queue.next, struct urb_priv, node); next 398 drivers/usb/host/uhci-q.c td = list_entry(urbp->td_list.next, struct uhci_td, list); next 415 drivers/usb/host/uhci-q.c pipe = list_entry(qh->queue.next, struct urb_priv, node)->urb->pipe; next 487 drivers/usb/host/uhci-q.c struct urb_priv *urbp = list_entry(qh->queue.next, next 489 drivers/usb/host/uhci-q.c struct uhci_td *td = list_entry(urbp->td_list.next, next 506 drivers/usb/host/uhci-q.c uhci->next_qh = list_entry(qh->node.next, struct uhci_qh, next 575 drivers/usb/host/uhci-q.c uhci->next_qh = list_entry(qh->node.next, struct uhci_qh, next 591 drivers/usb/host/uhci-q.c uhci->next_qh = list_entry(qh->node.next, struct uhci_qh, next 1213 drivers/usb/host/uhci-q.c if (td->list.next != urbp->td_list.prev) next 1259 drivers/usb/host/uhci-q.c unsigned frame, next; next 1279 drivers/usb/host/uhci-q.c next = uhci->frame_number + 10; next 1283 drivers/usb/host/uhci-q.c frame += (next - frame + qh->period - 1) & -qh->period; next 1289 drivers/usb/host/uhci-q.c next = uhci->frame_number + 1; next 1305 drivers/usb/host/uhci-q.c if (!uhci_frame_before_eq(next, frame)) { next 1309 drivers/usb/host/uhci-q.c frame += (next - frame + qh->period - 1) & next 1316 drivers/usb/host/uhci-q.c else if (!uhci_frame_before_eq(next, next 1323 drivers/usb/host/uhci-q.c next); next 1462 drivers/usb/host/uhci-q.c if (qh->queue.next == &urbp->node && !qh->is_stopped) { next 1535 drivers/usb/host/uhci-q.c urbp->node.next != &qh->queue) { next 1536 drivers/usb/host/uhci-q.c struct urb *nurb = list_entry(urbp->node.next, next 1582 drivers/usb/host/uhci-q.c urbp = list_entry(qh->queue.next, struct urb_priv, node); next 1641 drivers/usb/host/uhci-q.c urbp = list_entry(qh->queue.next, struct urb_priv, node); next 1643 drivers/usb/host/uhci-q.c struct uhci_td *td = list_entry(urbp->td_list.next, next 1691 drivers/usb/host/uhci-q.c urbp = list_entry(qh->queue.next, struct urb_priv, node); next 1692 drivers/usb/host/uhci-q.c td = list_entry(urbp->td_list.next, struct uhci_td, list); next 1762 drivers/usb/host/uhci-q.c uhci->next_qh = list_entry(uhci->skelqh[i]->node.next, next 1765 drivers/usb/host/uhci-q.c uhci->next_qh = list_entry(qh->node.next, next 1772 drivers/usb/host/uhci-q.c list_entry(qh->queue.next, struct urb_priv, node)); next 220 drivers/usb/host/xhci-dbgcap.c union xhci_trb *trb, *next; next 231 drivers/usb/host/xhci-dbgcap.c next = ++(ring->enqueue); next 232 drivers/usb/host/xhci-dbgcap.c if (TRB_TYPE_LINK_LE32(next->link.control)) { next 233 drivers/usb/host/xhci-dbgcap.c next->link.control ^= cpu_to_le32(TRB_CYCLE); next 41 drivers/usb/host/xhci-dbgtty.c req = list_entry(pool->next, struct dbc_request, list_pool); next 78 drivers/usb/host/xhci-dbgtty.c req = list_entry(pool->next, struct dbc_request, list_pool); next 166 drivers/usb/host/xhci-dbgtty.c req = list_entry(head->next, struct dbc_request, list_pool); next 220 drivers/usb/host/xhci-debugfs.c seg = seg->next; next 100 drivers/usb/host/xhci-ext-caps.h u32 next; next 119 drivers/usb/host/xhci-ext-caps.h next = XHCI_EXT_CAPS_NEXT(val); next 120 drivers/usb/host/xhci-ext-caps.h offset += next << 2; next 121 drivers/usb/host/xhci-ext-caps.h } while (next); next 63 drivers/usb/host/xhci-mem.c seg->next = NULL; next 83 drivers/usb/host/xhci-mem.c seg = first->next; next 85 drivers/usb/host/xhci-mem.c struct xhci_segment *next = seg->next; next 87 drivers/usb/host/xhci-mem.c seg = next; next 100 drivers/usb/host/xhci-mem.c struct xhci_segment *next, enum xhci_ring_type type) next 104 drivers/usb/host/xhci-mem.c if (!prev || !next) next 106 drivers/usb/host/xhci-mem.c prev->next = next; next 109 drivers/usb/host/xhci-mem.c cpu_to_le64(next->dma); next 133 drivers/usb/host/xhci-mem.c struct xhci_segment *next; next 138 drivers/usb/host/xhci-mem.c next = ring->enq_seg->next; next 140 drivers/usb/host/xhci-mem.c xhci_link_segments(xhci, last, next, ring->type); next 238 drivers/usb/host/xhci-mem.c seg = seg->next; next 250 drivers/usb/host/xhci-mem.c seg = seg->next; next 266 drivers/usb/host/xhci-mem.c seg = seg->next; next 332 drivers/usb/host/xhci-mem.c struct xhci_segment *next; next 334 drivers/usb/host/xhci-mem.c next = xhci_segment_alloc(xhci, cycle_state, max_packet, flags); next 335 drivers/usb/host/xhci-mem.c if (!next) { next 338 drivers/usb/host/xhci-mem.c next = prev->next; next 340 drivers/usb/host/xhci-mem.c prev = next; next 344 drivers/usb/host/xhci-mem.c xhci_link_segments(xhci, prev, next, type); next 346 drivers/usb/host/xhci-mem.c prev = next; next 440 drivers/usb/host/xhci-mem.c struct xhci_segment *next; next 442 drivers/usb/host/xhci-mem.c next = first->next; next 446 drivers/usb/host/xhci-mem.c first = next; next 807 drivers/usb/host/xhci-mem.c struct xhci_tt_bw_info *tt_info, *next; next 820 drivers/usb/host/xhci-mem.c list_for_each_entry_safe(tt_info, next, tt_list_head, tt_list) { next 940 drivers/usb/host/xhci-mem.c struct xhci_tt_bw_info *tt_info, *next; next 954 drivers/usb/host/xhci-mem.c list_for_each_entry_safe(tt_info, next, tt_list_head, tt_list) { next 1821 drivers/usb/host/xhci-mem.c seg = seg->next; next 1869 drivers/usb/host/xhci-mem.c list_del_init(ep->next); next 98 drivers/usb/host/xhci-ring.c return last_trb_on_seg(seg, trb) && (seg->next == ring->first_seg); next 145 drivers/usb/host/xhci-ring.c *seg = (*seg)->next; next 166 drivers/usb/host/xhci-ring.c ring->deq_seg = ring->deq_seg->next; next 177 drivers/usb/host/xhci-ring.c ring->deq_seg = ring->deq_seg->next; next 208 drivers/usb/host/xhci-ring.c union xhci_trb *next; next 214 drivers/usb/host/xhci-ring.c next = ++(ring->enqueue); next 217 drivers/usb/host/xhci-ring.c while (trb_is_link(next)) { next 236 drivers/usb/host/xhci-ring.c next->link.control &= cpu_to_le32(~TRB_CHAIN); next 237 drivers/usb/host/xhci-ring.c next->link.control |= cpu_to_le32(chain); next 241 drivers/usb/host/xhci-ring.c next->link.control ^= cpu_to_le32(TRB_CYCLE); next 244 drivers/usb/host/xhci-ring.c if (link_trb_toggles_cycle(next)) next 247 drivers/usb/host/xhci-ring.c ring->enq_seg = ring->enq_seg->next; next 249 drivers/usb/host/xhci-ring.c next = ring->enqueue; next 1019 drivers/usb/host/xhci-ring.c ep_ring->deq_seg = ep_ring->deq_seg->next; next 1031 drivers/usb/host/xhci-ring.c ep_ring->deq_seg = ep_ring->deq_seg->next; next 1820 drivers/usb/host/xhci-ring.c cur_seg = cur_seg->next; next 3004 drivers/usb/host/xhci-ring.c ep_ring->enq_seg = ep_ring->enq_seg->next; next 50 drivers/usb/host/xhci.c seg = seg->next; next 860 drivers/usb/host/xhci.c seg = seg->next; next 2372 drivers/usb/host/xhci.c ep_entry = bw_table->interval_bw[i].endpoints.next; next 1525 drivers/usb/host/xhci.h struct xhci_segment *next; next 31 drivers/usb/image/microtek.h struct mts_desc *next; next 523 drivers/usb/isp1760/isp1760-hcd.c qtd = list_entry(qtd->qtd_list.next, typeof(*qtd), qtd_list); next 850 drivers/usb/isp1760/isp1760-hcd.c if (usb_pipeint(list_entry(qh->qtd_list.next, struct isp1760_qtd, next 1144 drivers/usb/isp1760/isp1760-hcd.c qtd = list_entry(qtd->qtd_list.next, next 395 drivers/usb/mon/mon_main.c p = mon_buses.next; next 315 drivers/usb/mon/mon_text.c p = rp->e_list.next; next 669 drivers/usb/mon/mon_text.c p = rp->e_list.next; next 227 drivers/usb/mtu3/mtu3_qmu.c struct qmu_gpd *next; next 230 drivers/usb/mtu3/mtu3_qmu.c next = enq + 1; next 232 drivers/usb/mtu3/mtu3_qmu.c next = ring->start; next 235 drivers/usb/mtu3/mtu3_qmu.c return next == ring->dequeue; next 66 drivers/usb/musb/cppi_dma.c c->freelist = bd->next; next 75 drivers/usb/musb/cppi_dma.c bd->next = c->freelist; next 445 drivers/usb/musb/cppi_dma.c for (bd = rx->head; bd; bd = bd->next) next 622 drivers/usb/musb/cppi_dma.c if (++i < n_bds && bd->next) next 623 drivers/usb/musb/cppi_dma.c bd->hw_next = bd->next->dma; next 657 drivers/usb/musb/cppi_dma.c bd = bd->next; next 842 drivers/usb/musb/cppi_dma.c tail->next = bd; next 871 drivers/usb/musb/cppi_dma.c tail->next = NULL; next 885 drivers/usb/musb/cppi_dma.c for (d = rx->head; d; d = d->next) next 891 drivers/usb/musb/cppi_dma.c tail->next = bd; next 1016 drivers/usb/musb/cppi_dma.c bd = last ? last->next : rx->head; next 1023 drivers/usb/musb/cppi_dma.c i++, bd = bd->next) { next 1201 drivers/usb/musb/cppi_dma.c i++, bd = bd->next) { next 1528 drivers/usb/musb/cppi_dma.c struct cppi_descriptor *tmp = queue->next; next 64 drivers/usb/musb/cppi_dma.h struct cppi_descriptor *next; next 1247 drivers/usb/musb/musb_gadget.c if (!musb_ep->busy && &request->list == musb_ep->req_list.next) { next 1292 drivers/usb/musb/musb_gadget.c if (musb_ep->req_list.next != &req->list || musb_ep->busy) next 107 drivers/usb/musb/musb_gadget.h return container_of(queue->next, struct musb_request, list); next 50 drivers/usb/musb/musb_host.h return list_entry(q->next, struct musb_qh, ring); next 126 drivers/usb/musb/musb_host.h return list_entry(queue->next, struct urb, urb_list); next 967 drivers/usb/phy/phy-fsl-usb.c char *next = buf; next 974 drivers/usb/phy/phy-fsl-usb.c t = scnprintf(next, size, next 978 drivers/usb/phy/phy-fsl-usb.c next += t; next 981 drivers/usb/phy/phy-fsl-usb.c t = scnprintf(next, size, next 995 drivers/usb/phy/phy-fsl-usb.c next += t; next 998 drivers/usb/phy/phy-fsl-usb.c t = scnprintf(next, size, next 1002 drivers/usb/phy/phy-fsl-usb.c next += t; next 1005 drivers/usb/phy/phy-fsl-usb.c t = scnprintf(next, size, next 1037 drivers/usb/phy/phy-fsl-usb.c next += t; next 442 drivers/usb/renesas_usbhs/mod_host.c struct usbhsh_ep *uep, *next; next 444 drivers/usb/renesas_usbhs/mod_host.c list_for_each_entry_safe(uep, next, &udev->ep_list_head, ep_list) next 304 drivers/usb/serial/garmin_gps.c result = (struct garmin_packet *)garmin_data_p->pktlist.next; next 320 drivers/usb/serial/garmin_gps.c result = (struct garmin_packet *)garmin_data_p->pktlist.next; next 291 drivers/usb/serial/mos7720.c struct list_head *cursor, *next; next 322 drivers/usb/serial/mos7720.c list_for_each_safe(cursor, next, &mos_parport->deferred_urbs) next 71 drivers/vfio/mdev/mdev_core.c list_for_each_entry(parent, &parent_list, next) { next 195 drivers/vfio/mdev/mdev_core.c list_add(&parent->next, &parent_list); next 236 drivers/vfio/mdev/mdev_core.c list_del(&parent->next); next 258 drivers/vfio/mdev/mdev_core.c list_del(&mdev->next); next 287 drivers/vfio/mdev/mdev_core.c list_for_each_entry(tmp, &mdev_list, next) { next 303 drivers/vfio/mdev/mdev_core.c list_add(&mdev->next, &mdev_list); next 361 drivers/vfio/mdev/mdev_core.c list_for_each_entry(tmp, &mdev_list, next) { next 20 drivers/vfio/mdev/mdev_private.h struct list_head next; next 32 drivers/vfio/mdev/mdev_private.h struct list_head next; next 45 drivers/vfio/mdev/mdev_private.h struct list_head next; next 171 drivers/vfio/mdev/mdev_sysfs.c next) { next 172 drivers/vfio/mdev/mdev_sysfs.c list_del(<ype->next); next 177 drivers/vfio/mdev/mdev_sysfs.c list_add(&type->next, &parent->type_list); next 187 drivers/vfio/mdev/mdev_sysfs.c list_for_each_entry_safe(type, tmp, &parent->type_list, next) { next 188 drivers/vfio/mdev/mdev_sysfs.c list_del(&type->next); next 384 drivers/vfio/pci/vfio_pci.c &vdev->ioeventfds_list, next) { next 386 drivers/vfio/pci/vfio_pci.c list_del(&ioeventfd->next); next 1451 drivers/vfio/pci/vfio_pci_config.c u8 next; next 1459 drivers/vfio/pci/vfio_pci_config.c pos + PCI_CAP_LIST_NEXT, &next); next 1475 drivers/vfio/pci/vfio_pci_config.c *prev = next; next 1476 drivers/vfio/pci/vfio_pci_config.c pos = next; next 1497 drivers/vfio/pci/vfio_pci_config.c pos = next; next 33 drivers/vfio/pci/vfio_pci_private.h struct list_head next; next 333 drivers/vfio/pci/vfio_pci_rdwr.c list_for_each_entry(ioeventfd, &vdev->ioeventfds_list, next) { next 338 drivers/vfio/pci/vfio_pci_rdwr.c list_del(&ioeventfd->next); next 378 drivers/vfio/pci/vfio_pci_rdwr.c list_add(&ioeventfd->next, &vdev->ioeventfds_list); next 1819 drivers/vfio/vfio.c for (tmp = buf; tmp->next; tmp = buf + tmp->next) next 1822 drivers/vfio/vfio.c tmp->next = caps->size; next 1834 drivers/vfio/vfio.c for (tmp = buf; tmp->next; tmp = buf + tmp->next - offset) next 1835 drivers/vfio/vfio.c tmp->next += offset; next 43 drivers/vfio/vfio_iommu_spapr_tce.c struct list_head next; next 52 drivers/vfio/vfio_iommu_spapr_tce.c struct list_head next; next 96 drivers/vfio/vfio_iommu_spapr_tce.c list_del(&tcemem->next); next 117 drivers/vfio/vfio_iommu_spapr_tce.c list_for_each_entry(tcemem, &container->prereg_list, next) { next 148 drivers/vfio/vfio_iommu_spapr_tce.c list_for_each_entry(tcemem, &container->prereg_list, next) { next 167 drivers/vfio/vfio_iommu_spapr_tce.c list_add(&tcemem->next, &container->prereg_list); next 279 drivers/vfio/vfio_iommu_spapr_tce.c struct tce_iommu_group, next); next 351 drivers/vfio/vfio_iommu_spapr_tce.c struct tce_iommu_group, next); next 369 drivers/vfio/vfio_iommu_spapr_tce.c list_for_each_entry_safe(tcemem, tmtmp, &container->prereg_list, next) next 658 drivers/vfio/vfio_iommu_spapr_tce.c struct tce_iommu_group, next); next 683 drivers/vfio/vfio_iommu_spapr_tce.c list_for_each_entry(tcegrp, &container->group_list, next) { next 699 drivers/vfio/vfio_iommu_spapr_tce.c list_for_each_entry(tcegrp, &container->group_list, next) { next 723 drivers/vfio/vfio_iommu_spapr_tce.c list_for_each_entry(tcegrp, &container->group_list, next) { next 761 drivers/vfio/vfio_iommu_spapr_tce.c struct tce_iommu_group, next); next 816 drivers/vfio/vfio_iommu_spapr_tce.c struct tce_iommu_group, next); next 1048 drivers/vfio/vfio_iommu_spapr_tce.c list_for_each_entry(tcegrp, &container->group_list, next) { next 1266 drivers/vfio/vfio_iommu_spapr_tce.c list_for_each_entry(tcegrp, &container->group_list, next) { next 1311 drivers/vfio/vfio_iommu_spapr_tce.c list_add(&tcegrp->next, &container->group_list); next 1334 drivers/vfio/vfio_iommu_spapr_tce.c list_for_each_entry(tcegrp, &container->group_list, next) { next 1347 drivers/vfio/vfio_iommu_spapr_tce.c list_del(&tcegrp->next); next 77 drivers/vfio/vfio_iommu_type1.c struct list_head next; next 97 drivers/vfio/vfio_iommu_type1.c struct list_head next; next 666 drivers/vfio/vfio_iommu_type1.c struct vfio_regions *entry, *next; next 670 drivers/vfio/vfio_iommu_type1.c list_for_each_entry_safe(entry, next, regions, list) { next 776 drivers/vfio/vfio_iommu_type1.c struct vfio_domain, next); next 778 drivers/vfio/vfio_iommu_type1.c list_for_each_entry_continue(d, &iommu->domain_list, next) { next 786 drivers/vfio/vfio_iommu_type1.c phys_addr_t phys, next; next 801 drivers/vfio/vfio_iommu_type1.c next = iommu_iova_to_phys(domain->domain, iova + len); next 802 drivers/vfio/vfio_iommu_type1.c if (next != phys + len) next 851 drivers/vfio/vfio_iommu_type1.c list_for_each_entry(domain, &iommu->domain_list, next) next 991 drivers/vfio/vfio_iommu_type1.c list_for_each_entry(d, &iommu->domain_list, next) { next 1003 drivers/vfio/vfio_iommu_type1.c list_for_each_entry_continue_reverse(d, &iommu->domain_list, next) next 1196 drivers/vfio/vfio_iommu_type1.c d = list_first_entry(&iommu->domain_list, struct vfio_domain, next); next 1300 drivers/vfio/vfio_iommu_type1.c list_for_each_entry(g, &domain->group_list, next) { next 1497 drivers/vfio/vfio_iommu_type1.c struct vfio_iova *node, *next; next 1503 drivers/vfio/vfio_iommu_type1.c list_for_each_entry_safe(node, next, iova, list) { next 1516 drivers/vfio/vfio_iommu_type1.c list_for_each_entry_safe(node, next, iova, list) { next 1559 drivers/vfio/vfio_iommu_type1.c struct vfio_iova *n, *next; next 1570 drivers/vfio/vfio_iommu_type1.c list_for_each_entry_safe(n, next, iova, list) { next 1605 drivers/vfio/vfio_iommu_type1.c struct iommu_resv_region *n, *next; next 1607 drivers/vfio/vfio_iommu_type1.c list_for_each_entry_safe(n, next, resv_regions, list) { next 1615 drivers/vfio/vfio_iommu_type1.c struct vfio_iova *n, *next; next 1617 drivers/vfio/vfio_iommu_type1.c list_for_each_entry_safe(n, next, iova, list) { next 1668 drivers/vfio/vfio_iommu_type1.c list_for_each_entry(d, &iommu->domain_list, next) { next 1712 drivers/vfio/vfio_iommu_type1.c list_add(&group->next, next 1780 drivers/vfio/vfio_iommu_type1.c list_add(&group->next, &domain->group_list); next 1802 drivers/vfio/vfio_iommu_type1.c list_for_each_entry(d, &iommu->domain_list, next) { next 1807 drivers/vfio/vfio_iommu_type1.c list_add(&group->next, &d->group_list); next 1832 drivers/vfio/vfio_iommu_type1.c list_add(&domain->next, &iommu->domain_list); next 1919 drivers/vfio/vfio_iommu_type1.c list_for_each_entry(domain, &iommu->domain_list, next) { next 1954 drivers/vfio/vfio_iommu_type1.c list_for_each_entry(d, &iommu->domain_list, next) { next 1955 drivers/vfio/vfio_iommu_type1.c list_for_each_entry(g, &d->group_list, next) { next 1995 drivers/vfio/vfio_iommu_type1.c list_del(&group->next); next 2018 drivers/vfio/vfio_iommu_type1.c list_for_each_entry(domain, &iommu->domain_list, next) { next 2024 drivers/vfio/vfio_iommu_type1.c list_del(&group->next); next 2041 drivers/vfio/vfio_iommu_type1.c list_del(&domain->next); next 2094 drivers/vfio/vfio_iommu_type1.c &domain->group_list, next) { next 2097 drivers/vfio/vfio_iommu_type1.c list_del(&group->next); next 2119 drivers/vfio/vfio_iommu_type1.c &iommu->domain_list, next) { next 2121 drivers/vfio/vfio_iommu_type1.c list_del(&domain->next); next 2136 drivers/vfio/vfio_iommu_type1.c list_for_each_entry(domain, &iommu->domain_list, next) { next 2091 drivers/vhost/vhost.c unsigned int next; next 2098 drivers/vhost/vhost.c next = vhost16_to_cpu(vq, READ_ONCE(desc->next)); next 2099 drivers/vhost/vhost.c return next; next 172 drivers/vhost/vringh.c *up_next = vringh16_to_cpu(vrh, desc->next); next 378 drivers/vhost/vringh.c i = vringh16_to_cpu(vrh, desc.next); next 458 drivers/video/fbdev/aty/aty128fb.c struct fb_info *next; next 181 drivers/video/fbdev/core/fb_defio.c struct list_head *node, *next; next 197 drivers/video/fbdev/core/fb_defio.c list_for_each_safe(node, next, &fbdefio->pagelist) { next 733 drivers/video/fbdev/core/fbmem.c .next = fb_seq_next, next 1039 drivers/video/fbdev/core/fbmem.c if (info->modelist.prev && info->modelist.next && next 1627 drivers/video/fbdev/core/fbmem.c if (!fb_info->modelist.prev || !fb_info->modelist.next) next 1151 drivers/video/fbdev/core/modedb.c if (!head->prev || !head->next || list_empty(head)) next 1960 drivers/video/fbdev/matrox/matroxfb_base.c for (minfo = matroxfb_l(matroxfb_list.next); next 1962 drivers/video/fbdev/matrox/matroxfb_base.c minfo = matroxfb_l(minfo->next_fb.next)) { next 1980 drivers/video/fbdev/matrox/matroxfb_base.c for (minfo = matroxfb_l(matroxfb_list.next); next 1982 drivers/video/fbdev/matrox/matroxfb_base.c minfo = matroxfb_l(minfo->next_fb.next)) { next 2001 drivers/video/fbdev/matrox/matroxfb_base.c for (drv = matroxfb_driver_l(matroxfb_driver_list.next); next 2003 drivers/video/fbdev/matrox/matroxfb_base.c drv = matroxfb_driver_l(drv->node.next)) { next 240 drivers/video/fbdev/omap/hwa742.c req = list_entry(hwa742.free_req_list.next, next 275 drivers/video/fbdev/omap/hwa742.c req = list_entry(hwa742.pending_req_list.next, next 229 drivers/video/fbdev/omap2/omapfb/dss/display.c dssdev = list_entry(l->next, struct omap_dss_device, next 79 drivers/video/fbdev/pxa3xx-gcu.c struct pxa3xx_gcu_batch *next; next 192 drivers/video/fbdev/pxa3xx-gcu.c struct pxa3xx_gcu_batch *next; next 195 drivers/video/fbdev/pxa3xx-gcu.c next = running->next; next 196 drivers/video/fbdev/pxa3xx-gcu.c running->next = priv->free; next 198 drivers/video/fbdev/pxa3xx-gcu.c running = next; next 220 drivers/video/fbdev/pxa3xx-gcu.c ready = ready->next; next 408 drivers/video/fbdev/pxa3xx-gcu.c priv->free = buffer->next; next 416 drivers/video/fbdev/pxa3xx-gcu.c buffer->next = priv->free; next 432 drivers/video/fbdev/pxa3xx-gcu.c buffer->next = NULL; next 437 drivers/video/fbdev/pxa3xx-gcu.c priv->ready_last->next = buffer; next 545 drivers/video/fbdev/pxa3xx-gcu.c buffer->next = priv->free; next 555 drivers/video/fbdev/pxa3xx-gcu.c struct pxa3xx_gcu_batch *next, *buffer = priv->free; next 558 drivers/video/fbdev/pxa3xx-gcu.c next = buffer->next; next 564 drivers/video/fbdev/pxa3xx-gcu.c buffer = next; next 251 drivers/video/fbdev/riva/riva_hw.c int last, next, cur; next 292 drivers/video/fbdev/riva/riva_hw.c next = VIDEO; next 294 drivers/video/fbdev/riva/riva_hw.c next = MPORT; next 296 drivers/video/fbdev/riva/riva_hw.c next = GRAPHICS; next 302 drivers/video/fbdev/riva/riva_hw.c next = VIDEO; next 304 drivers/video/fbdev/riva/riva_hw.c next = GRAPHICS; next 306 drivers/video/fbdev/riva/riva_hw.c next = MPORT; next 311 drivers/video/fbdev/riva/riva_hw.c next = GRAPHICS; next 313 drivers/video/fbdev/riva/riva_hw.c next = VIDEO; next 315 drivers/video/fbdev/riva/riva_hw.c next = MPORT; next 320 drivers/video/fbdev/riva/riva_hw.c next = MPORT; next 322 drivers/video/fbdev/riva/riva_hw.c next = GRAPHICS; next 324 drivers/video/fbdev/riva/riva_hw.c next = VIDEO; next 329 drivers/video/fbdev/riva/riva_hw.c cur = next; next 650 drivers/video/fbdev/sis/sis.h struct sis_video_info *next; next 5867 drivers/video/fbdev/sis/sis_main.c while((countvideo = countvideo->next) != NULL) next 6494 drivers/video/fbdev/sis/sis_main.c ivideo->next = card_list; next 1828 drivers/video/fbdev/smscufx.c node = dev->urbs.list.next; /* have reserved one with sem */ next 1919 drivers/video/fbdev/smscufx.c entry = dev->urbs.list.next; next 1013 drivers/video/fbdev/udlfb.c struct dlfb_deferred_free *d = list_entry(dlfb->deferred_free.next, struct dlfb_deferred_free, list); next 1857 drivers/video/fbdev/udlfb.c node = dlfb->urbs.list.next; /* have reserved one with sem */ next 1952 drivers/video/fbdev/udlfb.c entry = dlfb->urbs.list.next; next 1086 drivers/video/fbdev/vermilion/vermilion.c list = global_no_mode.next; next 1124 drivers/video/fbdev/vermilion/vermilion.c list = global_no_mode.next; next 1137 drivers/video/fbdev/vermilion/vermilion.c struct vml_info *entry, *next; next 1146 drivers/video/fbdev/vermilion/vermilion.c list_for_each_entry_safe(entry, next, &global_has_mode, head) { next 901 drivers/video/fbdev/via/hw.c bool next = true; next 904 drivers/video/fbdev/via/hw.c while (next) { next 905 drivers/video/fbdev/via/hw.c next = false; next 913 drivers/video/fbdev/via/hw.c next = true; next 224 drivers/virtio/virtio_balloon.c struct page *page, *next; next 226 drivers/virtio/virtio_balloon.c list_for_each_entry_safe(page, next, pages, lru) { next 77 drivers/virtio/virtio_ring.c u16 next; /* The next desc state in a list. */ next 411 drivers/virtio/virtio_ring.c desc[i].next = cpu_to_virtio16(_vq->vdev, i + 1); next 491 drivers/virtio/virtio_ring.c i = virtio16_to_cpu(_vq->vdev, desc[i].next); next 504 drivers/virtio/virtio_ring.c i = virtio16_to_cpu(_vq->vdev, desc[i].next); next 533 drivers/virtio/virtio_ring.c vq->split.vring.desc[head].next); next 579 drivers/virtio/virtio_ring.c i = virtio16_to_cpu(_vq->vdev, desc[i].next); next 634 drivers/virtio/virtio_ring.c i = virtio16_to_cpu(vq->vq.vdev, vq->split.vring.desc[i].next); next 639 drivers/virtio/virtio_ring.c vq->split.vring.desc[i].next = cpu_to_virtio16(vq->vq.vdev, next 1064 drivers/virtio/virtio_ring.c vq->free_head = vq->packed.desc_state[id].next; next 1172 drivers/virtio/virtio_ring.c curr = vq->packed.desc_state[curr].next; next 1293 drivers/virtio/virtio_ring.c vq->packed.desc_state[state->last].next = vq->free_head; next 1302 drivers/virtio/virtio_ring.c curr = vq->packed.desc_state[curr].next; next 1654 drivers/virtio/virtio_ring.c vq->packed.desc_state[i].next = i + 1; next 2123 drivers/virtio/virtio_ring.c vq->split.vring.desc[i].next = cpu_to_virtio16(vdev, i + 1); next 48 drivers/w1/slaves/w1_ds2413.c next: next 65 drivers/w1/slaves/w1_ds2413.c goto next; next 182 drivers/xen/balloon.c page = list_entry(ballooned_pages.next, struct page, lru); next 198 drivers/xen/balloon.c struct list_head *next = page->lru.next; next 199 drivers/xen/balloon.c if (next == &ballooned_pages) next 201 drivers/xen/balloon.c return list_entry(next, struct page, lru); next 127 drivers/xen/gntalloc.c struct gntalloc_gref *gref, *next; next 167 drivers/xen/gntalloc.c list_for_each_entry_safe(gref, next, &queue_file, next_file) { next 270 drivers/xen/gntalloc.c gref = list_entry(priv->list.next, next 368 drivers/xen/gntalloc.c n = list_entry(gref->next_file.next, next 477 drivers/xen/gntalloc.c struct gntalloc_gref *gref, *next; next 489 drivers/xen/gntalloc.c next = list_entry(gref->next_gref.next, next 493 drivers/xen/gntalloc.c gref = next; next 553 drivers/xen/gntalloc.c gref = list_entry(gref->next_file.next, next 52 drivers/xen/gntdev-common.h struct list_head next; next 36 drivers/xen/gntdev-dmabuf.c struct list_head next; next 64 drivers/xen/gntdev-dmabuf.c struct list_head next; next 111 drivers/xen/gntdev-dmabuf.c list_add(&obj->next, &priv->exp_wait_list); next 122 drivers/xen/gntdev-dmabuf.c list_del(&obj->next); next 142 drivers/xen/gntdev-dmabuf.c list_for_each_entry(obj, &priv->exp_wait_list, next) next 156 drivers/xen/gntdev-dmabuf.c list_for_each_entry(gntdev_dmabuf, &priv->exp_list, next) next 319 drivers/xen/gntdev-dmabuf.c list_del(&gntdev_dmabuf->next); next 328 drivers/xen/gntdev-dmabuf.c list_del(&map->next); next 431 drivers/xen/gntdev-dmabuf.c list_add(&gntdev_dmabuf->next, &args->dmabuf_priv->exp_list); next 680 drivers/xen/gntdev-dmabuf.c list_add(&gntdev_dmabuf->next, &priv->imp_list); next 708 drivers/xen/gntdev-dmabuf.c list_for_each_entry_safe(gntdev_dmabuf, q, &priv->imp_list, next) { next 712 drivers/xen/gntdev-dmabuf.c list_del(&gntdev_dmabuf->next); next 752 drivers/xen/gntdev-dmabuf.c list_for_each_entry_safe(gntdev_dmabuf, q, &priv->imp_list, next) next 87 drivers/xen/gntdev.c list_for_each_entry(map, &priv->maps, next) next 209 drivers/xen/gntdev.c list_for_each_entry(map, &priv->maps, next) { next 211 drivers/xen/gntdev.c list_add_tail(&add->next, &map->next); next 216 drivers/xen/gntdev.c list_add_tail(&add->next, &priv->maps); next 227 drivers/xen/gntdev.c list_for_each_entry(map, &priv->maps, next) { next 254 drivers/xen/gntdev.c list_del(&map->next); next 530 drivers/xen/gntdev.c list_for_each_entry(map, &priv->maps, next) { next 536 drivers/xen/gntdev.c list_for_each_entry(map, &priv->freeable_maps, next) { next 557 drivers/xen/gntdev.c list_for_each_entry(map, &priv->maps, next) { next 566 drivers/xen/gntdev.c list_for_each_entry(map, &priv->freeable_maps, next) { next 642 drivers/xen/gntdev.c map = list_entry(priv->maps.next, next 643 drivers/xen/gntdev.c struct gntdev_grant_map, next); next 644 drivers/xen/gntdev.c list_del(&map->next); next 716 drivers/xen/gntdev.c list_del(&map->next); next 718 drivers/xen/gntdev.c list_add_tail(&map->next, &priv->freeable_maps); next 791 drivers/xen/gntdev.c list_for_each_entry(map, &priv->maps, next) { next 198 drivers/xen/grant-table.c struct gnttab_free_callback *callback, *next; next 204 drivers/xen/grant-table.c next = callback->next; next 206 drivers/xen/grant-table.c callback->next = NULL; next 209 drivers/xen/grant-table.c callback->next = gnttab_free_callback_list; next 212 drivers/xen/grant-table.c callback = next; next 625 drivers/xen/grant-table.c cb = cb->next; next 631 drivers/xen/grant-table.c callback->next = gnttab_free_callback_list; next 645 drivers/xen/grant-table.c for (pcb = &gnttab_free_callback_list; *pcb; pcb = &(*pcb)->next) { next 647 drivers/xen/grant-table.c *pcb = callback->next; next 115 drivers/xen/mcelog.c num = xen_mcelog.next; next 131 drivers/xen/mcelog.c xen_mcelog.next = 0; next 146 drivers/xen/mcelog.c if (xen_mcelog.next) next 201 drivers/xen/mcelog.c entry = xen_mcelog.next; next 216 drivers/xen/mcelog.c xen_mcelog.next++; next 171 drivers/xen/privcmd.c pos = pos->next; next 205 drivers/xen/privcmd.c pos = pos->next; next 457 drivers/xen/xen-pciback/pci_stub.c psdev = container_of(seized_devices.next, next 27 drivers/xen/xen-pciback/vpci.c return head->next; next 57 drivers/xen/xenbus/xenbus_client.c struct list_head next; next 598 drivers/xen/xenbus/xenbus_client.c list_add(&node->next, &xenbus_valloc_pages); next 718 drivers/xen/xenbus/xenbus_client.c list_add(&node->next, &xenbus_valloc_pages); next 744 drivers/xen/xenbus/xenbus_client.c list_for_each_entry(node, &xenbus_valloc_pages, next) { next 746 drivers/xen/xenbus/xenbus_client.c list_del(&node->next); next 832 drivers/xen/xenbus/xenbus_client.c list_for_each_entry(node, &xenbus_valloc_pages, next) { next 835 drivers/xen/xenbus/xenbus_client.c list_del(&node->next); next 148 drivers/xen/xenbus/xenbus_dev_frontend.c rb = list_entry(u->read_buffers.next, struct read_buffer, list); next 170 drivers/xen/xenbus/xenbus_dev_frontend.c rb = list_entry(u->read_buffers.next, next 220 drivers/xen/xenbus/xenbus_dev_frontend.c rb = list_entry(list->next, struct read_buffer, list); next 221 drivers/xen/xenbus/xenbus_dev_frontend.c list_del(list->next); next 883 drivers/xen/xenbus/xenbus_xs.c ent = watch_events.next; next 106 drivers/xen/xenfs/xensyms.c .next = xensyms_next, next 94 drivers/zorro/proc.c .next = zorro_seq_next, next 120 fs/affs/amigaffs.h __be32 next; next 587 fs/affs/file.c u32 tmp_next = be32_to_cpu(AFFS_DATA_HEAD(prev_bh)->next); next 593 fs/affs/file.c AFFS_DATA_HEAD(prev_bh)->next = cpu_to_be32(bh->b_blocknr); next 733 fs/affs/file.c AFFS_DATA_HEAD(bh)->next = 0; next 736 fs/affs/file.c u32 tmp_next = be32_to_cpu(AFFS_DATA_HEAD(prev_bh)->next); next 742 fs/affs/file.c AFFS_DATA_HEAD(prev_bh)->next = cpu_to_be32(bh->b_blocknr); next 767 fs/affs/file.c AFFS_DATA_HEAD(bh)->next = 0; next 770 fs/affs/file.c u32 tmp_next = be32_to_cpu(AFFS_DATA_HEAD(prev_bh)->next); next 776 fs/affs/file.c AFFS_DATA_HEAD(prev_bh)->next = cpu_to_be32(bh->b_blocknr); next 927 fs/affs/file.c tmp = be32_to_cpu(AFFS_DATA_HEAD(bh)->next); next 928 fs/affs/file.c AFFS_DATA_HEAD(bh)->next = 0; next 56 fs/afs/callback.c for (pp = &server->cb_volumes.first; *pp; pp = &(*pp)->next) { next 67 fs/afs/callback.c new_vi->srv_link.next = *pp; next 69 fs/afs/callback.c (*pp)->pprev = &new_vi->srv_link.next; next 605 fs/afs/cell.c for (p = &net->proc_cells.first; *p; p = &(*p)->next) { next 612 fs/afs/cell.c cell->proc_link.next = *p; next 613 fs/afs/cell.c rcu_assign_pointer(*p, &cell->proc_link.next); next 614 fs/afs/cell.c if (cell->proc_link.next) next 615 fs/afs/cell.c cell->proc_link.next->pprev = &cell->proc_link.next; next 356 fs/afs/dir.c unsigned offset, next, curr; next 367 fs/afs/dir.c offset = next next 369 fs/afs/dir.c next = offset + 1; next 378 fs/afs/dir.c next * sizeof(union afs_xdr_dirent); next 395 fs/afs/dir.c if (next >= AFS_DIR_SLOTS_PER_BLOCK) { next 400 fs/afs/dir.c offset, next, tmp, nlen); next 403 fs/afs/dir.c if (!(block->hdr.bitmap[next / 8] & next 404 fs/afs/dir.c (1 << (next % 8)))) { next 408 fs/afs/dir.c offset, next, tmp, nlen); next 414 fs/afs/dir.c next, tmp, nlen); next 415 fs/afs/dir.c next++; next 432 fs/afs/dir.c ctx->pos = blkoff + next * sizeof(union afs_xdr_dirent); next 112 fs/afs/flock.c struct file_lock *p, *_p, *next = NULL; next 131 fs/afs/flock.c if (next && next 132 fs/afs/flock.c (next->fl_type == F_WRLCK || p->fl_type == F_RDLCK)) next 134 fs/afs/flock.c next = p; next 140 fs/afs/flock.c if (next) { next 142 fs/afs/flock.c next->fl_u.afs.state = AFS_LOCK_YOUR_TRY; next 143 fs/afs/flock.c trace_afs_flock_op(vnode, next, afs_flock_op_wake); next 144 fs/afs/flock.c wake_up(&next->fl_wait); next 164 fs/afs/flock.c p = list_entry(vnode->pending_locks.next, next 602 fs/afs/flock.c ASSERTCMP(vnode->pending_locks.next, ==, &fl->fl_u.afs.link); next 786 fs/afs/inode.c struct afs_wb_key *wbk = list_entry(vnode->wb_keys.next, next 77 fs/afs/proc.c .next = afs_proc_cells_next, next 254 fs/afs/proc.c .next = afs_proc_cell_volumes_next, next 365 fs/afs/proc.c .next = afs_proc_cell_vlservers_next, next 418 fs/afs/proc.c .next = afs_proc_servers_next, next 473 fs/afs/proc.c .next = afs_proc_sysname_next, next 377 fs/afs/write.c p = vnode->wb_keys.next; next 389 fs/afs/write.c p = p->next; next 436 fs/afs/write.c p = wbk->vnode_link.next; next 700 fs/afs/write.c pgoff_t start, end, next; next 708 fs/afs/write.c ret = afs_writepages_region(mapping, wbc, start, end, &next); next 711 fs/afs/write.c &next); next 712 fs/afs/write.c mapping->writeback_index = next; next 715 fs/afs/write.c ret = afs_writepages_region(mapping, wbc, 0, end, &next); next 717 fs/afs/write.c mapping->writeback_index = next; next 721 fs/afs/write.c ret = afs_writepages_region(mapping, wbc, start, end, &next); next 841 fs/afs/write.c wbk = list_entry(graveyard.next, struct afs_wb_key, vnode_link); next 86 fs/autofs/autofs_i.h struct autofs_wait_queue *next; next 464 fs/autofs/expire.c goto next; next 472 fs/autofs/expire.c next: next 152 fs/autofs/root.c goto next; next 157 fs/autofs/root.c goto next; next 159 fs/autofs/root.c goto next; next 162 fs/autofs/root.c goto next; next 164 fs/autofs/root.c goto next; next 172 fs/autofs/root.c next: next 212 fs/autofs/root.c goto next; next 217 fs/autofs/root.c goto next; next 219 fs/autofs/root.c goto next; next 222 fs/autofs/root.c goto next; next 224 fs/autofs/root.c goto next; next 232 fs/autofs/root.c next: next 678 fs/autofs/root.c if (d_child->next == &parent->d_subdirs && next 31 fs/autofs/waitq.c nwq = wq->next; next 228 fs/autofs/waitq.c for (wq = sbi->queues; wq; wq = wq->next) { next 429 fs/autofs/waitq.c wq->next = sbi->queues; next 532 fs/autofs/waitq.c for (wql = &sbi->queues; (wq = *wql) != NULL; wql = &wq->next) { next 542 fs/autofs/waitq.c *wql = wq->next; /* Unlink from chain */ next 1662 fs/binfmt_elf.c struct elf_thread_core_info *next; next 1809 fs/binfmt_elf.c for (ct = &dump_task->mm->core_state->dumper; ct; ct = ct->next) { next 1818 fs/binfmt_elf.c t->next = info->thread; next 1825 fs/binfmt_elf.c t->next = info->thread->next; next 1826 fs/binfmt_elf.c info->thread->next = t; next 1833 fs/binfmt_elf.c for (t = info->thread; t != NULL; t = t->next) next 1892 fs/binfmt_elf.c t = t->next; next 1904 fs/binfmt_elf.c threads = t->next; next 2019 fs/binfmt_elf.c for (ct = current->mm->core_state->dumper.next; next 2020 fs/binfmt_elf.c ct; ct = ct->next) { next 2115 fs/binfmt_elf.c struct list_head *tmp = info->thread_list.next; next 1612 fs/binfmt_elf_fdpic.c for (ct = current->mm->core_state->dumper.next; next 1613 fs/binfmt_elf_fdpic.c ct; ct = ct->next) { next 1795 fs/binfmt_elf_fdpic.c struct list_head *tmp = thread_list.next; next 236 fs/btrfs/async-thread.c work = list_entry(list->next, struct btrfs_work, next 23 fs/btrfs/backref.c struct extent_inode_elem *next; next 56 fs/btrfs/backref.c e->next = *eie; next 69 fs/btrfs/backref.c eie_next = eie->next; next 251 fs/btrfs/backref.c while (eie && eie->next) next 252 fs/btrfs/backref.c eie = eie->next; next 257 fs/btrfs/backref.c eie->next = newref->inode_list; next 473 fs/btrfs/backref.c goto next; next 479 fs/btrfs/backref.c while (old->next) next 480 fs/btrfs/backref.c old = old->next; next 481 fs/btrfs/backref.c old->next = eie; next 485 fs/btrfs/backref.c next: next 1318 fs/btrfs/backref.c while (eie->next) next 1319 fs/btrfs/backref.c eie = eie->next; next 1320 fs/btrfs/backref.c eie->next = ref->inode_list; next 1872 fs/btrfs/backref.c for (eie = inode_list; eie; eie = eie->next) { next 535 fs/btrfs/block-group.c next: next 566 fs/btrfs/block-group.c goto next; next 587 fs/btrfs/block-group.c goto next; next 1297 fs/btrfs/block-group.c goto next; next 1306 fs/btrfs/block-group.c goto next; next 1318 fs/btrfs/block-group.c goto next; next 1408 fs/btrfs/block-group.c next: next 1883 fs/btrfs/block-group.c goto next; next 1899 fs/btrfs/block-group.c next: next 3102 fs/btrfs/block-group.c caching_ctl = list_entry(info->caching_block_groups.next, next 3163 fs/btrfs/block-group.c space_info = list_entry(info->space_info.next, next 466 fs/btrfs/compression.c goto next; next 476 fs/btrfs/compression.c goto next; next 529 fs/btrfs/compression.c next: next 875 fs/btrfs/compression.c ws = wsman->idle_ws.next; next 909 fs/btrfs/compression.c workspace = idle_ws->next; next 347 fs/btrfs/ctree.c struct rb_node *next; next 379 fs/btrfs/ctree.c for (node = rb_first(tm_root); node; node = next) { next 380 fs/btrfs/ctree.c next = rb_next(node); next 1224 fs/btrfs/ctree.c struct rb_node *next; next 1279 fs/btrfs/ctree.c next = rb_next(&tm->node); next 1280 fs/btrfs/ctree.c if (!next) next 1282 fs/btrfs/ctree.c tm = rb_entry(next, struct tree_mod_elem, node); next 5283 fs/btrfs/ctree.c next: next 5316 fs/btrfs/ctree.c goto next; next 5326 fs/btrfs/ctree.c goto next; next 5351 fs/btrfs/ctree.c struct extent_buffer *next; next 5365 fs/btrfs/ctree.c next = NULL; next 5430 fs/btrfs/ctree.c if (next) { next 5431 fs/btrfs/ctree.c btrfs_tree_unlock_rw(next, next_rw_lock); next 5432 fs/btrfs/ctree.c free_extent_buffer(next); next 5435 fs/btrfs/ctree.c next = c; next 5437 fs/btrfs/ctree.c ret = read_block_for_search(root, path, &next, level, next 5448 fs/btrfs/ctree.c ret = btrfs_try_tree_read_lock(next); next 5457 fs/btrfs/ctree.c free_extent_buffer(next); next 5464 fs/btrfs/ctree.c btrfs_tree_read_lock(next); next 5478 fs/btrfs/ctree.c path->nodes[level] = next; next 5485 fs/btrfs/ctree.c ret = read_block_for_search(root, path, &next, level, next 5496 fs/btrfs/ctree.c ret = btrfs_try_tree_read_lock(next); next 5499 fs/btrfs/ctree.c btrfs_tree_read_lock(next); next 215 fs/btrfs/delayed-inode.c p = delayed_root->node_list.next; next 229 fs/btrfs/delayed-inode.c struct btrfs_delayed_node *next = NULL; next 237 fs/btrfs/delayed-inode.c p = delayed_root->node_list.next; next 241 fs/btrfs/delayed-inode.c p = node->n_list.next; next 243 fs/btrfs/delayed-inode.c next = list_entry(p, struct btrfs_delayed_node, n_list); next 244 fs/btrfs/delayed-inode.c refcount_inc(&next->refs); next 248 fs/btrfs/delayed-inode.c return next; next 300 fs/btrfs/delayed-inode.c p = delayed_root->prepare_list.next; next 344 fs/btrfs/delayed-inode.c struct btrfs_delayed_item **next) next 377 fs/btrfs/delayed-inode.c if (next) { next 379 fs/btrfs/delayed-inode.c *next = NULL; next 381 fs/btrfs/delayed-inode.c *next = delayed_item; next 383 fs/btrfs/delayed-inode.c *next = rb_entry(node, struct btrfs_delayed_item, next 386 fs/btrfs/delayed-inode.c *next = NULL; next 538 fs/btrfs/delayed-inode.c struct btrfs_delayed_item *next = NULL; next 542 fs/btrfs/delayed-inode.c next = rb_entry(p, struct btrfs_delayed_item, rb_node); next 544 fs/btrfs/delayed-inode.c return next; next 699 fs/btrfs/delayed-inode.c struct btrfs_delayed_item *curr, *next; next 718 fs/btrfs/delayed-inode.c next = item; next 724 fs/btrfs/delayed-inode.c while (total_size + next->data_len + sizeof(struct btrfs_item) <= next 726 fs/btrfs/delayed-inode.c total_data_size += next->data_len; next 727 fs/btrfs/delayed-inode.c total_size += next->data_len + sizeof(struct btrfs_item); next 728 fs/btrfs/delayed-inode.c list_add_tail(&next->tree_list, &head); next 731 fs/btrfs/delayed-inode.c curr = next; next 732 fs/btrfs/delayed-inode.c next = __btrfs_next_delayed_item(curr); next 733 fs/btrfs/delayed-inode.c if (!next) next 736 fs/btrfs/delayed-inode.c if (!btrfs_is_continuous_delayed_item(curr, next)) next 766 fs/btrfs/delayed-inode.c list_for_each_entry(next, &head, tree_list) { next 767 fs/btrfs/delayed-inode.c keys[i] = next->key; next 768 fs/btrfs/delayed-inode.c data_size[i] = next->data_len; next 778 fs/btrfs/delayed-inode.c list_for_each_entry_safe(curr, next, &head, tree_list) { next 879 fs/btrfs/delayed-inode.c struct btrfs_delayed_item *curr, *next; next 895 fs/btrfs/delayed-inode.c next = item; next 902 fs/btrfs/delayed-inode.c while (btrfs_comp_cpu_keys(&next->key, &key) == 0) { next 903 fs/btrfs/delayed-inode.c list_add_tail(&next->tree_list, &head); next 906 fs/btrfs/delayed-inode.c curr = next; next 907 fs/btrfs/delayed-inode.c next = __btrfs_next_delayed_item(curr); next 908 fs/btrfs/delayed-inode.c if (!next) next 911 fs/btrfs/delayed-inode.c if (!btrfs_is_continuous_delayed_item(curr, next)) next 927 fs/btrfs/delayed-inode.c list_for_each_entry_safe(curr, next, &head, tree_list) { next 1641 fs/btrfs/delayed-inode.c struct btrfs_delayed_item *curr, *next; next 1643 fs/btrfs/delayed-inode.c list_for_each_entry_safe(curr, next, ins_list, readdir_list) { next 1649 fs/btrfs/delayed-inode.c list_for_each_entry_safe(curr, next, del_list, readdir_list) { next 1687 fs/btrfs/delayed-inode.c struct btrfs_delayed_item *curr, *next; next 1702 fs/btrfs/delayed-inode.c list_for_each_entry_safe(curr, next, ins_list, readdir_list) { next 436 fs/btrfs/delayed-ref.c struct btrfs_delayed_ref_node *next; next 443 fs/btrfs/delayed-ref.c next = rb_entry(node, struct btrfs_delayed_ref_node, ref_node); next 445 fs/btrfs/delayed-ref.c if (seq && next->seq >= seq) next 447 fs/btrfs/delayed-ref.c if (comp_refs(ref, next, false)) next 450 fs/btrfs/delayed-ref.c if (ref->action == next->action) { next 451 fs/btrfs/delayed-ref.c mod = next->ref_mod; next 453 fs/btrfs/delayed-ref.c if (ref->ref_mod < next->ref_mod) { next 454 fs/btrfs/delayed-ref.c swap(ref, next); next 457 fs/btrfs/delayed-ref.c mod = -next->ref_mod; next 460 fs/btrfs/delayed-ref.c drop_delayed_ref(trans, delayed_refs, head, next); next 2040 fs/btrfs/disk-io.c gang[0] = list_entry(fs_info->dead_roots.next, next 2727 fs/btrfs/extent-tree.c struct btrfs_caching_control *next; next 2733 fs/btrfs/extent-tree.c list_for_each_entry_safe(caching_ctl, next, next 4804 fs/btrfs/extent-tree.c struct extent_buffer *next; next 4827 fs/btrfs/extent-tree.c next = find_extent_buffer(fs_info, bytenr); next 4828 fs/btrfs/extent-tree.c if (!next) { next 4829 fs/btrfs/extent-tree.c next = btrfs_find_create_tree_block(fs_info, bytenr); next 4830 fs/btrfs/extent-tree.c if (IS_ERR(next)) next 4831 fs/btrfs/extent-tree.c return PTR_ERR(next); next 4833 fs/btrfs/extent-tree.c btrfs_set_buffer_lockdep_class(root->root_key.objectid, next, next 4837 fs/btrfs/extent-tree.c btrfs_tree_lock(next); next 4838 fs/btrfs/extent-tree.c btrfs_set_lock_blocking_write(next); next 4879 fs/btrfs/extent-tree.c if (!btrfs_buffer_uptodate(next, generation, 0)) { next 4880 fs/btrfs/extent-tree.c btrfs_tree_unlock(next); next 4881 fs/btrfs/extent-tree.c free_extent_buffer(next); next 4882 fs/btrfs/extent-tree.c next = NULL; next 4886 fs/btrfs/extent-tree.c if (!next) { next 4889 fs/btrfs/extent-tree.c next = read_tree_block(fs_info, bytenr, generation, level - 1, next 4891 fs/btrfs/extent-tree.c if (IS_ERR(next)) { next 4892 fs/btrfs/extent-tree.c return PTR_ERR(next); next 4893 fs/btrfs/extent-tree.c } else if (!extent_buffer_uptodate(next)) { next 4894 fs/btrfs/extent-tree.c free_extent_buffer(next); next 4897 fs/btrfs/extent-tree.c btrfs_tree_lock(next); next 4898 fs/btrfs/extent-tree.c btrfs_set_lock_blocking_write(next); next 4902 fs/btrfs/extent-tree.c ASSERT(level == btrfs_header_level(next)); next 4903 fs/btrfs/extent-tree.c if (level != btrfs_header_level(next)) { next 4908 fs/btrfs/extent-tree.c path->nodes[level] = next; next 4958 fs/btrfs/extent-tree.c ret = btrfs_qgroup_trace_subtree(trans, next, next 4988 fs/btrfs/extent-tree.c btrfs_tree_unlock(next); next 4989 fs/btrfs/extent-tree.c free_extent_buffer(next); next 69 fs/btrfs/extent_io.c state = list_entry(states.next, struct extent_state, leak_list); next 79 fs/btrfs/extent_io.c eb = list_entry(buffers.next, struct extent_buffer, leak_list); next 438 fs/btrfs/extent_io.c struct rb_node *next= NULL; next 441 fs/btrfs/extent_io.c ret = __etree_search(tree, offset, &next, NULL, p_ret, parent_ret); next 443 fs/btrfs/extent_io.c return next; next 586 fs/btrfs/extent_io.c struct rb_node *next = rb_next(&state->rb_node); next 587 fs/btrfs/extent_io.c if (next) next 588 fs/btrfs/extent_io.c return rb_entry(next, struct extent_state, rb_node); next 605 fs/btrfs/extent_io.c struct extent_state *next; next 624 fs/btrfs/extent_io.c next = next_state(state); next 634 fs/btrfs/extent_io.c next = next_state(state); next 636 fs/btrfs/extent_io.c return next; next 740 fs/btrfs/extent_io.c goto next; next 772 fs/btrfs/extent_io.c goto next; next 799 fs/btrfs/extent_io.c next: next 1579 fs/btrfs/extent_io.c struct rb_node *node, *prev = NULL, *next; next 1585 fs/btrfs/extent_io.c node = __etree_search(tree, start, &next, &prev, NULL, NULL); next 1586 fs/btrfs/extent_io.c if (!node && !next && !prev) { next 1594 fs/btrfs/extent_io.c } else if (!node && !next) { next 1604 fs/btrfs/extent_io.c node = next; next 2353 fs/btrfs/extent_io.c struct extent_state *state, *next; next 2366 fs/btrfs/extent_io.c next = next_state(state); next 2372 fs/btrfs/extent_io.c state = next; next 190 fs/btrfs/extent_map.c static int mergable_maps(struct extent_map *prev, struct extent_map *next) next 203 fs/btrfs/extent_map.c test_bit(EXTENT_FLAG_LOGGING, &next->flags)) next 211 fs/btrfs/extent_map.c if (!list_empty(&prev->list) || !list_empty(&next->list)) next 214 fs/btrfs/extent_map.c ASSERT(next->block_start != EXTENT_MAP_DELALLOC && next 217 fs/btrfs/extent_map.c if (extent_map_end(prev) == next->start && next 218 fs/btrfs/extent_map.c prev->flags == next->flags && next 219 fs/btrfs/extent_map.c prev->bdev == next->bdev && next 220 fs/btrfs/extent_map.c ((next->block_start == EXTENT_MAP_HOLE && next 222 fs/btrfs/extent_map.c (next->block_start == EXTENT_MAP_INLINE && next 224 fs/btrfs/extent_map.c (next->block_start < EXTENT_MAP_LAST_BYTE - 1 && next 225 fs/btrfs/extent_map.c next->block_start == extent_map_block_end(prev)))) { next 420 fs/btrfs/extent_map.c struct rb_node *next = NULL; next 423 fs/btrfs/extent_map.c rb_node = __tree_search(&tree->map.rb_root, start, &prev, &next); next 427 fs/btrfs/extent_map.c else if (next) next 428 fs/btrfs/extent_map.c rb_node = next; next 512 fs/btrfs/extent_map.c struct rb_node *next; next 514 fs/btrfs/extent_map.c next = rb_next(&em->rb_node); next 515 fs/btrfs/extent_map.c if (!next) next 517 fs/btrfs/extent_map.c return container_of(next, struct extent_map, rb_node); next 542 fs/btrfs/extent_map.c struct extent_map *next; next 550 fs/btrfs/extent_map.c next = existing; next 551 fs/btrfs/extent_map.c prev = prev_extent_map(next); next 554 fs/btrfs/extent_map.c next = next_extent_map(prev); next 559 fs/btrfs/extent_map.c end = next ? next->start : extent_map_end(em); next 214 fs/btrfs/file-item.c goto next; next 276 fs/btrfs/file-item.c next: next 413 fs/btrfs/file-item.c sums = list_entry(tmplist.next, struct btrfs_ordered_sum, list); next 486 fs/btrfs/file.c goto next; next 498 fs/btrfs/file.c next: next 644 fs/btrfs/file.c goto next; next 722 fs/btrfs/file.c next: next 647 fs/btrfs/free-space-cache.c goto next; next 649 fs/btrfs/free-space-cache.c goto next; next 660 fs/btrfs/free-space-cache.c next: next 934 fs/btrfs/free-space-cache.c cluster = list_entry(block_group->cluster_list.next, next 1103 fs/btrfs/free-space-cache.c struct btrfs_free_space *entry, *next; next 1107 fs/btrfs/free-space-cache.c list_for_each_entry_safe(entry, next, bitmap_list, list) { next 1132 fs/btrfs/free-space-cache.c struct btrfs_free_space *entry, *next; next 1134 fs/btrfs/free-space-cache.c list_for_each_entry_safe(entry, next, bitmap_list, list) next 1933 fs/btrfs/free-space-cache.c struct rb_node *next = rb_next(&bitmap_info->offset_index); next 1941 fs/btrfs/free-space-cache.c if (!next) next 1944 fs/btrfs/free-space-cache.c bitmap_info = rb_entry(next, struct btrfs_free_space, next 2076 fs/btrfs/free-space-cache.c cluster = list_entry(block_group->cluster_list.next, next 2614 fs/btrfs/free-space-cache.c while ((head = block_group->cluster_list.next) != next 3250 fs/btrfs/free-space-cache.c goto next; next 3266 fs/btrfs/free-space-cache.c next: next 3308 fs/btrfs/free-space-cache.c goto next; next 3317 fs/btrfs/free-space-cache.c goto next; next 3324 fs/btrfs/free-space-cache.c goto next; next 3341 fs/btrfs/free-space-cache.c next: next 103 fs/btrfs/inode-map.c goto next; next 115 fs/btrfs/inode-map.c next: next 767 fs/btrfs/inode.c async_extent = list_entry(async_chunk->extents.next, next 1316 fs/btrfs/inode.c sums = list_entry(list.next, struct btrfs_ordered_sum, list); next 3049 fs/btrfs/inode.c goto next; next 3053 fs/btrfs/inode.c goto next; next 3071 fs/btrfs/inode.c next: next 5255 fs/btrfs/inode.c goto next; next 5282 fs/btrfs/inode.c next: next 6155 fs/btrfs/inode.c goto next; next 6157 fs/btrfs/inode.c goto next; next 6185 fs/btrfs/inode.c next: next 7102 fs/btrfs/inode.c goto next; next 7133 fs/btrfs/inode.c next: next 7153 fs/btrfs/inode.c goto next; next 8133 fs/btrfs/inode.c goto next; next 8143 fs/btrfs/inode.c next: next 8237 fs/btrfs/inode.c goto next; next 8250 fs/btrfs/inode.c goto next; next 8259 fs/btrfs/inode.c next: next 10227 fs/btrfs/inode.c struct btrfs_delalloc_work *work, *next; next 10239 fs/btrfs/inode.c binode = list_entry(splice.next, struct btrfs_inode, next 10272 fs/btrfs/inode.c list_for_each_entry_safe(work, next, &works, list) { next 10547 fs/btrfs/inode.c goto next; next 10572 fs/btrfs/inode.c next: next 10777 fs/btrfs/inode.c struct rb_node *node, *next; next 10782 fs/btrfs/inode.c next = rb_next(node); next 10790 fs/btrfs/inode.c node = next; next 1144 fs/btrfs/ioctl.c struct extent_map *next; next 1151 fs/btrfs/ioctl.c next = defrag_lookup_extent(inode, em->start + em->len); next 1152 fs/btrfs/ioctl.c if (!next || next->block_start >= EXTENT_MAP_LAST_BYTE) next 1154 fs/btrfs/ioctl.c else if ((em->block_start + em->block_len == next->block_start) && next 1155 fs/btrfs/ioctl.c (em->block_len > SZ_128K && next->block_len > SZ_128K)) next 1158 fs/btrfs/ioctl.c free_extent_map(next); next 1497 fs/btrfs/ioctl.c unsigned long next; next 1502 fs/btrfs/ioctl.c next = DIV_ROUND_UP(skip, PAGE_SIZE); next 1503 fs/btrfs/ioctl.c i = max(i + 1, next); next 447 fs/btrfs/ordered-data.c cur = entry->list.next; next 526 fs/btrfs/ordered-data.c struct btrfs_ordered_extent *ordered, *next; next 564 fs/btrfs/ordered-data.c list_for_each_entry_safe(ordered, next, &works, work_list) { next 357 fs/btrfs/print-tree.c struct extent_buffer *next; next 360 fs/btrfs/print-tree.c next = read_tree_block(fs_info, btrfs_node_blockptr(c, i), next 363 fs/btrfs/print-tree.c if (IS_ERR(next)) { next 365 fs/btrfs/print-tree.c } else if (!extent_buffer_uptodate(next)) { next 366 fs/btrfs/print-tree.c free_extent_buffer(next); next 370 fs/btrfs/print-tree.c if (btrfs_is_leaf(next) && next 373 fs/btrfs/print-tree.c if (btrfs_header_level(next) != next 376 fs/btrfs/print-tree.c btrfs_print_tree(next, follow); next 377 fs/btrfs/print-tree.c free_extent_buffer(next); next 3815 fs/btrfs/qgroup.c struct btrfs_qgroup_swapped_block *next; next 3817 fs/btrfs/qgroup.c rbtree_postorder_for_each_entry_safe(entry, next, cur_root, next 4032 fs/btrfs/qgroup.c struct btrfs_qgroup_extent_record *next; next 4036 fs/btrfs/qgroup.c rbtree_postorder_for_each_entry_safe(entry, next, root, node) { next 433 fs/btrfs/raid56.c rbio = list_entry(table->stripe_cache.next, next 797 fs/btrfs/raid56.c struct btrfs_raid_bio *next; next 798 fs/btrfs/raid56.c struct list_head *head = rbio->plug_list.next; next 800 fs/btrfs/raid56.c next = list_entry(head, struct btrfs_raid_bio, next 805 fs/btrfs/raid56.c list_add(&next->hash_list, &h->hash_list); next 806 fs/btrfs/raid56.c refcount_inc(&next->refs); next 810 fs/btrfs/raid56.c if (next->operation == BTRFS_RBIO_READ_REBUILD) next 811 fs/btrfs/raid56.c start_async_work(next, read_rebuild_work); next 812 fs/btrfs/raid56.c else if (next->operation == BTRFS_RBIO_REBUILD_MISSING) { next 813 fs/btrfs/raid56.c steal_rbio(rbio, next); next 814 fs/btrfs/raid56.c start_async_work(next, read_rebuild_work); next 815 fs/btrfs/raid56.c } else if (next->operation == BTRFS_RBIO_WRITE) { next 816 fs/btrfs/raid56.c steal_rbio(rbio, next); next 817 fs/btrfs/raid56.c start_async_work(next, rmw_work); next 818 fs/btrfs/raid56.c } else if (next->operation == BTRFS_RBIO_PARITY_SCRUB) { next 819 fs/btrfs/raid56.c steal_rbio(rbio, next); next 820 fs/btrfs/raid56.c start_async_work(next, scrub_parity_work); next 859 fs/btrfs/raid56.c struct bio *next; next 862 fs/btrfs/raid56.c next = cur->bi_next; next 866 fs/btrfs/raid56.c cur = next; next 1700 fs/btrfs/raid56.c cur = list_entry(plug->rbio_list.next, next 217 fs/btrfs/relocation.c node = list_entry(cache->detached.next, next 223 fs/btrfs/relocation.c node = list_entry(cache->leaves.next, next 348 fs/btrfs/relocation.c edge = list_entry(node->upper.next, next 375 fs/btrfs/relocation.c edge = list_entry(edge->list[LOWER].next, next 429 fs/btrfs/relocation.c edge = list_entry(node->upper.next, struct backref_edge, next 490 fs/btrfs/relocation.c node = list_entry(cache->detached.next, next 496 fs/btrfs/relocation.c node = list_entry(cache->changed.next, next 754 fs/btrfs/relocation.c edge = list_entry(cur->upper.next, struct backref_edge, next 795 fs/btrfs/relocation.c goto next; next 827 fs/btrfs/relocation.c goto next; next 873 fs/btrfs/relocation.c goto next; next 881 fs/btrfs/relocation.c goto next; next 1015 fs/btrfs/relocation.c next: next 1034 fs/btrfs/relocation.c edge = list_entry(list.next, struct backref_edge, list[UPPER]); next 1058 fs/btrfs/relocation.c edge = list_entry(list.next, struct backref_edge, list[UPPER]); next 1115 fs/btrfs/relocation.c upper = list_entry(useless.next, struct backref_node, list); next 1125 fs/btrfs/relocation.c edge = list_entry(upper->lower.next, next 1149 fs/btrfs/relocation.c lower = list_entry(useless.next, next 1181 fs/btrfs/relocation.c lower = list_entry(useless.next, next 1278 fs/btrfs/relocation.c new_edge = list_entry(new_node->lower.next, next 2225 fs/btrfs/relocation.c struct btrfs_root *next; next 2229 fs/btrfs/relocation.c list_for_each_entry_safe(root, next, &rc->dirty_subvol_roots, next 2457 fs/btrfs/relocation.c reloc_root = list_entry(rc->reloc_roots.next, next 2491 fs/btrfs/relocation.c reloc_root = list_entry(list->next, struct btrfs_root, next 2525 fs/btrfs/relocation.c reloc_root = list_entry(reloc_roots.next, next 2617 fs/btrfs/relocation.c struct backref_node *next; next 2621 fs/btrfs/relocation.c next = node; next 2624 fs/btrfs/relocation.c next = walk_up_backref(next, edges, &index); next 2625 fs/btrfs/relocation.c root = next->root; next 2637 fs/btrfs/relocation.c if (next->new_bytenr != root->node->start) { next 2638 fs/btrfs/relocation.c BUG_ON(next->new_bytenr); next 2639 fs/btrfs/relocation.c BUG_ON(!list_empty(&next->list)); next 2640 fs/btrfs/relocation.c next->new_bytenr = root->node->start; next 2641 fs/btrfs/relocation.c next->root = root; next 2642 fs/btrfs/relocation.c list_add_tail(&next->list, next 2644 fs/btrfs/relocation.c __mark_block_processed(rc, next); next 2650 fs/btrfs/relocation.c next = walk_down_backref(edges, &index); next 2651 fs/btrfs/relocation.c if (!next || next->level <= node->level) next 2657 fs/btrfs/relocation.c next = node; next 2660 fs/btrfs/relocation.c rc->backref_cache.path[next->level] = next; next 2663 fs/btrfs/relocation.c next = edges[index]->node[UPPER]; next 2677 fs/btrfs/relocation.c struct backref_node *next; next 2683 fs/btrfs/relocation.c next = node; next 2686 fs/btrfs/relocation.c next = walk_up_backref(next, edges, &index); next 2687 fs/btrfs/relocation.c root = next->root; next 2697 fs/btrfs/relocation.c if (next != node) next 2700 fs/btrfs/relocation.c next = walk_down_backref(edges, &index); next 2701 fs/btrfs/relocation.c if (!next || next->level <= node->level) next 2715 fs/btrfs/relocation.c struct backref_node *next = node; next 2723 fs/btrfs/relocation.c while (next) { next 2726 fs/btrfs/relocation.c if (next->processed && (reserve || next != node)) next 2731 fs/btrfs/relocation.c if (list_empty(&next->upper)) next 2734 fs/btrfs/relocation.c edge = list_entry(next->upper.next, next 2737 fs/btrfs/relocation.c next = edge->node[UPPER]; next 2739 fs/btrfs/relocation.c next = walk_down_backref(edges, &index); next 2831 fs/btrfs/relocation.c goto next; next 2836 fs/btrfs/relocation.c goto next; next 2870 fs/btrfs/relocation.c goto next; next 2883 fs/btrfs/relocation.c goto next; next 2887 fs/btrfs/relocation.c goto next; next 2897 fs/btrfs/relocation.c goto next; next 2901 fs/btrfs/relocation.c goto next; next 2913 fs/btrfs/relocation.c goto next; next 2935 fs/btrfs/relocation.c next: next 2978 fs/btrfs/relocation.c node = list_entry(cache->pending[level].next, next 3020 fs/btrfs/relocation.c struct backref_node *next = node; next 3025 fs/btrfs/relocation.c while (next) { next 3028 fs/btrfs/relocation.c if (next->processed) next 3031 fs/btrfs/relocation.c __mark_block_processed(rc, next); next 3033 fs/btrfs/relocation.c if (list_empty(&next->upper)) next 3036 fs/btrfs/relocation.c edge = list_entry(next->upper.next, next 3039 fs/btrfs/relocation.c next = edge->node[UPPER]; next 3041 fs/btrfs/relocation.c next = walk_down_backref(edges, &index); next 3144 fs/btrfs/relocation.c struct tree_block *next; next 3155 fs/btrfs/relocation.c rbtree_postorder_for_each_entry_safe(block, next, blocks, rb_node) { next 3161 fs/btrfs/relocation.c rbtree_postorder_for_each_entry_safe(block, next, blocks, rb_node) { next 3170 fs/btrfs/relocation.c rbtree_postorder_for_each_entry_safe(block, next, blocks, rb_node) { next 3769 fs/btrfs/relocation.c goto next; next 3773 fs/btrfs/relocation.c goto next; next 3777 fs/btrfs/relocation.c goto next; next 3782 fs/btrfs/relocation.c goto next; next 3804 fs/btrfs/relocation.c next: next 3940 fs/btrfs/relocation.c next: next 3958 fs/btrfs/relocation.c goto next; next 3964 fs/btrfs/relocation.c goto next; next 3971 fs/btrfs/relocation.c goto next; next 4594 fs/btrfs/relocation.c reloc_root = list_entry(reloc_roots.next, next 4681 fs/btrfs/relocation.c sums = list_entry(list.next, struct btrfs_ordered_sum, list); next 2703 fs/btrfs/scrub.c struct scrub_page *curr, *next; next 2714 fs/btrfs/scrub.c list_for_each_entry_safe(curr, next, &sparity->spages, list) { next 2915 fs/btrfs/scrub.c goto next; next 2923 fs/btrfs/scrub.c goto next; next 2948 fs/btrfs/scrub.c goto next; next 3018 fs/btrfs/scrub.c next: next 3267 fs/btrfs/scrub.c goto next; next 3275 fs/btrfs/scrub.c goto next; next 3299 fs/btrfs/scrub.c goto next; next 3386 fs/btrfs/scrub.c next: next 2113 fs/btrfs/send.c nce = list_entry(sctx->name_cache_list.next, next 2125 fs/btrfs/send.c nce = list_entry(sctx->name_cache_list.next, next 2819 fs/btrfs/send.c cur = list_entry(head->next, struct recorded_ref, list); next 4102 fs/btrfs/send.c cur = list_entry(sctx->deleted_refs.next, struct recorded_ref, next 5219 fs/btrfs/send.c goto next; next 5319 fs/btrfs/send.c next: next 5661 fs/btrfs/send.c goto next; next 5679 fs/btrfs/send.c goto next; next 5682 fs/btrfs/send.c goto next; next 5686 fs/btrfs/send.c next: next 1126 fs/btrfs/transaction.c struct list_head *next; next 1164 fs/btrfs/transaction.c next = fs_info->dirty_cowonly_roots.next; next 1165 fs/btrfs/transaction.c list_del_init(next); next 1166 fs/btrfs/transaction.c root = list_entry(next, struct btrfs_root, dirty_list); next 1675 fs/btrfs/transaction.c struct btrfs_pending_snapshot *pending, *next; next 1679 fs/btrfs/transaction.c list_for_each_entry_safe(pending, next, head, list) { next 807 fs/btrfs/tree-log.c sums = list_entry(ordered_sums.next, next 1106 fs/btrfs/tree-log.c goto next; next 1147 fs/btrfs/tree-log.c next: next 2170 fs/btrfs/tree-log.c goto next; next 2183 fs/btrfs/tree-log.c next: next 2689 fs/btrfs/tree-log.c struct extent_buffer *next; next 2719 fs/btrfs/tree-log.c next = btrfs_find_create_tree_block(fs_info, bytenr); next 2720 fs/btrfs/tree-log.c if (IS_ERR(next)) next 2721 fs/btrfs/tree-log.c return PTR_ERR(next); next 2724 fs/btrfs/tree-log.c ret = wc->process_func(root, next, wc, ptr_gen, next 2727 fs/btrfs/tree-log.c free_extent_buffer(next); next 2733 fs/btrfs/tree-log.c ret = btrfs_read_buffer(next, ptr_gen, next 2736 fs/btrfs/tree-log.c free_extent_buffer(next); next 2741 fs/btrfs/tree-log.c btrfs_tree_lock(next); next 2742 fs/btrfs/tree-log.c btrfs_set_lock_blocking_write(next); next 2743 fs/btrfs/tree-log.c btrfs_clean_tree_block(next); next 2744 fs/btrfs/tree-log.c btrfs_wait_tree_block_writeback(next); next 2745 fs/btrfs/tree-log.c btrfs_tree_unlock(next); next 2747 fs/btrfs/tree-log.c if (test_and_clear_bit(EXTENT_BUFFER_DIRTY, &next->bflags)) next 2748 fs/btrfs/tree-log.c clear_extent_buffer_dirty(next); next 2757 fs/btrfs/tree-log.c free_extent_buffer(next); next 2761 fs/btrfs/tree-log.c free_extent_buffer(next); next 2764 fs/btrfs/tree-log.c ret = btrfs_read_buffer(next, ptr_gen, *level - 1, &first_key); next 2766 fs/btrfs/tree-log.c free_extent_buffer(next); next 2773 fs/btrfs/tree-log.c path->nodes[*level-1] = next; next 2774 fs/btrfs/tree-log.c *level = btrfs_header_level(next); next 2820 fs/btrfs/tree-log.c struct extent_buffer *next; next 2822 fs/btrfs/tree-log.c next = path->nodes[*level]; next 2825 fs/btrfs/tree-log.c btrfs_tree_lock(next); next 2826 fs/btrfs/tree-log.c btrfs_set_lock_blocking_write(next); next 2827 fs/btrfs/tree-log.c btrfs_clean_tree_block(next); next 2828 fs/btrfs/tree-log.c btrfs_wait_tree_block_writeback(next); next 2829 fs/btrfs/tree-log.c btrfs_tree_unlock(next); next 2831 fs/btrfs/tree-log.c if (test_and_clear_bit(EXTENT_BUFFER_DIRTY, &next->bflags)) next 2832 fs/btrfs/tree-log.c clear_extent_buffer_dirty(next); next 2902 fs/btrfs/tree-log.c struct extent_buffer *next; next 2904 fs/btrfs/tree-log.c next = path->nodes[orig_level]; next 2907 fs/btrfs/tree-log.c btrfs_tree_lock(next); next 2908 fs/btrfs/tree-log.c btrfs_set_lock_blocking_write(next); next 2909 fs/btrfs/tree-log.c btrfs_clean_tree_block(next); next 2910 fs/btrfs/tree-log.c btrfs_wait_tree_block_writeback(next); next 2911 fs/btrfs/tree-log.c btrfs_tree_unlock(next); next 2913 fs/btrfs/tree-log.c if (test_and_clear_bit(EXTENT_BUFFER_DIRTY, &next->bflags)) next 2914 fs/btrfs/tree-log.c clear_extent_buffer_dirty(next); next 2920 fs/btrfs/tree-log.c next->start, next->len); next 4071 fs/btrfs/tree-log.c struct btrfs_ordered_sum *sums = list_entry(ordered_sums.next, next 4130 fs/btrfs/tree-log.c struct btrfs_ordered_sum *sums = list_entry(ordered_sums.next, next 4436 fs/btrfs/tree-log.c em = list_entry(extents.next, struct extent_map, list); next 6254 fs/btrfs/tree-log.c goto next; next 6293 fs/btrfs/tree-log.c next: next 64 fs/btrfs/ulist.c struct ulist_node *next; next 66 fs/btrfs/ulist.c list_for_each_entry_safe(node, next, &ulist->nodes, list) { next 267 fs/btrfs/ulist.c if (uiter->cur_list && uiter->cur_list->next == &ulist->nodes) next 270 fs/btrfs/ulist.c uiter->cur_list = uiter->cur_list->next; next 272 fs/btrfs/ulist.c uiter->cur_list = ulist->nodes.next; next 353 fs/btrfs/volumes.c device = list_entry(fs_devices->devices.next, next 366 fs/btrfs/volumes.c fs_devices = list_entry(fs_uuids.next, next 1215 fs/btrfs/volumes.c struct btrfs_device *device, *next; next 1221 fs/btrfs/volumes.c list_for_each_entry_safe(device, next, &fs_devices->devices, dev_list) { next 1668 fs/btrfs/volumes.c goto next; next 1674 fs/btrfs/volumes.c goto next; next 1716 fs/btrfs/volumes.c next: next 7649 fs/btrfs/volumes.c struct btrfs_device *curr, *next; next 7662 fs/btrfs/volumes.c list_for_each_entry_safe(curr, next, &trans->dev_update_list, next 327 fs/btrfs/xattr.c goto next; next 339 fs/btrfs/xattr.c next: next 379 fs/btrfs/zlib.c goto next; next 396 fs/btrfs/zlib.c next: next 107 fs/btrfs/zstd.c struct list_head *pos, *next; next 116 fs/btrfs/zstd.c list_for_each_prev_safe(pos, next, &wsm.lru_list) { next 205 fs/btrfs/zstd.c workspace = container_of(wsm.idle_ws[i].next, next 237 fs/btrfs/zstd.c ws = wsm.idle_ws[i].next; next 686 fs/buffer.c bh = BH_ENTRY(list->next); next 770 fs/buffer.c __remove_assoc_queue(BH_ENTRY(list->next)); next 793 fs/buffer.c struct buffer_head *bh = BH_ENTRY(list->next); next 1472 fs/buffer.c struct buffer_head *head, *bh, *next; next 1489 fs/buffer.c next = bh->b_this_page; next 1503 fs/buffer.c bh = next; next 1607 fs/buffer.c goto next; next 1613 fs/buffer.c next: next 1776 fs/buffer.c struct buffer_head *next = bh->b_this_page; next 1782 fs/buffer.c bh = next; next 1830 fs/buffer.c struct buffer_head *next = bh->b_this_page; next 1837 fs/buffer.c bh = next; next 3232 fs/buffer.c struct buffer_head *next = bh->b_this_page; next 3236 fs/buffer.c bh = next; next 3285 fs/buffer.c struct buffer_head *next = bh->b_this_page; next 3287 fs/buffer.c bh = next; next 488 fs/cachefiles/namei.c struct dentry *dir, *next = NULL; next 531 fs/cachefiles/namei.c next = lookup_one_len(name, dir, nlen); next 533 fs/cachefiles/namei.c if (IS_ERR(next)) { next 534 fs/cachefiles/namei.c trace_cachefiles_lookup(object, next, NULL); next 538 fs/cachefiles/namei.c inode = d_backing_inode(next); next 539 fs/cachefiles/namei.c trace_cachefiles_lookup(object, next, inode); next 540 fs/cachefiles/namei.c _debug("next -> %p %s", next, inode ? "positive" : "negative"); next 548 fs/cachefiles/namei.c if (d_is_negative(next)) next 554 fs/cachefiles/namei.c if (d_is_negative(next)) { next 560 fs/cachefiles/namei.c ret = security_path_mkdir(&path, next, 0); next 564 fs/cachefiles/namei.c ret = vfs_mkdir(d_inode(dir), next, 0); next 567 fs/cachefiles/namei.c trace_cachefiles_mkdir(object, next, ret); next 571 fs/cachefiles/namei.c if (unlikely(d_unhashed(next))) { next 572 fs/cachefiles/namei.c dput(next); next 576 fs/cachefiles/namei.c ASSERT(d_backing_inode(next)); next 579 fs/cachefiles/namei.c next, d_backing_inode(next), d_backing_inode(next)->i_ino); next 581 fs/cachefiles/namei.c } else if (!d_can_lookup(next)) { next 583 fs/cachefiles/namei.c d_backing_inode(next)->i_ino); next 590 fs/cachefiles/namei.c if (d_is_negative(next)) { next 596 fs/cachefiles/namei.c ret = security_path_mknod(&path, next, S_IFREG, 0); next 600 fs/cachefiles/namei.c ret = vfs_create(d_inode(dir), next, S_IFREG, true); next 602 fs/cachefiles/namei.c trace_cachefiles_create(object, next, ret); next 606 fs/cachefiles/namei.c ASSERT(d_backing_inode(next)); next 609 fs/cachefiles/namei.c next, d_backing_inode(next), d_backing_inode(next)->i_ino); next 611 fs/cachefiles/namei.c } else if (!d_can_lookup(next) && next 612 fs/cachefiles/namei.c !d_is_reg(next) next 615 fs/cachefiles/namei.c d_backing_inode(next)->i_ino); next 626 fs/cachefiles/namei.c dir = next; next 627 fs/cachefiles/namei.c next = NULL; next 632 fs/cachefiles/namei.c object->dentry = next; next 637 fs/cachefiles/namei.c _debug("validate '%pd'", next); next 645 fs/cachefiles/namei.c ret = cachefiles_bury_object(cache, object, dir, next, next 648 fs/cachefiles/namei.c dput(next); next 649 fs/cachefiles/namei.c next = NULL; next 682 fs/cachefiles/namei.c path.dentry = next; next 736 fs/cachefiles/namei.c _debug("lookup error %ld", PTR_ERR(next)); next 737 fs/cachefiles/namei.c ret = PTR_ERR(next); next 740 fs/cachefiles/namei.c next = NULL; next 743 fs/cachefiles/namei.c dput(next); next 79 fs/cachefiles/proc.c .next = cachefiles_histogram_next, next 169 fs/cachefiles/rdwr.c monitor = list_entry(op->to_do.next, next 190 fs/cachefiles/rdwr.c goto next; next 208 fs/cachefiles/rdwr.c next: next 1272 fs/ceph/dir.c goto next; next 1297 fs/ceph/dir.c next: next 1326 fs/ceph/mds_client.c p = session->s_caps.next; next 1331 fs/ceph/mds_client.c p = p->next; next 1352 fs/ceph/mds_client.c p = p->next; next 1506 fs/ceph/mds_client.c cap = list_entry(session->s_caps.next, next 2632 fs/ceph/mds_client.c req = list_entry(tmp_list.next, next 213 fs/ceph/quota.c struct ceph_snap_realm *realm, *next; next 253 fs/ceph/quota.c next = realm->parent; next 254 fs/ceph/quota.c if (has_quota || !next) next 257 fs/ceph/quota.c ceph_get_snap_realm(mdsc, next); next 259 fs/ceph/quota.c realm = next; next 318 fs/ceph/quota.c struct ceph_snap_realm *realm, *next; next 391 fs/ceph/quota.c next = realm->parent; next 392 fs/ceph/quota.c if (exceeded || !next) next 394 fs/ceph/quota.c ceph_get_snap_realm(mdsc, next); next 396 fs/ceph/quota.c realm = next; next 35 fs/char_dev.c struct char_device_struct *next; next 56 fs/char_dev.c for (cd = chrdevs[major_to_index(offset)]; cd; cd = cd->next) { next 77 fs/char_dev.c for (cd = chrdevs[major_to_index(i)]; cd; cd = cd->next) next 134 fs/char_dev.c for (curr = chrdevs[i]; curr; prev = curr, curr = curr->next) { next 156 fs/char_dev.c cd->next = curr; next 159 fs/char_dev.c cd->next = prev->next; next 160 fs/char_dev.c prev->next = cd; next 178 fs/char_dev.c for (cp = &chrdevs[i]; *cp; cp = &(*cp)->next) next 185 fs/char_dev.c *cp = cd->next; next 204 fs/char_dev.c dev_t n, next; next 206 fs/char_dev.c for (n = from; n < to; n = next) { next 207 fs/char_dev.c next = MKDEV(MAJOR(n)+1, 0); next 208 fs/char_dev.c if (next > to) next 209 fs/char_dev.c next = to; next 211 fs/char_dev.c next - n, name); next 218 fs/char_dev.c for (n = from; n < to; n = next) { next 219 fs/char_dev.c next = MKDEV(MAJOR(n)+1, 0); next 220 fs/char_dev.c kfree(__unregister_chrdev_region(MAJOR(n), MINOR(n), next - n)); next 314 fs/char_dev.c dev_t n, next; next 316 fs/char_dev.c for (n = from; n < to; n = next) { next 317 fs/char_dev.c next = MKDEV(MAJOR(n)+1, 0); next 318 fs/char_dev.c if (next > to) next 319 fs/char_dev.c next = to; next 320 fs/char_dev.c kfree(__unregister_chrdev_region(MAJOR(n), MINOR(n), next - n)); next 440 fs/char_dev.c inode = container_of(cdev->list.next, struct inode, i_devices); next 1465 fs/cifs/dfs_cache.c goto next; next 1472 fs/cifs/dfs_cache.c next: next 1111 fs/cifs/file.c (lock->blist.next == &lock->blist)); next 1317 fs/cifs/file.c el = locks_to_send.next; next 2173 fs/cifs/file.c pgoff_t end, pgoff_t *index, pgoff_t *next, bool *done) next 2203 fs/cifs/file.c if (*next && (page->index != *next)) { next 2231 fs/cifs/file.c *next = page->index + 1; next 2314 fs/cifs/file.c pgoff_t next = 0, tofind, saved_index = index; next 2353 fs/cifs/file.c end, &index, &next, &done); next 2414 fs/cifs/file.c index = next; next 435 fs/cifs/smb2ops.c size_t next = 0; next 450 fs/cifs/smb2ops.c next = le32_to_cpu(p->Next); next 451 fs/cifs/smb2ops.c if (!next) { next 455 fs/cifs/smb2ops.c p = (struct network_interface_info_ioctl_rsp *)((u8 *)p+next); next 456 fs/cifs/smb2ops.c bytes_left -= next; next 534 fs/cifs/smb2ops.c next = le32_to_cpu(p->Next); next 535 fs/cifs/smb2ops.c if (!next) next 537 fs/cifs/smb2ops.c p = (struct network_interface_info_ioctl_rsp *)((u8 *)p+next); next 538 fs/cifs/smb2ops.c bytes_left -= next; next 3236 fs/cifs/smb2ops.c u64 next; next 3289 fs/cifs/smb2ops.c next = le64_to_cpu(out_data[num - 1].file_offset) + next 3291 fs/cifs/smb2ops.c len = len - (next - start); next 3292 fs/cifs/smb2ops.c start = next; next 1925 fs/cifs/smb2pdu.c unsigned int next; next 1948 fs/cifs/smb2pdu.c next = le32_to_cpu(cc->Next); next 1949 fs/cifs/smb2pdu.c if (!next) next 1951 fs/cifs/smb2pdu.c remaining -= next; next 1952 fs/cifs/smb2pdu.c cc = (struct create_context *)((char *)cc + next); next 748 fs/cifs/smbdirect.c send_wr.next = NULL; next 947 fs/cifs/smbdirect.c send_wr.next = NULL; next 1107 fs/cifs/smbdirect.c recv_wr.next = NULL; next 244 fs/coda/psdev.c req = list_entry(vcp->vc_pending.next, struct upc_req,uc_chain); next 1637 fs/configfs/dir.c for (p = q->next; p != &parent_sd->s_children; p = p->next) { next 1638 fs/configfs/dir.c struct configfs_dirent *next; next 1643 fs/configfs/dir.c next = list_entry(p, struct configfs_dirent, s_sibling); next 1644 fs/configfs/dir.c if (!next->s_element) next 1660 fs/configfs/dir.c dentry = next->s_dentry; next 1669 fs/configfs/dir.c name = configfs_get_name(next); next 1672 fs/configfs/dir.c if (!dir_emit(ctx, name, len, ino, dt_type(next))) next 1709 fs/configfs/dir.c p = sd->s_children.next; next 1711 fs/configfs/dir.c struct configfs_dirent *next; next 1712 fs/configfs/dir.c next = list_entry(p, struct configfs_dirent, next 1714 fs/configfs/dir.c if (next->s_element) next 1716 fs/configfs/dir.c p = p->next; next 446 fs/coredump.c core_state->dumper.next = NULL; next 466 fs/coredump.c ptr = core_state->dumper.next; next 469 fs/coredump.c ptr = ptr->next; next 478 fs/coredump.c struct core_thread *curr, *next; next 488 fs/coredump.c next = mm->core_state->dumper.next; next 489 fs/coredump.c while ((curr = next) != NULL) { next 490 fs/coredump.c next = curr->next; next 514 fs/dcache.c struct dentry *next; next 542 fs/dcache.c while (dentry->d_child.next != &parent->d_subdirs) { next 543 fs/dcache.c next = list_entry(dentry->d_child.next, struct dentry, d_child); next 544 fs/dcache.c if (likely(!(next->d_flags & DCACHE_DENTRY_CURSOR))) next 546 fs/dcache.c dentry->d_child.next = next->d_child.next; next 1271 fs/dcache.c struct list_head *next; next 1293 fs/dcache.c next = this_parent->d_subdirs.next; next 1295 fs/dcache.c while (next != &this_parent->d_subdirs) { next 1296 fs/dcache.c struct list_head *tmp = next; next 1298 fs/dcache.c next = tmp->next; next 1346 fs/dcache.c next = child->d_child.next; next 1347 fs/dcache.c if (next == &this_parent->d_subdirs) next 1349 fs/dcache.c child = list_entry(next, struct dentry, d_child); next 2374 fs/dcache.c goto next; next 2376 fs/dcache.c goto next; next 2379 fs/dcache.c goto next; next 2385 fs/dcache.c next: next 313 fs/dcookies.c struct list_head next; next 329 fs/dcookies.c list_add(&user->next, &dcookie_users); next 345 fs/dcookies.c list_del(&user->next); next 504 fs/dlm/debug_fs.c struct rb_node *next; next 518 fs/dlm/debug_fs.c next = rb_next(&rp->res_hashnode); next 520 fs/dlm/debug_fs.c if (next) { next 521 fs/dlm/debug_fs.c r = rb_entry(next, struct dlm_rsb, res_hashnode); next 551 fs/dlm/debug_fs.c next = rb_first(tree); next 552 fs/dlm/debug_fs.c r = rb_entry(next, struct dlm_rsb, res_hashnode); next 576 fs/dlm/debug_fs.c .next = table_seq_next, next 583 fs/dlm/debug_fs.c .next = table_seq_next, next 590 fs/dlm/debug_fs.c .next = table_seq_next, next 597 fs/dlm/debug_fs.c .next = table_seq_next, next 252 fs/dlm/dir.c list = r->res_root_list.next; next 254 fs/dlm/dir.c list = ls->ls_root_list.next; next 257 fs/dlm/dir.c for (offset = 0; list != &ls->ls_root_list; list = list->next) { next 1650 fs/dlm/lock.c struct rb_node *n, *next; next 1667 fs/dlm/lock.c for (n = rb_first(&ls->ls_rsbtbl[b].toss); n; n = next) { next 1668 fs/dlm/lock.c next = rb_next(n); next 2192 fs/dlm/lock.c struct dlm_lkb *first = list_entry(head->next, struct dlm_lkb, next 2793 fs/dlm/lock.c lkb = list_entry(r->res_lookup.next, struct dlm_lkb, next 6150 fs/dlm/lock.c lkb = list_entry(proc->locks.next, struct dlm_lkb, lkb_ownqueue); next 6223 fs/dlm/lock.c lkb = list_entry(proc->locks.next, struct dlm_lkb, next 1500 fs/dlm/lowcomms.c e = list_entry(con->writequeue.next, struct writequeue_entry, next 308 fs/dlm/member.c newlist->next = tmp; next 309 fs/dlm/member.c tmp->prev->next = newlist; next 367 fs/dlm/member.c memb = list_entry(head->next, struct dlm_member, list); next 392 fs/dlm/plock.c op = list_entry(send_list.next, struct plock_op, list); next 937 fs/dlm/recover.c struct rb_node *n, *next; next 944 fs/dlm/recover.c for (n = rb_first(&ls->ls_rsbtbl[i].toss); n; n = next) { next 945 fs/dlm/recover.c next = rb_next(n); next 77 fs/dlm/requestqueue.c e = list_entry(ls->ls_requestqueue.next, struct rq_entry, list); next 839 fs/dlm/user.c lkb = list_entry(proc->asts.next, struct dlm_lkb, lkb_cb_list); next 278 fs/erofs/utils.c p = erofs_sb_list.next; next 290 fs/erofs/utils.c p = p->next; next 301 fs/erofs/utils.c p = p->next; next 316 fs/erofs/zdata.c if (pcl->next == Z_EROFS_PCLUSTER_NIL) { next 318 fs/erofs/zdata.c if (cmpxchg(&pcl->next, Z_EROFS_PCLUSTER_NIL, next 322 fs/erofs/zdata.c *owned_head = &pcl->next; next 325 fs/erofs/zdata.c } else if (pcl->next == Z_EROFS_PCLUSTER_TAIL) { next 331 fs/erofs/zdata.c if (cmpxchg(&pcl->next, Z_EROFS_PCLUSTER_TAIL, next 355 fs/erofs/zdata.c if (clt->owned_head == &pcl->next || pcl == clt->tailpcl) { next 429 fs/erofs/zdata.c pcl->next = clt->owned_head; next 450 fs/erofs/zdata.c clt->owned_head = &pcl->next; next 947 fs/erofs/zdata.c WRITE_ONCE(pcl->next, Z_EROFS_PCLUSTER_NIL); next 971 fs/erofs/zdata.c pcl = container_of(owned, struct z_erofs_pcluster, next); next 972 fs/erofs/zdata.c owned = READ_ONCE(pcl->next); next 1179 fs/erofs/zdata.c WRITE_ONCE(pcl->next, Z_EROFS_PCLUSTER_TAIL_CLOSED); next 1182 fs/erofs/zdata.c WRITE_ONCE(*bypass_qtail, &pcl->next); next 1184 fs/erofs/zdata.c qtail[JQ_BYPASS] = &pcl->next; next 1241 fs/erofs/zdata.c pcl = container_of(owned_head, struct z_erofs_pcluster, next); next 1246 fs/erofs/zdata.c owned_head = cmpxchg(&pcl->next, Z_EROFS_PCLUSTER_TAIL, next 1292 fs/erofs/zdata.c qtail[JQ_SUBMIT] = &pcl->next; next 60 fs/erofs/zdata.h z_erofs_next_pcluster_t next; next 31 fs/erofs/zpvec.h struct page *curr, *next; next 56 fs/erofs/zpvec.h if (ctor->next) next 57 fs/erofs/zpvec.h return ctor->next; next 74 fs/erofs/zpvec.h struct page *next = z_erofs_pagevec_ctor_next_page(ctor, ctor->nr); next 78 fs/erofs/zpvec.h ctor->curr = next; next 79 fs/erofs/zpvec.h ctor->next = NULL; next 93 fs/erofs/zpvec.h ctor->curr = ctor->next = NULL; next 104 fs/erofs/zpvec.h ctor->next = z_erofs_pagevec_ctor_next_page(ctor, i); next 114 fs/erofs/zpvec.h if (!ctor->next && type) next 126 fs/erofs/zpvec.h if (type == (uintptr_t)ctor->next) { next 127 fs/erofs/zpvec.h ctor->next = page; next 141 fs/erofs/zpvec.h DBG_BUGON(!ctor->next); next 150 fs/erofs/zpvec.h if (*type == (uintptr_t)ctor->next) next 151 fs/erofs/zpvec.h ctor->next = tagptr_unfold_ptr(t); next 152 fs/eventpoll.c struct epitem *next; next 713 fs/eventpoll.c nepi = epi->next, epi->next = EP_UNACTIVE_PTR) { next 988 fs/eventpoll.c struct epitem *epi, *next; next 1004 fs/eventpoll.c list_for_each_entry_safe(epi, next, &file->f_ep_links, fllink) { next 1145 fs/eventpoll.c if (cmpxchg(&new->next, new, head) != new) next 1163 fs/eventpoll.c prev->next = new; next 1180 fs/eventpoll.c if (epi->next != EP_UNACTIVE_PTR) next 1184 fs/eventpoll.c if (cmpxchg(&epi->next, EP_UNACTIVE_PTR, NULL) != EP_UNACTIVE_PTR) next 1188 fs/eventpoll.c epi->next = xchg(&ep->ovflist, epi); next 1521 fs/eventpoll.c epi->next = EP_UNACTIVE_PTR; next 582 fs/ext2/balloc.c ext2_grpblk_t next; next 584 fs/ext2/balloc.c next = ext2_find_next_zero_bit(bh->b_data, maxblocks, start); next 585 fs/ext2/balloc.c if (next >= maxblocks) next 587 fs/ext2/balloc.c return next; next 605 fs/ext2/balloc.c ext2_grpblk_t here, next; next 632 fs/ext2/balloc.c next = (r - ((char *)bh->b_data)) << 3; next 634 fs/ext2/balloc.c if (next < maxblocks && next >= here) next 635 fs/ext2/balloc.c return next; next 783 fs/ext2/balloc.c struct rb_node *next; next 812 fs/ext2/balloc.c next = rb_next(&rsv->rsv_node); next 813 fs/ext2/balloc.c rsv = rb_entry(next,struct ext2_reserve_window_node,rsv_node); next 819 fs/ext2/balloc.c if (!next) next 1044 fs/ext2/balloc.c struct rb_node *next; next 1050 fs/ext2/balloc.c next = rb_next(&my_rsv->rsv_node); next 1052 fs/ext2/balloc.c if (!next) next 1055 fs/ext2/balloc.c next_rsv = rb_entry(next, struct ext2_reserve_window_node, rsv_node); next 152 fs/ext2/xattr.c struct ext2_xattr_entry *next; next 155 fs/ext2/xattr.c next = EXT2_XATTR_NEXT(entry); next 156 fs/ext2/xattr.c if ((char *)next >= end) next 392 fs/ext4/dir.c struct fname *next; next 405 fs/ext4/dir.c struct fname *fname, *next; next 407 fs/ext4/dir.c rbtree_postorder_for_each_entry_safe(fname, next, root, rb_hash) next 410 fs/ext4/dir.c fname = fname->next; next 480 fs/ext4/dir.c new_fn->next = fname->next; next 481 fs/ext4/dir.c fname->next = new_fn; next 529 fs/ext4/dir.c fname = fname->next; next 1965 fs/ext4/extents.c ext4_lblk_t next; next 2065 fs/ext4/extents.c next = EXT_MAX_BLOCKS; next 2067 fs/ext4/extents.c next = ext4_ext_next_leaf_block(path); next 2068 fs/ext4/extents.c if (next != EXT_MAX_BLOCKS) { next 2069 fs/ext4/extents.c ext_debug("next leaf block - %u\n", next); next 2071 fs/ext4/extents.c npath = ext4_find_extent(inode, next, NULL, 0); next 2183 fs/ext4/extents.c ext4_lblk_t next, next_del, start = 0, end = 0; next 2210 fs/ext4/extents.c next = ext4_ext_next_allocated_block(path); next 2230 fs/ext4/extents.c if (end >= next) next 2231 fs/ext4/extents.c end = next; next 2290 fs/ext4/extents.c if (next == next_del && next == EXT_MAX_BLOCKS) { next 2293 fs/ext4/extents.c next != EXT_MAX_BLOCKS)) { next 2297 fs/ext4/extents.c next, next_del); next 2329 fs/ext4/extents.c ext4_lblk_t next, end = block + num - 1; next 2336 fs/ext4/extents.c next = 0; next 2338 fs/ext4/extents.c if (!ext4_es_lookup_extent(inode, block, &next, &es)) next 2347 fs/ext4/extents.c if (next == 0) next 2359 fs/ext4/extents.c if (next == 0) next 2361 fs/ext4/extents.c block = next; next 2401 fs/ext4/extents.c ext4_lblk_t next; next 2404 fs/ext4/extents.c next = ext4_ext_next_allocated_block(path); next 2405 fs/ext4/extents.c BUG_ON(next == *lblk); next 2406 fs/ext4/extents.c len = next - *lblk; next 4236 fs/ext4/extents.c ext4_lblk_t next = ext4_ext_next_allocated_block(path); next 4237 fs/ext4/extents.c map->m_len = min(map->m_len, next - map->m_lblk); next 1696 fs/ext4/extents_status.c goto next; next 1699 fs/ext4/extents_status.c goto next; next 1705 fs/ext4/extents_status.c next: next 48 fs/ext4/fsync.c struct inode *next; next 59 fs/ext4/fsync.c next = igrab(d_inode(dentry->d_parent)); next 61 fs/ext4/fsync.c if (!next) next 64 fs/ext4/fsync.c inode = next; next 712 fs/ext4/ialloc.c next: next 723 fs/ext4/ialloc.c goto next; next 1102 fs/ext4/inode.c struct buffer_head *next; next 1106 fs/ext4/inode.c block_start = block_end, bh = next) { next 1107 fs/ext4/inode.c next = bh->b_this_page; next 1496 fs/ext4/mballoc.c int next = block; next 1522 fs/ext4/mballoc.c next = next - ex->fe_start; next 1523 fs/ext4/mballoc.c ex->fe_len -= next; next 1524 fs/ext4/mballoc.c ex->fe_start += next; next 1532 fs/ext4/mballoc.c next = (block + 1) * (1 << order); next 1533 fs/ext4/mballoc.c if (mb_test_bit(next, e4b->bd_bitmap)) next 1536 fs/ext4/mballoc.c order = mb_find_order_for_block(e4b, next); next 1538 fs/ext4/mballoc.c block = next >> order; next 2339 fs/ext4/mballoc.c .next = ext4_mb_seq_groups_next, next 3823 fs/ext4/mballoc.c unsigned int next; next 3839 fs/ext4/mballoc.c next = mb_find_next_bit(bitmap_bh->b_data, end, bit); next 3842 fs/ext4/mballoc.c (unsigned) next - bit, (unsigned) group); next 3843 fs/ext4/mballoc.c free += next - bit; next 3845 fs/ext4/mballoc.c trace_ext4_mballoc_discard(sb, NULL, group, bit, next - bit); next 3848 fs/ext4/mballoc.c next - bit); next 3849 fs/ext4/mballoc.c mb_free_blocks(pa->pa_inode, e4b, bit, next - bit); next 3850 fs/ext4/mballoc.c bit = next + 1; next 4035 fs/ext4/mballoc.c pa = list_entry(ei->i_prealloc_list.next, next 5182 fs/ext4/mballoc.c ext4_grpblk_t next, count = 0, free_count = 0; next 5208 fs/ext4/mballoc.c next = mb_find_next_bit(bitmap, max + 1, start); next 5210 fs/ext4/mballoc.c if ((next - start) >= minblocks) { next 5212 fs/ext4/mballoc.c next - start, group, &e4b); next 5216 fs/ext4/mballoc.c count += next - start; next 5218 fs/ext4/mballoc.c free_count += next - start; next 5219 fs/ext4/mballoc.c start = next + 1; next 5352 fs/ext4/mballoc.c ext4_grpblk_t next; next 5372 fs/ext4/mballoc.c next = mb_find_next_bit(bitmap, end + 1, start); next 5375 fs/ext4/mballoc.c error = formatter(sb, group, start, next - start, priv); next 5380 fs/ext4/mballoc.c start = next + 1; next 1528 fs/ext4/namei.c goto next; next 1559 fs/ext4/namei.c next: next 1787 fs/ext4/namei.c struct ext4_dir_entry_2 *next, *to, *prev, *de = (struct ext4_dir_entry_2 *) base; next 1792 fs/ext4/namei.c next = ext4_next_entry(de, blocksize); next 1801 fs/ext4/namei.c de = next; next 149 fs/ext4/page-io.c io, inode->i_ino, io->list.next, io->list.prev); next 179 fs/ext4/page-io.c after = cur->next; next 222 fs/ext4/page-io.c io = list_entry(unwritten.next, ext4_io_end_t, list); next 425 fs/ext4/super.c jce = list_entry(txn->t_private_list.next, next 3173 fs/ext4/super.c n = pos->next; next 190 fs/ext4/xattr.c struct ext4_xattr_entry *next = EXT4_XATTR_NEXT(e); next 191 fs/ext4/xattr.c if ((void *)next >= end) next 195 fs/ext4/xattr.c e = next; next 284 fs/ext4/xattr.c struct ext4_xattr_entry *entry, *next; next 291 fs/ext4/xattr.c for (entry = *pentry; !IS_LAST_ENTRY(entry); entry = next) { next 292 fs/ext4/xattr.c next = EXT4_XATTR_NEXT(entry); next 293 fs/ext4/xattr.c if ((void *) next >= end) { next 1565 fs/ext4/xattr.c struct ext4_xattr_entry *last, *next; next 1600 fs/ext4/xattr.c for (; !IS_LAST_ENTRY(last); last = next) { next 1601 fs/ext4/xattr.c next = EXT4_XATTR_NEXT(last); next 1602 fs/ext4/xattr.c if ((void *)next >= s->end) { next 610 fs/f2fs/data.c next: next 662 fs/f2fs/data.c goto next; next 1575 fs/f2fs/data.c next: next 1620 fs/f2fs/data.c goto next; next 565 fs/f2fs/dir.c next: next 578 fs/f2fs/dir.c goto next; next 172 fs/f2fs/extent_cache.c struct rb_node *cur = rb_first_cached(root), *next; next 179 fs/f2fs/extent_cache.c next = rb_next(cur); next 180 fs/f2fs/extent_cache.c if (!next) next 184 fs/f2fs/extent_cache.c next_re = rb_entry(next, struct rb_entry, rb_node); next 193 fs/f2fs/extent_cache.c cur = next; next 302 fs/f2fs/extent_cache.c struct rb_node *node, *next; next 308 fs/f2fs/extent_cache.c next = rb_next(node); next 311 fs/f2fs/extent_cache.c node = next; next 633 fs/f2fs/extent_cache.c struct extent_tree *et, *next; next 648 fs/f2fs/extent_cache.c list_for_each_entry_safe(et, next, &sbi->zombie_list, list) { next 1025 fs/f2fs/file.c goto next; next 1054 fs/f2fs/file.c next: next 87 fs/f2fs/gc.c goto next; next 94 fs/f2fs/gc.c goto next; next 113 fs/f2fs/gc.c next: next 392 fs/f2fs/gc.c goto next; next 398 fs/f2fs/gc.c goto next; next 403 fs/f2fs/gc.c goto next; next 405 fs/f2fs/gc.c goto next; next 413 fs/f2fs/gc.c next: next 2490 fs/f2fs/node.c struct free_nid *i, *next; next 2500 fs/f2fs/node.c list_for_each_entry_safe(i, next, &nm_i->free_nid_list, list) { next 332 fs/f2fs/recovery.c goto next; next 358 fs/f2fs/recovery.c goto next; next 368 fs/f2fs/recovery.c next: next 680 fs/f2fs/recovery.c goto next; next 708 fs/f2fs/recovery.c next: next 253 fs/f2fs/segment.c goto next; next 270 fs/f2fs/segment.c next: next 293 fs/f2fs/segment.c next: next 321 fs/f2fs/segment.c goto next; next 608 fs/f2fs/segment.c struct flush_cmd *cmd, *next; next 619 fs/f2fs/segment.c llist_for_each_entry_safe(cmd, next, next 682 fs/f2fs/segment.c struct flush_cmd *tmp, *next; next 686 fs/f2fs/segment.c llist_for_each_entry_safe(tmp, next, list, llnode) { next 1341 fs/f2fs/segment.c goto next; next 1373 fs/f2fs/segment.c next: next 1432 fs/f2fs/segment.c goto next; next 1444 fs/f2fs/segment.c next: next 1492 fs/f2fs/segment.c goto next; next 1516 fs/f2fs/segment.c next: next 1587 fs/f2fs/segment.c next: next 1611 fs/f2fs/segment.c goto next; next 1979 fs/f2fs/segment.c next: next 1989 fs/f2fs/segment.c goto next; next 2755 fs/f2fs/segment.c next: next 2797 fs/f2fs/segment.c goto next; next 3757 fs/f2fs/segment.c struct sit_entry_set *next = ses; next 3762 fs/f2fs/segment.c list_for_each_entry_continue(next, head, set_list) next 3763 fs/f2fs/segment.c if (ses->entry_cnt <= next->entry_cnt) next 3766 fs/f2fs/segment.c list_move_tail(&ses->set_list, &next->set_list); next 419 fs/f2fs/segment.h unsigned int next; next 425 fs/f2fs/segment.h next = find_next_bit(free_i->free_segmap, next 427 fs/f2fs/segment.h if (next >= start_segno + sbi->segs_per_sec) { next 452 fs/f2fs/segment.h unsigned int next; next 460 fs/f2fs/segment.h next = find_next_bit(free_i->free_segmap, next 462 fs/f2fs/segment.h if (next >= start_segno + sbi->segs_per_sec) { next 47 fs/f2fs/shrinker.c p = f2fs_list.next; next 53 fs/f2fs/shrinker.c p = p->next; next 68 fs/f2fs/shrinker.c p = p->next; next 88 fs/f2fs/shrinker.c p = f2fs_list.next; next 97 fs/f2fs/shrinker.c p = p->next; next 116 fs/f2fs/shrinker.c p = p->next; next 692 fs/f2fs/xattr.c struct f2fs_xattr_entry *next = XATTR_NEXT_ENTRY(here); next 695 fs/f2fs/xattr.c memmove(here, next, (char *)last - (char *)next); next 76 fs/fat/cache.c if (MSDOS_I(inode)->cache_lru.next != &cache->cache_list) next 192 fs/fat/cache.c cache = list_entry(i->cache_lru.next, next 90 fs/fat/dir.c next: next 106 fs/fat/dir.c goto next; next 121 fs/fat/fatent.c int next; next 125 fs/fat/fatent.c next = (*ent12_p[0] >> 4) | (*ent12_p[1] << 4); next 127 fs/fat/fatent.c next = (*ent12_p[1] << 8) | *ent12_p[0]; next 130 fs/fat/fatent.c next &= 0x0fff; next 131 fs/fat/fatent.c if (next >= BAD_FAT12) next 132 fs/fat/fatent.c next = FAT_ENT_EOF; next 133 fs/fat/fatent.c return next; next 138 fs/fat/fatent.c int next = le16_to_cpu(*fatent->u.ent16_p); next 140 fs/fat/fatent.c if (next >= BAD_FAT16) next 141 fs/fat/fatent.c next = FAT_ENT_EOF; next 142 fs/fat/fatent.c return next; next 147 fs/fat/fatent.c int next = le32_to_cpu(*fatent->u.ent32_p) & 0x0fffffff; next 149 fs/fat/fatent.c if (next >= BAD_FAT32) next 150 fs/fat/fatent.c next = FAT_ENT_EOF; next 151 fs/fat/fatent.c return next; next 52 fs/filesystems.c for (p = &file_systems; *p; p = &(*p)->next) next 81 fs/filesystems.c if (fs->next) next 115 fs/filesystems.c *tmp = fs->next; next 116 fs/filesystems.c fs->next = NULL; next 121 fs/filesystems.c tmp = &(*tmp)->next; next 144 fs/filesystems.c for (tmp=file_systems, index=0 ; tmp ; tmp=tmp->next, index++) { next 161 fs/filesystems.c for (tmp = file_systems; tmp; tmp = tmp->next, index--) next 181 fs/filesystems.c for (tmp = file_systems, index = 0 ; tmp ; tmp = tmp->next, index++) next 222 fs/filesystems.c tmp = tmp->next; next 239 fs/filesystems.c tmp = tmp->next; next 1180 fs/fs-writeback.c tail = wb_inode(wb->b_dirty.next); next 1942 fs/fs-writeback.c work = list_entry(wb->work_list.next, next 160 fs/fscache/cache.c cache = list_entry(fscache_cache_list.next, next 334 fs/fscache/cache.c object = list_entry(cache->object_list.next, next 85 fs/fscache/histogram.c .next = fscache_histogram_next, next 304 fs/fscache/object-list.c .next = fscache_objlist_next, next 880 fs/fscache/object.c dep = list_entry(object->dependents.next, next 336 fs/fscache/operation.c op = list_entry(object->pending_ops.next, next 428 fs/fscache/operation.c op = list_entry(object->pending_ops.next, next 578 fs/fscache/operation.c op = list_entry(cache->op_gc_list.next, next 1220 fs/fscache/page.c pgoff_t next; next 1231 fs/fscache/page.c next = 0; next 1233 fs/fscache/page.c if (!pagevec_lookup(&pvec, mapping, &next)) next 1244 fs/fscache/page.c } while (next); next 242 fs/fuse/dev.c fiq->forget_list_tail->next = forget; next 1004 fs/fuse/dev.c return fiq->forget_list_head.next != NULL; next 1055 fs/fuse/dev.c struct fuse_forget_link *head = fiq->forget_list_head.next; next 1060 fs/fuse/dev.c newhead = &(*newhead)->next; next 1062 fs/fuse/dev.c fiq->forget_list_head.next = *newhead; next 1064 fs/fuse/dev.c if (fiq->forget_list_head.next == NULL) next 1144 fs/fuse/dev.c head = forget->next; next 1161 fs/fuse/dev.c if (fc->minor < 16 || fiq->forget_list_head.next->next == NULL) next 1227 fs/fuse/dev.c req = list_entry(fiq->interrupts.next, struct fuse_req, next 1240 fs/fuse/dev.c req = list_entry(fiq->pending.next, struct fuse_req, list); next 2047 fs/fuse/dev.c req = list_entry(head->next, struct fuse_req, list); next 2095 fs/fuse/dev.c struct fuse_req *req, *next; next 2110 fs/fuse/dev.c list_for_each_entry_safe(req, next, &fpq->io, list) { next 362 fs/fuse/file.c struct fuse_writepage_args *next; next 1632 fs/fuse/file.c struct fuse_writepage_args *aux, *next; next 1672 fs/fuse/file.c for (aux = wpa->next; aux; aux = next) { next 1673 fs/fuse/file.c next = aux->next; next 1674 fs/fuse/file.c aux->next = NULL; next 1698 fs/fuse/file.c wpa = list_entry(fi->queued_writes.next, next 1715 fs/fuse/file.c while (wpa->next) { next 1718 fs/fuse/file.c struct fuse_writepage_args *next = wpa->next; next 1720 fs/fuse/file.c wpa->next = next->next; next 1721 fs/fuse/file.c next->next = NULL; next 1722 fs/fuse/file.c next->ia.ff = fuse_file_get(wpa->ia.ff); next 1723 fs/fuse/file.c list_add(&next->writepages_entry, &fi->writepages); next 1748 fs/fuse/file.c fuse_send_writepage(fc, next, inarg->offset + inarg->size); next 1763 fs/fuse/file.c ff = list_entry(fi->write_files.next, struct fuse_file, next 1845 fs/fuse/file.c wpa->next = NULL; next 1979 fs/fuse/file.c for (tmp = old_wpa->next; tmp; tmp = tmp->next) { next 1992 fs/fuse/file.c new_wpa->next = old_wpa->next; next 1993 fs/fuse/file.c old_wpa->next = new_wpa; next 2081 fs/fuse/file.c wpa->next = NULL; next 63 fs/fuse/fuse_i.h struct fuse_forget_link *next; next 455 fs/fuse/virtio_fs.c struct fuse_req *next; next 475 fs/fuse/virtio_fs.c list_for_each_entry_safe(req, next, &reqs, list) { next 2227 fs/gfs2/bmap.c jext = list_entry(jd->extent_list.next, struct gfs2_journal_extent, list); next 1009 fs/gfs2/dir.c struct gfs2_dirent *dent = NULL, *prev = NULL, *next = NULL, *new; next 1087 fs/gfs2/dir.c next = dent; next 1088 fs/gfs2/dir.c if (dirent_next(dip, obh, &next)) next 1089 fs/gfs2/dir.c next = NULL; next 1119 fs/gfs2/dir.c dent = next; next 998 fs/gfs2/file.c loff_t next = (offset + len - 1) >> sdp->sd_sb.sb_bsize_shift; next 1001 fs/gfs2/file.c next = (next + 1) << sdp->sd_sb.sb_bsize_shift; next 1005 fs/gfs2/file.c len = next - offset; next 284 fs/gfs2/glock.c const struct gfs2_holder *gh_head = list_entry(gl->gl_holders.next, const struct gfs2_holder, gh_list); next 610 fs/gfs2/glock.c gh = list_entry(gl->gl_holders.next, struct gfs2_holder, gh_list); next 831 fs/gfs2/glock.c gl->gl_node.next = NULL; next 1157 fs/gfs2/glock.c gh = list_entry(gl->gl_holders.next, struct gfs2_holder, gh_list); next 1552 fs/gfs2/glock.c gl = list_entry(list->next, struct gfs2_glock, gl_lru); next 1593 fs/gfs2/glock.c gl = list_entry(lru_list.next, struct gfs2_glock, gl_lru); next 2142 fs/gfs2/glock.c .next = gfs2_glock_seq_next, next 2149 fs/gfs2/glock.c .next = gfs2_glock_seq_next, next 2156 fs/gfs2/glock.c .next = gfs2_sbstats_seq_next, next 553 fs/gfs2/log.c ip = list_entry(sdp->sd_log_ordered.next, struct gfs2_inode, i_ordered); next 574 fs/gfs2/log.c ip = list_entry(sdp->sd_log_ordered.next, struct gfs2_inode, i_ordered); next 174 fs/gfs2/lops.c struct buffer_head *bh, *next; next 186 fs/gfs2/lops.c next = bh->b_this_page; next 189 fs/gfs2/lops.c bh = next; next 732 fs/gfs2/lops.c bd = list_entry(head->next, struct gfs2_bufdata, bd_list); next 902 fs/gfs2/lops.c bd = list_entry(head->next, struct gfs2_bufdata, bd_list); next 1081 fs/gfs2/lops.c bd = list_entry(head->next, struct gfs2_bufdata, bd_list); next 76 fs/gfs2/meta_io.c struct buffer_head *next = bh->b_this_page; next 81 fs/gfs2/meta_io.c bh = next; next 199 fs/gfs2/meta_io.c struct buffer_head *next = bh->b_this_page; next 202 fs/gfs2/meta_io.c bh = next; next 118 fs/gfs2/quota.c qd = list_entry(list->next, struct gfs2_quota_data, qd_lru); next 1506 fs/gfs2/quota.c ip = list_entry(sdp->sd_trunc_list.next, next 114 fs/gfs2/recovery.c rr = list_entry(head->next, struct gfs2_revoke_replay, rr_list); next 510 fs/gfs2/rgrp.c struct rb_node *n, *next; next 517 fs/gfs2/rgrp.c next = NULL; next 519 fs/gfs2/rgrp.c next = n->rb_left; next 521 fs/gfs2/rgrp.c next = n->rb_right; next 522 fs/gfs2/rgrp.c if (next == NULL) { next 532 fs/gfs2/rgrp.c n = next; next 1088 fs/gfs2/rgrp.c struct gfs2_rgrpd *next = gfs2_rgrpd_get_next(rgd); next 1095 fs/gfs2/rgrp.c if (next == NULL) next 1097 fs/gfs2/rgrp.c else if (next->rd_addr > rgd->rd_addr) next 1098 fs/gfs2/rgrp.c str->rg_skip = cpu_to_be32(next->rd_addr - rgd->rd_addr); next 65 fs/gfs2/super.c jd = list_entry(list.next, struct gfs2_jdesc, jd_list); next 448 fs/gfs2/super.c lfcc = list_entry(list.next, struct lfcc, list); next 195 fs/hfs/bfind.c idx = bnode->next; next 143 fs/hfs/bnode.c be32_to_cpu(desc.next), be32_to_cpu(desc.prev), next 182 fs/hfs/bnode.c tmp->next = node->next; next 183 fs/hfs/bnode.c cnid = cpu_to_be32(tmp->next); next 184 fs/hfs/bnode.c hfs_bnode_write(tmp, &cnid, offsetof(struct hfs_bnode_desc, next), 4); next 187 fs/hfs/bnode.c tree->leaf_head = node->next; next 189 fs/hfs/bnode.c if (node->next) { next 190 fs/hfs/bnode.c tmp = hfs_bnode_find(tree, node->next); next 201 fs/hfs/bnode.c if (!node->prev && !node->next) { next 340 fs/hfs/bnode.c node->next = be32_to_cpu(desc->next); next 246 fs/hfs/brec.c node->this, new_node->this, node->next); next 247 fs/hfs/brec.c new_node->next = node->next; next 253 fs/hfs/brec.c if (node->next) next 254 fs/hfs/brec.c next_node = hfs_bnode_find(tree, node->next); next 317 fs/hfs/brec.c node_desc.next = cpu_to_be32(new_node->next); next 326 fs/hfs/brec.c node->next = new_node->this; next 328 fs/hfs/brec.c node_desc.next = cpu_to_be32(node->next); next 487 fs/hfs/brec.c new_node->next = 0; next 491 fs/hfs/brec.c node_desc.next = cpu_to_be32(new_node->next); next 202 fs/hfs/btree.c prev->next = idx; next 204 fs/hfs/btree.c hfs_bnode_write(prev, &cnid, offsetof(struct hfs_bnode_desc, next), 4); next 209 fs/hfs/btree.c desc.next = 0; next 301 fs/hfs/btree.c nidx = node->next; next 340 fs/hfs/btree.c i = node->next; next 51 fs/hfs/btree.h u32 next; next 127 fs/hfs/btree.h __be32 next; /* (V) Number of the next node at this level */ next 264 fs/hfsplus/bfind.c idx = bnode->next; next 307 fs/hfsplus/bnode.c be32_to_cpu(desc.next), be32_to_cpu(desc.prev), next 346 fs/hfsplus/bnode.c tmp->next = node->next; next 347 fs/hfsplus/bnode.c cnid = cpu_to_be32(tmp->next); next 349 fs/hfsplus/bnode.c offsetof(struct hfs_bnode_desc, next), 4); next 352 fs/hfsplus/bnode.c tree->leaf_head = node->next; next 354 fs/hfsplus/bnode.c if (node->next) { next 355 fs/hfsplus/bnode.c tmp = hfs_bnode_find(tree, node->next); next 367 fs/hfsplus/bnode.c if (!node->prev && !node->next) next 508 fs/hfsplus/bnode.c node->next = be32_to_cpu(desc->next); next 250 fs/hfsplus/brec.c node->this, new_node->this, node->next); next 251 fs/hfsplus/brec.c new_node->next = node->next; next 257 fs/hfsplus/brec.c if (node->next) next 258 fs/hfsplus/brec.c next_node = hfs_bnode_find(tree, node->next); next 321 fs/hfsplus/brec.c node_desc.next = cpu_to_be32(new_node->next); next 330 fs/hfsplus/brec.c node->next = new_node->this; next 332 fs/hfsplus/brec.c node_desc.next = cpu_to_be32(node->next); next 491 fs/hfsplus/brec.c new_node->next = 0; next 495 fs/hfsplus/brec.c node_desc.next = cpu_to_be32(new_node->next); next 324 fs/hfsplus/btree.c prev->next = idx; next 326 fs/hfsplus/btree.c hfs_bnode_write(prev, &cnid, offsetof(struct hfs_bnode_desc, next), 4); next 331 fs/hfsplus/btree.c desc.next = 0; next 428 fs/hfsplus/btree.c nidx = node->next; next 468 fs/hfsplus/btree.c i = node->next; next 220 fs/hfsplus/dir.c goto next; next 256 fs/hfsplus/dir.c next: next 108 fs/hfsplus/hfsplus_fs.h u32 next; next 152 fs/hfsplus/hfsplus_raw.h __be32 next; next 98 fs/hfsplus/xattr.c desc->next = cpu_to_be32(be32_to_cpu(head->leaf_tail) + 1); next 277 fs/hostfs/hostfs_kern.c unsigned long long next, ino; next 288 fs/hostfs/hostfs_kern.c next = ctx->pos; next 289 fs/hostfs/hostfs_kern.c seek_dir(dir, next); next 290 fs/hostfs/hostfs_kern.c while ((name = read_dir(dir, &next, &ino, &len, &type)) != NULL) { next 293 fs/hostfs/hostfs_kern.c ctx->pos = next; next 423 fs/hugetlbfs/inode.c pgoff_t next, index; next 430 fs/hugetlbfs/inode.c next = start; next 431 fs/hugetlbfs/inode.c while (next < end) { next 435 fs/hugetlbfs/inode.c if (!pagevec_lookup_range(&pvec, mapping, &next, end - 1)) next 626 fs/inode.c struct inode *inode, *next; next 631 fs/inode.c list_for_each_entry_safe(inode, next, &sb->s_inodes, i_sb_list) { next 677 fs/inode.c struct inode *inode, *next; next 682 fs/inode.c list_for_each_entry_safe(inode, next, &sb->s_inodes, i_sb_list) { next 899 fs/inode.c int next = atomic_add_return(LAST_INO_BATCH, &shared_last_ino); next 901 fs/inode.c res = next - LAST_INO_BATCH; next 1757 fs/io_uring.c struct io_kiocb *poll_req, *next; next 1767 fs/io_uring.c list_for_each_entry_safe(poll_req, next, &ctx->cancel_list, list) { next 389 fs/iomap/buffered-io.c loff_t last = page_offset(list_entry(pages->next, struct page, lru)); next 1031 fs/jbd2/journal.c .next = jbd2_seq_info_next, next 70 fs/jbd2/recovery.c unsigned int max, nbufs, next; next 86 fs/jbd2/recovery.c for (next = start; next < max; next++) { next 87 fs/jbd2/recovery.c err = jbd2_journal_bmap(journal, next, &blocknr); next 91 fs/jbd2/recovery.c next); next 169 fs/jbd2/revoke.c record = (struct jbd2_revoke_record_s *) hash_list->next; next 175 fs/jbd2/revoke.c record = (struct jbd2_revoke_record_s *) record->hash.next; next 547 fs/jbd2/revoke.c hash_list->next; next 732 fs/jbd2/revoke.c record = (struct jbd2_revoke_record_s*) hash_list->next; next 2374 fs/jbd2/transaction.c struct buffer_head *head, *bh, *next; next 2395 fs/jbd2/transaction.c next = bh->b_this_page; next 2410 fs/jbd2/transaction.c bh = next; next 40 fs/jffs2/build.c if (ic->next) next 41 fs/jffs2/build.c return ic->next; next 61 fs/jffs2/build.c for(fd = ic->scan_dents; fd; fd = fd->next) { next 151 fs/jffs2/build.c dead_fds = fd->next; next 181 fs/jffs2/build.c ic->scan_dents = fd->next; next 229 fs/jffs2/build.c ic->scan_dents = fd->next; next 250 fs/jffs2/build.c struct jffs2_raw_node_ref *next = raw->next_in_ino; next 253 fs/jffs2/build.c raw = next; next 264 fs/jffs2/build.c ic->scan_dents = fd->next; next 292 fs/jffs2/build.c fd->next = *dead_fds; next 97 fs/jffs2/dir.c for (fd_list = dir_f->dents; fd_list && fd_list->nhash <= nhash; fd_list = fd_list->next) { next 133 fs/jffs2/dir.c for (fd = f->dents; fd; fd = fd->next) { next 593 fs/jffs2/dir.c for (fd = f->dents ; fd; fd = fd->next) { next 779 fs/jffs2/dir.c for (fd = victim_f->dents; fd; fd = fd->next) { next 114 fs/jffs2/erase.c jeb = list_entry(c->erase_complete_list.next, struct jffs2_eraseblock, list); next 127 fs/jffs2/erase.c jeb = list_entry(c->erase_pending_list.next, struct jffs2_eraseblock, list); next 304 fs/jffs2/fs.c for (fd=f->dents; fd; fd = fd->next) { next 97 fs/jffs2/gc.c ret = list_entry(nextlist->next, struct jffs2_eraseblock, list); next 159 fs/jffs2/gc.c for (ic = c->inocache_list[bucket]; ic; ic = ic->next) { next 566 fs/jffs2/gc.c for (fd = f->dents; fd; fd=fd->next) { next 999 fs/jffs2/gc.c *fdp = fd->next; next 1002 fs/jffs2/gc.c fdp = &(*fdp)->next; next 43 fs/jffs2/nodelist.c new->next = (*prev)->next; next 53 fs/jffs2/nodelist.c prev = &((*prev)->next); next 55 fs/jffs2/nodelist.c new->next = *prev; next 73 fs/jffs2/nodelist.c struct jffs2_node_frag *next = frag_next(frag); next 77 fs/jffs2/nodelist.c frag = next; next 395 fs/jffs2/nodelist.c struct jffs2_node_frag *next = frag_next(newfrag); next 397 fs/jffs2/nodelist.c if (next) { next 399 fs/jffs2/nodelist.c if (next->node) next 400 fs/jffs2/nodelist.c mark_ref_normal(next->node->raw); next 427 fs/jffs2/nodelist.c ret = ret->next; next 449 fs/jffs2/nodelist.c prev = &(*prev)->next; next 451 fs/jffs2/nodelist.c new->next = *prev; next 470 fs/jffs2/nodelist.c prev = &(*prev)->next; next 473 fs/jffs2/nodelist.c *prev = old->next; next 490 fs/jffs2/nodelist.c struct jffs2_inode_cache *this, *next; next 495 fs/jffs2/nodelist.c next = this->next; next 498 fs/jffs2/nodelist.c this = next; next 507 fs/jffs2/nodelist.c struct jffs2_raw_node_ref *this, *next; next 513 fs/jffs2/nodelist.c next = this[REFS_PER_BLOCK].next_in_ino; next 515 fs/jffs2/nodelist.c next = NULL; next 518 fs/jffs2/nodelist.c this = next; next 528 fs/jffs2/nodelist.c struct rb_node *next; next 534 fs/jffs2/nodelist.c next = fragtree->rb_node; next 536 fs/jffs2/nodelist.c while(next) { next 537 fs/jffs2/nodelist.c frag = rb_entry(next, struct jffs2_node_frag, rb); next 543 fs/jffs2/nodelist.c next = frag->rb.rb_right; next 545 fs/jffs2/nodelist.c next = frag->rb.rb_left; next 567 fs/jffs2/nodelist.c struct jffs2_node_frag *frag, *next; next 570 fs/jffs2/nodelist.c rbtree_postorder_for_each_entry_safe(frag, next, root, rb) { next 175 fs/jffs2/nodelist.h struct jffs2_inode_cache *next; next 257 fs/jffs2/nodelist.h struct jffs2_full_dirent *next; next 282 fs/jffs2/nodemgmt.c struct list_head *next; next 292 fs/jffs2/nodemgmt.c ejeb = list_entry(c->erasable_list.next, struct jffs2_eraseblock, list); next 334 fs/jffs2/nodemgmt.c next = c->free_list.next; next 335 fs/jffs2/nodemgmt.c list_del(next); next 336 fs/jffs2/nodemgmt.c c->nextblock = list_entry(next, struct jffs2_eraseblock, list); next 177 fs/jffs2/readinode.c struct rb_node *next; next 182 fs/jffs2/readinode.c next = tn_root->rb_node; next 184 fs/jffs2/readinode.c while (next) { next 185 fs/jffs2/readinode.c tn = rb_entry(next, struct jffs2_tmp_dnode_info, rb); next 188 fs/jffs2/readinode.c next = tn->rb.rb_right; next 190 fs/jffs2/readinode.c next = tn->rb.rb_left; next 303 fs/jffs2/readinode.c struct jffs2_tmp_dnode_info *next = tn_next(this); next 311 fs/jffs2/readinode.c this = next; next 546 fs/jffs2/readinode.c struct jffs2_tmp_dnode_info *tn, *next; next 548 fs/jffs2/readinode.c rbtree_postorder_for_each_entry_safe(tn, next, list, rb) { next 558 fs/jffs2/readinode.c struct jffs2_full_dirent *next; next 561 fs/jffs2/readinode.c next = fd->next; next 563 fs/jffs2/readinode.c fd = next; next 678 fs/jffs2/readinode.c fd->next = NULL; next 1420 fs/jffs2/readinode.c fds = fd->next; next 427 fs/jffs2/scan.c ref->next = c->xref_temp; next 1103 fs/jffs2/scan.c fd->next = NULL; next 1132 fs/jffs2/scan.c struct list_head *n = head->next; next 1136 fs/jffs2/scan.c n = n->next; next 68 fs/jffs2/summary.c s->sum_list_tail->u.next = (union jffs2_sum_mem *) item; next 128 fs/jffs2/summary.c temp->next = NULL; next 150 fs/jffs2/summary.c temp->next = NULL; next 171 fs/jffs2/summary.c temp->next = NULL; next 186 fs/jffs2/summary.c temp->next = NULL; next 202 fs/jffs2/summary.c s->sum_list_head = s->sum_list_head->u.next; next 276 fs/jffs2/summary.c temp->next = NULL; next 296 fs/jffs2/summary.c temp->next = NULL; next 326 fs/jffs2/summary.c temp->next = NULL; next 337 fs/jffs2/summary.c temp->next = NULL; next 476 fs/jffs2/summary.c fd->next = NULL; next 537 fs/jffs2/summary.c ref->next = c->xref_temp; next 792 fs/jffs2/summary.c c->summary->sum_list_head = temp->u.next; next 95 fs/jffs2/summary.h union jffs2_sum_mem *next; next 101 fs/jffs2/summary.h union jffs2_sum_mem *next; next 111 fs/jffs2/summary.h union jffs2_sum_mem *next; next 125 fs/jffs2/summary.h union jffs2_sum_mem *next; next 135 fs/jffs2/summary.h union jffs2_sum_mem *next; next 43 fs/jffs2/wbuf.c struct jffs2_inodirty *next; next 64 fs/jffs2/wbuf.c this = this->next; next 77 fs/jffs2/wbuf.c struct jffs2_inodirty *next = this->next; next 79 fs/jffs2/wbuf.c this = next; next 103 fs/jffs2/wbuf.c new->next = c->wbuf_inodes; next 110 fs/jffs2/wbuf.c struct list_head *this, *next; next 116 fs/jffs2/wbuf.c list_for_each_safe(this, next, &c->erasable_pending_wbuf_list) { next 213 fs/jffs2/wbuf.c for (fd = f->dents; fd; fd = fd->next) { next 278 fs/jffs2/wbuf.c struct jffs2_raw_node_ref *raw, *next, *first_raw = NULL; next 298 fs/jffs2/wbuf.c for (next = raw = jeb->first_node; next; raw = next) { next 299 fs/jffs2/wbuf.c next = ref_next(raw); next 302 fs/jffs2/wbuf.c (next && ref_offset(next) <= c->wbuf_ofs)) { next 609 fs/jffs2/write.c for (fd = dir_f->dents; fd; fd = fd->next) { next 641 fs/jffs2/write.c dead_f->dents = fd->next; next 570 fs/jffs2/xattr.c ref->next = ic->xref; next 586 fs/jffs2/xattr.c ref->next = c->xref_dead_list; next 607 fs/jffs2/xattr.c _ref = ref->next; next 622 fs/jffs2/xattr.c _ref = ref->next; next 648 fs/jffs2/xattr.c for (ref=ic->xref, pref=&ic->xref; ref; pref=&ref->next, ref=ref->next) { next 652 fs/jffs2/xattr.c *pref = ref->next; next 658 fs/jffs2/xattr.c for (cmp=ref->next, pcmp=&ref->next; cmp; pcmp=&cmp->next, cmp=cmp->next) { next 664 fs/jffs2/xattr.c *pcmp = cmp->next; next 673 fs/jffs2/xattr.c *pcmp = cmp->next; next 676 fs/jffs2/xattr.c *pref = ref->next; next 747 fs/jffs2/xattr.c _ref = ref->next; next 752 fs/jffs2/xattr.c _ref = ref->next; next 793 fs/jffs2/xattr.c _ref = ref->next; next 805 fs/jffs2/xattr.c for (tmp=xref_tmphash[i]; tmp; tmp=tmp->next) { next 822 fs/jffs2/xattr.c ref->next = xref_tmphash[i]; next 832 fs/jffs2/xattr.c _ref = ref->next; next 834 fs/jffs2/xattr.c ref->next = c->xref_dead_list; next 847 fs/jffs2/xattr.c ref->next = c->xref_dead_list; next 855 fs/jffs2/xattr.c ref->next = ic->xref; next 981 fs/jffs2/xattr.c for (ref=ic->xref, pref=&ic->xref; ref; pref=&ref->next, ref=ref->next) { next 994 fs/jffs2/xattr.c *pref = ref->next; next 1047 fs/jffs2/xattr.c for (ref=ic->xref, pref=&ic->xref; ref; pref=&ref->next, ref=ref->next) { next 1063 fs/jffs2/xattr.c *pref = ref->next; next 1119 fs/jffs2/xattr.c for (ref=ic->xref, pref=&ic->xref; ref; pref=&ref->next, ref=ref->next) { next 1126 fs/jffs2/xattr.c *pref = ref->next; next 1143 fs/jffs2/xattr.c *pref = ref->next; next 1145 fs/jffs2/xattr.c ref->next = c->xref_dead_list; next 1189 fs/jffs2/xattr.c *pref = ref->next; next 1193 fs/jffs2/xattr.c ref->next = ic->xref; next 1340 fs/jffs2/xattr.c for (tmp=c->xref_dead_list, ptmp=&c->xref_dead_list; tmp; ptmp=&tmp->next, tmp=tmp->next) { next 1342 fs/jffs2/xattr.c *ptmp = tmp->next; next 62 fs/jffs2/xattr.h struct jffs2_xattr_ref *next; /* chained from ic->xref_list */ next 483 fs/jfs/jfs_dtree.c static void free_index(tid_t tid, struct inode *ip, u32 index, u32 next) next 496 fs/jfs/jfs_dtree.c dirtab_slot->addr2 = cpu_to_le32(next); next 1398 fs/jfs/jfs_dtree.c nextbn = le64_to_cpu(sp->header.next); next 1399 fs/jfs/jfs_dtree.c rp->header.next = cpu_to_le64(nextbn); next 1401 fs/jfs/jfs_dtree.c sp->header.next = cpu_to_le64(rbn); next 1446 fs/jfs/jfs_dtree.c f->next = fsi; next 1447 fs/jfs/jfs_dtree.c f->next = -1; next 1563 fs/jfs/jfs_dtree.c f->next = fsi; next 1564 fs/jfs/jfs_dtree.c f->next = -1; next 1775 fs/jfs/jfs_dtree.c f->next = last; next 1788 fs/jfs/jfs_dtree.c f->next = fsi; next 1789 fs/jfs/jfs_dtree.c f->next = -1; next 1798 fs/jfs/jfs_dtree.c fsi = f->next; next 1801 fs/jfs/jfs_dtree.c f->next = n; next 1927 fs/jfs/jfs_dtree.c rp->header.next = 0; next 1960 fs/jfs/jfs_dtree.c f->next = fsi; next 1961 fs/jfs/jfs_dtree.c f->next = -1; next 1972 fs/jfs/jfs_dtree.c fsi = f->next; next 1975 fs/jfs/jfs_dtree.c f->next = n; next 2036 fs/jfs/jfs_dtree.c s->next = -1; next 2049 fs/jfs/jfs_dtree.c f->next = fsi; next 2050 fs/jfs/jfs_dtree.c f->next = -1; next 2114 fs/jfs/jfs_dtree.c || (p->header.next == 0)) next 2118 fs/jfs/jfs_dtree.c DT_GETPAGE(ip, le64_to_cpu(p->header.next), next 2485 fs/jfs/jfs_dtree.c if (p->header.next) { next 2486 fs/jfs/jfs_dtree.c nextbn = le64_to_cpu(p->header.next); next 2523 fs/jfs/jfs_dtree.c lp->header.next = cpu_to_le64(nxaddr); next 2705 fs/jfs/jfs_dtree.c if (p->header.next) next 2706 fs/jfs/jfs_dtree.c bn = le64_to_cpu(p->header.next); next 2744 fs/jfs/jfs_dtree.c nextbn = le64_to_cpu(p->header.next); next 2801 fs/jfs/jfs_dtree.c p->header.next = cpu_to_le64(nextbn); next 2887 fs/jfs/jfs_dtree.c f->next = fsi; next 2888 fs/jfs/jfs_dtree.c f->next = -1; next 3010 fs/jfs/jfs_dtree.c int i, next; next 3243 fs/jfs/jfs_dtree.c next = d->next; next 3244 fs/jfs/jfs_dtree.c while (next >= 0) { next 3245 fs/jfs/jfs_dtree.c t = (struct dtslot *) & p->slot[next]; next 3262 fs/jfs/jfs_dtree.c next = t->next; next 3277 fs/jfs/jfs_dtree.c bn = le64_to_cpu(p->header.next); next 3462 fs/jfs/jfs_dtree.c bn = le64_to_cpu(p->header.next); next 3495 fs/jfs/jfs_dtree.c bn = le64_to_cpu(p->header.next); next 3536 fs/jfs/jfs_dtree.c bn = le64_to_cpu(p->header.next); next 3608 fs/jfs/jfs_dtree.c si = ih->next; next 3635 fs/jfs/jfs_dtree.c si = t->next; next 3691 fs/jfs/jfs_dtree.c si = lh->next; next 3704 fs/jfs/jfs_dtree.c si = ih->next; next 3745 fs/jfs/jfs_dtree.c si = t->next; next 3846 fs/jfs/jfs_dtree.c si = lh->next; next 3855 fs/jfs/jfs_dtree.c si = ih->next; next 3880 fs/jfs/jfs_dtree.c si = t->next; next 3916 fs/jfs/jfs_dtree.c p->header.freelist = h->next; next 3929 fs/jfs/jfs_dtree.c lh->next = h->next; next 3944 fs/jfs/jfs_dtree.c ih->next = h->next; next 3964 fs/jfs/jfs_dtree.c p->header.freelist = t->next; next 4004 fs/jfs/jfs_dtree.c lh->next = -1; next 4006 fs/jfs/jfs_dtree.c ih->next = -1; next 4009 fs/jfs/jfs_dtree.c t->next = -1; next 4055 fs/jfs/jfs_dtree.c int ssi, next; /* src slot index */ next 4131 fs/jfs/jfs_dtree.c next = slh->next; next 4135 fs/jfs/jfs_dtree.c dlh->next = dsi; next 4143 fs/jfs/jfs_dtree.c next = sih->next; next 4146 fs/jfs/jfs_dtree.c dih->next = dsi; next 4150 fs/jfs/jfs_dtree.c s->next = sfsi; next 4162 fs/jfs/jfs_dtree.c while ((ssi = next) >= 0) { next 4197 fs/jfs/jfs_dtree.c d->next = dsi; next 4200 fs/jfs/jfs_dtree.c next = s->next; next 4201 fs/jfs/jfs_dtree.c s->next = sfsi; next 4212 fs/jfs/jfs_dtree.c dlh->next = -1; next 4214 fs/jfs/jfs_dtree.c dih->next = -1; next 4217 fs/jfs/jfs_dtree.c d->next = -1; next 4276 fs/jfs/jfs_dtree.c si = ((struct ldtentry *) t)->next; next 4278 fs/jfs/jfs_dtree.c si = ((struct idtentry *) t)->next; next 4279 fs/jfs/jfs_dtree.c t->next = si; next 4311 fs/jfs/jfs_dtree.c si = t->next; next 4321 fs/jfs/jfs_dtree.c t->next = p->header.freelist; next 4372 fs/jfs/jfs_dtree.c si = ((struct idtentry *) t)->next; next 4373 fs/jfs/jfs_dtree.c ((struct idtentry *) t)->next = -1; next 4406 fs/jfs/jfs_dtree.c si = t->next; next 4418 fs/jfs/jfs_dtree.c t->next = p->header.freelist; next 4452 fs/jfs/jfs_dtree.c si = t->next; next 4478 fs/jfs/jfs_dtree.c si = t->next; next 35 fs/jfs/jfs_dtree.h s8 next; /* 1: */ next 55 fs/jfs/jfs_dtree.h s8 next; /* 1: */ next 74 fs/jfs/jfs_dtree.h s8 next; /* 1: */ next 182 fs/jfs/jfs_dtree.h __le64 next; /* 8: next sibling */ next 463 fs/jfs/jfs_logmgr.c if ((i = linelock->next)) { next 760 fs/jfs/jfs_logmgr.c gcpn = list_entry(log->cqueue.next, struct tblock, cqueue)->pn; next 959 fs/jfs/jfs_logmgr.c lp = list_entry(log->synclist.next, next 1870 fs/jfs/jfs_logmgr.c struct lbuf *next = lbuf->l_freelist; next 1873 fs/jfs/jfs_logmgr.c lbuf = next; next 2378 fs/jfs/jfs_logmgr.c sbi = list_entry(log->sb_list.next, struct jfs_sb_info, log_list); next 191 fs/jfs/jfs_txnmgr.c TxAnchor.freelock = TxLock[lid].next; next 205 fs/jfs/jfs_txnmgr.c TxLock[lid].next = TxAnchor.freelock; next 276 fs/jfs/jfs_txnmgr.c TxBlock[k].next = k + 1; next 280 fs/jfs/jfs_txnmgr.c TxBlock[k].next = 0; next 304 fs/jfs/jfs_txnmgr.c TxLock[k].next = k + 1; next 305 fs/jfs/jfs_txnmgr.c TxLock[k].next = 0; next 398 fs/jfs/jfs_txnmgr.c if ((tblk->next == 0) && !(flag & COMMIT_FORCE)) { next 406 fs/jfs/jfs_txnmgr.c TxAnchor.freetid = tblk->next; next 418 fs/jfs/jfs_txnmgr.c tblk->next = tblk->last = tblk->xflag = tblk->flag = tblk->lsn = 0; next 523 fs/jfs/jfs_txnmgr.c assert(tblk->next == 0); next 528 fs/jfs/jfs_txnmgr.c tblk->next = TxAnchor.freetid; next 650 fs/jfs/jfs_txnmgr.c jfs_ip->atlhead = tlck->next; next 654 fs/jfs/jfs_txnmgr.c lid_to_tlock(last)->next != lid; next 655 fs/jfs/jfs_txnmgr.c last = lid_to_tlock(last)->next) { next 658 fs/jfs/jfs_txnmgr.c lid_to_tlock(last)->next = tlck->next; next 665 fs/jfs/jfs_txnmgr.c if (tblk->next) next 666 fs/jfs/jfs_txnmgr.c lid_to_tlock(tblk->last)->next = lid; next 668 fs/jfs/jfs_txnmgr.c tblk->next = lid; next 669 fs/jfs/jfs_txnmgr.c tlck->next = 0; next 733 fs/jfs/jfs_txnmgr.c if (tblk->next) next 734 fs/jfs/jfs_txnmgr.c lid_to_tlock(tblk->last)->next = lid; next 736 fs/jfs/jfs_txnmgr.c tblk->next = lid; next 737 fs/jfs/jfs_txnmgr.c tlck->next = 0; next 744 fs/jfs/jfs_txnmgr.c tlck->next = jfs_ip->atlhead; next 746 fs/jfs/jfs_txnmgr.c if (tlck->next == 0) { next 758 fs/jfs/jfs_txnmgr.c linelock->next = 0; next 871 fs/jfs/jfs_txnmgr.c for (lid = tblk->next; lid; lid = tlck->next) { next 899 fs/jfs/jfs_txnmgr.c lid_t lid, next, llid, k; next 911 fs/jfs/jfs_txnmgr.c for (lid = tblk->next; lid; lid = next) { next 913 fs/jfs/jfs_txnmgr.c next = tlck->next; next 950 fs/jfs/jfs_txnmgr.c llid = ((struct linelock *) & tlck->lock)->next; next 953 fs/jfs/jfs_txnmgr.c k = linelock->next; next 961 fs/jfs/jfs_txnmgr.c tblk->next = tblk->last = 0; next 1018 fs/jfs/jfs_txnmgr.c if (tblk->next) next 1019 fs/jfs/jfs_txnmgr.c lid_to_tlock(tblk->last)->next = lid; next 1021 fs/jfs/jfs_txnmgr.c tblk->next = lid; next 1022 fs/jfs/jfs_txnmgr.c tlck->next = 0; next 1029 fs/jfs/jfs_txnmgr.c tlck->next = jfs_ip->atlhead; next 1031 fs/jfs/jfs_txnmgr.c if (tlck->next == 0) { next 1043 fs/jfs/jfs_txnmgr.c maplock->next = 0; next 1071 fs/jfs/jfs_txnmgr.c linelock->next = 0; next 1079 fs/jfs/jfs_txnmgr.c linelock->next = tlock->next; next 1080 fs/jfs/jfs_txnmgr.c tlock->next = lid; next 1236 fs/jfs/jfs_txnmgr.c lid_to_tlock(jfs_ip->atltail)->next = tblk->next; next 1237 fs/jfs/jfs_txnmgr.c tblk->next = jfs_ip->atlhead; next 1379 fs/jfs/jfs_txnmgr.c for (lid = tblk->next; lid; lid = tlck->next) { next 1698 fs/jfs/jfs_txnmgr.c int next, lwm, hwm; next 1717 fs/jfs/jfs_txnmgr.c next = le16_to_cpu(p->header.nextindex); next 1751 fs/jfs/jfs_txnmgr.c if (lwm == next) next 1753 fs/jfs/jfs_txnmgr.c if (lwm > next) { next 1759 fs/jfs/jfs_txnmgr.c xadlock->count = next - lwm; next 1955 fs/jfs/jfs_txnmgr.c if (twm == next - 1) { next 1975 fs/jfs/jfs_txnmgr.c if (hwm >= next) { next 1986 fs/jfs/jfs_txnmgr.c cpu_to_le16(hwm - next + 1); next 1988 fs/jfs/jfs_txnmgr.c xtlck->header.offset = next; next 1989 fs/jfs/jfs_txnmgr.c xtlck->header.length = hwm - next + 1; next 2003 fs/jfs/jfs_txnmgr.c if (lwm < next) { next 2011 fs/jfs/jfs_txnmgr.c xadlock->count = next - lwm; next 2015 fs/jfs/jfs_txnmgr.c tlck->ip, mp, xadlock->count, lwm, next); next 2023 fs/jfs/jfs_txnmgr.c if (twm == next - 1) { next 2044 fs/jfs/jfs_txnmgr.c if (hwm >= next) { next 2051 fs/jfs/jfs_txnmgr.c xadlock->count = hwm - next + 1; next 2052 fs/jfs/jfs_txnmgr.c xadlock->xdlist = &p->xad[next]; next 2055 fs/jfs/jfs_txnmgr.c tlck->ip, mp, xadlock->count, next, hwm); next 2212 fs/jfs/jfs_txnmgr.c lid_t lid, next; next 2220 fs/jfs/jfs_txnmgr.c tlck = lid_to_tlock(tblk->next); next 2221 fs/jfs/jfs_txnmgr.c lid = tlck->next; next 2222 fs/jfs/jfs_txnmgr.c tlck->next = 0; next 2225 fs/jfs/jfs_txnmgr.c next = tlck->next; next 2226 fs/jfs/jfs_txnmgr.c tlck->next = tblk->next; next 2227 fs/jfs/jfs_txnmgr.c tblk->next = lid; next 2228 fs/jfs/jfs_txnmgr.c lid = next; next 2235 fs/jfs/jfs_txnmgr.c for (lid = tblk->next; lid; lid = next) { next 2237 fs/jfs/jfs_txnmgr.c next = tlck->next; next 2306 fs/jfs/jfs_txnmgr.c for (lid = tblk->next; lid; lid = tlck->next) { next 2591 fs/jfs/jfs_txnmgr.c while ((lid = xtlck->next) != 0) { next 2594 fs/jfs/jfs_txnmgr.c xtlck->next = tlck->next; next 2627 fs/jfs/jfs_txnmgr.c lid_t lid, next; next 2635 fs/jfs/jfs_txnmgr.c for (lid = tblk->next; lid; lid = next) { next 2637 fs/jfs/jfs_txnmgr.c next = tlck->next; next 2663 fs/jfs/jfs_txnmgr.c tblk->next = tblk->last = 0; next 2854 fs/jfs/jfs_txnmgr.c jfs_ip = list_entry(TxAnchor.anon_list.next, next 2925 fs/jfs/jfs_txnmgr.c jfs_ip = list_entry(TxAnchor.anon_list.next, next 35 fs/jfs/jfs_txnmgr.h lid_t next; /* index of first tlock of tid */ next 83 fs/jfs/jfs_txnmgr.h lid_t next; /* 2: index next lockword on tid locklist next 158 fs/jfs/jfs_txnmgr.h lid_t next; /* 2: next linelock */ next 174 fs/jfs/jfs_txnmgr.h lid_t next; /* 2: */ next 205 fs/jfs/jfs_txnmgr.h lid_t next; /* 2: */ next 233 fs/jfs/jfs_txnmgr.h lid_t next; /* 2: */ next 104 fs/jfs/jfs_xtree.c static int xtSearch(struct inode *ip, s64 xoff, s64 *next, int *cmpp, next 144 fs/jfs/jfs_xtree.c s64 next, size, xoff, xend; next 163 fs/jfs/jfs_xtree.c if ((rc = xtSearch(ip, lstart, &next, &cmp, &btstack, 0))) { next 182 fs/jfs/jfs_xtree.c if (next) next 183 fs/jfs/jfs_xtree.c *plen = min(next - lstart, llen); next 240 fs/jfs/jfs_xtree.c s64 next = 0; next 309 fs/jfs/jfs_xtree.c next = t64; next 355 fs/jfs/jfs_xtree.c *nextp = next; next 413 fs/jfs/jfs_xtree.c next = offsetXAD(&p->xad[index + 1]); next 414 fs/jfs/jfs_xtree.c goto next; next 430 fs/jfs/jfs_xtree.c next = offsetXAD(&p->xad[base]); next 465 fs/jfs/jfs_xtree.c *nextp = next; next 481 fs/jfs/jfs_xtree.c next: next 538 fs/jfs/jfs_xtree.c s64 next; next 552 fs/jfs/jfs_xtree.c if ((rc = xtSearch(ip, xoff, &next, &cmp, &btstack, XT_INSERT))) next 560 fs/jfs/jfs_xtree.c if ((cmp == 0) || (next && (xlen > next - xoff))) { next 1023 fs/jfs/jfs_xtree.c nextbn = le64_to_cpu(sp->header.next); next 1024 fs/jfs/jfs_xtree.c rp->header.next = cpu_to_le64(nextbn); next 1026 fs/jfs/jfs_xtree.c sp->header.next = cpu_to_le64(rbn); next 1266 fs/jfs/jfs_xtree.c rp->header.next = 0; next 2002 fs/jfs/jfs_xtree.c bn = le64_to_cpu(p->header.next); next 2176 fs/jfs/jfs_xtree.c s64 next; next 2191 fs/jfs/jfs_xtree.c if ((rc = xtSearch(ip, xoff, &next, &cmp, &btstack, XT_INSERT))) next 2202 fs/jfs/jfs_xtree.c if (next) next 2203 fs/jfs/jfs_xtree.c xlen = min(xlen, (int)(next - xoff)); next 2716 fs/jfs/jfs_xtree.c if (p->header.next) { next 2717 fs/jfs/jfs_xtree.c nextbn = le64_to_cpu(p->header.next); next 2747 fs/jfs/jfs_xtree.c lp->header.next = cpu_to_le64(nxaddr); next 2953 fs/jfs/jfs_xtree.c goto next; next 2975 fs/jfs/jfs_xtree.c next: next 3005 fs/jfs/jfs_xtree.c nextbn = le64_to_cpu(p->header.next); next 3045 fs/jfs/jfs_xtree.c p->header.next = le64_to_cpu(nextbn); next 3244 fs/jfs/jfs_xtree.c if (p->header.next) { next 3253 fs/jfs/jfs_xtree.c p->header.next = 0; next 3539 fs/jfs/jfs_xtree.c if (tblk->next == lid) next 3540 fs/jfs/jfs_xtree.c tblk->next = tlck->next; next 3542 fs/jfs/jfs_xtree.c for (prev = lid_to_tlock(tblk->next); next 3543 fs/jfs/jfs_xtree.c prev->next != lid; next 3544 fs/jfs/jfs_xtree.c prev = lid_to_tlock(prev->next)) { next 3545 fs/jfs/jfs_xtree.c assert(prev->next); next 3547 fs/jfs/jfs_xtree.c prev->next = tlck->next; next 3549 fs/jfs/jfs_xtree.c lid_to_tlock(tblk->last)->next = lid; next 3550 fs/jfs/jfs_xtree.c tlck->next = 0; next 73 fs/jfs/jfs_xtree.h __le64 next; /* 8: */ next 117 fs/kernfs/file.c void *next = ops->seq_start(sf, ppos); next 119 fs/kernfs/file.c if (next == ERR_PTR(-ENODEV)) next 120 fs/kernfs/file.c kernfs_seq_stop_active(sf, next); next 121 fs/kernfs/file.c return next; next 137 fs/kernfs/file.c void *next = ops->seq_next(sf, v, ppos); next 139 fs/kernfs/file.c if (next == ERR_PTR(-ENODEV)) next 140 fs/kernfs/file.c kernfs_seq_stop_active(sf, next); next 141 fs/kernfs/file.c return next; next 172 fs/kernfs/file.c .next = kernfs_seq_next, next 106 fs/libfs.c while ((p = p->next) != &dentry->d_subdirs) { next 189 fs/libfs.c struct dentry *next = NULL; next 202 fs/libfs.c while ((next = scan_positives(cursor, p, 1, next)) != NULL) { next 203 fs/libfs.c if (!dir_emit(ctx, next->d_name.name, next->d_name.len, next 204 fs/libfs.c d_inode(next)->i_ino, dt_type(d_inode(next)))) next 207 fs/libfs.c p = &next->d_child; next 210 fs/libfs.c if (next) next 211 fs/libfs.c list_move_tail(&cursor->d_child, &next->d_child); next 215 fs/libfs.c dput(next); next 234 fs/lockd/clntlock.c struct file_lock *fl, *next; next 260 fs/lockd/clntlock.c list_for_each_entry_safe(fl, next, &host->h_reclaim, fl_u.nfs_fl.list) { next 42 fs/lockd/host.c #define for_each_host_safe(host, next, chain, table) \ next 45 fs/lockd/host.c hlist_for_each_entry_safe((host), (next), \ next 657 fs/lockd/host.c struct hlist_node *next; next 671 fs/lockd/host.c for_each_host_safe(host, next, chain, nlm_server_hosts) { next 315 fs/lockd/svclock.c struct nlm_block *block, *next; next 319 fs/lockd/svclock.c list_for_each_entry_safe(block, next, &file->f_blocks, b_flist) { next 995 fs/lockd/svclock.c block = list_entry(nlm_blocked.next, struct nlm_block, b_list); next 256 fs/lockd/svcsubs.c struct hlist_node *next; next 262 fs/lockd/svcsubs.c hlist_for_each_entry_safe(file, next, &nlm_files[i], f_list) { next 2985 fs/locks.c .next = locks_next, next 139 fs/mbcache.c node = entry->e_hash_list.next; next 149 fs/mbcache.c node = node->next; next 394 fs/mbcache.c struct mb_cache_entry *entry, *next; next 402 fs/mbcache.c list_for_each_entry_safe(entry, next, &cache->c_list, e_list) { next 900 fs/namespace.c struct list_head *next = p->mnt_mounts.next; next 901 fs/namespace.c if (next == &p->mnt_mounts) { next 905 fs/namespace.c next = p->mnt_child.next; next 906 fs/namespace.c if (next != &p->mnt_parent->mnt_mounts) next 911 fs/namespace.c return list_entry(next, struct mount, mnt_child); next 1293 fs/namespace.c .next = m_next, next 2882 fs/namespace.c struct mount *mnt, *next; next 2897 fs/namespace.c list_for_each_entry_safe(mnt, next, mounts, mnt_expire) { next 2923 fs/namespace.c struct list_head *next; next 2927 fs/namespace.c next = this_parent->mnt_mounts.next; next 2929 fs/namespace.c while (next != &this_parent->mnt_mounts) { next 2930 fs/namespace.c struct list_head *tmp = next; next 2933 fs/namespace.c next = tmp->next; next 2953 fs/namespace.c next = this_parent->mnt_child.next; next 3893 fs/namespace.c goto next; next 3900 fs/namespace.c next: ; next 235 fs/nfs/blocklayout/extent_tree.c struct pnfs_block_extent *next = ext_tree_next(be); next 239 fs/nfs/blocklayout/extent_tree.c be = next; next 1102 fs/nfs/client.c .next = nfs_server_list_next, next 1114 fs/nfs/client.c .next = nfs_volume_list_next, next 525 fs/nfs/delegation.c server = list_entry_rcu(clp->cl_superblocks.next, next 532 fs/nfs/delegation.c delegation = list_entry_rcu(server->delegations.next, next 2185 fs/nfs/dir.c cache = list_entry(head->next, struct nfs_access_entry, lru); next 2195 fs/nfs/dir.c struct nfs_inode *nfsi, *next; next 2200 fs/nfs/dir.c list_for_each_entry_safe(nfsi, next, &nfs_access_lru_list, access_cache_inode_lru) { next 2209 fs/nfs/dir.c cache = list_entry(nfsi->access_cache_entry_lru.next, next 401 fs/nfs/direct.c struct nfs_page *req = nfs_list_entry(hdr->pages.next); next 422 fs/nfs/direct.c req = nfs_list_entry(head->next); next 663 fs/nfs/direct.c req = nfs_list_entry(failed.next); next 684 fs/nfs/direct.c req = nfs_list_entry(data->pages.next); next 761 fs/nfs/direct.c struct nfs_page *req = nfs_list_entry(hdr->pages.next); next 792 fs/nfs/direct.c req = nfs_list_entry(hdr->pages.next); next 813 fs/nfs/direct.c req = nfs_list_entry(head->next); next 157 fs/nfs/nfs4client.c dss = list_entry(clp->cl_ds_clients.next, next 171 fs/nfs/nfs4client.c cp_state = list_entry(clp->pending_cb_stateids.next, next 47 fs/nfs/nfs4session.c p = &(*p)->next; next 51 fs/nfs/nfs4session.c *p = slot->next; next 137 fs/nfs/nfs4session.c p = &slot->next; next 284 fs/nfs/nfs4session.c p = &(*p)->next; next 22 fs/nfs/nfs4session.h struct nfs4_slot *next; next 1094 fs/nfs/nfs4state.c struct nfs_seqid *next; next 1096 fs/nfs/nfs4state.c next = list_first_entry(&sequence->list, next 1098 fs/nfs/nfs4state.c rpc_wake_up_queued_task(&sequence->wait, next->task); next 49 fs/nfs/pagelist.c hdr->req = nfs_list_entry(mirror->pg_list.next); next 279 fs/nfs/pagelist.c struct nfs_page *tmp, *next; next 286 fs/nfs/pagelist.c next = tmp->wb_this_page; next 291 fs/nfs/pagelist.c tmp = next; next 792 fs/nfs/pagelist.c req = nfs_list_entry(head->next); next 1251 fs/nfs/pagelist.c struct nfs_page *req = nfs_list_entry(pages.next); next 419 fs/nfs/pnfs.c struct pnfs_layout_segment *lseg, *next; next 422 fs/nfs/pnfs.c list_for_each_entry_safe(lseg, next, &lo->plh_segs, pls_list) next 677 fs/nfs/pnfs.c struct pnfs_layout_segment *lseg, *next; next 684 fs/nfs/pnfs.c list_for_each_entry_safe(lseg, next, &lo->plh_segs, pls_list) next 703 fs/nfs/pnfs.c struct pnfs_layout_segment *lseg, *next; next 705 fs/nfs/pnfs.c list_for_each_entry_safe(lseg, next, &lo->plh_return_segs, pls_list) { next 774 fs/nfs/pnfs.c struct pnfs_layout_hdr *lo, *next; next 777 fs/nfs/pnfs.c list_for_each_entry_safe(lo, next, &server->layouts, plh_layouts) { next 817 fs/nfs/pnfs.c lo = list_entry(layout_list->next, struct pnfs_layout_hdr, next 1356 fs/nfs/pnfs.c struct pnfs_layout_segment *lseg, *next; next 1402 fs/nfs/pnfs.c list_for_each_entry_safe(lseg, next, &lo->plh_segs, pls_list) { next 2397 fs/nfs/pnfs.c struct pnfs_layout_segment *lseg, *next; next 2404 fs/nfs/pnfs.c list_for_each_entry_safe(lseg, next, &lo->plh_segs, pls_list) next 167 fs/nfs/read.c struct nfs_page *req = nfs_list_entry(hdr->pages.next); next 226 fs/nfs/read.c req = nfs_list_entry(head->next); next 1004 fs/nfs/write.c struct nfs_page *req = nfs_list_entry(hdr->pages.next); next 1020 fs/nfs/write.c goto next; next 1024 fs/nfs/write.c next: next 1438 fs/nfs/write.c req = nfs_list_entry(head->next); next 1747 fs/nfs/write.c struct nfs_page *first = nfs_list_entry(head->next); next 1785 fs/nfs/write.c req = nfs_list_entry(page_list->next); next 1848 fs/nfs/write.c req = nfs_list_entry(data->pages.next); next 1864 fs/nfs/write.c goto next; next 1875 fs/nfs/write.c goto next; next 1881 fs/nfs/write.c next: next 1235 fs/nfsd/export.c .next = cache_seq_next_rcu, next 658 fs/nfsd/filecache.c struct hlist_node *next; next 669 fs/nfsd/filecache.c hlist_for_each_entry_safe(nf, next, &nfb->nfb_head, nf_node) { next 1868 fs/nfsd/nfs4proc.c struct nfsd4_op *next = &argp->ops[resp->opcnt]; next 1885 fs/nfsd/nfs4proc.c if (next->opnum == OP_ILLEGAL) next 1887 fs/nfsd/nfs4proc.c nextd = OPDESC(next); next 1392 fs/nfsd/nfs4state.c stp = list_entry(open_stp->st_locks.next, next 1943 fs/nfsd/nfs4state.c ses = list_entry(clp->cl_sessions.next, struct nfsd4_session, next 2013 fs/nfsd/nfs4state.c dp = list_entry(clp->cl_delegations.next, struct nfs4_delegation, dl_perclnt); next 2019 fs/nfsd/nfs4state.c dp = list_entry(reaplist.next, struct nfs4_delegation, dl_recall_lru); next 2024 fs/nfsd/nfs4state.c dp = list_entry(clp->cl_revoked.next, struct nfs4_delegation, dl_recall_lru); next 2029 fs/nfsd/nfs4state.c oo = list_entry(clp->cl_openowners.next, struct nfs4_openowner, oo_perclient); next 2497 fs/nfsd/nfs4state.c .next = states_next, next 5217 fs/nfsd/nfs4state.c struct list_head *pos, *next, reaplist; next 5231 fs/nfsd/nfs4state.c list_for_each_safe(pos, next, &nn->client_lru) { next 5246 fs/nfsd/nfs4state.c list_for_each_safe(pos, next, &reaplist) { next 5254 fs/nfsd/nfs4state.c list_for_each_safe(pos, next, &nn->del_recall_lru) { next 6955 fs/nfsd/nfs4state.c crp = list_entry(nn->reclaim_str_hashtbl[i].next, next 7084 fs/nfsd/nfs4state.c struct nfs4_client *clp, *next; next 7093 fs/nfsd/nfs4state.c list_for_each_entry_safe(clp, next, &nn->client_lru, cl_lru) { next 7102 fs/nfsd/nfs4state.c list_for_each_entry_safe(clp, next, &reaplist, cl_lru) next 7210 fs/nfsd/nfs4state.c struct nfs4_ol_stateid *stp, *next; next 7212 fs/nfsd/nfs4state.c list_for_each_entry_safe(stp, next, reaplist, st_locks) { next 7269 fs/nfsd/nfs4state.c struct nfs4_openowner *oop, *next; next 7277 fs/nfsd/nfs4state.c list_for_each_entry_safe(oop, next, &clp->cl_openowners, oo_perclient) { next 7341 fs/nfsd/nfs4state.c struct nfs4_openowner *oop, *next; next 7343 fs/nfsd/nfs4state.c list_for_each_entry_safe(oop, next, reaplist, oo_perclient) { next 7400 fs/nfsd/nfs4state.c struct nfs4_delegation *dp, *next; next 7408 fs/nfsd/nfs4state.c list_for_each_entry_safe(dp, next, &clp->cl_delegations, dl_perclnt) { next 7470 fs/nfsd/nfs4state.c struct nfs4_delegation *dp, *next; next 7472 fs/nfsd/nfs4state.c list_for_each_entry_safe(dp, next, reaplist, dl_recall_lru) { next 7530 fs/nfsd/nfs4state.c struct nfs4_delegation *dp, *next; next 7532 fs/nfsd/nfs4state.c list_for_each_entry_safe(dp, next, reaplist, dl_recall_lru) { next 7576 fs/nfsd/nfs4state.c struct nfs4_client *clp, *next; next 7585 fs/nfsd/nfs4state.c list_for_each_entry_safe(clp, next, &nn->client_lru, cl_lru) { next 7682 fs/nfsd/nfs4state.c clp = list_entry(nn->conf_id_hashtbl[i].next, struct nfs4_client, cl_idhash); next 7691 fs/nfsd/nfs4state.c clp = list_entry(nn->unconf_id_hashtbl[i].next, struct nfs4_client, cl_idhash); next 7757 fs/nfsd/nfs4state.c struct list_head *pos, *next, reaplist; next 7765 fs/nfsd/nfs4state.c list_for_each_safe(pos, next, &nn->del_recall_lru) { next 7771 fs/nfsd/nfs4state.c list_for_each_safe(pos, next, &reaplist) { next 235 fs/nfsd/nfs4xdr.c tb->next = argp->to_free; next 2065 fs/nfsd/nfs4xdr.c char *str, *end, *next; next 2083 fs/nfsd/nfs4xdr.c next = end + 1; next 2084 fs/nfsd/nfs4xdr.c if (*end && (!*next || *next == sep)) { next 2105 fs/nfsd/nfs4xdr.c end = next; next 4091 fs/nfsd/nfs4xdr.c struct nfsd4_test_stateid_id *stateid, *next; next 4099 fs/nfsd/nfs4xdr.c list_for_each_entry_safe(stateid, next, &test_stateid->ts_stateid_list, ts_id_list) { next 4545 fs/nfsd/nfs4xdr.c args->to_free = tb->next; next 872 fs/nfsd/nfssvc.c struct svc_xprt *perm_sock = list_entry(rqstp->rq_server->sv_permsocks.next, typeof(struct svc_xprt), xpt_list); next 641 fs/nfsd/xdr4.h struct svcxdr_tmpbuf *next; next 483 fs/nilfs2/cpfile.c __u64 curr = *cnop, next; next 525 fs/nilfs2/cpfile.c next = le64_to_cpu(cp->cp_snapshot_list.ssl_next); next 526 fs/nilfs2/cpfile.c if (next == 0) next 529 fs/nilfs2/cpfile.c next_blkoff = nilfs_cpfile_get_blkoff(cpfile, next); next 533 fs/nilfs2/cpfile.c ret = nilfs_cpfile_get_checkpoint_block(cpfile, next, next 541 fs/nilfs2/cpfile.c curr = next; next 742 fs/nilfs2/cpfile.c __u64 next, prev; next 767 fs/nilfs2/cpfile.c next = le64_to_cpu(list->ssl_next); next 774 fs/nilfs2/cpfile.c if (next != 0) { next 775 fs/nilfs2/cpfile.c ret = nilfs_cpfile_get_checkpoint_block(cpfile, next, 0, next 795 fs/nilfs2/cpfile.c cpfile, next, next_bh, kaddr); next 802 fs/nilfs2/cpfile.c list->ssl_next = cpu_to_le64(next); next 288 fs/nilfs2/mdt.c __u64 next; next 302 fs/nilfs2/mdt.c ret = nilfs_bmap_seek_key(NILFS_I(inode)->i_bmap, start + 1, &next); next 304 fs/nilfs2/mdt.c if (next <= end) { next 305 fs/nilfs2/mdt.c ret = nilfs_mdt_read_block(inode, next, true, out_bh); next 307 fs/nilfs2/mdt.c *blkoff = next; next 539 fs/nilfs2/recovery.c goto next; next 552 fs/nilfs2/recovery.c next: next 92 fs/nilfs2/segbuf.c segbuf->sb_sum.next = nilfs_get_segment_start_blocknr(nilfs, nextnum); next 149 fs/nilfs2/segbuf.c bh_sum = list_entry(segbuf->sb_segsum_buffers.next, next 158 fs/nilfs2/segbuf.c raw_sum->ss_next = cpu_to_le64(segbuf->sb_sum.next); next 177 fs/nilfs2/segbuf.c bh = list_entry(segbuf->sb_segsum_buffers.next, struct buffer_head, next 205 fs/nilfs2/segbuf.c bh = list_entry(segbuf->sb_segsum_buffers.next, struct buffer_head, next 41 fs/nilfs2/segbuf.h sector_t next; next 87 fs/nilfs2/segbuf.h #define NILFS_NEXT_SEGBUF(segbuf) NILFS_LIST_SEGBUF((segbuf)->sb_list.next) next 90 fs/nilfs2/segbuf.h #define NILFS_FIRST_SEGBUF(head) NILFS_LIST_SEGBUF((head)->next) next 91 fs/nilfs2/segbuf.h #define NILFS_SEGBUF_IS_LAST(segbuf, head) ((segbuf)->sb_list.next == (head)) next 98 fs/nilfs2/segbuf.h (list_entry((head)->next, struct buffer_head, b_assoc_buffers)) next 100 fs/nilfs2/segbuf.h (list_entry((bh)->b_assoc_buffers.next, struct buffer_head, \ next 102 fs/nilfs2/segbuf.h #define NILFS_SEGBUF_BH_IS_LAST(bh, head) ((bh)->b_assoc_buffers.next == head) next 239 fs/nls/nls_base.c if (nls->next) next 249 fs/nls/nls_base.c tmp = &(*tmp)->next; next 251 fs/nls/nls_base.c nls->next = tables; next 265 fs/nls/nls_base.c *tmp = nls->next; next 269 fs/nls/nls_base.c tmp = &(*tmp)->next; next 279 fs/nls/nls_base.c for (nls = tables; nls; nls = nls->next) { next 255 fs/notify/fsnotify.c node = srcu_dereference(mark->obj_list.next, next 824 fs/notify/mark.c struct fsnotify_mark *mark, *next; next 834 fs/notify/mark.c list_for_each_entry_safe(mark, next, &private_destroy_list, g_list) { next 861 fs/ntfs/aops.c struct buffer_head *next = bh->b_this_page; next 866 fs/ntfs/aops.c bh = next; next 7502 fs/ocfs2/alloc.c int ret = 0, count = 0, next; next 7515 fs/ocfs2/alloc.c next = ocfs2_find_next_bit(bitmap, max, start); next 7517 fs/ocfs2/alloc.c if ((next - start) >= minbits) { next 7519 fs/ocfs2/alloc.c start, next - start); next 7524 fs/ocfs2/alloc.c count += next - start; next 7526 fs/ocfs2/alloc.c start = next + 1; next 437 fs/ocfs2/aops.c struct buffer_head *next; next 441 fs/ocfs2/aops.c block_start = block_end, bh = next) next 443 fs/ocfs2/aops.c next = bh->b_this_page; next 766 fs/ocfs2/cluster/heartbeat.c event = list_entry(o2hb_node_events.next, next 161 fs/ocfs2/cluster/netdebug.c .next = nst_seq_next, next 375 fs/ocfs2/cluster/netdebug.c .next = sc_seq_next, next 717 fs/ocfs2/dir.c goto next; next 725 fs/ocfs2/dir.c goto next; next 740 fs/ocfs2/dir.c next: next 3459 fs/ocfs2/dir.c goto next; next 3469 fs/ocfs2/dir.c next: next 608 fs/ocfs2/dlm/dlmdebug.c .next = lockres_seq_next, next 435 fs/ocfs2/dlm/dlmdomain.c iter = res->hash_node.next; next 2932 fs/ocfs2/dlm/dlmmaster.c struct dlm_lock *lock, *next; next 2939 fs/ocfs2/dlm/dlmmaster.c list_for_each_entry_safe(lock, next, queue, list) { next 142 fs/ocfs2/dlm/dlmrecovery.c struct dlm_work_item *item, *next; next 155 fs/ocfs2/dlm/dlmrecovery.c list_for_each_entry_safe(item, next, &tmp_list, list) { next 775 fs/ocfs2/dlm/dlmrecovery.c struct dlm_reco_node_data *ndata, *next; next 782 fs/ocfs2/dlm/dlmrecovery.c list_for_each_entry_safe(ndata, next, &tmplist, list) { next 1047 fs/ocfs2/dlm/dlmrecovery.c struct dlm_lock_resource *res, *next; next 1051 fs/ocfs2/dlm/dlmrecovery.c list_for_each_entry_safe(res, next, &dlm->reco.resources, recovering) { next 2058 fs/ocfs2/dlm/dlmrecovery.c struct dlm_lock *lock, *next; next 2077 fs/ocfs2/dlm/dlmrecovery.c list_for_each_entry_safe(lock, next, queue, list) { next 2142 fs/ocfs2/dlm/dlmrecovery.c struct dlm_lock_resource *res, *next; next 2146 fs/ocfs2/dlm/dlmrecovery.c list_for_each_entry_safe(res, next, &dlm->reco.resources, recovering) { next 2264 fs/ocfs2/dlm/dlmrecovery.c struct dlm_lock *lock, *next; next 2278 fs/ocfs2/dlm/dlmrecovery.c list_for_each_entry_safe(lock, next, &res->granted, list) { next 2287 fs/ocfs2/dlm/dlmrecovery.c list_for_each_entry_safe(lock, next, &res->converting, list) { next 2296 fs/ocfs2/dlm/dlmrecovery.c list_for_each_entry_safe(lock, next, &res->blocked, list) { next 293 fs/ocfs2/dlm/dlmthread.c lockres = list_entry(dlm->purge_list.next, next 368 fs/ocfs2/dlm/dlmthread.c target = list_entry(res->converting.next, struct dlm_lock, list); next 439 fs/ocfs2/dlm/dlmthread.c target = list_entry(res->blocked.next, struct dlm_lock, list); next 582 fs/ocfs2/dlm/dlmthread.c lock = list_entry(dlm->pending_asts.next, next 626 fs/ocfs2/dlm/dlmthread.c lock = list_entry(dlm->pending_basts.next, next 712 fs/ocfs2/dlm/dlmthread.c res = list_entry(dlm->dirty_list.next, next 3232 fs/ocfs2/dlmglue.c .next = ocfs2_dlm_seq_next, next 4354 fs/ocfs2/dlmglue.c lockres = list_entry(osb->blocked_lock_list.next, next 1485 fs/ocfs2/file.c goto next; next 1495 fs/ocfs2/file.c next: next 850 fs/ocfs2/move_extents.c goto next; next 861 fs/ocfs2/move_extents.c goto next; next 884 fs/ocfs2/move_extents.c next: next 239 fs/ocfs2/quota_local.c struct ocfs2_quota_chunk *pos, *next; next 241 fs/ocfs2/quota_local.c list_for_each_entry_safe(pos, next, head, qc_chunk) { next 318 fs/ocfs2/quota_local.c struct ocfs2_recovery_chunk *next; next 321 fs/ocfs2/quota_local.c list_for_each_entry_safe(rchunk, next, head, rc_list) { next 471 fs/ocfs2/quota_local.c struct ocfs2_recovery_chunk *rchunk, *next; next 476 fs/ocfs2/quota_local.c list_for_each_entry_safe(rchunk, next, &(rec->r_list[type]), rc_list) { next 1360 fs/ocfs2/refcounttree.c struct ocfs2_refcount_rec *next) next 1363 fs/ocfs2/refcounttree.c ocfs2_get_ref_rec_low_cpos(next)) next 469 fs/ocfs2/reservations.c struct rb_node *prev, *next; next 529 fs/ocfs2/reservations.c next = rb_first(root); next 530 fs/ocfs2/reservations.c next_resv = rb_entry(next, struct ocfs2_alloc_reservation, next 565 fs/ocfs2/reservations.c next = rb_next(prev); next 566 fs/ocfs2/reservations.c if (next) { next 567 fs/ocfs2/reservations.c next_resv = rb_entry(next, next 585 fs/ocfs2/reservations.c trace_ocfs2_resv_find_window_next(next ? next_resv->r_start: -1, next 586 fs/ocfs2/reservations.c next ? ocfs2_resv_end(next_resv) : -1); next 606 fs/ocfs2/reservations.c if (!next) next 609 fs/ocfs2/reservations.c prev = next; next 1200 fs/ocfs2/super.c .next = NULL next 164 fs/omfs/dir.c __be64 *entry, next; next 184 fs/omfs/dir.c next = oi->i_sibling; next 198 fs/omfs/dir.c *entry = next; next 37 fs/omfs/file.c u64 next, last; next 45 fs/omfs/file.c next = inode->i_ino; next 52 fs/omfs/file.c bh = omfs_bread(inode->i_sb, next); next 61 fs/omfs/file.c if (omfs_is_bad(sbi, (struct omfs_header *) bh->b_data, next)) next 69 fs/omfs/file.c last = next; next 70 fs/omfs/file.c next = be64_to_cpu(oe->e_next); next 89 fs/omfs/file.c if (next == ~0) next 92 fs/omfs/file.c bh = omfs_bread(inode->i_sb, next); next 221 fs/omfs/file.c sector_t next, offset; next 239 fs/omfs/file.c next = inode->i_ino; next 243 fs/omfs/file.c if (omfs_is_bad(sbi, (struct omfs_header *) bh->b_data, next)) next 247 fs/omfs/file.c next = be64_to_cpu(oe->e_next); next 262 fs/omfs/file.c if (next == ~0) next 266 fs/omfs/file.c bh = omfs_bread(inode->i_sb, next); next 140 fs/openpromfs/inode.c .next = property_next, next 228 fs/openpromfs/inode.c prop = prop->next; next 311 fs/openpromfs/inode.c prop = prop->next; next 320 fs/openpromfs/inode.c prop = prop->next; next 61 fs/orangefs/devorangefs-req.c struct orangefs_kernel_op_s *op, *next; next 68 fs/orangefs/devorangefs-req.c next, next 11 fs/orangefs/dir.c struct orangefs_dir_part *next; next 141 fs/orangefs/dir.c if (part->next) next 142 fs/orangefs/dir.c part = part->next; next 148 fs/orangefs/dir.c new->next = NULL; next 154 fs/orangefs/dir.c part->next = new; next 221 fs/orangefs/dir.c goto next; next 224 fs/orangefs/dir.c goto next; next 235 fs/orangefs/dir.c next: next 251 fs/orangefs/dir.c while (part->next && count) { next 253 fs/orangefs/dir.c part = part->next; next 276 fs/orangefs/dir.c part = part->next; next 293 fs/orangefs/dir.c struct orangefs_dir_part *next = part->next; next 295 fs/orangefs/dir.c part = next; next 390 fs/orangefs/dir.c struct orangefs_dir_part *next = part->next; next 392 fs/orangefs/dir.c part = next; next 111 fs/orangefs/orangefs-debugfs.c .next = help_next, next 212 fs/orangefs/orangefs-mod.c struct orangefs_kernel_op_s *next; next 216 fs/orangefs/orangefs-mod.c next, next 866 fs/overlayfs/copy_up.c struct dentry *next; next 872 fs/overlayfs/copy_up.c next = dget(dentry); next 875 fs/overlayfs/copy_up.c parent = dget_parent(next); next 880 fs/overlayfs/copy_up.c dput(next); next 881 fs/overlayfs/copy_up.c next = parent; next 884 fs/overlayfs/copy_up.c err = ovl_copy_up_one(parent, next, flags); next 887 fs/overlayfs/copy_up.c dput(next); next 107 fs/overlayfs/export.c struct dentry *next, *parent = NULL; next 120 fs/overlayfs/export.c next = dget(dentry); next 122 fs/overlayfs/export.c parent = dget_parent(next); next 123 fs/overlayfs/export.c if (WARN_ON(parent == next)) { next 133 fs/overlayfs/export.c err = ovl_encode_maybe_copy_up(next); next 142 fs/overlayfs/export.c dput(next); next 143 fs/overlayfs/export.c next = parent; next 147 fs/overlayfs/export.c dput(next); next 501 fs/overlayfs/export.c struct dentry *next, *parent = NULL; next 508 fs/overlayfs/export.c next = dget(real); next 510 fs/overlayfs/export.c parent = dget_parent(next); next 516 fs/overlayfs/export.c ancestor = ovl_lookup_real_inode(sb, next, layer); next 530 fs/overlayfs/export.c if (parent == next) { next 535 fs/overlayfs/export.c dput(next); next 536 fs/overlayfs/export.c next = parent; next 540 fs/overlayfs/export.c dput(next); next 562 fs/overlayfs/export.c struct dentry *next, *this; next 571 fs/overlayfs/export.c next = dget(real); next 573 fs/overlayfs/export.c parent = dget_parent(next); next 598 fs/overlayfs/export.c if (parent == next) { next 603 fs/overlayfs/export.c dput(next); next 604 fs/overlayfs/export.c next = parent; next 608 fs/overlayfs/export.c this = ovl_lookup_real_one(connected, next, layer); next 634 fs/overlayfs/export.c dput(next); next 294 fs/overlayfs/namei.c const char *next = strchrnul(s, '/'); next 295 fs/overlayfs/namei.c size_t thislen = next - s; next 296 fs/overlayfs/namei.c bool end = !next[0]; next 303 fs/overlayfs/namei.c d->name.len - rem, next, &base); next 367 fs/overlayfs/readdir.c int idx, next; next 369 fs/overlayfs/readdir.c for (idx = 0; idx != -1; idx = next) { next 370 fs/overlayfs/readdir.c next = ovl_path_next(idx, dentry, &realpath); next 373 fs/overlayfs/readdir.c if (next != -1) { next 777 fs/overlayfs/readdir.c od->cursor = p->l_node.next; next 1122 fs/overlayfs/readdir.c goto next; next 1147 fs/overlayfs/readdir.c next: next 1506 fs/overlayfs/super.c struct dentry *next = dentry, *parent; next 1512 fs/overlayfs/super.c parent = dget_parent(next); next 1515 fs/overlayfs/super.c while (!err && parent != next) { next 1522 fs/overlayfs/super.c next = parent; next 1523 fs/overlayfs/super.c parent = dget_parent(next); next 1524 fs/overlayfs/super.c dput(next); next 911 fs/overlayfs/util.c char *s, *next, *buf = NULL; next 922 fs/overlayfs/util.c for (s = buf; *s++ == '/'; s = next) { next 923 fs/overlayfs/util.c next = strchrnul(s, '/'); next 924 fs/overlayfs/util.c if (s == next) next 19 fs/pnode.c return list_entry(p->mnt_share.next, struct mount, mnt_share); next 24 fs/pnode.c return list_entry(p->mnt_slave_list.next, struct mount, mnt_slave); next 34 fs/pnode.c return list_entry(p->mnt_slave.next, struct mount, mnt_slave); next 160 fs/pnode.c struct mount *next = next_peer(m); next 161 fs/pnode.c return (next == origin) ? NULL : next; next 162 fs/pnode.c } else if (m->mnt_slave.next != &master->mnt_slave_list) next 187 fs/pnode.c struct mount *next; next 190 fs/pnode.c next = next_peer(m); next 192 fs/pnode.c if (next == origin) next 194 fs/pnode.c } else if (m->mnt_slave.next != &next->mnt_slave) next 196 fs/pnode.c m = next; next 201 fs/pnode.c if (m->mnt_slave.next != &master->mnt_slave_list) next 206 fs/pnode.c if (master->mnt_slave.next == &m->mnt_slave) next 751 fs/proc/array.c .next = children_seq_next, next 2333 fs/proc/base.c .next = timers_next, next 78 fs/proc/consoles.c return con->next; next 88 fs/proc/consoles.c .next = c_next, next 49 fs/proc/devices.c .next = devinfo_next, next 160 fs/proc/generic.c const char *cp = name, *next; next 169 fs/proc/generic.c next = strchr(cp, '/'); next 170 fs/proc/generic.c if (!next) next 173 fs/proc/generic.c len = next - cp; next 305 fs/proc/generic.c struct proc_dir_entry *next; next 315 fs/proc/generic.c next = pde_subdir_next(de); next 317 fs/proc/generic.c de = next; next 691 fs/proc/generic.c struct proc_dir_entry *root = NULL, *de, *next; next 711 fs/proc/generic.c next = pde_subdir_first(de); next 712 fs/proc/generic.c if (next) { next 713 fs/proc/generic.c rb_erase(&next->subdir_node, &de->subdir); next 714 fs/proc/generic.c de = next; next 720 fs/proc/generic.c next = de->parent; next 722 fs/proc/generic.c next->nlink--; next 729 fs/proc/generic.c de = next; next 32 fs/proc/interrupts.c .next = int_seq_next, next 107 fs/proc/nommu.c .next = nommu_region_list_next, next 75 fs/proc/proc_tty.c if (&p->tty_drivers == tty_drivers.next) { next 96 fs/proc/proc_tty.c dev_t next = MKDEV(MAJOR(from)+1, 0); next 97 fs/proc/proc_tty.c show_tty_range(m, p, from, next - from); next 98 fs/proc/proc_tty.c from = next; next 124 fs/proc/proc_tty.c .next = t_next, next 97 fs/proc/stat.c unsigned int i, next = 0; next 100 fs/proc/stat.c show_irq_gap(p, i - next); next 102 fs/proc/stat.c next = i + 1; next 104 fs/proc/stat.c show_irq_gap(p, nr_irqs - next); next 205 fs/proc/task_mmu.c struct vm_area_struct *next; next 208 fs/proc/task_mmu.c next = m_next_vma(priv, v); next 209 fs/proc/task_mmu.c if (!next) next 211 fs/proc/task_mmu.c return next; next 372 fs/proc/task_mmu.c .next = m_next, next 909 fs/proc/task_mmu.c .next = m_next, next 1896 fs/proc/task_mmu.c .next = m_next, next 253 fs/proc/task_nommu.c .next = m_next, next 120 fs/pstore/inode.c .next = pstore_ftrace_seq_next, next 1046 fs/quota/dquot.c act_head = tofree_head->next; next 1049 fs/quota/dquot.c act_head = act_head->next; next 121 fs/reiserfs/bitmap.c int off, int *next) next 127 fs/reiserfs/bitmap.c *next = tmp; next 130 fs/reiserfs/bitmap.c (*next) = off + 1; /* inc offset to avoid looping. */ next 150 fs/reiserfs/bitmap.c int end, next; next 194 fs/reiserfs/bitmap.c next = end; next 203 fs/reiserfs/bitmap.c if (unfm && is_block_in_journal(s, bmap_n, end, &next)) next 269 fs/reiserfs/bitmap.c *beg = next; next 553 fs/reiserfs/bitmap.c ei = list_entry(plist->next, struct reiserfs_inode_info, next 195 fs/reiserfs/inode.c struct buffer_head *head, *next, *bh; next 201 fs/reiserfs/inode.c next = bh->b_this_page; next 205 fs/reiserfs/inode.c bh = next; next 2672 fs/reiserfs/inode.c struct buffer_head *next = bh->b_this_page; next 2678 fs/reiserfs/inode.c bh = next; next 2731 fs/reiserfs/inode.c struct buffer_head *next = bh->b_this_page; next 2738 fs/reiserfs/inode.c bh = next; next 3161 fs/reiserfs/inode.c struct buffer_head *head, *bh, *next; next 3180 fs/reiserfs/inode.c next = bh->b_this_page; next 3195 fs/reiserfs/inode.c bh = next; next 161 fs/reiserfs/journal.c struct list_head *entry = journal->j_bitmap_nodes.next; next 261 fs/reiserfs/journal.c struct list_head *next = journal->j_bitmap_nodes.next; next 264 fs/reiserfs/journal.c while (next != &journal->j_bitmap_nodes) { next 265 fs/reiserfs/journal.c bn = list_entry(next, struct reiserfs_bitmap_node, list); next 266 fs/reiserfs/journal.c list_del(next); next 269 fs/reiserfs/journal.c next = journal->j_bitmap_nodes.next; next 359 fs/reiserfs/journal.c head[0].next = head + 1; next 362 fs/reiserfs/journal.c head[i].next = head + (i + 1); /* if last one, overwrite it after the if */ next 364 fs/reiserfs/journal.c head[num_cnodes - 1].next = NULL; next 385 fs/reiserfs/journal.c if (cn->next) { next 386 fs/reiserfs/journal.c cn->next->prev = NULL; next 388 fs/reiserfs/journal.c journal->j_cnode_free_list = cn->next; next 406 fs/reiserfs/journal.c cn->next = journal->j_cnode_free_list; next 588 fs/reiserfs/journal.c jl = JOURNAL_LIST_ENTRY(entry->next); next 809 fs/reiserfs/journal.c jh = JH_ENTRY(list->next); next 940 fs/reiserfs/journal.c entry = entry->next; next 1222 fs/reiserfs/journal.c cn = cn->next; next 1311 fs/reiserfs/journal.c entry = journal->j_journal_list.next; next 1509 fs/reiserfs/journal.c cn = cn->next; next 1559 fs/reiserfs/journal.c cn = cn->next; next 1649 fs/reiserfs/journal.c goto next; next 1673 fs/reiserfs/journal.c next: next 1674 fs/reiserfs/journal.c cn = cn->next; next 1712 fs/reiserfs/journal.c cn = cn->next; next 1755 fs/reiserfs/journal.c entry = jl->j_list.next; next 1815 fs/reiserfs/journal.c if (tjl->j_list.next == &journal->j_journal_list) next 1817 fs/reiserfs/journal.c tjl = JOURNAL_LIST_ENTRY(tjl->j_list.next); next 3360 fs/reiserfs/journal.c cn->next = NULL; next 3364 fs/reiserfs/journal.c journal->j_last->next = cn; next 3429 fs/reiserfs/journal.c cn->prev->next = cn->next; next 3431 fs/reiserfs/journal.c if (cn->next) { next 3432 fs/reiserfs/journal.c cn->next->prev = cn->prev; next 3435 fs/reiserfs/journal.c journal->j_first = cn->next; next 3976 fs/reiserfs/journal.c struct reiserfs_journal_cnode *cn, *next, *jl_cn; next 4116 fs/reiserfs/journal.c for (i = 0, cn = journal->j_first; cn; cn = cn->next, i++) { next 4127 fs/reiserfs/journal.c jl_cn->next = NULL; next 4129 fs/reiserfs/journal.c last_cn->next = jl_cn; next 4219 fs/reiserfs/journal.c next = cn->next; next 4221 fs/reiserfs/journal.c cn = next; next 1400 fs/reiserfs/lbalance.c int prev, next; next 1404 fs/reiserfs/lbalance.c next = next 1416 fs/reiserfs/lbalance.c if (next && next >= deh_location(&deh[i])) next 238 fs/reiserfs/reiserfs.h struct reiserfs_journal_cnode *next; /* next in transaction list */ next 2943 fs/reiserfs/reiserfs.h int bit_nr, int searchall, b_blocknr_t *next); next 1505 fs/reiserfs/stree.c struct buffer_head *next; next 1516 fs/reiserfs/stree.c next = bh->b_this_page; next 1530 fs/reiserfs/stree.c bh = next; next 196 fs/romfs/super.c nextfh = be32_to_cpu(ri.next); next 249 fs/romfs/super.c if ((be32_to_cpu(ri.next) & ROMFH_TYPE) == ROMFH_HRD) next 256 fs/romfs/super.c offset = be32_to_cpu(ri.next) & ROMFH_MASK; next 297 fs/romfs/super.c nextfh = be32_to_cpu(ri.next); next 98 fs/select.c struct poll_table_page * next; next 154 fs/select.c p = p->next; next 176 fs/select.c new_table->next = table; next 827 fs/select.c struct poll_list *next; next 894 fs/select.c for (walk = list; walk != NULL; walk = walk->next) { next 978 fs/select.c walk->next = NULL; next 992 fs/select.c walk = walk->next = kmalloc(struct_size(walk, entries, len), next 1004 fs/select.c for (walk = head; walk; walk = walk->next) { next 1015 fs/select.c walk = head->next; next 1018 fs/select.c walk = walk->next; next 122 fs/seq_file.c p = m->op->next(m, p, &m->index); next 235 fs/seq_file.c p = m->op->next(m, p, &m->index); next 258 fs/seq_file.c p = m->op->next(m, p, &m->index); next 574 fs/seq_file.c op->next = single_next; next 921 fs/seq_file.c lh = ((struct list_head *)v)->next; next 979 fs/seq_file.c return node->next; next 1050 fs/seq_file.c return rcu_dereference(node->next); next 1094 fs/seq_file.c if (node->next) next 1095 fs/seq_file.c return node->next; next 714 fs/ubifs/commit.c i = list_entry(list.next, struct idx_node, list); next 2306 fs/ubifs/debug.c for (cur = head->next; cur->next != head; cur = cur->next) { next 2312 fs/ubifs/debug.c sb = container_of(cur->next, struct ubifs_scan_node, list); next 2373 fs/ubifs/debug.c for (cur = head->next; cur->next != head; cur = cur->next) { next 2379 fs/ubifs/debug.c sb = container_of(cur->next, struct ubifs_scan_node, list); next 546 fs/ubifs/gc.c snod = list_entry(sleb->nodes.next, struct ubifs_scan_node, list); next 978 fs/ubifs/gc.c idx_gc = list_entry(c->idx_gc.next, struct ubifs_gced_idx_leb, next 999 fs/ubifs/gc.c idx_gc = list_entry(c->idx_gc.next, struct ubifs_gced_idx_leb, list); next 522 fs/ubifs/log.c bud = list_entry(c->old_buds.next, struct ubifs_bud, list); next 787 fs/ubifs/lprops.c lprops = list_entry(c->empty_list.next, struct ubifs_lprops, list); next 810 fs/ubifs/lprops.c lprops = list_entry(c->freeable_list.next, struct ubifs_lprops, list); next 834 fs/ubifs/lprops.c lprops = list_entry(c->frdi_idx_list.next, struct ubifs_lprops, list); next 878 fs/ubifs/recovery.c snod = list_entry(sleb->nodes.next, next 1082 fs/ubifs/recovery.c ucleb = list_entry(c->unclean_leb_list.next, next 516 fs/ubifs/replay.c struct ubifs_bud *next; next 550 fs/ubifs/replay.c next = list_entry(bud->list.next, struct ubifs_bud, list); next 551 fs/ubifs/replay.c if (!list_is_last(&next->list, &jh->buds_list)) next 554 fs/ubifs/replay.c err = ubifs_leb_read(c, next->lnum, (char *)&data, next->start, 4, 1); next 877 fs/ubifs/replay.c b = list_entry(c->replay_buds.next, struct bud_entry, list); next 1017 fs/ubifs/replay.c snod = list_entry(sleb->nodes.next, struct ubifs_scan_node, list); next 361 fs/ubifs/scan.c node = list_entry(head->next, struct ubifs_scan_node, list); next 156 fs/ubifs/shrinker.c p = ubifs_infos.next; next 168 fs/ubifs/shrinker.c p = p->next; next 178 fs/ubifs/shrinker.c p = p->next; next 191 fs/ubifs/shrinker.c p = p->next; next 878 fs/ubifs/super.c orph = list_entry(c->orph_list.next, struct ubifs_orphan, list); next 1138 fs/ubifs/super.c ucleb = list_entry(c->unclean_leb_list.next, next 1146 fs/ubifs/super.c bud = list_entry(c->old_buds.next, struct ubifs_bud, list); next 1018 fs/udf/inode.c int next = laarr[start].extLocation.logicalBlockNum + next 1024 fs/udf/inode.c next, (UDF_DEFAULT_PREALLOC_BLOCKS > length ? next 1036 fs/udf/inode.c laarr[c + 1].extLocation.logicalBlockNum = next; next 57 fs/udf/unicode.c unicode_t next; next 72 fs/udf/unicode.c next = str_i[str_i_idx++] << 8; next 73 fs/udf/unicode.c next |= str_i[str_i_idx++]; next 74 fs/udf/unicode.c if ((next & SURROGATE_MASK) != SURROGATE_PAIR || next 75 fs/udf/unicode.c !(next & SURROGATE_LOW)) { next 82 fs/udf/unicode.c (next & SURROGATE_CHAR_MASK); next 350 fs/unicode/mkutf8data.c struct tree *next; next 1050 fs/unicode/mkutf8data.c struct tree *next; next 1094 fs/unicode/mkutf8data.c next = tree->next; next 1096 fs/unicode/mkutf8data.c assert(next); next 1097 fs/unicode/mkutf8data.c n = next->root; next 1116 fs/unicode/mkutf8data.c next = next->next; next 1650 fs/unicode/mkutf8data.c trees[trees_count-2].next = &trees[trees_count-1]; next 1654 fs/unicode/mkutf8data.c trees[i].next = &trees[trees_count-2]; next 1656 fs/unicode/mkutf8data.c trees[i+1].next = &trees[trees_count-1]; next 1470 fs/userfaultfd.c goto next; next 1482 fs/userfaultfd.c next: next 1632 fs/userfaultfd.c goto next; next 1644 fs/userfaultfd.c next: next 5820 fs/xfs/libxfs/xfs_bmap.c struct xfs_bmbt_irec got, next; next 5872 fs/xfs/libxfs/xfs_bmap.c if (xfs_iext_peek_next_extent(ifp, &icur, &next)) { next 5873 fs/xfs/libxfs/xfs_bmap.c if (new_startoff + got.br_blockcount > next.br_startoff) { next 5884 fs/xfs/libxfs/xfs_bmap.c if (xfs_bmse_can_merge(&got, &next, offset_shift_fsb)) next 1870 fs/xfs/libxfs/xfs_ialloc.c goto next; next 1892 fs/xfs/libxfs/xfs_ialloc.c next: next 125 fs/xfs/libxfs/xfs_iext_tree.c struct xfs_iext_leaf *next; next 241 fs/xfs/libxfs/xfs_iext_tree.c cur->leaf->next) { next 242 fs/xfs/libxfs/xfs_iext_tree.c cur->leaf = cur->leaf->next; next 571 fs/xfs/libxfs/xfs_iext_tree.c if (leaf->next) next 572 fs/xfs/libxfs/xfs_iext_tree.c leaf->next->prev = new; next 573 fs/xfs/libxfs/xfs_iext_tree.c new->next = leaf->next; next 575 fs/xfs/libxfs/xfs_iext_tree.c leaf->next = new; next 700 fs/xfs/libxfs/xfs_iext_tree.c struct xfs_iext_node *next = parent->ptrs[*pos + 1]; next 701 fs/xfs/libxfs/xfs_iext_tree.c int nr_next = xfs_iext_node_nr_entries(next, 0), i; next 710 fs/xfs/libxfs/xfs_iext_tree.c node->keys[nr_entries + i] = next->keys[i]; next 711 fs/xfs/libxfs/xfs_iext_tree.c node->ptrs[nr_entries + i] = next->ptrs[i]; next 715 fs/xfs/libxfs/xfs_iext_tree.c return next; next 818 fs/xfs/libxfs/xfs_iext_tree.c if (leaf->next) { next 819 fs/xfs/libxfs/xfs_iext_tree.c int nr_next = xfs_iext_leaf_nr_entries(ifp, leaf->next, 0), i; next 829 fs/xfs/libxfs/xfs_iext_tree.c leaf->next->recs[i]; next 832 fs/xfs/libxfs/xfs_iext_tree.c if (cur->leaf == leaf->next) { next 837 fs/xfs/libxfs/xfs_iext_tree.c offset = xfs_iext_leaf_key(leaf->next, 0); next 838 fs/xfs/libxfs/xfs_iext_tree.c leaf = leaf->next; next 846 fs/xfs/libxfs/xfs_iext_tree.c leaf->prev->next = leaf->next; next 847 fs/xfs/libxfs/xfs_iext_tree.c if (leaf->next) next 848 fs/xfs/libxfs/xfs_iext_tree.c leaf->next->prev = leaf->prev; next 891 fs/xfs/libxfs/xfs_iext_tree.c if (ifp->if_height > 1 && leaf->next) next 892 fs/xfs/libxfs/xfs_iext_tree.c cur->leaf = leaf->next; next 944 fs/xfs/libxfs/xfs_iext_tree.c if (ifp->if_height == 1 || !cur->leaf->next) next 946 fs/xfs/libxfs/xfs_iext_tree.c cur->leaf = cur->leaf->next; next 129 fs/xfs/scrub/bitmap.c lp = bitmap->list.next; next 143 fs/xfs/scrub/bitmap.c lp = lp->next; next 171 fs/xfs/scrub/bitmap.c lp = lp->next; next 175 fs/xfs/scrub/bitmap.c lp = lp->next; next 195 fs/xfs/scrub/bitmap.c lp = lp->next; next 90 fs/xfs/xfs_aops.c struct bio *last = ioend->io_bio, *next; next 94 fs/xfs/xfs_aops.c for (bio = &ioend->io_inline_bio; bio; bio = next) { next 103 fs/xfs/xfs_aops.c next = NULL; next 105 fs/xfs/xfs_aops.c next = bio->bi_private; next 297 fs/xfs/xfs_aops.c struct xfs_ioend *next) next 299 fs/xfs/xfs_aops.c if (ioend->io_bio->bi_status != next->io_bio->bi_status) next 301 fs/xfs/xfs_aops.c if ((ioend->io_fork == XFS_COW_FORK) ^ (next->io_fork == XFS_COW_FORK)) next 304 fs/xfs/xfs_aops.c (next->io_state == XFS_EXT_UNWRITTEN)) next 306 fs/xfs/xfs_aops.c if (ioend->io_offset + ioend->io_size != next->io_offset) next 320 fs/xfs/xfs_aops.c struct xfs_ioend *next) next 323 fs/xfs/xfs_aops.c ioend->io_append_trans = next->io_append_trans; next 324 fs/xfs/xfs_aops.c next->io_append_trans = NULL; next 326 fs/xfs/xfs_aops.c xfs_setfilesize_ioend(next, -ECANCELED); next 876 fs/xfs/xfs_aops.c struct xfs_ioend *ioend, *next; next 951 fs/xfs/xfs_aops.c list_for_each_entry_safe(ioend, next, &submit_list, io_list) { next 494 fs/xfs/xfs_log_cil.c struct xfs_log_vec *next = lv->lv_next; next 496 fs/xfs/xfs_log_cil.c lv = next; next 1001 fs/xfs/xfs_log_cil.c struct xfs_log_item *lip, *next; next 1036 fs/xfs/xfs_log_cil.c list_for_each_entry_safe(lip, next, &tp->t_items, li_trans) { next 2619 fs/xfs/xfs_log_recover.c goto next; next 2625 fs/xfs/xfs_log_recover.c goto next; next 2633 fs/xfs/xfs_log_recover.c goto next; next 2641 fs/xfs/xfs_log_recover.c next: next 4109 fs/xfs/xfs_log_recover.c struct xlog_recover_item *next; next 4121 fs/xfs/xfs_log_recover.c list_for_each_entry_safe(item, next, &trans->r_itemq, ri_list) { next 232 fs/xfs/xfs_mru_cache.c struct xfs_mru_cache_elem *elem, *next; next 236 fs/xfs/xfs_mru_cache.c list_for_each_entry_safe(elem, next, &mru->reap_list, list_node) { next 249 fs/xfs/xfs_mru_cache.c list_for_each_entry_safe(elem, next, &tmp, list_node) { next 270 fs/xfs/xfs_mru_cache.c unsigned long now, next; next 277 fs/xfs/xfs_mru_cache.c next = _xfs_mru_cache_migrate(mru, jiffies); next 280 fs/xfs/xfs_mru_cache.c mru->queued = next; next 283 fs/xfs/xfs_mru_cache.c if (next <= now) next 284 fs/xfs/xfs_mru_cache.c next = 0; next 286 fs/xfs/xfs_mru_cache.c next -= now; next 287 fs/xfs/xfs_mru_cache.c queue_delayed_work(xfs_mru_reap_wq, &mru->work, next); next 1424 fs/xfs/xfs_reflink.c goto next; next 1457 fs/xfs/xfs_reflink.c next: next 1495 fs/xfs/xfs_reflink.c goto next; next 1509 fs/xfs/xfs_reflink.c next: next 239 fs/xfs/xfs_rtalloc.c xfs_rtblock_t next; /* next rtblock to try */ next 254 fs/xfs/xfs_rtalloc.c error = xfs_rtcheck_range(mp, tp, i, maxlen, 1, &next, &stat); next 280 fs/xfs/xfs_rtalloc.c thislen = next - i; next 289 fs/xfs/xfs_rtalloc.c if (next < end) { next 290 fs/xfs/xfs_rtalloc.c error = xfs_rtfind_forw(mp, tp, next, end, &i); next 326 fs/xfs/xfs_rtalloc.c *nextp = next; next 353 fs/xfs/xfs_rtalloc.c xfs_rtblock_t next; /* next block to try (dummy) */ next 359 fs/xfs/xfs_rtalloc.c error = xfs_rtcheck_range(mp, tp, bno, maxlen, 1, &next, &isfree); next 378 fs/xfs/xfs_rtalloc.c maxlen = next - bno; next 772 fs/xfs/xfs_trans.c struct xfs_log_item *lip, *next; next 776 fs/xfs/xfs_trans.c list_for_each_entry_safe(lip, next, &tp->t_items, li_trans) { next 54 fs/xfs/xfs_trans_ail.c next_lip = list_entry(lip->li_ail.next, struct xfs_log_item, li_ail); next 97 fs/xfs/xfs_trans_ail.c if (lip->li_ail.next == &ailp->ail_head) next 36 include/asm-generic/mmu_context.h struct mm_struct *next, next 21 include/asm-generic/switch_to.h #define switch_to(prev, next, last) \ next 23 include/asm-generic/switch_to.h ((last) = __switch_to((prev), (next))); \ next 210 include/asm-generic/tlb.h struct mmu_gather_batch *next; next 384 include/drm/drm_bridge.h struct drm_bridge *next; next 92 include/drm/drm_dp_mst_helper.h struct list_head next; next 396 include/drm/drm_dp_mst_helper.h struct list_head next; next 431 include/drm/drm_dp_mst_helper.h struct list_head next; next 61 include/drm/drm_legacy.h struct drm_buf *next; /**< Kernel-only: used for free list */ next 274 include/drm/drm_mm.h return mm->hole_stack.next; next 371 include/drm/drm_mm.h #define drm_mm_for_each_node_safe(entry, next, mm) \ next 372 include/drm/drm_mm.h list_for_each_entry_safe(entry, next, drm_mm_nodes(mm), node_list) next 35 include/drm/spsc_queue.h struct spsc_node *next; next 69 include/drm/spsc_queue.h node->next = NULL; next 73 include/drm/spsc_queue.h tail = (struct spsc_node **)atomic_long_xchg(&queue->tail, (long)&node->next); next 91 include/drm/spsc_queue.h struct spsc_node *next, *node; next 101 include/drm/spsc_queue.h next = READ_ONCE(node->next); next 102 include/drm/spsc_queue.h WRITE_ONCE(queue->head, next); next 104 include/drm/spsc_queue.h if (unlikely(!next)) { next 108 include/drm/spsc_queue.h (long)&node->next, (long) &queue->head) != (long)&node->next) { next 112 include/drm/spsc_queue.h } while (unlikely(!(queue->head = READ_ONCE(node->next)))); next 87 include/keys/rxrpc-type.h struct rxrpc_key_token *next; /* the next token in the list */ next 21 include/linux/adb.h struct adb_request *next; next 70 include/linux/agp_backend.h struct agp_memory *next; next 92 include/linux/agpgart.h struct agp_client *next; next 100 include/linux/agpgart.h struct agp_controller *next; next 115 include/linux/agpgart.h struct agp_file_private *next; next 14 include/linux/atalk.h struct atalk_route *next; next 32 include/linux/atalk.h struct atalk_iface *next; next 279 include/linux/bpf_verifier.h struct bpf_verifier_state_list *next; next 375 include/linux/compat.h compat_uptr_t next; next 157 include/linux/console.h struct console *next; next 164 include/linux/console.h for (con = console_drivers; con != NULL; con = con->next) next 83 include/linux/debugobjects.h unsigned int expect, unsigned int next); next 24 include/linux/dm-io.h struct page_list *next; next 508 include/linux/dmaengine.h struct dma_async_tx_descriptor *next; next 555 include/linux/dmaengine.h static inline void txd_chain(struct dma_async_tx_descriptor *txd, struct dma_async_tx_descriptor *next) next 583 include/linux/dmaengine.h static inline void txd_chain(struct dma_async_tx_descriptor *txd, struct dma_async_tx_descriptor *next) next 585 include/linux/dmaengine.h txd->next = next; next 586 include/linux/dmaengine.h next->parent = txd; next 594 include/linux/dmaengine.h txd->next = NULL; next 602 include/linux/dmaengine.h return txd->next; next 1783 include/linux/efi.h phys_addr_t next; // pa of next struct instance next 23 include/linux/frontswap.h struct frontswap_ops *next; /* private pointer to next ops */ next 2233 include/linux/fs.h struct file_system_type * next; next 193 include/linux/ftrace.h struct ftrace_ops __rcu *next; next 30 include/linux/hdlc.h struct hdlc_proto *next; /* next protocol in the list */ next 34 include/linux/huge_mm.h pmd_t *pmd, unsigned long addr, unsigned long next); next 48 include/linux/if_pppox.h struct pppox_sock *next; /* for hash table */ next 71 include/linux/if_tap.h struct list_head next; next 15 include/linux/if_tunnel.h for (pos = rcu_dereference(start); pos; pos = rcu_dereference(pos->next)) next 147 include/linux/if_vlan.h struct vlan_priority_tci_mapping *next; next 206 include/linux/if_vlan.h mp = mp->next; next 78 include/linux/igmp.h struct ip_mc_list *next; next 114 include/linux/interrupt.h struct irqaction *next; next 594 include/linux/interrupt.h struct tasklet_struct *next; next 105 include/linux/jump_label.h struct static_key_mod *next; next 28 include/linux/list.h WRITE_ONCE(list->next, list); next 35 include/linux/list.h struct list_head *next); next 40 include/linux/list.h struct list_head *next) next 58 include/linux/list.h struct list_head *next) next 60 include/linux/list.h if (!__list_add_valid(new, prev, next)) next 63 include/linux/list.h next->prev = new; next 64 include/linux/list.h new->next = next; next 66 include/linux/list.h WRITE_ONCE(prev->next, new); next 79 include/linux/list.h __list_add(new, head, head->next); next 103 include/linux/list.h static inline void __list_del(struct list_head * prev, struct list_head * next) next 105 include/linux/list.h next->prev = prev; next 106 include/linux/list.h WRITE_ONCE(prev->next, next); next 119 include/linux/list.h __list_del(entry->prev, entry->next); next 134 include/linux/list.h __list_del(entry->prev, entry->next); next 140 include/linux/list.h entry->next = LIST_POISON1; next 154 include/linux/list.h new->next = old->next; next 155 include/linux/list.h new->next->prev = new; next 157 include/linux/list.h new->prev->next = new; next 230 include/linux/list.h first->prev->next = last->next; next 231 include/linux/list.h last->next->prev = first->prev; next 233 include/linux/list.h head->prev->next = first; next 236 include/linux/list.h last->next = head; next 259 include/linux/list.h return list->next == head; next 268 include/linux/list.h return READ_ONCE(head->next) == head; next 286 include/linux/list.h struct list_head *next = head->next; next 287 include/linux/list.h return (next == head) && (next == head->prev); next 299 include/linux/list.h first = head->next; next 328 include/linux/list.h return !list_empty(head) && (head->next == head->prev); next 334 include/linux/list.h struct list_head *new_first = entry->next; next 335 include/linux/list.h list->next = head->next; next 336 include/linux/list.h list->next->prev = list; next 338 include/linux/list.h entry->next = list; next 339 include/linux/list.h head->next = new_first; next 363 include/linux/list.h (head->next != entry && head != entry)) next 389 include/linux/list.h if (head->next == entry) { next 393 include/linux/list.h list->next = head->next; next 394 include/linux/list.h list->next->prev = list; next 396 include/linux/list.h list->prev->next = list; next 397 include/linux/list.h head->next = entry; next 403 include/linux/list.h struct list_head *next) next 405 include/linux/list.h struct list_head *first = list->next; next 409 include/linux/list.h prev->next = first; next 411 include/linux/list.h last->next = next; next 412 include/linux/list.h next->prev = last; next 424 include/linux/list.h __list_splice(list, head, head->next); next 450 include/linux/list.h __list_splice(list, head, head->next); next 490 include/linux/list.h list_entry((ptr)->next, type, member) next 513 include/linux/list.h struct list_head *pos__ = READ_ONCE(head__->next); \ next 523 include/linux/list.h list_entry((pos)->member.next, typeof(*(pos)), member) next 539 include/linux/list.h for (pos = (head)->next; pos != (head); pos = pos->next) next 556 include/linux/list.h for (pos = (head)->next, n = pos->next; pos != (head); \ next 557 include/linux/list.h pos = n, n = pos->next) next 743 include/linux/list.h h->next = NULL; next 759 include/linux/list.h struct hlist_node *next = n->next; next 762 include/linux/list.h WRITE_ONCE(*pprev, next); next 763 include/linux/list.h if (next) next 764 include/linux/list.h next->pprev = pprev; next 770 include/linux/list.h n->next = LIST_POISON1; next 785 include/linux/list.h n->next = first; next 787 include/linux/list.h first->pprev = &n->next; next 794 include/linux/list.h struct hlist_node *next) next 796 include/linux/list.h n->pprev = next->pprev; next 797 include/linux/list.h n->next = next; next 798 include/linux/list.h next->pprev = &n->next; next 805 include/linux/list.h n->next = prev->next; next 806 include/linux/list.h prev->next = n; next 807 include/linux/list.h n->pprev = &prev->next; next 809 include/linux/list.h if (n->next) next 810 include/linux/list.h n->next->pprev = &n->next; next 816 include/linux/list.h n->pprev = &n->next; next 821 include/linux/list.h return h->pprev == &h->next; next 831 include/linux/list.h return !n->next && n->pprev == &h->first; next 850 include/linux/list.h for (pos = (head)->first; pos ; pos = pos->next) next 853 include/linux/list.h for (pos = (head)->first; pos && ({ n = pos->next; 1; }); \ next 870 include/linux/list.h pos = hlist_entry_safe((pos)->member.next, typeof(*(pos)), member)) next 878 include/linux/list.h for (pos = hlist_entry_safe((pos)->member.next, typeof(*(pos)), member);\ next 880 include/linux/list.h pos = hlist_entry_safe((pos)->member.next, typeof(*(pos)), member)) next 889 include/linux/list.h pos = hlist_entry_safe((pos)->member.next, typeof(*(pos)), member)) next 900 include/linux/list.h pos && ({ n = pos->member.next; 1; }); \ next 39 include/linux/list_bl.h struct hlist_bl_node *next, **pprev; next 46 include/linux/list_bl.h h->next = NULL; next 82 include/linux/list_bl.h n->next = first; next 84 include/linux/list_bl.h first->pprev = &n->next; next 90 include/linux/list_bl.h struct hlist_bl_node *next) next 92 include/linux/list_bl.h struct hlist_bl_node **pprev = next->pprev; next 95 include/linux/list_bl.h n->next = next; next 96 include/linux/list_bl.h next->pprev = &n->next; next 107 include/linux/list_bl.h n->next = prev->next; next 108 include/linux/list_bl.h n->pprev = &prev->next; next 109 include/linux/list_bl.h prev->next = n; next 111 include/linux/list_bl.h if (n->next) next 112 include/linux/list_bl.h n->next->pprev = &n->next; next 117 include/linux/list_bl.h struct hlist_bl_node *next = n->next; next 125 include/linux/list_bl.h ((unsigned long)next | next 127 include/linux/list_bl.h if (next) next 128 include/linux/list_bl.h next->pprev = pprev; next 134 include/linux/list_bl.h n->next = LIST_POISON1; next 173 include/linux/list_bl.h pos = pos->next) next 185 include/linux/list_bl.h pos && ({ n = pos->next; 1; }) && \ next 26 include/linux/list_nulls.h struct hlist_nulls_node *next, **pprev; next 74 include/linux/list_nulls.h n->next = first; next 78 include/linux/list_nulls.h WRITE_ONCE(first->pprev, &n->next); next 83 include/linux/list_nulls.h struct hlist_nulls_node *next = n->next; next 86 include/linux/list_nulls.h WRITE_ONCE(*pprev, next); next 87 include/linux/list_nulls.h if (!is_a_nulls(next)) next 88 include/linux/list_nulls.h WRITE_ONCE(next->pprev, pprev); next 109 include/linux/list_nulls.h pos = pos->next) next 121 include/linux/list_nulls.h pos = pos->next) next 59 include/linux/llist.h struct llist_node *next; next 115 include/linux/llist.h for ((pos) = (node); pos; (pos) = (pos)->next) next 134 include/linux/llist.h for ((pos) = (node); (pos) && ((n) = (pos)->next, true); (pos) = (n)) next 154 include/linux/llist.h (pos) = llist_entry((pos)->member.next, typeof(*(pos)), member)) next 176 include/linux/llist.h (n = llist_entry(pos->member.next, typeof(*n), member), true); \ next 194 include/linux/llist.h return node->next; next 727 include/linux/mlx5/device.h __be64 next; next 240 include/linux/mlx5/driver.h struct mlx5_cmd_mailbox *next; next 320 include/linux/mlx5/driver.h struct mlx5_cmd_mailbox *next; next 107 include/linux/mm_types.h struct page *next; next 360 include/linux/mm_types.h struct core_thread *next; next 25 include/linux/mmc/sdio_func.h struct sdio_func_tuple *next; next 13 include/linux/msg.h struct msg_msgseg *next; next 92 include/linux/n_r3964.h struct r3964_client_info *next; next 110 include/linux/n_r3964.h struct r3964_message *next; next 124 include/linux/n_r3964.h struct r3964_block_header *next; next 2580 include/linux/netdevice.h lh = dev->dev_list.next; next 2597 include/linux/netdevice.h net_device_entry(net->dev_base_head.next); next 4279 include/linux/netdevice.h for (iter = (dev)->adj_list.lower.next, \ next 4294 include/linux/netdevice.h for (iter = (dev)->adj_list.lower.next, \ next 314 include/linux/netfilter.h struct sk_buff *skb, *next; next 318 include/linux/netfilter.h list_for_each_entry_safe(skb, next, head, list) { next 34 include/linux/nls.h struct nls_table *next; next 56 include/linux/notifier.h struct notifier_block __rcu *next; next 35 include/linux/of.h struct property *next; next 356 include/linux/of.h for (pp = dn->properties; pp != NULL; pp = pp->next) next 10 include/linux/osq_lock.h struct optimistic_spin_node *next, *prev; next 58 include/linux/page_ext.h void *next = curr; next 59 include/linux/page_ext.h next += page_ext_size; next 60 include/linux/page_ext.h return next; next 30 include/linux/pagewalk.h unsigned long next, struct mm_walk *walk); next 32 include/linux/pagewalk.h unsigned long next, struct mm_walk *walk); next 34 include/linux/pagewalk.h unsigned long next, struct mm_walk *walk); next 35 include/linux/pagewalk.h int (*pte_hole)(unsigned long addr, unsigned long next, next 38 include/linux/pagewalk.h unsigned long addr, unsigned long next, next 40 include/linux/pagewalk.h int (*test_walk)(unsigned long addr, unsigned long next, next 147 include/linux/parport.h struct pardevice *next; next 279 include/linux/pci.h struct hlist_node next; next 79 include/linux/perf_event.h struct perf_raw_frag *next; next 897 include/linux/perf_event.h struct task_struct *next); next 1162 include/linux/perf_event.h struct task_struct *next) next 1167 include/linux/perf_event.h __perf_event_task_sched_out(prev, next); next 1358 include/linux/perf_event.h struct task_struct *next) { } next 282 include/linux/plist.h return list_entry(head->node_list.next, next 371 include/linux/pnp.h struct pnp_id *next; next 296 include/linux/preempt.h struct task_struct *next); next 11 include/linux/raid/xor.h struct xor_block_template *next; next 33 include/linux/rculist.h WRITE_ONCE(list->next, list); next 41 include/linux/rculist.h #define list_next_rcu(list) (*((struct list_head __rcu **)(&(list)->next))) next 68 include/linux/rculist.h struct list_head *prev, struct list_head *next) next 70 include/linux/rculist.h if (!__list_add_valid(new, prev, next)) next 73 include/linux/rculist.h new->next = next; next 76 include/linux/rculist.h next->prev = new; next 97 include/linux/rculist.h __list_add_rcu(new, head, head->next); next 191 include/linux/rculist.h new->next = old->next; next 194 include/linux/rculist.h new->next->prev = new; next 219 include/linux/rculist.h struct list_head *next, next 222 include/linux/rculist.h struct list_head *first = list->next; next 250 include/linux/rculist.h last->next = next; next 253 include/linux/rculist.h next->prev = last; next 268 include/linux/rculist.h __list_splice_init_rcu(list, head, head->next, sync); next 334 include/linux/rculist.h struct list_head *__next = READ_ONCE(__ptr->next); \ next 354 include/linux/rculist.h struct list_head *__next = READ_ONCE(__ptr->next); \ next 372 include/linux/rculist.h pos = list_entry_rcu((head)->next, typeof(*pos), member); \ next 374 include/linux/rculist.h pos = list_entry_rcu(pos->member.next, typeof(*pos), member)) next 406 include/linux/rculist.h for (pos = list_entry_lockless((head)->next, typeof(*pos), member); \ next 408 include/linux/rculist.h pos = list_entry_lockless(pos->member.next, typeof(*pos), member)) next 429 include/linux/rculist.h for (pos = list_entry_rcu(pos->member.next, typeof(*pos), member); \ next 431 include/linux/rculist.h pos = list_entry_rcu(pos->member.next, typeof(*pos), member)) next 452 include/linux/rculist.h pos = list_entry_rcu(pos->member.next, typeof(*(pos)), member)) next 489 include/linux/rculist.h struct hlist_node *next = old->next; next 491 include/linux/rculist.h new->next = next; next 494 include/linux/rculist.h if (next) next 495 include/linux/rculist.h new->next->pprev = &new->next; next 503 include/linux/rculist.h #define hlist_next_rcu(node) (*((struct hlist_node __rcu **)(&(node)->next))) next 530 include/linux/rculist.h n->next = first; next 534 include/linux/rculist.h first->pprev = &n->next; next 562 include/linux/rculist.h for (i = h->first; i; i = i->next) next 566 include/linux/rculist.h n->next = last->next; next 567 include/linux/rculist.h n->pprev = &last->next; next 593 include/linux/rculist.h struct hlist_node *next) next 595 include/linux/rculist.h n->pprev = next->pprev; next 596 include/linux/rculist.h n->next = next; next 598 include/linux/rculist.h next->pprev = &n->next; next 622 include/linux/rculist.h n->next = prev->next; next 623 include/linux/rculist.h n->pprev = &prev->next; next 625 include/linux/rculist.h if (n->next) next 626 include/linux/rculist.h n->next->pprev = &n->next; next 107 include/linux/rculist_bl.h n->next = first; next 109 include/linux/rculist_bl.h first->pprev = &n->next; next 127 include/linux/rculist_bl.h pos = rcu_dereference_raw(pos->next)) next 45 include/linux/rculist_nulls.h (*((struct hlist_nulls_node __rcu __force **)&(node)->next)) next 96 include/linux/rculist_nulls.h n->next = first; next 100 include/linux/rculist_nulls.h WRITE_ONCE(first->pprev, &n->next); next 128 include/linux/rculist_nulls.h for (i = h->first; !is_a_nulls(i); i = i->next) next 132 include/linux/rculist_nulls.h n->next = last->next; next 133 include/linux/rculist_nulls.h n->pprev = &last->next; next 18 include/linux/rhashtable-types.h struct rhash_head __rcu *next; next 23 include/linux/rhashtable-types.h struct rhlist_head __rcu *next; next 424 include/linux/rhashtable.h pos = rht_dereference_bucket((pos)->next, tbl, hash)) next 448 include/linux/rhashtable.h pos = rht_dereference_bucket((pos)->next, tbl, hash)) next 475 include/linux/rhashtable.h #define rht_for_each_entry_safe(tpos, pos, next, tbl, hash, member) \ next 477 include/linux/rhashtable.h next = !rht_is_a_nulls(pos) ? \ next 478 include/linux/rhashtable.h rht_dereference_bucket(pos->next, tbl, hash) : NULL; \ next 480 include/linux/rhashtable.h pos = next, \ next 481 include/linux/rhashtable.h next = !rht_is_a_nulls(pos) ? \ next 482 include/linux/rhashtable.h rht_dereference_bucket(pos->next, tbl, hash) : NULL) next 499 include/linux/rhashtable.h pos = rcu_dereference_raw(pos->next)) next 515 include/linux/rhashtable.h pos = rcu_dereference_raw(pos->next)) next 534 include/linux/rhashtable.h pos = rht_dereference_bucket_rcu(pos->next, tbl, hash)) next 562 include/linux/rhashtable.h for (pos = list; pos; pos = rcu_dereference_raw(pos->next)) next 576 include/linux/rhashtable.h pos = rcu_dereference_raw(pos->next)) next 748 include/linux/rhashtable.h pprev = &head->next; next 761 include/linux/rhashtable.h RCU_INIT_POINTER(list->next, plist); next 762 include/linux/rhashtable.h head = rht_dereference_bucket(head->next, tbl, hash); next 763 include/linux/rhashtable.h RCU_INIT_POINTER(list->rhead.next, head); next 786 include/linux/rhashtable.h RCU_INIT_POINTER(obj->next, head); next 791 include/linux/rhashtable.h RCU_INIT_POINTER(list->next, NULL); next 1019 include/linux/rhashtable.h pprev = &he->next; next 1025 include/linux/rhashtable.h lpprev = &list->next; next 1026 include/linux/rhashtable.h list = rht_dereference_bucket(list->next, next 1033 include/linux/rhashtable.h list = rht_dereference_bucket(list->next, tbl, hash); next 1039 include/linux/rhashtable.h obj = rht_dereference_bucket(obj->next, tbl, hash); next 1043 include/linux/rhashtable.h list = rht_dereference_bucket(list->next, tbl, hash); next 1045 include/linux/rhashtable.h RCU_INIT_POINTER(list->rhead.next, obj); next 1172 include/linux/rhashtable.h pprev = &he->next; next 1176 include/linux/rhashtable.h rcu_assign_pointer(obj_new->next, obj_old->next); next 157 include/linux/rmap.h struct vm_area_struct *next) next 159 include/linux/rmap.h VM_BUG_ON_VMA(vma->anon_vma != next->anon_vma, vma); next 160 include/linux/rmap.h unlink_anon_vmas(next); next 265 include/linux/sbitmap.h goto next; next 282 include/linux/sbitmap.h next: next 621 include/linux/sched.h struct wake_q_node *next; next 380 include/linux/sched/mm.h struct mm_struct *next, next 563 include/linux/sched/signal.h list_entry_rcu((p)->tasks.next, struct task_struct, tasks) next 653 include/linux/sched/signal.h return list_entry_rcu(p->thread_group.next, next 35 include/linux/seq_file.h void * (*next) (struct seq_file *m, void *v, loff_t *pos); next 291 include/linux/skbuff.h struct sk_buff *next; next 689 include/linux/skbuff.h struct sk_buff *next; next 1480 include/linux/skbuff.h skb->next = NULL; next 1497 include/linux/skbuff.h return list->next == (const struct sk_buff *) list; next 1509 include/linux/skbuff.h return READ_ONCE(list->next) == (const struct sk_buff *) list; next 1523 include/linux/skbuff.h return skb->next == (const struct sk_buff *) list; next 1554 include/linux/skbuff.h return skb->next; next 1748 include/linux/skbuff.h struct sk_buff *skb = list_->next; next 1763 include/linux/skbuff.h return list_->next; next 1778 include/linux/skbuff.h struct sk_buff *next = skb->next; next 1780 include/linux/skbuff.h if (next == (struct sk_buff *)list_) next 1781 include/linux/skbuff.h next = NULL; next 1782 include/linux/skbuff.h return next; next 1831 include/linux/skbuff.h list->prev = list->next = (struct sk_buff *)list; next 1863 include/linux/skbuff.h struct sk_buff *prev, struct sk_buff *next, next 1869 include/linux/skbuff.h WRITE_ONCE(newsk->next, next); next 1871 include/linux/skbuff.h WRITE_ONCE(next->prev, newsk); next 1872 include/linux/skbuff.h WRITE_ONCE(prev->next, newsk); next 1878 include/linux/skbuff.h struct sk_buff *next) next 1880 include/linux/skbuff.h struct sk_buff *first = list->next; next 1884 include/linux/skbuff.h WRITE_ONCE(prev->next, first); next 1886 include/linux/skbuff.h WRITE_ONCE(last->next, next); next 1887 include/linux/skbuff.h WRITE_ONCE(next->prev, last); next 1899 include/linux/skbuff.h __skb_queue_splice(list, (struct sk_buff *) head, head->next); next 1915 include/linux/skbuff.h __skb_queue_splice(list, (struct sk_buff *) head, head->next); next 1968 include/linux/skbuff.h __skb_insert(newsk, prev, prev->next, list); next 1975 include/linux/skbuff.h struct sk_buff *next, next 1978 include/linux/skbuff.h __skb_insert(newsk, next->prev, next, list); next 2022 include/linux/skbuff.h struct sk_buff *next, *prev; next 2025 include/linux/skbuff.h next = skb->next; next 2027 include/linux/skbuff.h skb->next = skb->prev = NULL; next 2028 include/linux/skbuff.h WRITE_ONCE(next->prev, prev); next 2029 include/linux/skbuff.h WRITE_ONCE(prev->next, next); next 3405 include/linux/skbuff.h for (skb = (queue)->next; \ next 3407 include/linux/skbuff.h skb = skb->next) next 3410 include/linux/skbuff.h for (skb = (queue)->next, tmp = skb->next; \ next 3412 include/linux/skbuff.h skb = tmp, tmp = skb->next) next 3416 include/linux/skbuff.h skb = skb->next) next 3431 include/linux/skbuff.h for (tmp = skb->next; \ next 3433 include/linux/skbuff.h skb = tmp, tmp = skb->next) next 3461 include/linux/skbuff.h for (iter = skb_shinfo(skb)->frag_list; iter; iter = iter->next) next 58 include/linux/slub_def.h slub_percpu_partial(c) = (p)->next; \ next 358 include/linux/suspend.h struct pbe *next; next 224 include/linux/swap.h unsigned int next; /* Likely next allocation offset */ next 76 include/linux/timer.h .entry = { .next = TIMER_ENTRY_STATIC }, \ next 20 include/linux/trace.h struct trace_export __rcu *next; next 60 include/linux/tty.h struct tty_buffer *next; next 182 include/linux/types.h struct list_head *next, *prev; next 190 include/linux/types.h struct hlist_node *next, **pprev; next 220 include/linux/types.h struct callback_head *next; next 45 include/linux/uprobes.h struct uprobe_consumer *next; next 90 include/linux/uprobes.h struct return_instance *next; /* keep as stack */ next 20 include/linux/user-return-notifier.h struct task_struct *next) next 24 include/linux/user-return-notifier.h set_tsk_thread_flag(next, TIF_USER_RETURN_NOTIFY); next 40 include/linux/user-return-notifier.h struct task_struct *next) next 42 include/linux/vmalloc.h struct vm_struct *next; next 528 include/linux/xarray.h struct xa_limit, u32 *next, gfp_t); next 909 include/linux/xarray.h struct xa_limit limit, u32 *next, gfp_t gfp) next 914 include/linux/xarray.h err = __xa_alloc_cyclic(xa, id, entry, limit, next, gfp); next 942 include/linux/xarray.h struct xa_limit limit, u32 *next, gfp_t gfp) next 947 include/linux/xarray.h err = __xa_alloc_cyclic(xa, id, entry, limit, next, gfp); next 975 include/linux/xarray.h struct xa_limit limit, u32 *next, gfp_t gfp) next 980 include/linux/xarray.h err = __xa_alloc_cyclic(xa, id, entry, limit, next, gfp); next 87 include/media/dmxdev.h struct list_head next; next 66 include/media/drv-intf/saa7146_vv.h struct saa7146_buf *next); next 198 include/media/drv-intf/saa7146_vv.h void saa7146_set_capture(struct saa7146_dev *dev, struct saa7146_buf *buf, struct saa7146_buf *next); next 84 include/media/dvb_demux.h struct dvb_demux_filter *next; next 289 include/media/v4l2-ctrls.h struct v4l2_ctrl_ref *next; next 34 include/net/atmclip.h struct clip_vcc *next; /* next VCC */ next 50 include/net/atmclip.h struct net_device *next; /* next CLIP interface */ next 189 include/net/ax25.h struct ax25_route *next; next 231 include/net/ax25.h struct ax25_dev *next; next 372 include/net/ax25.h struct ax25_protocol *next; next 70 include/net/bond_alb.h u32 next; /* The next Hash table entry index, assigned next 60 include/net/bonding.h netdev_adjacent_get_private(bond_slave_list(bond)->next) : \ next 93 include/net/if_inet6.h struct ipv6_mc_socklist __rcu *next; next 117 include/net/if_inet6.h struct ifmcaddr6 *next; next 122 include/net/ip.h struct ip_ra_chain __rcu *next; next 181 include/net/ip.h iter->frag = skb->next; next 47 include/net/ip6_tunnel.h struct ip6_tnl __rcu *next; /* next tunnel in list */ next 96 include/net/ip_tunnels.h struct ip_tunnel_prl_entry __rcu *next; next 105 include/net/ip_tunnels.h struct ip_tunnel __rcu *next; next 418 include/net/ip_vs.h struct ip_vs_protocol *next; next 478 include/net/ip_vs.h struct ip_vs_proto_data *next; next 172 include/net/ipv6.h iter->frag = skb->next; next 273 include/net/ipv6.h struct ip6_ra_chain *next; next 313 include/net/ipv6.h struct ip6_flowlabel __rcu *next; next 338 include/net/ipv6.h struct ipv6_fl_socklist __rcu *next; next 135 include/net/neighbour.h struct neighbour __rcu *next; next 172 include/net/neighbour.h struct pneigh_entry *next; next 296 include/net/neighbour.h n = rcu_dereference_bh(n->next)) { next 157 include/net/netlabel.h struct netlbl_lsm_catmap *next; next 334 include/net/netlabel.h catmap = catmap->next; next 92 include/net/rose.h struct rose_neigh *next; next 109 include/net/rose.h struct rose_node *next; next 118 include/net/rose.h struct rose_route *next; next 32 include/net/sch_generic.h struct qdisc_rate_table *next; next 236 include/net/sch_generic.h struct Qdisc_ops *next; next 351 include/net/sch_generic.h struct tcf_proto __rcu *next; next 959 include/net/sch_generic.h skb->next = NULL; next 960 include/net/sch_generic.h last->next = skb; next 979 include/net/sch_generic.h skb->next = qh->head; next 992 include/net/sch_generic.h qh->head = skb->next; next 996 include/net/sch_generic.h skb->next = NULL; next 1019 include/net/sch_generic.h skb->next = *to_free; next 1027 include/net/sch_generic.h skb->prev->next = *to_free; next 1029 include/net/sch_generic.h skb->next = *to_free; next 388 include/net/sctp/sctp.h result = list->next; next 415 include/net/sctp/sctp.h return (head->next != head) && (head->next == head->prev); next 1398 include/net/sctp/structs.h struct sctp_stream_out_ext *next; next 131 include/net/slhc_vj.h struct cstate *next; /* next in ring (xmit) */ next 594 include/net/sock.h return hlist_entry_safe(sk->sk_node.next, struct sock, sk_node); next 599 include/net/sock.h return (!is_a_nulls(sk->sk_nulls_node.next)) ? next 600 include/net/sock.h hlist_nulls_entry(sk->sk_nulls_node.next, next 910 include/net/sock.h sk->sk_backlog.tail->next = skb; next 913 include/net/sock.h skb->next = NULL; next 1063 include/net/sock.h if (offsetof(struct sock, sk_node.next) != 0) next 1064 include/net/sock.h memset(sk, 0, offsetof(struct sock, sk_node.next)); next 1796 include/net/tcp.h if (sk->sk_write_queue.next == skb) next 142 include/net/udp.h for (skb = skb_shinfo(skb)->frag_list; skb; skb = skb->next) { next 236 include/net/vxlan.h struct list_head next; /* vxlan's per namespace list */ next 991 include/net/xfrm.h struct xfrm_if __rcu *next; /* next interface in list */ next 1402 include/net/xfrm.h struct xfrm4_protocol __rcu *next; next 1412 include/net/xfrm.h struct xfrm6_protocol __rcu *next; next 1421 include/net/xfrm.h struct xfrm_tunnel __rcu *next; next 1429 include/net/xfrm.h struct xfrm6_tunnel __rcu *next; next 490 include/rdma/ib_mad.h struct ib_mad_send_buf *next; next 1325 include/rdma/ib_verbs.h struct ib_send_wr *next; next 1396 include/rdma/ib_verbs.h struct ib_recv_wr *next; next 2096 include/rdma/ib_verbs.h struct ib_flow_spec_list *next; next 101 include/rdma/rdmavt_mr.h u32 next; /* next unused index (speeds search) */ next 334 include/rdma/rdmavt_qp.h struct rvt_qp __rcu *next; /* link list for QPN hash table */ next 726 include/rdma/rdmavt_qp.h qp = rcu_dereference(qp->next)) next 1575 include/sound/emu10k1.h struct snd_emu10k1_fx8010_irq *next; next 183 include/sound/gus.h struct snd_gf1_mem_block *next; next 203 include/sound/gus.h struct snd_gf1_dma_block *next; next 258 include/sound/opl3.h struct fm_patch *next; next 458 include/sound/pcm.h struct snd_pcm_substream *next; next 1386 include/sound/pcm.h for (s = info->pcm->streams[info->stream].substream; s; s = s->next) next 21 include/sound/pcm_oss.h struct snd_pcm_oss_setup *next; next 20 include/sound/soundfont.h struct snd_sf_zone *next; /* Link to next */ next 38 include/sound/soundfont.h struct snd_sf_sample *next; next 45 include/sound/soundfont.h struct snd_soundfont *next; /* Link to next */ next 142 include/trace/events/sched.h struct task_struct *next), next 144 include/trace/events/sched.h TP_ARGS(preempt, prev, next), next 157 include/trace/events/sched.h memcpy(__entry->next_comm, next->comm, TASK_COMM_LEN); next 162 include/trace/events/sched.h __entry->next_pid = next->pid; next 163 include/trace/events/sched.h __entry->next_prio = next->prio; next 116 include/uapi/drm/drm.h unsigned char next; next 151 include/uapi/drm/i810_drm.h unsigned char next, prev; /* indices to form a circular LRU */ next 180 include/uapi/drm/via_drm.h unsigned char next, prev; /* indices to form a circular LRU */ next 168 include/uapi/linux/dm-ioctl.h __u32 next; next 193 include/uapi/linux/dm-ioctl.h __u32 next; /* offset to the next record from next 202 include/uapi/linux/dm-ioctl.h __u32 next; next 353 include/uapi/linux/fd.h struct floppy_raw_cmd *next; /* used for chaining of raw cmd's next 59 include/uapi/linux/futex.h struct robust_list __user *next; next 36 include/uapi/linux/romfs_fs.h __be32 next; /* low 4 bits see ROMFH_ */ next 78 include/uapi/linux/vfio.h __u32 next; /* Offset of next capability */ next 1130 include/uapi/linux/videodev2.h struct v4l2_clip __user *next; next 98 include/uapi/linux/virtio_ring.h __virtio16 next; next 59 include/xen/grant_table.h struct gnttab_free_callback *next; next 370 include/xen/interface/xen-mca.h unsigned next; next 371 init/do_mounts.c char *p, *next; next 374 init/do_mounts.c for (p = page-1; p; p = next) { next 375 init/do_mounts.c next = strchr(++p, '\n'); next 51 init/initramfs.c struct hash *next; next 66 init/initramfs.c for (p = head + hash(major, minor, ino); *p; p = &(*p)->next) { next 85 init/initramfs.c q->next = NULL; next 96 init/initramfs.c *p = q->next; next 207 init/initramfs.c static void __init read_into(char *buf, unsigned size, enum state next) next 212 init/initramfs.c state = next; next 216 init/initramfs.c next_state = next; next 802 init/main.c struct list_head next; next 829 init/main.c list_add(&entry->next, &blacklisted_initcalls); next 854 init/main.c list_for_each_entry(entry, &blacklisted_initcalls, next) { next 193 ipc/msg.c if (mss->list.next) next 206 ipc/msg.c mss->list.next = NULL; next 38 ipc/msgutil.c struct msg_msgseg *next; next 57 ipc/msgutil.c msg->next = NULL; next 61 ipc/msgutil.c pseg = &msg->next; next 72 ipc/msgutil.c seg->next = NULL; next 73 ipc/msgutil.c pseg = &seg->next; next 99 ipc/msgutil.c for (seg = msg->next; seg != NULL; seg = seg->next) { next 130 ipc/msgutil.c for (dst_pseg = dst->next, src_pseg = src->next; next 132 ipc/msgutil.c dst_pseg = dst_pseg->next, src_pseg = src_pseg->next) { next 159 ipc/msgutil.c for (seg = msg->next; seg != NULL; seg = seg->next) { next 175 ipc/msgutil.c seg = msg->next; next 178 ipc/msgutil.c struct msg_msgseg *tmp = seg->next; next 2319 ipc/sem.c un = list_entry_rcu(ulp->list_proc.next, next 1635 ipc/shm.c struct vm_area_struct *next; next 1669 ipc/shm.c next = vma->vm_next; next 1695 ipc/shm.c vma = next; next 1698 ipc/shm.c vma = next; next 1708 ipc/shm.c next = vma->vm_next; next 1715 ipc/shm.c vma = next; next 859 ipc/util.c .next = sysvipc_proc_next, next 546 kernel/audit_tree.c struct audit_krule *rule, *next; next 549 kernel/audit_tree.c list_for_each_entry_safe(rule, next, &tree->rules, rlist) { next 620 kernel/audit_tree.c for (p = tree->chunks.next; p != &tree->chunks; p = q) { next 622 kernel/audit_tree.c q = p->next; next 685 kernel/audit_tree.c while (cursor.next != &tree_list) { next 692 kernel/audit_tree.c tree = container_of(cursor.next, struct audit_tree, list); next 770 kernel/audit_tree.c victim = list_entry(prune_list.next, next 898 kernel/audit_tree.c while (cursor.next != &tree_list) { next 902 kernel/audit_tree.c tree = container_of(cursor.next, struct audit_tree, list); next 987 kernel/audit_tree.c victim = list_entry(list->next, struct audit_tree, list); next 1016 kernel/audit_tree.c owner = list_entry(chunk->trees.next, next 100 kernel/auditsc.c struct audit_aux_data *next; next 129 kernel/auditsc.c struct audit_tree_refs *next; next 222 kernel/auditsc.c p = p->next; next 241 kernel/auditsc.c p->next = ctx->trees; next 262 kernel/auditsc.c for (q = p; q != ctx->trees; q = q->next, n = 31) { next 280 kernel/auditsc.c q = p->next; next 292 kernel/auditsc.c for (p = ctx->first_trees; p != ctx->trees; p = p->next) { next 863 kernel/auditsc.c struct audit_names *n, *next; next 865 kernel/auditsc.c list_for_each_entry_safe(n, next, &context->names_list, list) { next 883 kernel/auditsc.c context->aux = aux->next; next 887 kernel/auditsc.c context->aux_pids = aux->next; next 1478 kernel/auditsc.c for (aux = context->aux; aux; aux = aux->next) { next 1531 kernel/auditsc.c for (aux = context->aux_pids; aux; aux = aux->next) { next 2414 kernel/auditsc.c axp->d.next = ctx->aux_pids; next 2453 kernel/auditsc.c ax->d.next = context->aux; next 257 kernel/bpf/arraymap.c u32 *next = (u32 *)next_key; next 260 kernel/bpf/arraymap.c *next = 0; next 267 kernel/bpf/arraymap.c *next = index + 1; next 170 kernel/bpf/bpf_lru_list.c struct list_head *cur, *last, *next = inactive; next 177 kernel/bpf/bpf_lru_list.c last = l->next_inactive_rotation->next; next 179 kernel/bpf/bpf_lru_list.c last = last->next; next 189 kernel/bpf/bpf_lru_list.c next = cur->prev; next 194 kernel/bpf/bpf_lru_list.c cur = next; next 198 kernel/bpf/bpf_lru_list.c l->next_inactive_rotation = next; next 579 kernel/bpf/cpumap.c u32 *next = next_key; next 582 kernel/bpf/cpumap.c *next = 0; next 588 kernel/bpf/cpumap.c *next = index + 1; next 240 kernel/bpf/devmap.c struct hlist_node *next; next 244 kernel/bpf/devmap.c hlist_for_each_entry_safe(dev, next, head, index_hlist) { next 277 kernel/bpf/devmap.c u32 *next = next_key; next 280 kernel/bpf/devmap.c *next = 0; next 286 kernel/bpf/devmap.c *next = index + 1; next 308 kernel/bpf/devmap.c u32 idx, *next = next_key; next 326 kernel/bpf/devmap.c *next = next_dev->idx; next 341 kernel/bpf/devmap.c *next = next_dev->idx; next 751 kernel/bpf/devmap.c struct hlist_node *next; next 755 kernel/bpf/devmap.c hlist_for_each_entry_safe(dev, next, head, index_hlist) { next 244 kernel/bpf/hashtab.c BUILD_BUG_ON(offsetof(struct htab_elem, fnode.next) != next 244 kernel/bpf/inode.c .next = map_seq_next, next 240 kernel/bpf/local_storage.c struct bpf_cgroup_storage_key *next = _next_key; next 262 kernel/bpf/local_storage.c next->attach_type = storage->key.attach_type; next 263 kernel/bpf/local_storage.c next->cgroup_inode_id = storage->key.cgroup_inode_id; next 32 kernel/bpf/percpu_freelist.c node->next = head->first; next 96 kernel/bpf/percpu_freelist.c head->first = node->next; next 19 kernel/bpf/percpu_freelist.h struct pcpu_freelist_node *next; next 339 kernel/bpf/reuseport_array.c u32 *next = (u32 *)next_key; next 342 kernel/bpf/reuseport_array.c *next = 0; next 349 kernel/bpf/reuseport_array.c *next = index + 1; next 168 kernel/bpf/verifier.c struct bpf_verifier_stack_elem *next; next 794 kernel/bpf/verifier.c elem = head->next; next 816 kernel/bpf/verifier.c elem->next = env->head; next 1164 kernel/bpf/verifier.c goto next; next 1166 kernel/bpf/verifier.c goto next; next 1172 kernel/bpf/verifier.c next: next 6881 kernel/bpf/verifier.c goto next; next 6884 kernel/bpf/verifier.c goto next; next 6887 kernel/bpf/verifier.c goto next; next 6889 kernel/bpf/verifier.c next: next 6890 kernel/bpf/verifier.c sl = sl->next; next 7342 kernel/bpf/verifier.c goto next; next 7410 kernel/bpf/verifier.c *pprev = sl->next; next 7425 kernel/bpf/verifier.c sl->next = env->free_list; next 7431 kernel/bpf/verifier.c next: next 7432 kernel/bpf/verifier.c pprev = &sl->next; next 7477 kernel/bpf/verifier.c new_sl->next = *explored_state(env, insn_idx); next 9246 kernel/bpf/verifier.c sln = sl->next; next 9259 kernel/bpf/verifier.c sln = sl->next; next 150 kernel/bpf/xskmap.c u32 *next = next_key; next 153 kernel/bpf/xskmap.c *next = 0; next 159 kernel/bpf/xskmap.c *next = index + 1; next 1019 kernel/cgroup/cgroup.c l1 = l1->next; next 1020 kernel/cgroup/cgroup.c l2 = l2->next; next 4151 kernel/cgroup/cgroup.c unsigned long next = last + CGROUP_FILE_NOTIFY_MIN_INTV; next 4153 kernel/cgroup/cgroup.c if (time_in_range(jiffies, last, next)) { next 4154 kernel/cgroup/cgroup.c timer_reduce(&cfile->notify_timer, next); next 4183 kernel/cgroup/cgroup.c struct cgroup_subsys_state *next; next 4208 kernel/cgroup/cgroup.c next = list_entry_rcu(parent->children.next, struct cgroup_subsys_state, sibling); next 4210 kernel/cgroup/cgroup.c next = list_entry_rcu(pos->sibling.next, struct cgroup_subsys_state, sibling); next 4212 kernel/cgroup/cgroup.c list_for_each_entry_rcu(next, &parent->children, sibling) next 4213 kernel/cgroup/cgroup.c if (next->serial_nr > pos->serial_nr) next 4221 kernel/cgroup/cgroup.c if (&next->sibling != &parent->children) next 4222 kernel/cgroup/cgroup.c return next; next 4251 kernel/cgroup/cgroup.c struct cgroup_subsys_state *next; next 4260 kernel/cgroup/cgroup.c next = css_next_child(NULL, pos); next 4261 kernel/cgroup/cgroup.c if (next) next 4262 kernel/cgroup/cgroup.c return next; next 4266 kernel/cgroup/cgroup.c next = css_next_child(pos, pos->parent); next 4267 kernel/cgroup/cgroup.c if (next) next 4268 kernel/cgroup/cgroup.c return next; next 4346 kernel/cgroup/cgroup.c struct cgroup_subsys_state *next; next 4359 kernel/cgroup/cgroup.c next = css_next_child(pos, pos->parent); next 4360 kernel/cgroup/cgroup.c if (next) next 4361 kernel/cgroup/cgroup.c return css_leftmost_descendant(next); next 4401 kernel/cgroup/cgroup.c l = it->tcset_pos->next; next 4414 kernel/cgroup/cgroup.c l = l->next; next 4465 kernel/cgroup/cgroup.c it->task_pos = cset->tasks.next; next 4468 kernel/cgroup/cgroup.c it->task_pos = cset->mg_tasks.next; next 4471 kernel/cgroup/cgroup.c it->task_pos = cset->dying_tasks.next; next 4509 kernel/cgroup/cgroup.c it->task_pos = it->task_pos->next; next 4529 kernel/cgroup/cgroup.c it->task_pos = it->task_pos->next; next 4532 kernel/cgroup/cgroup.c it->task_pos = it->mg_tasks_head->next; next 4536 kernel/cgroup/cgroup.c it->task_pos = it->dying_tasks_head->next; next 209 kernel/crash_core.c goto next; next 217 kernel/crash_core.c next: next 444 kernel/crash_core.c VMCOREINFO_OFFSET(list_head, next); next 950 kernel/debug/debug_core.c .next = NULL, next 714 kernel/debug/kdb/kdb_io.c c = c->next; next 778 kernel/debug/kdb/kdb_io.c c = c->next; next 693 kernel/debug/kdb/kdb_support.c u32 next; /* offset of next header from start of pool */ next 768 kernel/debug/kdb/kdb_support.c if (!h->next) next 771 kernel/debug/kdb/kdb_support.c h = (struct debug_alloc_header *)(debug_alloc_pool + h->next); next 777 kernel/debug/kdb/kdb_support.c if (best->next == 0 && bestprev == NULL && rem < dah_overhead) next 785 kernel/debug/kdb/kdb_support.c h->next = best->next; next 787 kernel/debug/kdb/kdb_support.c h_offset = best->next; next 792 kernel/debug/kdb/kdb_support.c bestprev->next = h_offset; next 825 kernel/debug/kdb/kdb_support.c h->next = dah_first; next 833 kernel/debug/kdb/kdb_support.c if (!prev->next || prev->next > h_offset) next 836 kernel/debug/kdb/kdb_support.c (debug_alloc_pool + prev->next); next 846 kernel/debug/kdb/kdb_support.c h->next = prev->next; next 847 kernel/debug/kdb/kdb_support.c prev->next = h_offset; next 850 kernel/debug/kdb/kdb_support.c if (h_offset + dah_overhead + h->size == h->next) { next 851 kernel/debug/kdb/kdb_support.c struct debug_alloc_header *next; next 852 kernel/debug/kdb/kdb_support.c next = (struct debug_alloc_header *) next 853 kernel/debug/kdb/kdb_support.c (debug_alloc_pool + h->next); next 854 kernel/debug/kdb/kdb_support.c h->size += dah_overhead + next->size; next 855 kernel/debug/kdb/kdb_support.c h->next = next->next; next 856 kernel/debug/kdb/kdb_support.c memset(next, POISON_FREE, dah_overhead - 1); next 857 kernel/debug/kdb/kdb_support.c *((char *)next + dah_overhead - 1) = POISON_END; next 899 kernel/debug/kdb/kdb_support.c (debug_alloc_pool + h_free->next); next 900 kernel/debug/kdb/kdb_support.c } while (h_free->next); next 650 kernel/dma/debug.c entry = list_entry(free_entries.next, struct dma_debug_entry, list); next 848 kernel/events/core.c struct task_struct *next) next 860 kernel/events/core.c cgrp2 = perf_cgroup_from_task(next, NULL); next 1017 kernel/events/core.c struct task_struct *next) next 1040 kernel/events/core.c perf_cgroup_switch(struct task_struct *task, struct task_struct *next) next 1647 kernel/events/core.c struct perf_event *next; next 1649 kernel/events/core.c next = rb_entry_safe(rb_next(&event->group_node), typeof(*event), group_node); next 1650 kernel/events/core.c if (next && next->cpu == event->cpu) next 1651 kernel/events/core.c return next; next 3172 kernel/events/core.c struct task_struct *next) next 3188 kernel/events/core.c next_ctx = next->perf_event_ctxp[ctxn]; next 3212 kernel/events/core.c WRITE_ONCE(ctx->task, next); next 3225 kernel/events/core.c RCU_INIT_POINTER(next->perf_event_ctxp[ctxn], ctx); next 3276 kernel/events/core.c struct task_struct *next, next 3282 kernel/events/core.c if (prev == next) next 3319 kernel/events/core.c struct task_struct *next) next 3324 kernel/events/core.c perf_pmu_sched_task(task, next, false); next 3327 kernel/events/core.c perf_event_switch(task, next, false); next 3330 kernel/events/core.c perf_event_context_sched_out(task, ctxn, next); next 3338 kernel/events/core.c perf_cgroup_sched_out(task, next); next 3741 kernel/events/core.c goto next; next 3763 kernel/events/core.c next: next 6423 kernel/events/core.c frag = frag->next; next 6620 kernel/events/core.c frag = frag->next; next 11573 kernel/events/core.c struct perf_event *child_event, *next; next 11623 kernel/events/core.c list_for_each_entry_safe(child_event, next, &child_ctx->event_list, event_entry) next 769 kernel/events/uprobes.c uc->next = uprobe->consumers; next 785 kernel/events/uprobes.c for (con = &uprobe->consumers; *con; con = &(*con)->next) { next 787 kernel/events/uprobes.c *con = uc->next; next 892 kernel/events/uprobes.c for (uc = uprobe->consumers; uc; uc = uc->next) { next 959 kernel/events/uprobes.c struct map_info *next; next 966 kernel/events/uprobes.c struct map_info *next = info->next; next 968 kernel/events/uprobes.c return next; next 995 kernel/events/uprobes.c prev->next = NULL; next 1006 kernel/events/uprobes.c prev = prev->next; next 1007 kernel/events/uprobes.c info->next = curr; next 1021 kernel/events/uprobes.c curr = curr->next; next 1030 kernel/events/uprobes.c info->next = prev; next 1234 kernel/events/uprobes.c for (con = uprobe->consumers; con && con != uc ; con = con->next) next 1713 kernel/events/uprobes.c struct return_instance *next = ri->next; next 1716 kernel/events/uprobes.c return next; next 1769 kernel/events/uprobes.c for (o = o_utask->return_instances; o; o = o->next) { next 1776 kernel/events/uprobes.c n->next = NULL; next 1779 kernel/events/uprobes.c p = &n->next; next 1925 kernel/events/uprobes.c ri->next = utask->return_instances; next 2086 kernel/events/uprobes.c for (uc = uprobe->consumers; uc; uc = uc->next) { next 2118 kernel/events/uprobes.c for (uc = uprobe->consumers; uc; uc = uc->next) { next 2131 kernel/events/uprobes.c ri = ri->next; /* can't be NULL if chained */ next 2140 kernel/events/uprobes.c struct return_instance *ri, *next; next 2158 kernel/events/uprobes.c next = find_next_ret_chain(ri); next 2159 kernel/events/uprobes.c valid = !next || arch_uretprobe_is_alive(next, RP_CHECK_RET, regs); next 2167 kernel/events/uprobes.c } while (ri != next); next 459 kernel/exit.c self.next = xchg(&core_state->dumper.next, &self); next 212 kernel/fail_function.c .next = fei_seq_next, next 929 kernel/fork.c tsk->wake_q.next = NULL; next 918 kernel/futex.c struct list_head *next, *head = &curr->pi_state_list; next 932 kernel/futex.c next = head->next; next 933 kernel/futex.c pi_state = list_entry(next, struct futex_pi_state, list); next 962 kernel/futex.c if (head->next != next) { next 1675 kernel/futex.c struct futex_q *this, *next; next 1695 kernel/futex.c plist_for_each_entry_safe(this, next, &hb->chain, list) { next 1777 kernel/futex.c struct futex_q *this, *next; next 1825 kernel/futex.c plist_for_each_entry_safe(this, next, &hb1->chain, list) { next 1839 kernel/futex.c plist_for_each_entry_safe(this, next, &hb2->chain, list) { next 2023 kernel/futex.c struct futex_q *this, *next; next 2199 kernel/futex.c plist_for_each_entry_safe(this, next, &hb1->chain, list) { next 3710 kernel/futex.c if (fetch_robust_entry(&entry, &head->list.next, &pi)) next 3730 kernel/futex.c rc = fetch_robust_entry(&next_entry, &entry->next, &next_pi); next 4011 kernel/futex.c if (compat_fetch_robust_entry(&uentry, &entry, &head->list.next, &pi)) next 4033 kernel/futex.c (compat_uptr_t __user *)&entry->next, &next_pi); next 135 kernel/gcov/fs.c .next = gcov_seq_next, next 568 kernel/gcov/fs.c char *next; next 577 kernel/gcov/fs.c for (curr = filename; (next = strchr(curr, '/')); curr = next + 1) { next 578 kernel/gcov/fs.c if (curr == next) next 580 kernel/gcov/fs.c *next = 0; next 75 kernel/gcov/gcc_3_4.c struct gcov_info *next; next 114 kernel/gcov/gcc_3_4.c return info->next; next 123 kernel/gcov/gcc_3_4.c info->next = gcov_info_head; next 135 kernel/gcov/gcc_3_4.c prev->next = info->next; next 137 kernel/gcov/gcc_3_4.c gcov_info_head = info->next; next 89 kernel/gcov/gcc_4_7.c struct gcov_info *next; next 127 kernel/gcov/gcc_4_7.c return info->next; next 136 kernel/gcov/gcc_4_7.c info->next = gcov_info_head; next 148 kernel/gcov/gcc_4_7.c prev->next = info->next; next 150 kernel/gcov/gcc_4_7.c gcov_info_head = info->next; next 282 kernel/gcov/gcc_4_7.c dup->next = NULL; next 162 kernel/irq/internals.h for (act = desc->action; act; act = act->next) next 56 kernel/irq/ipi.c unsigned int next; next 69 kernel/irq/ipi.c next = cpumask_next_zero(offset, dest); next 70 kernel/irq/ipi.c if (next < nr_cpu_ids) next 71 kernel/irq/ipi.c next = cpumask_next(next, dest); next 72 kernel/irq/ipi.c if (next < nr_cpu_ids) { next 254 kernel/irq/irqdesc.c for (action = desc->action; action != NULL; action = action->next) { next 1437 kernel/irq/manage.c old_ptr = &old->next; next 1709 kernel/irq/manage.c action_ptr = &action->next; next 1713 kernel/irq/manage.c *action_ptr = action->next; next 527 kernel/irq/proc.c while ((action = action->next) != NULL) next 511 kernel/jump_label.c struct static_key_mod *next; next 535 kernel/jump_label.c key->next = mod; next 561 kernel/jump_label.c for (mod = static_key_mod(key); mod; mod = mod->next) { next 650 kernel/jump_label.c jlm2->next = NULL; next 656 kernel/jump_label.c jlm->next = static_key_mod(key); next 689 kernel/jump_label.c prev = &key->next; next 693 kernel/jump_label.c prev = &jlm->next; next 694 kernel/jump_label.c jlm = jlm->next; next 701 kernel/jump_label.c if (prev == &key->next) next 702 kernel/jump_label.c static_key_set_mod(key, jlm->next); next 704 kernel/jump_label.c *prev = jlm->next; next 710 kernel/jump_label.c if (jlm->next == NULL) { next 625 kernel/kallsyms.c .next = s_next, next 342 kernel/kexec_core.c struct page *page, *next; next 344 kernel/kexec_core.c list_for_each_entry_safe(page, next, list, lru) { next 216 kernel/kprobes.c struct kprobe_insn_page *kip, *next; next 221 kernel/kprobes.c list_for_each_entry_safe(kip, next, &c->pages, list) { next 1275 kernel/kprobes.c struct hlist_node *next; next 1277 kernel/kprobes.c hlist_for_each_entry_safe(ri, next, &rp->free_instances, hlist) { next 1287 kernel/kprobes.c struct hlist_node *next; next 1294 kernel/kprobes.c hlist_for_each_entry_safe(ri, next, head, hlist) { next 1786 kernel/kprobes.c ap = list_entry(p->list.next, struct kprobe, list); next 2403 kernel/kprobes.c .next = kprobe_seq_next, next 2451 kernel/kprobes.c .next = kprobe_blacklist_seq_next, next 591 kernel/kthread.c create = list_entry(kthread_create_list.next, next 969 kernel/kthread.c kthread_insert_work(worker, &fwork.work, work->node.next); next 972 kernel/kthread.c worker->work_list.next); next 101 kernel/livepatch/patch.c func = list_entry_rcu(func->stack_node.next, next 2034 kernel/locking/lockdep.c struct held_lock *next, next 2054 kernel/locking/lockdep.c print_lock(next); next 2061 kernel/locking/lockdep.c print_lock_name(hlock_class(next)); next 2080 kernel/locking/lockdep.c hlock_class(prev), hlock_class(next)); next 2228 kernel/locking/lockdep.c struct held_lock *next) next 2261 kernel/locking/lockdep.c that.class = hlock_class(next); next 2298 kernel/locking/lockdep.c prev, next, next 2320 kernel/locking/lockdep.c struct held_lock *prev, struct held_lock *next) next 2335 kernel/locking/lockdep.c struct lock_class *next = hlock_class(nxt); next 2345 kernel/locking/lockdep.c __print_lock_name(next); next 2353 kernel/locking/lockdep.c struct held_lock *next) next 2365 kernel/locking/lockdep.c print_lock(next); next 2370 kernel/locking/lockdep.c print_deadlock_scenario(next, prev); next 2386 kernel/locking/lockdep.c check_deadlock(struct task_struct *curr, struct held_lock *next) next 2395 kernel/locking/lockdep.c if (prev->instance == next->nest_lock) next 2398 kernel/locking/lockdep.c if (hlock_class(prev) != hlock_class(next)) next 2405 kernel/locking/lockdep.c if ((next->read == 2) && prev->read) next 2415 kernel/locking/lockdep.c print_deadlock_bug(curr, prev, next); next 2445 kernel/locking/lockdep.c struct held_lock *next, int distance, next 2451 kernel/locking/lockdep.c if (!hlock_class(prev)->key || !hlock_class(next)->key) { next 2462 kernel/locking/lockdep.c WARN_ONCE(!debug_locks_silent && !hlock_class(next)->key, next 2464 kernel/locking/lockdep.c hlock_class(next), next 2465 kernel/locking/lockdep.c hlock_class(next)->name); next 2479 kernel/locking/lockdep.c ret = check_noncircular(next, prev, trace); next 2483 kernel/locking/lockdep.c if (!check_irq_usage(curr, prev, next)) next 2494 kernel/locking/lockdep.c if (next->read == 2 || prev->read == 2) next 2505 kernel/locking/lockdep.c if (entry->class == hlock_class(next)) { next 2516 kernel/locking/lockdep.c ret = check_redundant(prev, next); next 2531 kernel/locking/lockdep.c ret = add_lock_to_list(hlock_class(next), hlock_class(prev), next 2533 kernel/locking/lockdep.c next->acquire_ip, distance, *trace); next 2538 kernel/locking/lockdep.c ret = add_lock_to_list(hlock_class(prev), hlock_class(next), next 2539 kernel/locking/lockdep.c &hlock_class(next)->locks_before, next 2540 kernel/locking/lockdep.c next->acquire_ip, distance, *trace); next 2554 kernel/locking/lockdep.c check_prevs_add(struct task_struct *curr, struct held_lock *next) next 2584 kernel/locking/lockdep.c int ret = check_prev_add(curr, hlock, next, distance, next 4891 kernel/locking/lockdep.c WARN_ON_ONCE(!class->lock_entry.next); next 4895 kernel/locking/lockdep.c WARN_ON_ONCE(!class->lock_entry.next); next 99 kernel/locking/lockdep_proc.c .next = l_next, next 157 kernel/locking/lockdep_proc.c .next = lc_next, next 582 kernel/locking/lockdep_proc.c .next = ls_next, next 19 kernel/locking/mcs_spinlock.h struct mcs_spinlock *next; next 71 kernel/locking/mcs_spinlock.h node->next = NULL; next 91 kernel/locking/mcs_spinlock.h WRITE_ONCE(prev->next, node); next 104 kernel/locking/mcs_spinlock.h struct mcs_spinlock *next = READ_ONCE(node->next); next 106 kernel/locking/mcs_spinlock.h if (likely(!next)) { next 113 kernel/locking/mcs_spinlock.h while (!(next = READ_ONCE(node->next))) next 118 kernel/locking/mcs_spinlock.h arch_mcs_spin_unlock_contended(&next->locked); next 76 kernel/locking/mutex-debug.c DEBUG_LOCKS_WARN_ON(!lock->wait_list.prev && !lock->wait_list.next); next 1224 kernel/locking/mutex.c struct task_struct *next = NULL; next 1269 kernel/locking/mutex.c next = waiter->task; next 1272 kernel/locking/mutex.c wake_q_add(&wake_q, next); next 1276 kernel/locking/mutex.c __mutex_handoff(lock, next); next 14 kernel/locking/mutex.h __list_del((waiter)->list.prev, (waiter)->list.next) next 46 kernel/locking/osq_lock.c struct optimistic_spin_node *next = NULL; next 78 kernel/locking/osq_lock.c if (node->next) { next 79 kernel/locking/osq_lock.c next = xchg(&node->next, NULL); next 80 kernel/locking/osq_lock.c if (next) next 87 kernel/locking/osq_lock.c return next; next 93 kernel/locking/osq_lock.c struct optimistic_spin_node *prev, *next; next 98 kernel/locking/osq_lock.c node->next = NULL; next 126 kernel/locking/osq_lock.c WRITE_ONCE(prev->next, node); next 160 kernel/locking/osq_lock.c if (prev->next == node && next 161 kernel/locking/osq_lock.c cmpxchg(&prev->next, node, NULL) == node) next 188 kernel/locking/osq_lock.c next = osq_wait_next(lock, node, prev); next 189 kernel/locking/osq_lock.c if (!next) next 200 kernel/locking/osq_lock.c WRITE_ONCE(next->prev, prev); next 201 kernel/locking/osq_lock.c WRITE_ONCE(prev->next, next); next 208 kernel/locking/osq_lock.c struct optimistic_spin_node *node, *next; next 222 kernel/locking/osq_lock.c next = xchg(&node->next, NULL); next 223 kernel/locking/osq_lock.c if (next) { next 224 kernel/locking/osq_lock.c WRITE_ONCE(next->locked, 1); next 228 kernel/locking/osq_lock.c next = osq_wait_next(lock, node, NULL); next 229 kernel/locking/osq_lock.c if (next) next 230 kernel/locking/osq_lock.c WRITE_ONCE(next->locked, 1); next 316 kernel/locking/qspinlock.c struct mcs_spinlock *prev, *next, *node; next 434 kernel/locking/qspinlock.c node->next = NULL; next 460 kernel/locking/qspinlock.c next = NULL; next 470 kernel/locking/qspinlock.c WRITE_ONCE(prev->next, node); next 481 kernel/locking/qspinlock.c next = READ_ONCE(node->next); next 482 kernel/locking/qspinlock.c if (next) next 483 kernel/locking/qspinlock.c prefetchw(next); next 549 kernel/locking/qspinlock.c if (!next) next 550 kernel/locking/qspinlock.c next = smp_cond_load_relaxed(&node->next, (VAL)); next 552 kernel/locking/qspinlock.c arch_mcs_spin_unlock_contended(&next->locked); next 553 kernel/locking/qspinlock.c pv_kick_node(lock, next); next 4376 kernel/module.c .next = m_next, next 29 kernel/notifier.c nl = &((*nl)->next); next 31 kernel/notifier.c n->next = *nl; next 44 kernel/notifier.c nl = &((*nl)->next); next 46 kernel/notifier.c n->next = *nl; next 56 kernel/notifier.c rcu_assign_pointer(*nl, n->next); next 59 kernel/notifier.c nl = &((*nl)->next); next 86 kernel/notifier.c next_nb = rcu_dereference_raw(nb->next); next 78 kernel/padata.c padata = list_entry(local_list.next, next 192 kernel/padata.c padata = list_entry(reorder->list.next, struct padata_priv, list); next 304 kernel/padata.c padata = list_entry(local_list.next, next 272 kernel/pid_namespace.c int ret, next; next 283 kernel/pid_namespace.c next = idr_get_cursor(&pid_ns->idr) - 1; next 285 kernel/pid_namespace.c tmp.data = &next; next 288 kernel/pid_namespace.c idr_set_cursor(&pid_ns->idr, next + 1); next 122 kernel/power/snapshot.c struct linked_page *next; next 180 kernel/power/snapshot.c safe_pages_list = safe_pages_list->next; next 208 kernel/power/snapshot.c lp->next = safe_pages_list; next 239 kernel/power/snapshot.c struct linked_page *lp = list->next; next 287 kernel/power/snapshot.c lp->next = ca->chain; next 550 kernel/power/snapshot.c bm->cur.zone = list_entry(bm->zones.next, struct mem_zone_bm_rtree, next 552 kernel/power/snapshot.c bm->cur.node = list_entry(bm->cur.zone->leaves.next, next 848 kernel/power/snapshot.c bm->cur.node = list_entry(bm->cur.node->list.next, next 858 kernel/power/snapshot.c bm->cur.zone = list_entry(bm->cur.zone->list.next, next 860 kernel/power/snapshot.c bm->cur.node = list_entry(bm->cur.zone->leaves.next, next 937 kernel/power/snapshot.c p_list = lp->next; next 2246 kernel/power/snapshot.c struct highmem_pbe *next; next 2388 kernel/power/snapshot.c safe_pages_list = safe_pages_list->next; next 2391 kernel/power/snapshot.c pbe->next = highmem_pblist; next 2503 kernel/power/snapshot.c lp->next = safe_pages_list; next 2517 kernel/power/snapshot.c lp->next = safe_pages_list; next 2569 kernel/power/snapshot.c safe_pages_list = safe_pages_list->next; next 2570 kernel/power/snapshot.c pbe->next = restore_pblist; next 2729 kernel/power/snapshot.c pbe = pbe->next; next 87 kernel/power/swap.c struct swap_map_page_list *next; next 955 kernel/power/swap.c handle->maps = handle->maps->next; next 985 kernel/power/swap.c last->next = tmp; next 1026 kernel/power/swap.c handle->maps = handle->maps->next; next 1347 kernel/printk/printk.c const char *next = memchr(text, '\n', text_size); next 1350 kernel/printk/printk.c if (next) { next 1351 kernel/printk/printk.c text_len = next - text; next 1352 kernel/printk/printk.c next++; next 1353 kernel/printk/printk.c text_size -= next - text; next 1372 kernel/printk/printk.c text = next; next 2783 kernel/printk/printk.c newcon->next = console_drivers; next 2785 kernel/printk/printk.c if (newcon->next) next 2786 kernel/printk/printk.c newcon->next->flags &= ~CON_CONSDEV; next 2788 kernel/printk/printk.c newcon->next = console_drivers->next; next 2789 kernel/printk/printk.c console_drivers->next = newcon; next 2858 kernel/printk/printk.c console_drivers=console->next; next 2861 kernel/printk/printk.c for (a=console_drivers->next, b=console_drivers ; next 2862 kernel/printk/printk.c a; b=a, a=b->next) { next 2864 kernel/printk/printk.c b->next = a->next; next 35 kernel/rcu/rcu_segcblist.c rclp->tail = &rhp->next; next 61 kernel/rcu/rcu_segcblist.c rhp->next = NULL; next 63 kernel/rcu/rcu_segcblist.c srclp->tail = &rhp->next; next 84 kernel/rcu/rcu_segcblist.c rclp->head = rhp->next; next 262 kernel/rcu/rcu_segcblist.c rhp->next = NULL; next 264 kernel/rcu/rcu_segcblist.c WRITE_ONCE(rsclp->tails[RCU_NEXT_TAIL], &rhp->next); next 288 kernel/rcu/rcu_segcblist.c rhp->next = NULL; next 294 kernel/rcu/rcu_segcblist.c WRITE_ONCE(rsclp->tails[i], &rhp->next); next 267 kernel/rcu/rcutorture.c p = rcu_torture_freelist.next; next 136 kernel/rcu/srcutiny.c lh = lh->next; next 164 kernel/rcu/srcutiny.c rhp->next = NULL; next 167 kernel/rcu/srcutiny.c ssp->srcu_cb_tail = &rhp->next; next 79 kernel/rcu/tiny.c struct rcu_head *next, *list; next 99 kernel/rcu/tiny.c next = list->next; next 100 kernel/rcu/tiny.c prefetch(next); next 105 kernel/rcu/tiny.c list = next; next 139 kernel/rcu/tiny.c head->next = NULL; next 143 kernel/rcu/tiny.c rcu_ctrlblk.curtail = &head->next; next 2577 kernel/rcu/tree.c head->next = NULL; next 371 kernel/rcu/tree_exp.h rnp->exp_tasks = rnp->blkd_tasks.next; next 404 kernel/rcu/tree_plugin.h np = t->rcu_node_entry.next; next 666 kernel/rcu/tree_plugin.h WRITE_ONCE(rnp->gp_tasks, rnp->blkd_tasks.next); next 527 kernel/rcu/update.c rhp->next = NULL; next 532 kernel/rcu/update.c rcu_tasks_cbs_tail = &rhp->next; next 636 kernel/rcu/update.c struct rcu_head *next; next 777 kernel/rcu/update.c next = list->next; next 781 kernel/rcu/update.c list = next; next 134 kernel/resource.c .next = r_next, next 616 kernel/resource.c goto next; next 638 kernel/resource.c next: if (!this || this->end == root->end) next 783 kernel/resource.c struct resource *first, *next; next 801 kernel/resource.c for (next = first; ; next = next->sibling) { next 803 kernel/resource.c if (next->start < new->start || next->end > new->end) next 804 kernel/resource.c return next; next 805 kernel/resource.c if (!next->sibling) next 807 kernel/resource.c if (next->sibling->start > new->end) next 812 kernel/resource.c new->sibling = next->sibling; next 815 kernel/resource.c next->sibling = NULL; next 816 kernel/resource.c for (next = first; next; next = next->sibling) next 817 kernel/resource.c next->parent = new; next 822 kernel/resource.c next = parent->child; next 823 kernel/resource.c while (next->sibling != first) next 824 kernel/resource.c next = next->sibling; next 825 kernel/resource.c next->sibling = new; next 210 kernel/sched/autogroup.c static unsigned long next = INITIAL_JIFFIES; next 226 kernel/sched/autogroup.c if (!capable(CAP_SYS_ADMIN) && time_before(jiffies, next)) next 229 kernel/sched/autogroup.c next = HZ / 10 + jiffies; next 426 kernel/sched/core.c if (unlikely(cmpxchg_relaxed(&node->next, NULL, WAKE_Q_TAIL))) next 433 kernel/sched/core.c head->lastp = &node->next; next 488 kernel/sched/core.c node = node->next; next 489 kernel/sched/core.c task->wake_q.next = NULL; next 3036 kernel/sched/core.c struct task_struct *next) next 3041 kernel/sched/core.c notifier->ops->sched_out(notifier, next); next 3046 kernel/sched/core.c struct task_struct *next) next 3049 kernel/sched/core.c __fire_sched_out_preempt_notifiers(curr, next); next 3060 kernel/sched/core.c struct task_struct *next) next 3066 kernel/sched/core.c static inline void prepare_task(struct task_struct *next) next 3073 kernel/sched/core.c next->on_cpu = 1; next 3095 kernel/sched/core.c prepare_lock_switch(struct rq *rq, struct task_struct *next, struct rq_flags *rf) next 3107 kernel/sched/core.c rq->lock.owner = next; next 3127 kernel/sched/core.c # define prepare_arch_switch(next) do { } while (0) next 3149 kernel/sched/core.c struct task_struct *next) next 3152 kernel/sched/core.c sched_info_switch(rq, prev, next); next 3153 kernel/sched/core.c perf_event_task_sched_out(prev, next); next 3155 kernel/sched/core.c fire_sched_out_preempt_notifiers(prev, next); next 3156 kernel/sched/core.c prepare_task(next); next 3157 kernel/sched/core.c prepare_arch_switch(next); next 3265 kernel/sched/core.c struct callback_head *head, *next; next 3274 kernel/sched/core.c next = head->next; next 3275 kernel/sched/core.c head->next = NULL; next 3276 kernel/sched/core.c head = next; next 3330 kernel/sched/core.c struct task_struct *next, struct rq_flags *rf) next 3332 kernel/sched/core.c prepare_task_switch(rq, prev, next); next 3348 kernel/sched/core.c if (!next->mm) { // to kernel next 3349 kernel/sched/core.c enter_lazy_tlb(prev->active_mm, next); next 3351 kernel/sched/core.c next->active_mm = prev->active_mm; next 3357 kernel/sched/core.c membarrier_switch_mm(rq, prev->active_mm, next->mm); next 3366 kernel/sched/core.c switch_mm_irqs_off(prev->active_mm, next->mm, next); next 3377 kernel/sched/core.c prepare_lock_switch(rq, next, rf); next 3380 kernel/sched/core.c switch_to(prev, next, prev); next 3999 kernel/sched/core.c struct task_struct *prev, *next; next 4047 kernel/sched/core.c next = pick_next_task(rq, prev, &rf); next 4051 kernel/sched/core.c if (likely(prev != next)) { next 4057 kernel/sched/core.c RCU_INIT_POINTER(rq->curr, next); next 4074 kernel/sched/core.c trace_sched_switch(preempt, prev, next); next 4077 kernel/sched/core.c rq = context_switch(rq, prev, next, &rf); next 6208 kernel/sched/core.c struct task_struct *next; next 6211 kernel/sched/core.c next = class->pick_next_task(rq, NULL, NULL); next 6212 kernel/sched/core.c if (next) { next 6213 kernel/sched/core.c next->sched_class->put_prev_task(rq, next); next 6214 kernel/sched/core.c return next; next 6233 kernel/sched/core.c struct task_struct *next, *stop = rq->stop; next 6263 kernel/sched/core.c next = __pick_migrate_task(rq); next 6275 kernel/sched/core.c raw_spin_lock(&next->pi_lock); next 6283 kernel/sched/core.c if (WARN_ON(task_rq(next) != rq || !task_on_rq_queued(next))) { next 6284 kernel/sched/core.c raw_spin_unlock(&next->pi_lock); next 6289 kernel/sched/core.c dest_cpu = select_fallback_rq(dead_rq->cpu, next); next 6290 kernel/sched/core.c rq = __migrate_task(rq, rf, next, dest_cpu); next 6297 kernel/sched/core.c raw_spin_unlock(&next->pi_lock); next 362 kernel/sched/deadline.c dl_rq->earliest_dl.curr = dl_rq->earliest_dl.next = 0; next 468 kernel/sched/deadline.c dl_rq->earliest_dl.next = p->dl.deadline; next 487 kernel/sched/deadline.c dl_rq->earliest_dl.next = rb_entry(next_node, next 1355 kernel/sched/deadline.c dl_rq->earliest_dl.next = 0; next 2163 kernel/sched/deadline.c src_rq->dl.earliest_dl.next)) next 2435 kernel/sched/deadline.c .next = &rt_sched_class, next 800 kernel/sched/debug.c .next = sched_debug_next, next 612 kernel/sched/fair.c struct rb_node *next = rb_next(&se->run_node); next 614 kernel/sched/fair.c if (!next) next 617 kernel/sched/fair.c return rb_entry(next, struct sched_entity, run_node); next 2497 kernel/sched/fair.c work->next = work; next 2642 kernel/sched/fair.c p->numa_work.next = &p->numa_work; next 2681 kernel/sched/fair.c if ((curr->flags & (PF_EXITING | PF_KTHREAD)) || work->next != work) next 3171 kernel/sched/fair.c struct cfs_rq *prev, struct cfs_rq *next) next 3196 kernel/sched/fair.c n_last_update_time_copy = next->load_last_update_time_copy; next 3201 kernel/sched/fair.c n_last_update_time = next->avg.last_update_time; next 3208 kernel/sched/fair.c n_last_update_time = next->avg.last_update_time; next 4036 kernel/sched/fair.c if (cfs_rq->next != se) next 4039 kernel/sched/fair.c cfs_rq->next = NULL; next 4059 kernel/sched/fair.c if (cfs_rq->next == se) next 4246 kernel/sched/fair.c if (cfs_rq->next && wakeup_preempt_entity(cfs_rq->next, left) < 1) next 4247 kernel/sched/fair.c se = cfs_rq->next; next 4634 kernel/sched/fair.c goto next; next 4650 kernel/sched/fair.c next: next 5021 kernel/sched/fair.c if (!cfs_b->throttled_cfs_rq.next) next 5661 kernel/sched/fair.c } while (group = group->next, group != sd->groups); next 6388 kernel/sched/fair.c for (; pd; pd = pd->next) { next 6679 kernel/sched/fair.c cfs_rq_of(se)->next = se; next 7168 kernel/sched/fair.c (&p->se == cfs_rq_of(&p->se)->next || next 7408 kernel/sched/fair.c goto next; next 7413 kernel/sched/fair.c goto next; next 7416 kernel/sched/fair.c goto next; next 7442 kernel/sched/fair.c next: next 7874 kernel/sched/fair.c group = group->next; next 8291 kernel/sched/fair.c sg = sg->next; next 9121 kernel/sched/fair.c unsigned long interval, next; next 9125 kernel/sched/fair.c next = sd->last_balance + interval; next 9127 kernel/sched/fair.c if (time_after(*next_balance, next)) next 9128 kernel/sched/fair.c *next_balance = next; next 10466 kernel/sched/fair.c .next = &idle_sched_class, next 388 kernel/sched/idle.c static void set_next_task_idle(struct rq *rq, struct task_struct *next, bool first) next 397 kernel/sched/idle.c struct task_struct *next = rq->idle; next 402 kernel/sched/idle.c set_next_task_idle(rq, next, true); next 404 kernel/sched/idle.c return next; next 91 kernel/sched/rt.c rt_rq->highest_prio.next = MAX_RT_PRIO; next 381 kernel/sched/rt.c if (p->prio < rq->rt.highest_prio.next) next 382 kernel/sched/rt.c rq->rt.highest_prio.next = p->prio; next 393 kernel/sched/rt.c rq->rt.highest_prio.next = p->prio; next 395 kernel/sched/rt.c rq->rt.highest_prio.next = MAX_RT_PRIO; next 460 kernel/sched/rt.c tg = list_entry_rcu(tg->list.next, next 664 kernel/sched/rt.c goto next; next 682 kernel/sched/rt.c next: next 1543 kernel/sched/rt.c struct sched_rt_entity *next = NULL; next 1551 kernel/sched/rt.c next = list_entry(queue->next, struct sched_rt_entity, run_list); next 1553 kernel/sched/rt.c return next; next 1924 kernel/sched/rt.c int next; next 1958 kernel/sched/rt.c next = atomic_read_acquire(&rd->rto_loop_next); next 1960 kernel/sched/rt.c if (rd->rto_loop == next) next 1963 kernel/sched/rt.c rd->rto_loop = next; next 2091 kernel/sched/rt.c if (src_rq->rt.highest_prio.next >= next 2289 kernel/sched/rt.c unsigned long next; next 2296 kernel/sched/rt.c next = DIV_ROUND_UP(min(soft, hard), USEC_PER_SEC/HZ); next 2297 kernel/sched/rt.c if (p->rt.timeout > next) { next 2341 kernel/sched/rt.c if (rt_se->run_list.prev != rt_se->run_list.next) { next 2361 kernel/sched/rt.c .next = &fair_sched_class, next 482 kernel/sched/sched.h struct cfs_rq *prev, struct cfs_rq *next); next 485 kernel/sched/sched.h struct cfs_rq *prev, struct cfs_rq *next) { } next 516 kernel/sched/sched.h struct sched_entity *next; next 605 kernel/sched/sched.h int next; /* next highest */ next 653 kernel/sched/sched.h u64 next; next 722 kernel/sched/sched.h struct perf_domain *next; next 1321 kernel/sched/sched.h if (unlikely(head->next)) next 1325 kernel/sched/sched.h head->next = rq->balance_callback; next 1411 kernel/sched/sched.h struct sched_group *next; /* Must be a circular list */ next 1709 kernel/sched/sched.h const struct sched_class *next; next 1786 kernel/sched/sched.h static inline void set_next_task(struct rq *rq, struct task_struct *next) next 1788 kernel/sched/sched.h WARN_ON_ONCE(rq->curr != next); next 1789 kernel/sched/sched.h next->sched_class->set_next_task(rq, next, false); next 1799 kernel/sched/sched.h for (class = (_from); class != (_to); class = class->next) next 118 kernel/sched/stats.c .next = schedstat_next, next 226 kernel/sched/stats.h __sched_info_switch(struct rq *rq, struct task_struct *prev, struct task_struct *next) next 236 kernel/sched/stats.h if (next != rq->idle) next 237 kernel/sched/stats.h sched_info_arrive(rq, next); next 241 kernel/sched/stats.h sched_info_switch(struct rq *rq, struct task_struct *prev, struct task_struct *next) next 244 kernel/sched/stats.h __sched_info_switch(rq, prev, next); next 252 kernel/sched/stats.h # define sched_info_arrive(rq, next) do { } while (0) next 253 kernel/sched/stats.h # define sched_info_switch(rq, t, next) do { } while (0) next 122 kernel/sched/stop_task.c .next = &dl_sched_class, next 98 kernel/sched/topology.c group = group->next; next 162 kernel/sched/topology.c if (sd->groups != sd->groups->next) next 185 kernel/sched/topology.c if (parent->groups == parent->groups->next) { next 240 kernel/sched/topology.c tmp = pd->next; next 251 kernel/sched/topology.c pd = pd->next; next 289 kernel/sched/topology.c pd = pd->next; next 385 kernel/sched/topology.c tmp->next = pd; next 567 kernel/sched/topology.c tmp = sg->next; next 971 kernel/sched/topology.c last->next = sg; next 973 kernel/sched/topology.c last->next = first; next 1128 kernel/sched/topology.c last->next = sg; next 1131 kernel/sched/topology.c last->next = first; next 1159 kernel/sched/topology.c goto next; next 1169 kernel/sched/topology.c next: next 1170 kernel/sched/topology.c sg = sg->next; next 1799 kernel/sched/topology.c sg->next = sg; next 70 kernel/sched/wait.c wait_queue_entry_t *curr, *next; next 86 kernel/sched/wait.c list_for_each_entry_safe_from(curr, next, &wq_head->head, entry) { next 100 kernel/sched/wait.c (&next->entry != &wq_head->head)) { next 102 kernel/sched/wait.c list_add_tail(&bookmark->entry, &next->entry); next 466 kernel/signal.c q = list_entry(queue->list.next, struct sigqueue , list); next 726 kernel/signal.c goto next; next 730 kernel/signal.c next: next 480 kernel/softirq.c t->next = NULL; next 482 kernel/softirq.c head->tail = &(t->next); next 516 kernel/softirq.c list = list->next; next 531 kernel/softirq.c t->next = NULL; next 533 kernel/softirq.c tl_head->tail = &t->next; next 552 kernel/softirq.c t->next = NULL; next 632 kernel/softirq.c for (i = &per_cpu(tasklet_vec, cpu).head; *i; i = &(*i)->next) { next 634 kernel/softirq.c *i = t->next; next 36 kernel/task_work.c work->next = head; next 73 kernel/task_work.c pprev = &work->next; next 74 kernel/task_work.c else if (cmpxchg(pprev, work, work->next) == work) next 93 kernel/task_work.c struct callback_head *work, *head, *next; next 112 kernel/task_work.c next = work->next; next 114 kernel/task_work.c work = next; next 274 kernel/time/alarmtimer.c struct timerqueue_node *next; next 278 kernel/time/alarmtimer.c next = timerqueue_getnext(&base->timerqueue); next 280 kernel/time/alarmtimer.c if (!next) next 282 kernel/time/alarmtimer.c delta = ktime_sub(next->expires, base->gettime()); next 284 kernel/time/alarmtimer.c expires = next->expires; next 348 kernel/time/clockevents.c dev = list_entry(clockevents_released.next, next 512 kernel/time/hrtimer.c struct timerqueue_node *next; next 515 kernel/time/hrtimer.c next = timerqueue_getnext(&base->active); next 516 kernel/time/hrtimer.c timer = container_of(next, struct hrtimer, node); next 519 kernel/time/hrtimer.c next = timerqueue_iterate_next(next); next 520 kernel/time/hrtimer.c if (!next) next 523 kernel/time/hrtimer.c timer = container_of(next, struct hrtimer, node); next 504 kernel/time/ntp.c struct timespec64 next; next 506 kernel/time/ntp.c ktime_get_real_ts64(&next); next 508 kernel/time/ntp.c next.tv_sec = 659; next 516 kernel/time/ntp.c next.tv_sec = 0; next 520 kernel/time/ntp.c next.tv_nsec = target_nsec - next.tv_nsec; next 521 kernel/time/ntp.c if (next.tv_nsec <= 0) next 522 kernel/time/ntp.c next.tv_nsec += NSEC_PER_SEC; next 523 kernel/time/ntp.c if (next.tv_nsec >= NSEC_PER_SEC) { next 524 kernel/time/ntp.c next.tv_sec++; next 525 kernel/time/ntp.c next.tv_nsec -= NSEC_PER_SEC; next 529 kernel/time/ntp.c timespec64_to_jiffies(&next)); next 766 kernel/time/posix-cpu-timers.c struct timerqueue_node *next; next 769 kernel/time/posix-cpu-timers.c while ((next = timerqueue_getnext(head))) { next 773 kernel/time/posix-cpu-timers.c ctmr = container_of(next, struct cpu_timer, node); next 1116 kernel/time/posix-cpu-timers.c struct k_itimer *timer, *next; next 1156 kernel/time/posix-cpu-timers.c list_for_each_entry_safe(timer, next, &firing, it.cpu.elist) { next 1040 kernel/time/posix-timers.c tmr = list_entry(sig->posix_timers.next, struct k_itimer, list); next 334 kernel/time/tick-broadcast.c ktime_t next = ktime_add(dev->next_event, tick_period); next 336 kernel/time/tick-broadcast.c clockevents_program_event(dev, next, true); next 106 kernel/time/tick-common.c ktime_t next = dev->next_event; next 127 kernel/time/tick-common.c next = ktime_add(next, tick_period); next 129 kernel/time/tick-common.c if (!clockevents_program_event(dev, next, false)) next 161 kernel/time/tick-common.c ktime_t next; next 165 kernel/time/tick-common.c next = tick_next_period; next 171 kernel/time/tick-common.c if (!clockevents_program_event(dev, next, false)) next 173 kernel/time/tick-common.c next = ktime_add(next, tick_period); next 1228 kernel/time/tick-sched.c ktime_t next; next 1242 kernel/time/tick-sched.c next = tick_init_jiffy_update(); next 1244 kernel/time/tick-sched.c hrtimer_set_expires(&ts->sched_timer, next); next 613 kernel/time/timer.c timer->entry.next == TIMER_ENTRY_STATIC); next 818 kernel/time/timer.c entry->next = LIST_POISON2; next 1508 kernel/time/timer.c unsigned long clk, next, adj; next 1511 kernel/time/timer.c next = base->clk + NEXT_TIMER_MAX_DELTA; next 1520 kernel/time/timer.c if (time_before(tmp, next)) next 1521 kernel/time/timer.c next = tmp; next 1563 kernel/time/timer.c return next; next 1689 kernel/time/timer.c unsigned long next = __next_timer_interrupt(base); next 1695 kernel/time/timer.c if (time_after(next, now)) { next 1703 kernel/time/timer.c base->clk = next; next 77 kernel/time/timer_list.c unsigned long next = 0, i; next 93 kernel/time/timer_list.c while (curr && i < next) { next 105 kernel/time/timer_list.c next++; next 364 kernel/time/timer_list.c .next = timer_list_next, next 525 kernel/trace/bpf_trace.c .next = ctx_size ? &frag : NULL, next 390 kernel/trace/fgraph.c struct task_struct *prev, struct task_struct *next) next 407 kernel/trace/fgraph.c if (!next->ftrace_timestamp) next 414 kernel/trace/fgraph.c timestamp -= next->ftrace_timestamp; next 416 kernel/trace/fgraph.c for (index = next->curr_ret_stack; index >= 0; index--) next 417 kernel/trace/fgraph.c next->ret_stack[index].calltime += timestamp; next 206 kernel/trace/ftrace.c } else if (rcu_dereference_protected(ftrace_ops_list->next, next 267 kernel/trace/ftrace.c rcu_assign_pointer(ops->next, *list); next 289 kernel/trace/ftrace.c rcu_dereference_protected(ops->next, next 295 kernel/trace/ftrace.c for (p = list; *p != &ftrace_list_end; p = &(*p)->next) next 302 kernel/trace/ftrace.c *p = (*p)->next; next 400 kernel/trace/ftrace.c struct ftrace_profile_page *next; next 442 kernel/trace/ftrace.c pg = pg->next; next 581 kernel/trace/ftrace.c pg = pg->next; next 621 kernel/trace/ftrace.c pg->next = (void *)get_zeroed_page(GFP_KERNEL); next 622 kernel/trace/ftrace.c if (!pg->next) next 624 kernel/trace/ftrace.c pg = pg->next; next 634 kernel/trace/ftrace.c pg = pg->next; next 744 kernel/trace/ftrace.c if (!stat->pages->next) next 746 kernel/trace/ftrace.c stat->pages = stat->pages->next; next 1100 kernel/trace/ftrace.c struct ftrace_page *next; next 1515 kernel/trace/ftrace.c for (pg = ftrace_pages_start; pg; pg = pg->next) { \ next 1558 kernel/trace/ftrace.c for (pg = ftrace_pages_start; pg; pg = pg->next) { next 1613 kernel/trace/ftrace.c ops != &ftrace_list_end; ops = ops->next) { next 2452 kernel/trace/ftrace.c iter->pg = iter->pg->next; next 2471 kernel/trace/ftrace.c iter->pg = iter->pg->next; next 2476 kernel/trace/ftrace.c iter->pg = iter->pg->next; next 2930 kernel/trace/ftrace.c for (pg = new_pgs; pg; pg = pg->next) { next 3026 kernel/trace/ftrace.c pg->next = kzalloc(sizeof(*pg), GFP_KERNEL); next 3027 kernel/trace/ftrace.c if (!pg->next) next 3030 kernel/trace/ftrace.c pg = pg->next; next 3040 kernel/trace/ftrace.c start_pg = pg->next; next 3075 kernel/trace/ftrace.c struct list_head *next; next 3091 kernel/trace/ftrace.c next = func_probes->next; next 3092 kernel/trace/ftrace.c iter->probe = list_entry(next, struct ftrace_func_probe, list); next 3111 kernel/trace/ftrace.c if (iter->probe->list.next == func_probes) next 3113 kernel/trace/ftrace.c next = iter->probe->list.next; next 3114 kernel/trace/ftrace.c iter->probe = list_entry(next, struct ftrace_func_probe, list); next 3131 kernel/trace/ftrace.c hnd = hnd->next; next 3208 kernel/trace/ftrace.c iter->mod_list = iter->mod_list->next; next 3285 kernel/trace/ftrace.c if (iter->pg->next) { next 3286 kernel/trace/ftrace.c iter->pg = iter->pg->next; next 3481 kernel/trace/ftrace.c .next = t_next, next 4643 kernel/trace/ftrace.c char *func, *command, *next = buff; next 4647 kernel/trace/ftrace.c func = strsep(&next, ":"); next 4649 kernel/trace/ftrace.c if (!next) { next 4660 kernel/trace/ftrace.c command = strsep(&next, ":"); next 4665 kernel/trace/ftrace.c ret = p->func(tr, hash, func, command, next, enable); next 5214 kernel/trace/ftrace.c .next = g_next, next 5614 kernel/trace/ftrace.c if (WARN_ON(ftrace_pages->next)) { next 5616 kernel/trace/ftrace.c while (ftrace_pages->next) next 5617 kernel/trace/ftrace.c ftrace_pages = ftrace_pages->next; next 5620 kernel/trace/ftrace.c ftrace_pages->next = start_pg; next 5638 kernel/trace/ftrace.c if (WARN_ON(!pg->next)) next 5640 kernel/trace/ftrace.c pg = pg->next; next 5648 kernel/trace/ftrace.c WARN_ON(pg->next); next 5692 kernel/trace/ftrace.c #define next_to_ftrace_page(p) container_of(p, struct ftrace_page, next) next 5701 kernel/trace/ftrace.c for (ops = ftrace_ops_list; ops != &ftrace_list_end; ops = ops->next) { next 5809 kernel/trace/ftrace.c *last_pg = pg->next; next 5811 kernel/trace/ftrace.c pg->next = tmp_page; next 5814 kernel/trace/ftrace.c last_pg = &pg->next; next 5826 kernel/trace/ftrace.c tmp_page = pg->next; next 6146 kernel/trace/ftrace.c for (pg = ftrace_pages_start; pg; last_pg = &pg->next, pg = *last_pg) { next 6166 kernel/trace/ftrace.c *last_pg = pg->next; next 6170 kernel/trace/ftrace.c pg = container_of(last_pg, struct ftrace_page, next); next 6424 kernel/trace/ftrace.c struct task_struct *prev, struct task_struct *next) next 6432 kernel/trace/ftrace.c trace_ignore_this_task(pid_list, next)); next 6566 kernel/trace/ftrace.c .next = fpid_next, next 24 kernel/trace/ftrace_internal.h while (likely(op = rcu_dereference_raw_check((op)->next)) && \ next 874 kernel/trace/ring_buffer.c val = (unsigned long)list->next; next 893 kernel/trace/ring_buffer.c return rb_list_head(list->next) != &page->list; next 904 kernel/trace/ring_buffer.c ptr = (unsigned long *)&list->next; next 928 kernel/trace/ring_buffer.c unsigned long *ptr = (unsigned long *)&list->next; next 961 kernel/trace/ring_buffer.c ret = cmpxchg((unsigned long *)&list->next, next 1001 kernel/trace/ring_buffer.c struct list_head *p = rb_list_head((*bpage)->list.next); next 1019 kernel/trace/ring_buffer.c if (RB_WARN_ON(cpu_buffer, rb_list_head(list->prev->next) != list)) next 1047 kernel/trace/ring_buffer.c unsigned long *ptr = (unsigned long *)&old->list.prev->next; next 1142 kernel/trace/ring_buffer.c if (RB_WARN_ON(cpu_buffer, rb_list_head(list->next) != list->next)) next 1165 kernel/trace/ring_buffer.c if (RB_WARN_ON(cpu_buffer, head->next->prev != head)) next 1167 kernel/trace/ring_buffer.c if (RB_WARN_ON(cpu_buffer, head->prev->next != head)) next 1175 kernel/trace/ring_buffer.c bpage->list.next->prev != &bpage->list)) next 1178 kernel/trace/ring_buffer.c bpage->list.prev->next != &bpage->list)) next 1275 kernel/trace/ring_buffer.c cpu_buffer->pages = pages.next; next 1524 kernel/trace/ring_buffer.c tail_page = rb_list_head(tail_page->next); next 1528 kernel/trace/ring_buffer.c first_page = list_entry(rb_list_head(to_remove->next), next 1532 kernel/trace/ring_buffer.c to_remove = rb_list_head(to_remove)->next; next 1536 kernel/trace/ring_buffer.c next_page = rb_list_head(to_remove)->next; next 1543 kernel/trace/ring_buffer.c tail_page->next = (struct list_head *)((unsigned long)next_page | next 1639 kernel/trace/ring_buffer.c first_page = pages->next; next 1645 kernel/trace/ring_buffer.c last_page->next = head_page_with_bit; next 1648 kernel/trace/ring_buffer.c r = cmpxchg(&prev_page->next, head_page_with_bit, first_page); next 3730 kernel/trace/ring_buffer.c cpu_buffer->reader_page->list.next = rb_list_head(reader->list.next); next 3779 kernel/trace/ring_buffer.c rb_list_head(reader->list.next)->prev = &cpu_buffer->reader_page->list; next 140 kernel/trace/trace.c union trace_eval_map_item *next; next 1758 kernel/trace/trace.c for (t = trace_types; t; t = t->next) { next 1760 kernel/trace/trace.c *last = t->next; next 1763 kernel/trace/trace.c last = &t->next; next 1819 kernel/trace/trace.c for (t = trace_types; t; t = t->next) { next 1851 kernel/trace/trace.c type->next = trace_types; next 2293 kernel/trace/trace.c struct task_struct *next, int flags) next 2305 kernel/trace/trace.c done &= !(flags & TRACE_RECORD_CMDLINE) || trace_save_cmdline(next); next 2307 kernel/trace/trace.c done &= !(flags & TRACE_RECORD_TGID) || trace_save_tgid(next); next 2674 kernel/trace/trace.c export = rcu_dereference_raw_check(export->next); next 2683 kernel/trace/trace.c rcu_assign_pointer(export->next, *list); next 2698 kernel/trace/trace.c for (p = list; *p != NULL; p = &(*p)->next) next 2705 kernel/trace/trace.c rcu_assign_pointer(*p, (*p)->next); next 3301 kernel/trace/trace.c struct trace_entry *ent, *next = NULL; next 3333 kernel/trace/trace.c if (ent && (!next || ts < next_ts)) { next 3334 kernel/trace/trace.c next = ent; next 3353 kernel/trace/trace.c return next; next 4065 kernel/trace/trace.c .next = s_next, next 4316 kernel/trace/trace.c t = t->next; next 4330 kernel/trace/trace.c t = get_tracer_for_array(tr, t->next); next 4363 kernel/trace/trace.c if (t->next) next 4373 kernel/trace/trace.c .next = t_next, next 5079 kernel/trace/trace.c .next = saved_tgids_next, next 5158 kernel/trace/trace.c .next = saved_cmdlines_next, next 5259 kernel/trace/trace.c if (ptr->tail.next) { next 5260 kernel/trace/trace.c ptr = ptr->tail.next; next 5326 kernel/trace/trace.c .next = eval_map_next, next 5386 kernel/trace/trace.c if (!ptr->tail.next) next 5388 kernel/trace/trace.c ptr = ptr->tail.next; next 5391 kernel/trace/trace.c ptr->tail.next = map_array; next 5654 kernel/trace/trace.c for (t = trace_types; t; t = t->next) { next 7083 kernel/trace/trace.c struct tracing_log_err *err, *next; next 7086 kernel/trace/trace.c list_for_each_entry_safe(err, next, &tr->err_log, list) { next 7148 kernel/trace/trace.c .next = tracing_err_log_seq_next, next 8372 kernel/trace/trace.c for (t = trace_types; t; t = t->next) next 8741 kernel/trace/trace.c last = &map->tail.next; next 8742 kernel/trace/trace.c map = map->tail.next; next 8747 kernel/trace/trace.c *last = trace_eval_jmp_to_tail(map)->tail.next; next 8839 kernel/trace/trace.c .next = NULL, next 498 kernel/trace/trace.h struct tracer *next; next 130 kernel/trace/trace_dynevent.c .next = dyn_event_seq_next, next 190 kernel/trace/trace_events.c struct ftrace_event_field *field, *next; next 194 kernel/trace/trace_events.c list_for_each_entry_safe(field, next, head, link) { next 542 kernel/trace/trace_events.c struct task_struct *prev, struct task_struct *next) next 551 kernel/trace/trace_events.c trace_ignore_this_task(pid_list, next)); next 556 kernel/trace/trace_events.c struct task_struct *prev, struct task_struct *next) next 564 kernel/trace/trace_events.c trace_ignore_this_task(pid_list, next)); next 1288 kernel/trace/trace_events.c .next = f_next, next 1661 kernel/trace/trace_events.c .next = t_next, next 1668 kernel/trace/trace_events.c .next = s_next, next 1675 kernel/trace/trace_events.c .next = p_next, next 2878 kernel/trace/trace_events.c struct trace_event_file *file, *next; next 2880 kernel/trace/trace_events.c list_for_each_entry_safe(file, next, &tr->events, list) next 448 kernel/trace/trace_events_filter.c const char *next = ptr++; next 450 kernel/trace/trace_events_filter.c if (isspace(*next)) next 453 kernel/trace/trace_events_filter.c switch (*next) { next 462 kernel/trace/trace_events_filter.c if (!is_not(next)) next 469 kernel/trace/trace_events_filter.c parse_error(pe, FILT_ERR_TOO_MANY_PREDS, next - str); next 476 kernel/trace/trace_events_filter.c len = parse_pred(next, data, ptr - str, pe, &prog[N].pred); next 481 kernel/trace/trace_events_filter.c ptr = next + len; next 487 kernel/trace/trace_events_filter.c next = ptr++; next 488 kernel/trace/trace_events_filter.c if (isspace(*next)) next 491 kernel/trace/trace_events_filter.c switch (*next) { next 498 kernel/trace/trace_events_filter.c if (next[1] == next[0]) { next 505 kernel/trace/trace_events_filter.c next - str); next 515 kernel/trace/trace_events_filter.c if (*next == '&') { /* #8 */ next 523 kernel/trace/trace_events_filter.c if (*next == '|') { /* #10 */ next 527 kernel/trace/trace_events_filter.c if (!*next) /* #11 */ next 1493 kernel/trace/trace_events_hist.c .next = dyn_event_seq_next, next 171 kernel/trace/trace_events_trigger.c .next = trigger_next, next 219 kernel/trace/trace_events_trigger.c char *command, *next = buff; next 223 kernel/trace/trace_events_trigger.c command = strsep(&next, ": \t"); next 229 kernel/trace/trace_events_trigger.c ret = p->func(p, file, buff, command, next); next 426 kernel/trace/trace_functions_graph.c struct ftrace_graph_ret_entry *next; next 434 kernel/trace/trace_functions_graph.c next = &data->ret; next 456 kernel/trace/trace_functions_graph.c next = ring_buffer_event_data(event); next 469 kernel/trace/trace_functions_graph.c if (next->ent.type == TRACE_GRAPH_RET) next 470 kernel/trace/trace_functions_graph.c data->ret = *next; next 472 kernel/trace/trace_functions_graph.c data->ret.ent.type = next->ent.type; next 476 kernel/trace/trace_functions_graph.c if (next->ent.type != TRACE_GRAPH_RET) next 479 kernel/trace/trace_functions_graph.c if (curr->ent.pid != next->ent.pid || next 480 kernel/trace/trace_functions_graph.c curr->graph_ent.func != next->ret.func) next 487 kernel/trace/trace_functions_graph.c return next; next 953 kernel/trace/trace_kprobe.c .next = dyn_event_seq_next, next 1011 kernel/trace/trace_kprobe.c .next = dyn_event_seq_next, next 154 kernel/trace/trace_printk.c if (mod_fmt->list.next == &trace_bprintk_fmt_list) next 157 kernel/trace/trace_printk.c mod_fmt = container_of(mod_fmt->list.next, typeof(*mod_fmt), list); next 344 kernel/trace/trace_printk.c .next = t_next, next 25 kernel/trace/trace_sched_switch.c struct task_struct *prev, struct task_struct *next) next 34 kernel/trace/trace_sched_switch.c tracing_record_taskinfo_sched_switch(prev, next, flags); next 377 kernel/trace/trace_sched_wakeup.c struct task_struct *next, next 393 kernel/trace/trace_sched_wakeup.c entry->next_pid = next->pid; next 394 kernel/trace/trace_sched_wakeup.c entry->next_prio = next->prio; next 395 kernel/trace/trace_sched_wakeup.c entry->next_state = task_state_index(next); next 396 kernel/trace/trace_sched_wakeup.c entry->next_cpu = task_cpu(next); next 432 kernel/trace/trace_sched_wakeup.c struct task_struct *prev, struct task_struct *next) next 455 kernel/trace/trace_sched_wakeup.c if (next != wakeup_task) next 470 kernel/trace/trace_sched_wakeup.c if (unlikely(!tracer_enabled || next != wakeup_task)) next 477 kernel/trace/trace_sched_wakeup.c tracing_sched_switch_trace(wakeup_trace, prev, next, flags, pc); next 472 kernel/trace/trace_stack.c .next = t_next, next 229 kernel/trace/trace_stat.c .next = stat_seq_next, next 757 kernel/trace/trace_uprobe.c .next = dyn_event_seq_next, next 812 kernel/trace/trace_uprobe.c .next = dyn_event_seq_next, next 81 kernel/tracepoint.c early_probes = tmp->next; next 102 kernel/tracepoint.c tp_probes->rcu.next = early_probes; next 691 kernel/user_namespace.c .next = m_next, next 698 kernel/user_namespace.c .next = m_next, next 705 kernel/user_namespace.c .next = m_next, next 871 kernel/workqueue.c struct worker *next, *worker = kthread_data(task); next 903 kernel/workqueue.c next = first_idle_worker(pool); next 904 kernel/workqueue.c if (next) next 905 kernel/workqueue.c wake_up_process(next->task); next 1774 kernel/workqueue.c (worker->hentry.next || worker->hentry.pprev))) next 2678 kernel/workqueue.c head = worker->scheduled.next; next 2682 kernel/workqueue.c head = target->entry.next; next 2858 kernel/workqueue.c struct wq_flusher *next, *tmp; next 2861 kernel/workqueue.c list_for_each_entry_safe(next, tmp, &wq->flusher_queue, list) { next 2862 kernel/workqueue.c if (next->flush_color != wq->flush_color) next 2864 kernel/workqueue.c list_del_init(&next->list); next 2865 kernel/workqueue.c complete(&next->done); next 2902 kernel/workqueue.c WARN_ON_ONCE(wq->flush_color != next->flush_color); next 2904 kernel/workqueue.c list_del_init(&next->list); next 2905 kernel/workqueue.c wq->first_flusher = next; next 4174 kernel/workqueue.c WARN(!ret && (wq->pwqs.next != &wq->dfl_pwq->pwqs_node || next 482 lib/842/842_compress.c u64 last, next, pad, total; next 517 lib/842/842_compress.c next = get_unaligned((u64 *)p->in); next 528 lib/842/842_compress.c if (next == last) { next 536 lib/842/842_compress.c if (next == last) /* reached max repeat bits */ next 540 lib/842/842_compress.c if (next == 0) next 549 lib/842/842_compress.c last = next; next 205 lib/cmdline.c char *next; next 243 lib/cmdline.c next = args + i + 1; next 245 lib/cmdline.c next = args + i; next 248 lib/cmdline.c return skip_spaces(next); next 79 lib/cpumask.c int next; next 82 lib/cpumask.c next = cpumask_next(n, mask); next 84 lib/cpumask.c if (wrap && n < start && next >= start) { next 87 lib/cpumask.c } else if (next >= nr_cpumask_bits) { next 93 lib/cpumask.c return next; next 890 lib/debugobjects.c unsigned int expect, unsigned int next) next 909 lib/debugobjects.c obj->astate = next; next 734 lib/dynamic_debug.c iter->table = list_entry(ddebug_tables.next, next 757 lib/dynamic_debug.c iter->table = list_entry(iter->table->link.next, next 851 lib/dynamic_debug.c .next = ddebug_proc_next, next 984 lib/dynamic_debug.c struct ddebug_table *dt = list_entry(ddebug_tables.next, next 199 lib/error-inject.c .next = ei_seq_next, next 395 lib/idr.c next: next 433 lib/idr.c goto next; next 107 lib/ioremap.c unsigned long next; next 113 lib/ioremap.c next = pmd_addr_end(addr, end); next 115 lib/ioremap.c if (ioremap_try_huge_pmd(pmd, addr, next, phys_addr, prot)) next 118 lib/ioremap.c if (ioremap_pte_range(pmd, addr, next, phys_addr, prot)) next 120 lib/ioremap.c } while (pmd++, phys_addr += (next - addr), addr = next, addr != end); next 150 lib/ioremap.c unsigned long next; next 156 lib/ioremap.c next = pud_addr_end(addr, end); next 158 lib/ioremap.c if (ioremap_try_huge_pud(pud, addr, next, phys_addr, prot)) next 161 lib/ioremap.c if (ioremap_pmd_range(pud, addr, next, phys_addr, prot)) next 163 lib/ioremap.c } while (pud++, phys_addr += (next - addr), addr = next, addr != end); next 193 lib/ioremap.c unsigned long next; next 199 lib/ioremap.c next = p4d_addr_end(addr, end); next 201 lib/ioremap.c if (ioremap_try_huge_p4d(p4d, addr, next, phys_addr, prot)) next 204 lib/ioremap.c if (ioremap_pud_range(p4d, addr, next, phys_addr, prot)) next 206 lib/ioremap.c } while (p4d++, phys_addr += (next - addr), addr = next, addr != end); next 215 lib/ioremap.c unsigned long next; next 224 lib/ioremap.c next = pgd_addr_end(addr, end); next 225 lib/ioremap.c err = ioremap_p4d_range(pgd, addr, next, phys_addr, prot); next 228 lib/ioremap.c } while (pgd++, phys_addr += (next - addr), addr = next, addr != end); next 329 lib/iov_iter.c int next = pipe->curbuf + pipe->nrbufs; next 334 lib/iov_iter.c if (unlikely(idx != ((next - 1) & (pipe->buffers - 1)))) next 341 lib/iov_iter.c if (idx != (next & (pipe->buffers - 1))) next 568 lib/iov_iter.c __wsum next = csum_partial_copy_nocheck(from, to, len, 0); next 569 lib/iov_iter.c return csum_block_add(sum, next, off); next 1408 lib/iov_iter.c __wsum sum, next; next 1417 lib/iov_iter.c next = csum_and_copy_from_user(v.iov_base, next 1421 lib/iov_iter.c sum = csum_block_add(sum, next, off); next 1448 lib/iov_iter.c __wsum sum, next; next 1459 lib/iov_iter.c next = csum_and_copy_from_user(v.iov_base, next 1464 lib/iov_iter.c sum = csum_block_add(sum, next, off); next 1492 lib/iov_iter.c __wsum sum, next; next 1505 lib/iov_iter.c next = csum_and_copy_to_user((from += v.iov_len) - v.iov_len, next 1509 lib/iov_iter.c sum = csum_block_add(sum, next, off); next 105 lib/irq_poll.c iop = list_entry(list->next, struct irq_poll, list); next 379 lib/klist.c struct klist_node *next; next 385 lib/klist.c next = to_klist_node(last->n_node.next); next 389 lib/klist.c next = to_klist_node(i->i_klist->k_list.next); next 392 lib/klist.c while (next != to_klist_node(&i->i_klist->k_list)) { next 393 lib/klist.c if (likely(!knode_dead(next))) { next 394 lib/klist.c kref_get(&next->n_ref); next 395 lib/klist.c i->i_cur = next; next 398 lib/klist.c next = to_klist_node(next->n_node.next); next 102 lib/kobject_uevent.c const char *next = buf; next 104 lib/kobject_uevent.c while (next <= buf_end && *next != delim) next 105 lib/kobject_uevent.c if (!isalnum(*next++)) next 108 lib/kobject_uevent.c if (next == buf) next 111 lib/kobject_uevent.c return next; next 118 lib/kobject_uevent.c const char *next, *buf_end, *key; next 141 lib/kobject_uevent.c next = buf + UUID_STRING_LEN; next 144 lib/kobject_uevent.c while (next <= buf_end) { next 145 lib/kobject_uevent.c if (*next != ' ') next 149 lib/kobject_uevent.c key = ++next; next 153 lib/kobject_uevent.c buf = next; next 154 lib/kobject_uevent.c next = action_arg_word_end(buf, buf_end, '='); next 155 lib/kobject_uevent.c if (!next || next > buf_end || *next != '=') next 157 lib/kobject_uevent.c key_len = next - buf; next 160 lib/kobject_uevent.c if (++next > buf_end) next 163 lib/kobject_uevent.c buf = next; next 164 lib/kobject_uevent.c next = action_arg_word_end(buf, buf_end, ' '); next 165 lib/kobject_uevent.c if (!next) next 169 lib/kobject_uevent.c key_len, key, (int) (next - buf), buf)) next 21 lib/list_debug.c struct list_head *next) next 23 lib/list_debug.c if (CHECK_DATA_CORRUPTION(next->prev != prev, next 25 lib/list_debug.c prev, next->prev, next) || next 26 lib/list_debug.c CHECK_DATA_CORRUPTION(prev->next != next, next 28 lib/list_debug.c next, prev->next, prev) || next 29 lib/list_debug.c CHECK_DATA_CORRUPTION(new == prev || new == next, next 31 lib/list_debug.c new, prev, next)) next 40 lib/list_debug.c struct list_head *prev, *next; next 43 lib/list_debug.c next = entry->next; next 45 lib/list_debug.c if (CHECK_DATA_CORRUPTION(next == LIST_POISON1, next 51 lib/list_debug.c CHECK_DATA_CORRUPTION(prev->next != entry, next 53 lib/list_debug.c entry, prev->next) || next 54 lib/list_debug.c CHECK_DATA_CORRUPTION(next->prev != entry, next 56 lib/list_debug.c entry, next->prev)) next 28 lib/list_sort.c tail = &a->next; next 29 lib/list_sort.c a = a->next; next 36 lib/list_sort.c tail = &b->next; next 37 lib/list_sort.c b = b->next; next 64 lib/list_sort.c tail->next = a; next 67 lib/list_sort.c a = a->next; next 71 lib/list_sort.c tail->next = b; next 74 lib/list_sort.c b = b->next; next 83 lib/list_sort.c tail->next = b; next 96 lib/list_sort.c b = b->next; next 100 lib/list_sort.c tail->next = head; next 192 lib/list_sort.c struct list_head *list = head->next, *pending = NULL; next 199 lib/list_sort.c head->prev->next = NULL; next 239 lib/list_sort.c list = list->next; next 240 lib/list_sort.c pending->next = NULL; next 248 lib/list_sort.c struct list_head *next = pending->prev; next 250 lib/list_sort.c if (!next) next 253 lib/list_sort.c pending = next; next 32 lib/llist.c new_last->next = first = READ_ONCE(head->first); next 55 lib/llist.c struct llist_node *entry, *old_entry, *next; next 62 lib/llist.c next = READ_ONCE(entry->next); next 63 lib/llist.c entry = cmpxchg(&head->first, old_entry, next); next 85 lib/llist.c head = head->next; next 86 lib/llist.c tmp->next = new_head; next 327 lib/lru_cache.c n = lc->free.next; next 46 lib/lzo/lzo1x_compress.c next: next 298 lib/lzo/lzo1x_compress.c goto next; next 44 lib/lzo/lzo1x_decompress_safe.c size_t t, next; next 68 lib/lzo/lzo1x_decompress_safe.c next = t; next 121 lib/lzo/lzo1x_decompress_safe.c next = t & 3; next 132 lib/lzo/lzo1x_decompress_safe.c next = t & 3; next 139 lib/lzo/lzo1x_decompress_safe.c next = t & 3; next 163 lib/lzo/lzo1x_decompress_safe.c next = get_unaligned_le16(ip); next 165 lib/lzo/lzo1x_decompress_safe.c m_pos -= next >> 2; next 166 lib/lzo/lzo1x_decompress_safe.c next &= 3; next 169 lib/lzo/lzo1x_decompress_safe.c next = get_unaligned_le16(ip); next 170 lib/lzo/lzo1x_decompress_safe.c if (((next & 0xfffc) == 0xfffc) && next 180 lib/lzo/lzo1x_decompress_safe.c next &= 3; next 202 lib/lzo/lzo1x_decompress_safe.c next = get_unaligned_le16(ip); next 205 lib/lzo/lzo1x_decompress_safe.c m_pos -= next >> 2; next 206 lib/lzo/lzo1x_decompress_safe.c next &= 3; next 227 lib/lzo/lzo1x_decompress_safe.c state = next; next 229 lib/lzo/lzo1x_decompress_safe.c op += next; next 230 lib/lzo/lzo1x_decompress_safe.c ip += next; next 253 lib/lzo/lzo1x_decompress_safe.c state = next; next 254 lib/lzo/lzo1x_decompress_safe.c t = next; next 120 lib/mpi/mpi-internal.h struct karatsuba_ctx *next; next 374 lib/mpi/mpih-mul.c if (!ctx->next) { next 375 lib/mpi/mpih-mul.c ctx->next = kzalloc(sizeof *ctx, GFP_KERNEL); next 376 lib/mpi/mpih-mul.c if (!ctx->next) next 382 lib/mpi/mpih-mul.c ctx->next) < 0) next 401 lib/mpi/mpih-mul.c for (ctx = ctx->next; ctx; ctx = ctx2) { next 402 lib/mpi/mpih-mul.c ctx2 = ctx->next; next 79 lib/oid_registry.c goto next; next 83 lib/oid_registry.c goto next; next 87 lib/oid_registry.c next: next 35 lib/plist.c WARN(n->prev != p || p->next != n, next 39 lib/plist.c t, t->next, t->prev, next 40 lib/plist.c p, p->next, p->prev, next 41 lib/plist.c n, n->next, n->prev); next 46 lib/plist.c struct list_head *prev = top, *next = top->next; next 48 lib/plist.c plist_check_prev_next(top, prev, next); next 49 lib/plist.c while (next != top) { next 50 lib/plist.c prev = next; next 51 lib/plist.c next = prev->next; next 52 lib/plist.c plist_check_prev_next(top, prev, next); next 94 lib/plist.c iter = list_entry(iter->prio_list.next, next 117 lib/plist.c if (node->node_list.next != &head->node_list) { next 118 lib/plist.c struct plist_node *next; next 120 lib/plist.c next = list_entry(node->node_list.next, next 124 lib/plist.c if (list_empty(&next->prio_list)) next 125 lib/plist.c list_add(&next->prio_list, &node->prio_list); next 204 lib/plist.c BUG_ON(prio_pos->prio_list.next != &node_pos->prio_list); next 209 lib/plist.c BUG_ON(prio_pos->prio_list.next != &first->prio_list); next 222 lib/rhashtable.c struct rhash_head *head, *next, *entry; next 234 lib/rhashtable.c next = rht_dereference_bucket(entry->next, old_tbl, old_hash); next 236 lib/rhashtable.c if (rht_is_a_nulls(next)) next 239 lib/rhashtable.c pprev = &entry->next; next 251 lib/rhashtable.c RCU_INIT_POINTER(entry->next, head); next 256 lib/rhashtable.c rcu_assign_pointer(*pprev, next); next 259 lib/rhashtable.c rht_assign_locked(bkt, next); next 501 lib/rhashtable.c pprev = &head->next; next 511 lib/rhashtable.c RCU_INIT_POINTER(list->next, plist); next 512 lib/rhashtable.c head = rht_dereference_bucket(head->next, tbl, hash); next 513 lib/rhashtable.c RCU_INIT_POINTER(list->rhead.next, head); next 560 lib/rhashtable.c RCU_INIT_POINTER(obj->next, head); next 565 lib/rhashtable.c RCU_INIT_POINTER(list->next, NULL); next 754 lib/rhashtable.c list = rcu_dereference(list->next)) { next 800 lib/rhashtable.c goto next; next 802 lib/rhashtable.c list = rcu_dereference(list->next); next 812 lib/rhashtable.c next: next 860 lib/rhashtable.c if (!rhlist || !(list = rcu_dereference(list->next))) { next 861 lib/rhashtable.c p = rcu_dereference(p->next); next 1103 lib/rhashtable.c list = rht_dereference(list->next, ht); next 1137 lib/rhashtable.c struct rhash_head *pos, *next; next 1141 lib/rhashtable.c next = !rht_is_a_nulls(pos) ? next 1142 lib/rhashtable.c rht_dereference(pos->next, ht) : NULL; next 1144 lib/rhashtable.c pos = next, next 1145 lib/rhashtable.c next = !rht_is_a_nulls(pos) ? next 1146 lib/rhashtable.c rht_dereference(pos->next, ht) : NULL) next 195 lib/scatterlist.c struct scatterlist *sgl, *next; next 213 lib/scatterlist.c next = sg_chain_ptr(&sgl[curr_max_ents - 1]); next 218 lib/scatterlist.c next = NULL; next 226 lib/scatterlist.c sgl = next; next 60 lib/stackdepot.c struct stack_record *next; /* Link in the hashtable */ next 183 lib/stackdepot.c for (found = bucket; found; found = found->next) { next 280 lib/stackdepot.c new->next = *bucket; next 100 lib/test_list_sort.c for (cur = head.next; cur->next != &head; cur = cur->next) { next 104 lib/test_list_sort.c if (cur->next->prev != cur) { next 109 lib/test_list_sort.c cmp_result = cmp(NULL, cur, cur->next); next 116 lib/test_list_sort.c el1 = container_of(cur->next, struct debug_el, list); next 497 lib/test_rhashtable.c struct rhash_head *pos, *next; next 501 lib/test_rhashtable.c next = !rht_is_a_nulls(pos) ? rht_dereference(pos->next, ht) : NULL; next 512 lib/test_rhashtable.c list = rht_dereference(list->next, ht); next 520 lib/test_rhashtable.c pos = next, next 521 lib/test_rhashtable.c next = !rht_is_a_nulls(pos) ? next 522 lib/test_rhashtable.c rht_dereference(pos->next, ht) : NULL; next 209 lib/test_xarray.c unsigned long next = base + (1UL << order); next 216 lib/test_xarray.c XA_BUG_ON(xa, xa_store_index(xa, next, GFP_KERNEL)); next 219 lib/test_xarray.c for (i = base; i < next; i++) { next 244 lib/test_xarray.c XA_BUG_ON(xa, xa_get_mark(xa, next, XA_MARK_0)); next 245 lib/test_xarray.c XA_BUG_ON(xa, xa_get_mark(xa, next, XA_MARK_1)); next 246 lib/test_xarray.c XA_BUG_ON(xa, xa_get_mark(xa, next, XA_MARK_2)); next 248 lib/test_xarray.c xa_erase_index(xa, next); next 781 lib/test_xarray.c u32 next = 0; next 787 lib/test_xarray.c &next, GFP_KERNEL) != 0); next 790 lib/test_xarray.c next = 0x3ffd; next 792 lib/test_xarray.c &next, GFP_KERNEL) != 0); next 804 lib/test_xarray.c &next, GFP_KERNEL) != (id == 1)); next 812 lib/test_xarray.c next = UINT_MAX; next 814 lib/test_xarray.c xa_limit_32b, &next, GFP_KERNEL) != 0); next 817 lib/test_xarray.c xa_limit_32b, &next, GFP_KERNEL) != 1); next 820 lib/test_xarray.c xa_limit_32b, &next, GFP_KERNEL) != 0); next 913 lib/test_xarray.c unsigned long next = 4 << order; next 917 lib/test_xarray.c XA_BUG_ON(xa, xa_store_index(xa, next, GFP_KERNEL) != NULL); next 918 lib/test_xarray.c XA_BUG_ON(xa, xa_store_index(xa, next + 1, GFP_KERNEL) != NULL); next 927 lib/test_xarray.c XA_BUG_ON(xa, (index < multi) || (index >= next)); next 929 lib/test_xarray.c xa_mk_value(next)); next 930 lib/test_xarray.c XA_BUG_ON(xa, index != next); next 931 lib/test_xarray.c XA_BUG_ON(xa, xa_find_after(xa, &index, next, XA_PRESENT) != NULL); next 932 lib/test_xarray.c XA_BUG_ON(xa, index != next); next 935 lib/test_xarray.c xa_erase_index(xa, next); next 936 lib/test_xarray.c xa_erase_index(xa, next + 1); next 85 lib/timerqueue.c struct rb_node *next; next 89 lib/timerqueue.c next = rb_next(&node->node); next 90 lib/timerqueue.c if (!next) next 92 lib/timerqueue.c return container_of(next, struct timerqueue_node, node); next 79 lib/ts_bm.c goto next; next 85 lib/ts_bm.c next: bs = bm->bad_shift[text[shift-i]]; next 137 lib/ts_fsm.c struct ts_fsm_token *cur = NULL, *next; next 169 lib/ts_fsm.c next = &fsm->tokens[tok_idx + 1]; next 171 lib/ts_fsm.c next = NULL; next 199 lib/ts_fsm.c if (next == NULL) next 205 lib/ts_fsm.c while (!match_token(next, data[block_idx])) { next 222 lib/ts_fsm.c while (!match_token(next, data[block_idx])) { next 3134 lib/vsprintf.c char *next; next 3347 lib/vsprintf.c simple_strtol(str, &next, base) : next 3348 lib/vsprintf.c simple_strtoll(str, &next, base); next 3351 lib/vsprintf.c simple_strtoul(str, &next, base) : next 3352 lib/vsprintf.c simple_strtoull(str, &next, base); next 3354 lib/vsprintf.c if (field_width > 0 && next - str > field_width) { next 3357 lib/vsprintf.c while (next - str > field_width) { next 3362 lib/vsprintf.c --next; next 3403 lib/vsprintf.c if (!next) next 3405 lib/vsprintf.c str = next; next 776 lib/xarray.c void *first, *next; next 794 lib/xarray.c next = first; next 814 lib/xarray.c if (xa_is_node(next) && (!node || node->shift)) next 815 lib/xarray.c xas_free_nodes(xas, xa_to_node(next)); next 818 lib/xarray.c count += !next - !entry; next 829 lib/xarray.c next = xa_entry_locked(xas->xa, node, ++offset); next 830 lib/xarray.c if (!xa_is_sibling(next)) { next 833 lib/xarray.c first = next; next 1664 lib/xarray.c struct xa_limit limit, u32 *next, gfp_t gfp) next 1669 lib/xarray.c limit.min = max(min, *next); next 1684 lib/xarray.c *next = *id + 1; next 1685 lib/xarray.c if (*next == 0) next 2022 lib/xarray.c node->array, node->private_list.prev, node->private_list.next); next 179 lib/zlib_deflate/deflate.c char *next; next 208 lib/zlib_deflate/deflate.c next = (char *) mem; next 209 lib/zlib_deflate/deflate.c next += sizeof(*mem); next 210 lib/zlib_deflate/deflate.c mem->window_memory = (Byte *) next; next 211 lib/zlib_deflate/deflate.c next += zlib_deflate_window_memsize(windowBits); next 212 lib/zlib_deflate/deflate.c mem->prev_memory = (Pos *) next; next 213 lib/zlib_deflate/deflate.c next += zlib_deflate_prev_memsize(windowBits); next 214 lib/zlib_deflate/deflate.c mem->head_memory = (Pos *) next; next 215 lib/zlib_deflate/deflate.c next += zlib_deflate_head_memsize(memLevel); next 216 lib/zlib_deflate/deflate.c mem->overlay_memory = next; next 38 lib/zlib_inflate/inflate.c state->lencode = state->distcode = state->next = state->codes; next 170 lib/zlib_inflate/inflate.c next = strm->next_in; \ next 181 lib/zlib_inflate/inflate.c strm->next_in = next; \ next 200 lib/zlib_inflate/inflate.c hold += (unsigned long)(*next++) << bits; \ next 320 lib/zlib_inflate/inflate.c const unsigned char *next; /* next input */ next 441 lib/zlib_inflate/inflate.c memcpy(put, next, copy); next 443 lib/zlib_inflate/inflate.c next += copy; next 477 lib/zlib_inflate/inflate.c state->next = state->codes; next 478 lib/zlib_inflate/inflate.c state->lencode = (code const *)(state->next); next 480 lib/zlib_inflate/inflate.c ret = zlib_inflate_table(CODES, state->lens, 19, &(state->next), next 543 lib/zlib_inflate/inflate.c state->next = state->codes; next 544 lib/zlib_inflate/inflate.c state->lencode = (code const *)(state->next); next 546 lib/zlib_inflate/inflate.c ret = zlib_inflate_table(LENS, state->lens, state->nlen, &(state->next), next 553 lib/zlib_inflate/inflate.c state->distcode = (code const *)(state->next); next 556 lib/zlib_inflate/inflate.c &(state->next), &(state->distbits), state->work); next 106 lib/zlib_inflate/inflate.h code *next; /* next available space in codes[] */ next 40 lib/zlib_inflate/inftrees.c code *next; /* next available space in table */ next 190 lib/zlib_inflate/inftrees.c next = *table; /* current table to fill in */ next 224 lib/zlib_inflate/inftrees.c next[(huff >> drop) + fill] = this; next 252 lib/zlib_inflate/inftrees.c next += min; /* here min is 1 << curr */ next 273 lib/zlib_inflate/inftrees.c (*table)[low].val = (unsigned short)(next - *table); next 292 lib/zlib_inflate/inftrees.c next = *table; next 297 lib/zlib_inflate/inftrees.c next[huff >> drop] = this; next 588 mm/backing-dev.c blkcg_cgwb_list->next && memcg_cgwb_list->next) { next 747 mm/backing-dev.c struct bdi_writeback *wb, *next; next 750 mm/backing-dev.c list_for_each_entry_safe(wb, next, memcg_cgwb_list, memcg_node) next 752 mm/backing-dev.c memcg_cgwb_list->next = NULL; /* prevent new wb's */ next 764 mm/backing-dev.c struct bdi_writeback *wb, *next; next 767 mm/backing-dev.c list_for_each_entry_safe(wb, next, &blkcg->cgwb_list, blkcg_node) next 769 mm/backing-dev.c blkcg->cgwb_list.next = NULL; /* prevent new wb's */ next 55 mm/compaction.c struct page *page, *next; next 58 mm/compaction.c list_for_each_entry_safe(page, next, freelist, lru) { next 72 mm/compaction.c struct page *page, *next; next 75 mm/compaction.c list_for_each_entry_safe(page, next, list, lru) { next 1562 mm/compaction.c freepage = list_entry(cc->freepages.next, struct page, lru); next 69 mm/dmapool.c char *next; next 73 mm/dmapool.c next = buf; next 76 mm/dmapool.c temp = scnprintf(next, size, "poolinfo - 0.1\n"); next 78 mm/dmapool.c next += temp; next 93 mm/dmapool.c temp = scnprintf(next, size, "%-16s %4u %4zu %4zu %2u\n", next 98 mm/dmapool.c next += temp; next 211 mm/dmapool.c unsigned int next = offset + pool->size; next 212 mm/dmapool.c if (unlikely((next + pool->size) >= next_boundary)) { next 213 mm/dmapool.c next = next_boundary; next 216 mm/dmapool.c *(int *)(page->vaddr + offset) = next; next 217 mm/dmapool.c offset = next; next 288 mm/dmapool.c page = list_entry(pool->page_list.next, next 2617 mm/filemap.c goto next; next 2624 mm/filemap.c goto next; next 2626 mm/filemap.c goto next; next 2657 mm/filemap.c goto next; next 2662 mm/filemap.c next: next 33 mm/frontswap.c for ((ops) = frontswap_ops; (ops); (ops) = (ops)->next) next 140 mm/frontswap.c ops->next = frontswap_ops; next 141 mm/frontswap.c } while (cmpxchg(&frontswap_ops, ops->next, ops) != ops->next); next 2047 mm/gup.c unsigned long next; next 2051 mm/gup.c next = hugepte_addr_end(addr, end, sz); next 2054 mm/gup.c } while (ptep++, addr = next, addr != end); next 2190 mm/gup.c unsigned long next; next 2197 mm/gup.c next = pmd_addr_end(addr, end); next 2211 mm/gup.c if (!gup_huge_pmd(pmd, pmdp, addr, next, flags, next 2221 mm/gup.c PMD_SHIFT, next, flags, pages, nr)) next 2223 mm/gup.c } else if (!gup_pte_range(pmd, addr, next, flags, pages, nr)) next 2225 mm/gup.c } while (pmdp++, addr = next, addr != end); next 2233 mm/gup.c unsigned long next; next 2240 mm/gup.c next = pud_addr_end(addr, end); next 2244 mm/gup.c if (!gup_huge_pud(pud, pudp, addr, next, flags, next 2249 mm/gup.c PUD_SHIFT, next, flags, pages, nr)) next 2251 mm/gup.c } else if (!gup_pmd_range(pud, addr, next, flags, pages, nr)) next 2253 mm/gup.c } while (pudp++, addr = next, addr != end); next 2261 mm/gup.c unsigned long next; next 2268 mm/gup.c next = p4d_addr_end(addr, end); next 2274 mm/gup.c P4D_SHIFT, next, flags, pages, nr)) next 2276 mm/gup.c } else if (!gup_pud_range(p4d, addr, next, flags, pages, nr)) next 2278 mm/gup.c } while (p4dp++, addr = next, addr != end); next 2286 mm/gup.c unsigned long next; next 2293 mm/gup.c next = pgd_addr_end(addr, end); next 2297 mm/gup.c if (!gup_huge_pgd(pgd, pgdp, addr, next, flags, next 2302 mm/gup.c PGDIR_SHIFT, next, flags, pages, nr)) next 2304 mm/gup.c } else if (!gup_p4d_range(pgd, addr, next, flags, pages, nr)) next 2306 mm/gup.c } while (pgdp++, addr = next, addr != end); next 26 mm/gup_benchmark.c unsigned long i, nr_pages, addr, next; next 42 mm/gup_benchmark.c for (addr = gup->addr; addr < gup->addr + gup->size; addr = next) { next 46 mm/gup_benchmark.c next = addr + nr * PAGE_SIZE; next 47 mm/gup_benchmark.c if (next > gup->addr + gup->size) { next 48 mm/gup_benchmark.c next = gup->addr + gup->size; next 49 mm/gup_benchmark.c nr = (next - addr) / PAGE_SIZE; next 661 mm/hmm.c unsigned long addr = start, next; next 713 mm/hmm.c next = pmd_addr_end(addr, end); next 714 mm/hmm.c ret = hmm_vma_walk_pmd(pmdp, addr, next, walk); next 717 mm/hmm.c } while (pmdp++, addr = next, addr != end); next 1695 mm/huge_memory.c pmd_t *pmd, unsigned long addr, unsigned long next) next 1734 mm/huge_memory.c if (next - addr != HPAGE_PMD_SIZE) { next 2400 mm/huge_memory.c struct vm_area_struct *next = vma->vm_next; next 2401 mm/huge_memory.c unsigned long nstart = next->vm_start; next 2404 mm/huge_memory.c (nstart & HPAGE_PMD_MASK) >= next->vm_start && next 2405 mm/huge_memory.c (nstart & HPAGE_PMD_MASK) + HPAGE_PMD_SIZE <= next->vm_end) next 2406 mm/huge_memory.c split_huge_pmd_address(next, nstart, false, NULL); next 2907 mm/huge_memory.c LIST_HEAD(list), *pos, *next; next 2918 mm/huge_memory.c list_for_each_safe(pos, next, &ds_queue->split_queue) { next 2933 mm/huge_memory.c list_for_each_safe(pos, next, &list) { next 2936 mm/huge_memory.c goto next; next 2941 mm/huge_memory.c next: next 2988 mm/huge_memory.c goto next; next 2991 mm/huge_memory.c goto next; next 2998 mm/huge_memory.c next: next 1344 mm/hugetlb.c node = node->next; next 1573 mm/hugetlb.c list_entry(h->hugepage_freelists[node].next, next 2370 mm/hugetlb.c struct page *page, *next; next 2372 mm/hugetlb.c list_for_each_entry_safe(page, next, freel, lru) { next 117 mm/kasan/init.c unsigned long next; next 120 mm/kasan/init.c next = pmd_addr_end(addr, end); next 140 mm/kasan/init.c zero_pte_populate(pmd, addr, next); next 141 mm/kasan/init.c } while (pmd++, addr = next, addr != end); next 150 mm/kasan/init.c unsigned long next; next 153 mm/kasan/init.c next = pud_addr_end(addr, end); next 177 mm/kasan/init.c zero_pmd_populate(pud, addr, next); next 178 mm/kasan/init.c } while (pud++, addr = next, addr != end); next 187 mm/kasan/init.c unsigned long next; next 190 mm/kasan/init.c next = p4d_addr_end(addr, end); next 218 mm/kasan/init.c zero_pud_populate(p4d, addr, next); next 219 mm/kasan/init.c } while (p4d++, addr = next, addr != end); next 236 mm/kasan/init.c unsigned long next; next 239 mm/kasan/init.c next = pgd_addr_end(addr, end); next 291 mm/kasan/init.c zero_p4d_populate(pgd, addr, next); next 292 mm/kasan/init.c } while (pgd++, addr = next, addr != end); next 360 mm/kasan/init.c unsigned long next; next 362 mm/kasan/init.c for (; addr < end; addr = next, pte++) { next 363 mm/kasan/init.c next = (addr + PAGE_SIZE) & PAGE_MASK; next 364 mm/kasan/init.c if (next > end) next 365 mm/kasan/init.c next = end; next 379 mm/kasan/init.c unsigned long next; next 381 mm/kasan/init.c for (; addr < end; addr = next, pmd++) { next 384 mm/kasan/init.c next = pmd_addr_end(addr, end); next 391 mm/kasan/init.c IS_ALIGNED(next, PMD_SIZE)) next 396 mm/kasan/init.c kasan_remove_pte_table(pte, addr, next); next 404 mm/kasan/init.c unsigned long next; next 406 mm/kasan/init.c for (; addr < end; addr = next, pud++) { next 409 mm/kasan/init.c next = pud_addr_end(addr, end); next 416 mm/kasan/init.c IS_ALIGNED(next, PUD_SIZE)) next 422 mm/kasan/init.c kasan_remove_pmd_table(pmd, addr, next); next 430 mm/kasan/init.c unsigned long next; next 432 mm/kasan/init.c for (; addr < end; addr = next, p4d++) { next 435 mm/kasan/init.c next = p4d_addr_end(addr, end); next 442 mm/kasan/init.c IS_ALIGNED(next, P4D_SIZE)) next 447 mm/kasan/init.c kasan_remove_pud_table(pud, addr, next); next 454 mm/kasan/init.c unsigned long addr, end, next; next 465 mm/kasan/init.c for (; addr < end; addr = next) { next 468 mm/kasan/init.c next = pgd_addr_end(addr, end); next 476 mm/kasan/init.c IS_ALIGNED(next, PGDIR_SIZE)) next 482 mm/kasan/init.c kasan_remove_p4d_table(p4d, addr, next); next 114 mm/kasan/kasan.h struct qlist_node *next; next 67 mm/kasan/quarantine.c q->tail->next = qlink; next 69 mm/kasan/quarantine.c qlink->next = NULL; next 84 mm/kasan/quarantine.c to->tail->next = from->head; next 165 mm/kasan/quarantine.c struct qlist_node *next = qlink->next; next 168 mm/kasan/quarantine.c qlink = next; next 276 mm/kasan/quarantine.c struct qlist_node *next = curr->next; next 284 mm/kasan/quarantine.c curr = next; next 1919 mm/khugepaged.c mm_slot = list_entry(khugepaged_scan.mm_head.next, next 2015 mm/khugepaged.c if (mm_slot->mm_node.next != &khugepaged_scan.mm_head) { next 2017 mm/khugepaged.c mm_slot->mm_node.next, next 1303 mm/kmemleak.c void *next; next 1306 mm/kmemleak.c next = min(start + MAX_SCAN_SIZE, end); next 1307 mm/kmemleak.c scan_block(start, next, NULL); next 1308 mm/kmemleak.c start = next; next 1337 mm/kmemleak.c void *next; next 1340 mm/kmemleak.c next = min(start + MAX_SCAN_SIZE, end); next 1341 mm/kmemleak.c scan_block(start, next, object); next 1343 mm/kmemleak.c start = next; next 1373 mm/kmemleak.c object = list_entry(gray_list.next, typeof(*object), gray_list); next 1381 mm/kmemleak.c tmp = list_entry(object->gray_list.next, typeof(*object), next 1693 mm/kmemleak.c .next = kmemleak_seq_next, next 639 mm/ksm.c if (rmap_item->hlist.next) next 940 mm/ksm.c struct stable_node *stable_node, *next; next 956 mm/ksm.c list_for_each_entry_safe(stable_node, next, &migrate_nodes, list) { next 972 mm/ksm.c ksm_scan.mm_slot = list_entry(ksm_mm_head.mm_list.next, next 995 mm/ksm.c ksm_scan.mm_slot = list_entry(mm_slot->mm_list.next, next 1422 mm/ksm.c BUG_ON(stable_node->hlist.first->next); next 2020 mm/ksm.c if (rmap_item->hlist.next) next 2256 mm/ksm.c struct stable_node *stable_node, *next; next 2259 mm/ksm.c list_for_each_entry_safe(stable_node, next, next 2273 mm/ksm.c slot = list_entry(slot->mm_list.next, struct mm_slot, mm_list); next 2342 mm/ksm.c ksm_scan.mm_slot = list_entry(slot->mm_list.next, next 2768 mm/ksm.c struct stable_node *stable_node, *next; next 2786 mm/ksm.c list_for_each_entry_safe(stable_node, next, &migrate_nodes, list) { next 320 mm/madvise.c unsigned long next = pmd_addr_end(addr, end); next 343 mm/madvise.c if (next - addr != HPAGE_PMD_SIZE) { next 573 mm/madvise.c unsigned long next; next 575 mm/madvise.c next = pmd_addr_end(addr, end); next 577 mm/madvise.c if (madvise_free_huge_pmd(tlb, vma, pmd, addr, next)) next 578 mm/madvise.c goto next; next 691 mm/madvise.c next: next 518 mm/memblock.c struct memblock_region *next = &type->regions[i + 1]; next 520 mm/memblock.c if (this->base + this->size != next->base || next 522 mm/memblock.c memblock_get_region_node(next) || next 523 mm/memblock.c this->flags != next->flags) { next 524 mm/memblock.c BUG_ON(this->base + this->size > next->base); next 529 mm/memblock.c this->size += next->size; next 531 mm/memblock.c memmove(next, next + 1, (type->cnt - (i + 2)) * sizeof(*next)); next 880 mm/memcontrol.c unsigned long val, next; next 883 mm/memcontrol.c next = __this_cpu_read(memcg->vmstats_percpu->targets[target]); next 885 mm/memcontrol.c if ((long)(next - val) < 0) { next 888 mm/memcontrol.c next = val + THRESHOLDS_EVENTS_TARGET; next 891 mm/memcontrol.c next = val + SOFTLIMIT_EVENTS_TARGET; next 894 mm/memcontrol.c next = val + NUMAINFO_EVENTS_TARGET; next 899 mm/memcontrol.c __this_cpu_write(memcg->vmstats_percpu->targets[target], next); next 6745 mm/memcontrol.c struct list_head *next; next 6753 mm/memcontrol.c next = page_list->next; next 6757 mm/memcontrol.c page = list_entry(next, struct page, lru); next 6758 mm/memcontrol.c next = page->lru.next; next 6761 mm/memcontrol.c } while (next != page_list); next 364 mm/memory-failure.c struct to_kill *tk, *next; next 366 mm/memory-failure.c list_for_each_entry_safe (tk, next, to_kill, nd) { next 209 mm/memory.c unsigned long next; next 215 mm/memory.c next = pmd_addr_end(addr, end); next 219 mm/memory.c } while (pmd++, addr = next, addr != end); next 243 mm/memory.c unsigned long next; next 249 mm/memory.c next = pud_addr_end(addr, end); next 252 mm/memory.c free_pmd_range(tlb, pud, addr, next, floor, ceiling); next 253 mm/memory.c } while (pud++, addr = next, addr != end); next 277 mm/memory.c unsigned long next; next 283 mm/memory.c next = p4d_addr_end(addr, end); next 286 mm/memory.c free_pud_range(tlb, p4d, addr, next, floor, ceiling); next 287 mm/memory.c } while (p4d++, addr = next, addr != end); next 313 mm/memory.c unsigned long next; next 363 mm/memory.c next = pgd_addr_end(addr, end); next 366 mm/memory.c free_p4d_range(tlb, pgd, addr, next, floor, ceiling); next 367 mm/memory.c } while (pgd++, addr = next, addr != end); next 374 mm/memory.c struct vm_area_struct *next = vma->vm_next; next 386 mm/memory.c floor, next ? next->vm_start : ceiling); next 391 mm/memory.c while (next && next->vm_start <= vma->vm_end + PMD_SIZE next 392 mm/memory.c && !is_vm_hugetlb_page(next)) { next 393 mm/memory.c vma = next; next 394 mm/memory.c next = vma->vm_next; next 399 mm/memory.c floor, next ? next->vm_start : ceiling); next 401 mm/memory.c vma = next; next 852 mm/memory.c unsigned long next; next 859 mm/memory.c next = pmd_addr_end(addr, end); next 863 mm/memory.c VM_BUG_ON_VMA(next-addr != HPAGE_PMD_SIZE, vma); next 875 mm/memory.c vma, addr, next)) next 877 mm/memory.c } while (dst_pmd++, src_pmd++, addr = next, addr != end); next 886 mm/memory.c unsigned long next; next 893 mm/memory.c next = pud_addr_end(addr, end); next 897 mm/memory.c VM_BUG_ON_VMA(next-addr != HPAGE_PUD_SIZE, vma); next 909 mm/memory.c vma, addr, next)) next 911 mm/memory.c } while (dst_pud++, src_pud++, addr = next, addr != end); next 920 mm/memory.c unsigned long next; next 927 mm/memory.c next = p4d_addr_end(addr, end); next 931 mm/memory.c vma, addr, next)) next 933 mm/memory.c } while (dst_p4d++, src_p4d++, addr = next, addr != end); next 941 mm/memory.c unsigned long next; next 989 mm/memory.c next = pgd_addr_end(addr, end); next 993 mm/memory.c vma, addr, next))) { next 997 mm/memory.c } while (dst_pgd++, src_pgd++, addr = next, addr != end); next 1145 mm/memory.c unsigned long next; next 1149 mm/memory.c next = pmd_addr_end(addr, end); next 1151 mm/memory.c if (next - addr != HPAGE_PMD_SIZE) next 1154 mm/memory.c goto next; next 1165 mm/memory.c goto next; next 1166 mm/memory.c next = zap_pte_range(tlb, vma, pmd, addr, next, details); next 1167 mm/memory.c next: next 1169 mm/memory.c } while (pmd++, addr = next, addr != end); next 1180 mm/memory.c unsigned long next; next 1184 mm/memory.c next = pud_addr_end(addr, end); next 1186 mm/memory.c if (next - addr != HPAGE_PUD_SIZE) { next 1190 mm/memory.c goto next; next 1195 mm/memory.c next = zap_pmd_range(tlb, vma, pud, addr, next, details); next 1196 mm/memory.c next: next 1198 mm/memory.c } while (pud++, addr = next, addr != end); next 1209 mm/memory.c unsigned long next; next 1213 mm/memory.c next = p4d_addr_end(addr, end); next 1216 mm/memory.c next = zap_pud_range(tlb, vma, p4d, addr, next, details); next 1217 mm/memory.c } while (p4d++, addr = next, addr != end); next 1228 mm/memory.c unsigned long next; next 1234 mm/memory.c next = pgd_addr_end(addr, end); next 1237 mm/memory.c next = zap_p4d_range(tlb, vma, pgd, addr, next, details); next 1238 mm/memory.c } while (pgd++, addr = next, addr != end); next 1822 mm/memory.c unsigned long next; next 1831 mm/memory.c next = pmd_addr_end(addr, end); next 1832 mm/memory.c err = remap_pte_range(mm, pmd, addr, next, next 1836 mm/memory.c } while (pmd++, addr = next, addr != end); next 1845 mm/memory.c unsigned long next; next 1853 mm/memory.c next = pud_addr_end(addr, end); next 1854 mm/memory.c err = remap_pmd_range(mm, pud, addr, next, next 1858 mm/memory.c } while (pud++, addr = next, addr != end); next 1867 mm/memory.c unsigned long next; next 1875 mm/memory.c next = p4d_addr_end(addr, end); next 1876 mm/memory.c err = remap_pud_range(mm, p4d, addr, next, next 1880 mm/memory.c } while (p4d++, addr = next, addr != end); next 1900 mm/memory.c unsigned long next; next 1941 mm/memory.c next = pgd_addr_end(addr, end); next 1942 mm/memory.c err = remap_p4d_range(mm, pgd, addr, next, next 1946 mm/memory.c } while (pgd++, addr = next, addr != end); next 2040 mm/memory.c unsigned long next; next 2049 mm/memory.c next = pmd_addr_end(addr, end); next 2050 mm/memory.c err = apply_to_pte_range(mm, pmd, addr, next, fn, data); next 2053 mm/memory.c } while (pmd++, addr = next, addr != end); next 2062 mm/memory.c unsigned long next; next 2069 mm/memory.c next = pud_addr_end(addr, end); next 2070 mm/memory.c err = apply_to_pmd_range(mm, pud, addr, next, fn, data); next 2073 mm/memory.c } while (pud++, addr = next, addr != end); next 2082 mm/memory.c unsigned long next; next 2089 mm/memory.c next = p4d_addr_end(addr, end); next 2090 mm/memory.c err = apply_to_pud_range(mm, p4d, addr, next, fn, data); next 2093 mm/memory.c } while (p4d++, addr = next, addr != end); next 2105 mm/memory.c unsigned long next; next 2114 mm/memory.c next = pgd_addr_end(addr, end); next 2115 mm/memory.c err = apply_to_p4d_range(mm, pgd, addr, next, fn, data); next 2118 mm/memory.c } while (pgd++, addr = next, addr != end); next 734 mm/mempolicy.c struct vm_area_struct *next; next 750 mm/mempolicy.c for (; vma && vma->vm_start < end; prev = vma, vma = next) { next 751 mm/mempolicy.c next = vma->vm_next; next 765 mm/mempolicy.c next = vma->vm_next; next 1828 mm/mempolicy.c unsigned next; next 1831 mm/mempolicy.c next = next_node_in(me->il_prev, policy->v.nodes); next 1832 mm/mempolicy.c if (next < MAX_NUMNODES) next 1833 mm/mempolicy.c me->il_prev = next; next 1834 mm/mempolicy.c return next; next 2518 mm/mempolicy.c struct rb_node *next = rb_next(&n->nd); next 2541 mm/mempolicy.c if (!next) next 2543 mm/mempolicy.c n = rb_entry(next, struct sp_node, nd); next 2647 mm/mempolicy.c struct rb_node *next; next 2652 mm/mempolicy.c next = rb_first(&p->root); next 2653 mm/mempolicy.c while (next) { next 2654 mm/mempolicy.c n = rb_entry(next, struct sp_node, nd); next 2655 mm/mempolicy.c next = rb_next(&n->nd); next 2257 mm/migrate.c goto next; next 2270 mm/migrate.c goto next; next 2282 mm/migrate.c goto next; next 2292 mm/migrate.c goto next; next 2338 mm/migrate.c next: next 500 mm/mlock.c goto next; next 505 mm/mlock.c next: next 173 mm/mmap.c struct vm_area_struct *next = vma->vm_next; next 182 mm/mmap.c return next; next 192 mm/mmap.c struct vm_area_struct *next; next 260 mm/mmap.c next = find_vma(mm, oldbrk); next 261 mm/mmap.c if (next && newbrk + PAGE_SIZE > vm_start_gap(next)) next 683 mm/mmap.c struct vm_area_struct *next; next 686 mm/mmap.c next = vma->vm_next; next 688 mm/mmap.c prev->vm_next = next; next 692 mm/mmap.c prev->vm_next = next; next 694 mm/mmap.c mm->mmap = next; next 696 mm/mmap.c if (next) next 697 mm/mmap.c next->vm_prev = prev; next 722 mm/mmap.c struct vm_area_struct *next = vma->vm_next, *orig_vma = vma; next 731 mm/mmap.c if (next && !insert) { next 734 mm/mmap.c if (end >= next->vm_end) { next 741 mm/mmap.c if (next == expand) { next 746 mm/mmap.c VM_WARN_ON(end != next->vm_end); next 753 mm/mmap.c VM_WARN_ON(file != next->vm_file); next 754 mm/mmap.c swap(vma, next); next 761 mm/mmap.c remove_next = 1 + (end > next->vm_end); next 763 mm/mmap.c end != next->vm_next->vm_end); next 765 mm/mmap.c end != next->vm_end); next 767 mm/mmap.c end = next->vm_end; next 770 mm/mmap.c exporter = next; next 777 mm/mmap.c if (remove_next == 2 && !next->anon_vma) next 778 mm/mmap.c exporter = next->vm_next; next 780 mm/mmap.c } else if (end > next->vm_start) { next 785 mm/mmap.c adjust_next = (end - next->vm_start) >> PAGE_SHIFT; next 786 mm/mmap.c exporter = next; next 797 mm/mmap.c importer = next; next 824 mm/mmap.c uprobe_munmap(next, next->vm_start, next->vm_end); next 840 mm/mmap.c anon_vma = next->anon_vma; next 842 mm/mmap.c VM_WARN_ON(adjust_next && next->anon_vma && next 843 mm/mmap.c anon_vma != next->anon_vma); next 847 mm/mmap.c anon_vma_interval_tree_pre_update_vma(next); next 854 mm/mmap.c vma_interval_tree_remove(next, root); next 867 mm/mmap.c next->vm_start += adjust_next << PAGE_SHIFT; next 868 mm/mmap.c next->vm_pgoff += adjust_next; next 873 mm/mmap.c vma_interval_tree_insert(next, root); next 884 mm/mmap.c __vma_unlink_prev(mm, next, vma); next 895 mm/mmap.c __vma_unlink_common(mm, next, NULL, false, vma); next 897 mm/mmap.c __remove_shared_vm_struct(next, file, mapping); next 909 mm/mmap.c if (!next) next 912 mm/mmap.c vma_gap_update(next); next 919 mm/mmap.c anon_vma_interval_tree_post_update_vma(next); next 929 mm/mmap.c uprobe_mmap(next); next 934 mm/mmap.c uprobe_munmap(next, next->vm_start, next->vm_end); next 937 mm/mmap.c if (next->anon_vma) next 938 mm/mmap.c anon_vma_merge(vma, next); next 940 mm/mmap.c mpol_put(vma_policy(next)); next 941 mm/mmap.c vm_area_free(next); next 954 mm/mmap.c next = vma->vm_next; next 966 mm/mmap.c next = vma; next 970 mm/mmap.c end = next->vm_end; next 973 mm/mmap.c else if (next) next 974 mm/mmap.c vma_gap_update(next); next 1143 mm/mmap.c struct vm_area_struct *area, *next; next 1154 mm/mmap.c next = prev->vm_next; next 1156 mm/mmap.c next = mm->mmap; next 1157 mm/mmap.c area = next; next 1159 mm/mmap.c next = next->vm_next; next 1177 mm/mmap.c if (next && end == next->vm_start && next 1178 mm/mmap.c mpol_equal(policy, vma_policy(next)) && next 1179 mm/mmap.c can_vma_merge_before(next, vm_flags, next 1184 mm/mmap.c next->anon_vma, NULL)) { next 1187 mm/mmap.c next->vm_end, prev->vm_pgoff, NULL, next 1201 mm/mmap.c if (next && end == next->vm_start && next 1202 mm/mmap.c mpol_equal(policy, vma_policy(next)) && next 1203 mm/mmap.c can_vma_merge_before(next, vm_flags, next 1208 mm/mmap.c addr, prev->vm_pgoff, NULL, next); next 1210 mm/mmap.c err = __vma_adjust(area, addr, next->vm_end, next 1211 mm/mmap.c next->vm_pgoff - pglen, NULL, next); next 1217 mm/mmap.c area = next; next 2342 mm/mmap.c struct vm_area_struct *next; next 2362 mm/mmap.c next = vma->vm_next; next 2363 mm/mmap.c if (next && next->vm_start < gap_addr && next 2364 mm/mmap.c (next->vm_flags & (VM_WRITE|VM_READ|VM_EXEC))) { next 2365 mm/mmap.c if (!(next->vm_flags & VM_GROWSUP)) next 2609 mm/mmap.c struct vm_area_struct *next = prev ? prev->vm_next : mm->mmap; next 2617 mm/mmap.c next ? next->vm_start : USER_PGTABLES_CEILING); next 2928 mm/mmap.c struct vm_area_struct *next; next 2930 mm/mmap.c for (next = vma->vm_next; next; next = next->vm_next) { next 2932 mm/mmap.c if (next->vm_start != next->vm_prev->vm_end) next 2935 mm/mmap.c if (next->vm_file != vma->vm_file) next 2938 mm/mmap.c if (next->vm_flags != vma->vm_flags) next 2941 mm/mmap.c if (start + size <= next->vm_end) next 2945 mm/mmap.c if (!next) next 21 mm/mmu_gather.c if (batch->next) { next 22 mm/mmu_gather.c tlb->active = batch->next; next 34 mm/mmu_gather.c batch->next = NULL; next 38 mm/mmu_gather.c tlb->active->next = batch; next 48 mm/mmu_gather.c for (batch = &tlb->local; batch && batch->nr; batch = batch->next) { next 57 mm/mmu_gather.c struct mmu_gather_batch *batch, *next; next 59 mm/mmu_gather.c for (batch = tlb->local.next; batch; batch = next) { next 60 mm/mmu_gather.c next = batch->next; next 63 mm/mmu_gather.c tlb->local.next = NULL; next 216 mm/mmu_gather.c tlb->local.next = NULL; next 194 mm/mprotect.c unsigned long next; next 205 mm/mprotect.c next = pmd_addr_end(addr, end); next 217 mm/mprotect.c goto next; next 228 mm/mprotect.c if (next - addr != HPAGE_PMD_SIZE) { next 241 mm/mprotect.c goto next; next 246 mm/mprotect.c this_pages = change_pte_range(vma, pmd, addr, next, newprot, next 249 mm/mprotect.c next: next 251 mm/mprotect.c } while (pmd++, addr = next, addr != end); next 266 mm/mprotect.c unsigned long next; next 271 mm/mprotect.c next = pud_addr_end(addr, end); next 274 mm/mprotect.c pages += change_pmd_range(vma, pud, addr, next, newprot, next 276 mm/mprotect.c } while (pud++, addr = next, addr != end); next 286 mm/mprotect.c unsigned long next; next 291 mm/mprotect.c next = p4d_addr_end(addr, end); next 294 mm/mprotect.c pages += change_pud_range(vma, p4d, addr, next, newprot, next 296 mm/mprotect.c } while (p4d++, addr = next, addr != end); next 307 mm/mprotect.c unsigned long next; next 316 mm/mprotect.c next = pgd_addr_end(addr, end); next 319 mm/mprotect.c pages += change_p4d_range(vma, pgd, addr, next, newprot, next 321 mm/mprotect.c } while (pgd++, addr = next, addr != end); next 346 mm/mprotect.c unsigned long next, struct mm_walk *walk) next 353 mm/mprotect.c unsigned long addr, unsigned long next, next 360 mm/mprotect.c static int prot_none_test(unsigned long addr, unsigned long next, next 245 mm/mremap.c unsigned long extent, next, old_end; next 258 mm/mremap.c next = (old_addr + PMD_SIZE) & PMD_MASK; next 260 mm/mremap.c extent = next - old_addr; next 306 mm/mremap.c next = (new_addr + PMD_SIZE) & PMD_MASK; next 307 mm/mremap.c if (extent > next - new_addr) next 308 mm/mremap.c extent = next - new_addr; next 315 mm/oom_kill.c goto next; next 319 mm/oom_kill.c goto next; next 329 mm/oom_kill.c goto next; next 344 mm/oom_kill.c goto next; next 352 mm/oom_kill.c next: next 3082 mm/page_alloc.c struct page *page, *next; next 3087 mm/page_alloc.c list_for_each_entry_safe(page, next, list, lru) { next 3095 mm/page_alloc.c list_for_each_entry_safe(page, next, list, lru) { next 6952 mm/page_alloc.c phys_addr_t next = 0; next 6960 mm/page_alloc.c if (next < start) next 6961 mm/page_alloc.c pgcnt += zero_pfn_range(PFN_DOWN(next), PFN_UP(start)); next 6962 mm/page_alloc.c next = end; next 6972 mm/page_alloc.c pgcnt += zero_pfn_range(PFN_DOWN(next), next 33 mm/pagewalk.c unsigned long next; next 40 mm/pagewalk.c next = pmd_addr_end(addr, end); next 43 mm/pagewalk.c err = ops->pte_hole(addr, next, walk); next 53 mm/pagewalk.c err = ops->pmd_entry(pmd, addr, next, walk); next 67 mm/pagewalk.c err = walk_pte_range(pmd, addr, next, walk); next 70 mm/pagewalk.c } while (pmd++, addr = next, addr != end); next 79 mm/pagewalk.c unsigned long next; next 86 mm/pagewalk.c next = pud_addr_end(addr, end); next 89 mm/pagewalk.c err = ops->pte_hole(addr, next, walk); next 99 mm/pagewalk.c err = ops->pud_entry(pud, addr, next, walk); next 112 mm/pagewalk.c err = walk_pmd_range(pud, addr, next, walk); next 115 mm/pagewalk.c } while (pud++, addr = next, addr != end); next 124 mm/pagewalk.c unsigned long next; next 130 mm/pagewalk.c next = p4d_addr_end(addr, end); next 133 mm/pagewalk.c err = ops->pte_hole(addr, next, walk); next 139 mm/pagewalk.c err = walk_pud_range(p4d, addr, next, walk); next 142 mm/pagewalk.c } while (p4d++, addr = next, addr != end); next 151 mm/pagewalk.c unsigned long next; next 157 mm/pagewalk.c next = pgd_addr_end(addr, end); next 160 mm/pagewalk.c err = ops->pte_hole(addr, next, walk); next 166 mm/pagewalk.c err = walk_p4d_range(pgd, addr, next, walk); next 169 mm/pagewalk.c } while (pgd++, addr = next, addr != end); next 187 mm/pagewalk.c unsigned long next; next 195 mm/pagewalk.c next = hugetlb_entry_end(h, addr, end); next 199 mm/pagewalk.c err = ops->hugetlb_entry(pte, hmask, addr, next, walk); next 201 mm/pagewalk.c err = ops->pte_hole(addr, next, walk); next 205 mm/pagewalk.c } while (addr = next, addr != end); next 306 mm/pagewalk.c unsigned long next; next 326 mm/pagewalk.c next = end; next 329 mm/pagewalk.c next = min(end, vma->vm_start); next 332 mm/pagewalk.c next = min(end, vma->vm_end); next 335 mm/pagewalk.c err = walk_page_test(start, next, &walk); next 349 mm/pagewalk.c err = __walk_page_range(start, next, &walk); next 352 mm/pagewalk.c } while (start = next, start < end); next 1594 mm/percpu.c struct pcpu_chunk *chunk, *next; next 1656 mm/percpu.c list_for_each_entry_safe(chunk, next, &pcpu_slot[slot], list) { next 1838 mm/percpu.c struct pcpu_chunk *chunk, *next; next 1848 mm/percpu.c list_for_each_entry_safe(chunk, next, free_head, list) { next 1860 mm/percpu.c list_for_each_entry_safe(chunk, next, &to_free, list) { next 379 mm/rmap.c struct anon_vma_chain *avc, *next; next 386 mm/rmap.c list_for_each_entry_safe(avc, next, &vma->anon_vma_chain, same_vma) { next 413 mm/rmap.c list_for_each_entry_safe(avc, next, &vma->anon_vma_chain, same_vma) { next 106 mm/shmem.c pgoff_t next; /* the next page offset to be fallocated */ next 463 mm/shmem.c LIST_HEAD(list), *pos, *next; next 475 mm/shmem.c list_for_each_safe(pos, next, &sbinfo->shrinklist) { next 485 mm/shmem.c goto next; next 493 mm/shmem.c goto next; next 497 mm/shmem.c next: next 503 mm/shmem.c list_for_each_safe(pos, next, &to_remove) { next 510 mm/shmem.c list_for_each_safe(pos, next, &list) { next 633 mm/shmem.c next: next 637 mm/shmem.c goto next; next 1251 mm/shmem.c struct shmem_inode_info *info, *next; next 1258 mm/shmem.c list_for_each_entry_safe(info, next, &shmem_swaplist, swaplist) { next 1277 mm/shmem.c next = list_next_entry(info, swaplist); next 1343 mm/shmem.c index < shmem_falloc->next) next 2024 mm/shmem.c vmf->pgoff < shmem_falloc->next) { next 2754 mm/shmem.c shmem_falloc.next = (unmap_end + 1) >> PAGE_SHIFT; next 2794 mm/shmem.c shmem_falloc.next = start; next 2828 mm/shmem.c shmem_falloc.next++; next 2754 mm/slab.c void *next = *list; next 2757 mm/slab.c while (next) { next 2758 mm/slab.c objp = next - obj_offset(cachep); next 2759 mm/slab.c next = *(void **)next; next 3780 mm/slab.c if (!cachep->list.next) { next 4009 mm/slab.c goto next; next 4024 mm/slab.c next: next 1477 mm/slab_common.c if (p == slab_root_caches.next) next 1544 mm/slab_common.c if (p == memcg->kmem_caches.next) next 1566 mm/slab_common.c .next = slab_next, next 146 mm/slob.c static void set_slob(slob_t *s, slobidx_t size, slob_t *next) next 149 mm/slob.c slobidx_t offset = next - base; next 174 mm/slob.c slobidx_t next; next 177 mm/slob.c next = -s[0].units; next 179 mm/slob.c next = s[1].units; next 180 mm/slob.c return base+next; next 261 mm/slob.c slob_t *next; next 264 mm/slob.c next = slob_next(cur); next 265 mm/slob.c set_slob(aligned, avail - delta, next); next 272 mm/slob.c next = slob_next(cur); next 275 mm/slob.c set_slob(prev, slob_units(prev), next); next 277 mm/slob.c sp->freelist = next; next 283 mm/slob.c set_slob(cur + units, avail - units, next); next 385 mm/slob.c slob_t *prev, *next, *b = (slob_t *)block; next 442 mm/slob.c next = slob_next(prev); next 443 mm/slob.c while (b > next) { next 444 mm/slob.c prev = next; next 445 mm/slob.c next = slob_next(prev); next 448 mm/slob.c if (!slob_last(prev) && b + units == next) { next 449 mm/slob.c units += slob_units(next); next 450 mm/slob.c set_slob(b, units, slob_next(next)); next 452 mm/slob.c set_slob(b, units, next); next 1432 mm/slub.c void *next = *head; next 1441 mm/slub.c object = next; next 1442 mm/slub.c next = get_freepointer(s, object); next 1575 mm/slub.c void *next; next 1594 mm/slub.c next = next_freelist_entry(s, page, &pos, start, page_limit, next 1596 mm/slub.c next = setup_object(s, page, next); next 1597 mm/slub.c set_freepointer(s, cur, next); next 1598 mm/slub.c cur = next; next 1621 mm/slub.c void *start, *p, *next; next 1674 mm/slub.c next = p + s->size; next 1675 mm/slub.c next = setup_object(s, page, next); next 1676 mm/slub.c set_freepointer(s, p, next); next 1677 mm/slub.c p = next; next 2192 mm/slub.c c->partial = page->next; next 2220 mm/slub.c page->next = discard_page; next 2233 mm/slub.c discard_page = discard_page->next; next 2286 mm/slub.c page->next = oldpage; next 5844 mm/slub.c struct saved_alias *next; next 5867 mm/slub.c al->next = alias_list; next 5898 mm/slub.c alias_list = alias_list->next; next 612 mm/swapfile.c cluster->next = cluster_next(&cluster->index) * next 632 mm/swapfile.c tmp = cluster->next; next 652 mm/swapfile.c cluster->next = tmp + 1; next 995 mm/swapfile.c struct swap_info_struct *si, *next; next 1019 mm/swapfile.c plist_for_each_entry_safe(si, next, &swap_avail_heads[node], avail_lists[node]) { next 1065 mm/swapfile.c if (plist_node_empty(&next->avail_lists[node])) next 1983 mm/swapfile.c unsigned long next; next 1989 mm/swapfile.c next = pmd_addr_end(addr, end); next 1992 mm/swapfile.c ret = unuse_pte_range(vma, pmd, addr, next, type, next 1996 mm/swapfile.c } while (pmd++, addr = next, addr != end); next 2006 mm/swapfile.c unsigned long next; next 2011 mm/swapfile.c next = pud_addr_end(addr, end); next 2014 mm/swapfile.c ret = unuse_pmd_range(vma, pud, addr, next, type, next 2018 mm/swapfile.c } while (pud++, addr = next, addr != end); next 2028 mm/swapfile.c unsigned long next; next 2033 mm/swapfile.c next = p4d_addr_end(addr, end); next 2036 mm/swapfile.c ret = unuse_pud_range(vma, p4d, addr, next, type, next 2040 mm/swapfile.c } while (p4d++, addr = next, addr != end); next 2048 mm/swapfile.c unsigned long addr, end, next; next 2056 mm/swapfile.c next = pgd_addr_end(addr, end); next 2059 mm/swapfile.c ret = unuse_p4d_range(vma, pgd, addr, next, type, next 2063 mm/swapfile.c } while (pgd++, addr = next, addr != end); next 2153 mm/swapfile.c (p = p->next) != &init_mm.mmlist) { next 2239 mm/swapfile.c struct list_head *p, *next; next 2246 mm/swapfile.c list_for_each_safe(p, next, &init_mm.mmlist) next 2780 mm/swapfile.c .next = swap_next, next 3649 mm/swapfile.c page = list_entry(head->lru.next, struct page, lru); next 3661 mm/swapfile.c page = list_entry(page->lru.next, struct page, lru); next 3667 mm/swapfile.c page = list_entry(page->lru.next, struct page, lru); next 3693 mm/swapfile.c page = list_entry(page->lru.next, struct page, lru); next 3729 mm/swapfile.c struct page *page, *next; next 3731 mm/swapfile.c list_for_each_entry_safe(page, next, &head->lru, lru) { next 3743 mm/swapfile.c struct swap_info_struct *si, *next; next 3758 mm/swapfile.c plist_for_each_entry_safe(si, next, &swap_avail_heads[node], next 276 mm/util.c struct vm_area_struct *next; next 280 mm/util.c next = prev->vm_next; next 285 mm/util.c next = rb_entry(rb_parent, next 288 mm/util.c next = NULL; next 290 mm/util.c vma->vm_next = next; next 291 mm/util.c if (next) next 292 mm/util.c next->vm_prev = vma; next 78 mm/vmalloc.c unsigned long next; next 82 mm/vmalloc.c next = pmd_addr_end(addr, end); next 87 mm/vmalloc.c vunmap_pte_range(pmd, addr, next); next 88 mm/vmalloc.c } while (pmd++, addr = next, addr != end); next 94 mm/vmalloc.c unsigned long next; next 98 mm/vmalloc.c next = pud_addr_end(addr, end); next 103 mm/vmalloc.c vunmap_pmd_range(pud, addr, next); next 104 mm/vmalloc.c } while (pud++, addr = next, addr != end); next 110 mm/vmalloc.c unsigned long next; next 114 mm/vmalloc.c next = p4d_addr_end(addr, end); next 119 mm/vmalloc.c vunmap_pud_range(p4d, addr, next); next 120 mm/vmalloc.c } while (p4d++, addr = next, addr != end); next 126 mm/vmalloc.c unsigned long next; next 131 mm/vmalloc.c next = pgd_addr_end(addr, end); next 134 mm/vmalloc.c vunmap_p4d_range(pgd, addr, next); next 135 mm/vmalloc.c } while (pgd++, addr = next, addr != end); next 168 mm/vmalloc.c unsigned long next; next 174 mm/vmalloc.c next = pmd_addr_end(addr, end); next 175 mm/vmalloc.c if (vmap_pte_range(pmd, addr, next, prot, pages, nr)) next 177 mm/vmalloc.c } while (pmd++, addr = next, addr != end); next 185 mm/vmalloc.c unsigned long next; next 191 mm/vmalloc.c next = pud_addr_end(addr, end); next 192 mm/vmalloc.c if (vmap_pmd_range(pud, addr, next, prot, pages, nr)) next 194 mm/vmalloc.c } while (pud++, addr = next, addr != end); next 202 mm/vmalloc.c unsigned long next; next 208 mm/vmalloc.c next = p4d_addr_end(addr, end); next 209 mm/vmalloc.c if (vmap_pud_range(p4d, addr, next, prot, pages, nr)) next 211 mm/vmalloc.c } while (p4d++, addr = next, addr != end); next 225 mm/vmalloc.c unsigned long next; next 233 mm/vmalloc.c next = pgd_addr_end(addr, end); next 234 mm/vmalloc.c err = vmap_p4d_range(pgd, addr, next, prot, pages, &nr); next 237 mm/vmalloc.c } while (pgd++, addr = next, addr != end); next 497 mm/vmalloc.c return (&parent->rb_right == link ? list->next : list); next 691 mm/vmalloc.c struct list_head *next; next 705 mm/vmalloc.c next = get_va_next_sibling(parent, link); next 706 mm/vmalloc.c if (unlikely(next == NULL)) next 716 mm/vmalloc.c if (next != head) { next 717 mm/vmalloc.c sibling = list_entry(next, struct vmap_area, list); next 740 mm/vmalloc.c if (next->prev != head) { next 741 mm/vmalloc.c sibling = list_entry(next->prev, struct vmap_area, list); next 1824 mm/vmalloc.c for (p = &vmlist; (tmp = *p) != NULL; p = &tmp->next) { next 1831 mm/vmalloc.c vm->next = *p; next 1925 mm/vmalloc.c for (tmp = vmlist; tmp; tmp = tmp->next) { next 3558 mm/vmalloc.c .next = s_next, next 1546 mm/vmscan.c struct page *page, *next; next 1549 mm/vmscan.c list_for_each_entry_safe(page, next, page_list, lru) { next 1533 mm/vmstat.c .next = frag_next, next 1540 mm/vmstat.c .next = frag_next, next 1656 mm/vmstat.c .next = frag_next, next 1742 mm/vmstat.c .next = vmstat_next, next 2065 mm/vmstat.c .next = frag_next, next 2117 mm/vmstat.c .next = frag_next, next 1201 mm/z3fold.c goto next; next 1206 mm/z3fold.c goto next; next 1211 mm/z3fold.c goto next; next 1213 mm/z3fold.c next: next 536 mm/zbud.c goto next; next 541 mm/zbud.c goto next; next 543 mm/zbud.c next: next 244 mm/zsmalloc.c unsigned long next; next 938 mm/zsmalloc.c struct page *page, *next; next 949 mm/zsmalloc.c next = page = get_first_page(zspage); next 952 mm/zsmalloc.c next = get_next_page(page); next 957 mm/zsmalloc.c page = next; next 1000 mm/zsmalloc.c link->next = freeobj++ << OBJ_TAG_BITS; next 1011 mm/zsmalloc.c link->next = freeobj++ << OBJ_TAG_BITS; next 1017 mm/zsmalloc.c link->next = -1UL << OBJ_TAG_BITS; next 1431 mm/zsmalloc.c set_freeobj(zspage, link->next >> OBJ_TAG_BITS); next 1533 mm/zsmalloc.c link->next = get_freeobj(zspage) << OBJ_TAG_BITS; next 382 net/802/garp.c struct rb_node *node, *next; next 386 net/802/garp.c next = node ? rb_next(node) : NULL, node != NULL; next 387 net/802/garp.c node = next) { next 567 net/802/mrp.c struct rb_node *node, *next; next 571 net/802/mrp.c next = node ? rb_next(node) : NULL, node != NULL; next 572 net/802/mrp.c node = next) { next 189 net/8021q/vlan_dev.c mp = mp->next; next 198 net/8021q/vlan_dev.c np->next = mp; next 598 net/8021q/vlan_dev.c vlan->egress_priority_map[i] = pm->next; next 256 net/8021q/vlan_netlink.c pm = pm->next) { next 68 net/8021q/vlanproc.c .next = vlan_seq_next, next 285 net/8021q/vlanproc.c mp = mp->next; next 458 net/9p/trans_fd.c req = list_entry(m->unsent_req_list.next, struct p9_req_t, next 860 net/9p/trans_fd.c m, m->mux_list.prev, m->mux_list.next); next 401 net/9p/trans_rdma.c wr.next = NULL; next 493 net/9p/trans_rdma.c wr.next = NULL; next 67 net/appletalk/aarp.c struct aarp_entry *next; next 260 net/appletalk/aarp.c *n = (*n)->next; next 263 net/appletalk/aarp.c n = &((*n)->next); next 279 net/appletalk/aarp.c *n = (*n)->next; next 283 net/appletalk/aarp.c n = &((*n)->next); next 300 net/appletalk/aarp.c *n = (*n)->next; next 303 net/appletalk/aarp.c n = &((*n)->next); next 357 net/appletalk/aarp.c *n = (*n)->next; next 404 net/appletalk/aarp.c list = list->next; next 513 net/appletalk/aarp.c entry->next = proxies[hash]; next 641 net/appletalk/aarp.c a->next = unresolved[hash]; next 693 net/appletalk/aarp.c *list = a->next; next 696 net/appletalk/aarp.c a->next = resolved[hash]; next 706 net/appletalk/aarp.c list = &((*list)->next); next 928 net/appletalk/aarp.c for (entry = table[ct]; entry; entry = entry->next) { next 975 net/appletalk/aarp.c else if (entry->next) next 976 net/appletalk/aarp.c entry = entry->next; next 1036 net/appletalk/aarp.c .next = aarp_seq_next, next 21 net/appletalk/atalk_proc.c for (i = atalk_interfaces; pos && i; i = i->next) next 48 net/appletalk/atalk_proc.c i = i->next; next 82 net/appletalk/atalk_proc.c for (r = atalk_routes; pos && r; r = r->next) next 109 net/appletalk/atalk_proc.c r = r->next; next 191 net/appletalk/atalk_proc.c .next = atalk_seq_interface_next, next 198 net/appletalk/atalk_proc.c .next = atalk_seq_route_next, next 205 net/appletalk/atalk_proc.c .next = atalk_seq_socket_next, next 209 net/appletalk/ddp.c *iface = tmp->next; next 214 net/appletalk/ddp.c iface = &tmp->next; next 234 net/appletalk/ddp.c iface->next = atalk_interfaces; next 351 net/appletalk/ddp.c for (iface = atalk_interfaces; iface; iface = iface->next) { next 399 net/appletalk/ddp.c for (iface = atalk_interfaces; iface; iface = iface->next) { next 434 net/appletalk/ddp.c for (r = atalk_routes; r; r = r->next) { next 515 net/appletalk/ddp.c for (rt = atalk_routes; rt; rt = rt->next) { next 531 net/appletalk/ddp.c for (iface = atalk_interfaces; iface; iface = iface->next) { next 559 net/appletalk/ddp.c rt->next = atalk_routes; next 589 net/appletalk/ddp.c *r = tmp->next; next 594 net/appletalk/ddp.c r = &tmp->next; next 614 net/appletalk/ddp.c *r = tmp->next; next 618 net/appletalk/ddp.c r = &tmp->next; next 288 net/atm/br2684.c return list_empty(&brdev->brvccs) ? NULL : list_entry_brvcc(brdev->brvccs.next); /* 1 vcc/dev right now */ next 376 net/atm/br2684.c brdev->brvccs.next != brdev->brvccs.prev) /* >1 VCC */ next 379 net/atm/br2684.c brvcc = list_entry_brvcc(brdev->brvccs.next); next 816 net/atm/br2684.c .next = br2684_seq_next, next 852 net/atm/br2684.c net_dev = list_entry_brdev(br2684_devs.next); next 855 net/atm/br2684.c brvcc = list_entry_brvcc(brdev->brvccs.next); next 81 net/atm/clip.c clip_vcc->next = entry->vccs; next 97 net/atm/clip.c for (walk = &entry->vccs; *walk; walk = &(*walk)->next) next 101 net/atm/clip.c *walk = clip_vcc->next; /* atomic */ next 128 net/atm/clip.c for (cv = entry->vccs; cv; cv = cv->next) { next 517 net/atm/clip.c for (dev = clip_devs; dev; dev = PRIV(dev)->next) next 522 net/atm/clip.c for (dev = clip_devs; dev; dev = PRIV(dev)->next) next 539 net/atm/clip.c clip_priv->next = clip_devs; next 801 net/atm/clip.c curr = curr->next; next 862 net/atm/clip.c .next = neigh_seq_next, next 897 net/atm/clip.c struct net_device *dev, *next; next 911 net/atm/clip.c next = PRIV(dev)->next; next 914 net/atm/clip.c dev = next; next 502 net/atm/common.c dev = list_entry(atm_devs.next, next 844 net/atm/lec.c for (; e; e = e->next) { next 981 net/atm/lec.c next); next 991 net/atm/lec.c .next = lec_seq_next, next 1288 net/atm/lec.c hlist_add_head(&entry->next, tmp); next 1305 net/atm/lec.c hlist_del(&to_remove->next); next 1318 net/atm/lec.c &priv->lec_arp_tables[i], next) { next 1362 net/atm/lec.c &priv->lec_arp_tables[i], next) { next 1393 net/atm/lec.c hlist_for_each_entry(rulla, &priv->lec_no_forward, next) { next 1418 net/atm/lec.c hlist_for_each_entry(rulla, &priv->lec_arp_empty_ones, next) { next 1443 net/atm/lec.c hlist_for_each_entry(rulla, &priv->mcast_fwds, next) { next 1477 net/atm/lec.c struct hlist_node *next; next 1489 net/atm/lec.c hlist_for_each_entry_safe(entry, next, next 1490 net/atm/lec.c &priv->lec_arp_tables[i], next) { next 1497 net/atm/lec.c hlist_for_each_entry_safe(entry, next, next 1498 net/atm/lec.c &priv->lec_arp_empty_ones, next) { next 1501 net/atm/lec.c hlist_del(&entry->next); next 1506 net/atm/lec.c hlist_for_each_entry_safe(entry, next, next 1507 net/atm/lec.c &priv->lec_no_forward, next) { next 1510 net/atm/lec.c hlist_del(&entry->next); next 1515 net/atm/lec.c hlist_for_each_entry_safe(entry, next, &priv->mcast_fwds, next) { next 1518 net/atm/lec.c hlist_del(&entry->next); next 1538 net/atm/lec.c hlist_for_each_entry(entry, head, next) { next 1556 net/atm/lec.c INIT_HLIST_NODE(&to_return->next); next 1602 net/atm/lec.c hlist_del(&to_remove->next); next 1674 net/atm/lec.c struct hlist_node *next; next 1684 net/atm/lec.c hlist_for_each_entry_safe(entry, next, next 1685 net/atm/lec.c &priv->lec_arp_tables[i], next) { next 1811 net/atm/lec.c struct hlist_node *next; next 1818 net/atm/lec.c hlist_for_each_entry_safe(entry, next, next 1819 net/atm/lec.c &priv->lec_arp_tables[i], next) { next 1843 net/atm/lec.c struct hlist_node *next; next 1858 net/atm/lec.c hlist_for_each_entry_safe(entry, next, next 1859 net/atm/lec.c &priv->lec_arp_empty_ones, next) { next 1861 net/atm/lec.c hlist_del(&entry->next); next 1905 net/atm/lec.c &priv->lec_arp_tables[i], next) { next 1969 net/atm/lec.c hlist_add_head(&entry->next, &priv->mcast_fwds); next 1997 net/atm/lec.c hlist_add_head(&entry->next, &priv->lec_no_forward); next 2015 net/atm/lec.c &priv->lec_arp_tables[i], next) { next 2078 net/atm/lec.c hlist_add_head(&entry->next, &priv->lec_arp_empty_ones); next 2099 net/atm/lec.c &priv->lec_arp_tables[i], next) { next 2133 net/atm/lec.c &priv->lec_arp_tables[i], next) { next 2184 net/atm/lec.c struct hlist_node *next; next 2194 net/atm/lec.c hlist_for_each_entry_safe(entry, next, next 2195 net/atm/lec.c &priv->lec_arp_tables[i], next) { next 2205 net/atm/lec.c hlist_for_each_entry_safe(entry, next, next 2206 net/atm/lec.c &priv->lec_arp_empty_ones, next) { next 2210 net/atm/lec.c hlist_del(&entry->next); next 2215 net/atm/lec.c hlist_for_each_entry_safe(entry, next, next 2216 net/atm/lec.c &priv->lec_no_forward, next) { next 2220 net/atm/lec.c hlist_del(&entry->next); next 2225 net/atm/lec.c hlist_for_each_entry_safe(entry, next, &priv->mcast_fwds, next) { next 2229 net/atm/lec.c hlist_del(&entry->next); next 2243 net/atm/lec.c struct hlist_node *next; next 2249 net/atm/lec.c hlist_for_each_entry_safe(entry, next, next 2250 net/atm/lec.c &priv->lec_arp_empty_ones, next) { next 2262 net/atm/lec.c hlist_del(&entry->next); next 15 net/atm/lec_arpc.h struct hlist_node next; /* Linked entry list */ next 136 net/atm/mpc.c mpc = mpc->next; next 150 net/atm/mpc.c mpc = mpc->next; next 164 net/atm/mpc.c mpc = mpc->next; next 196 net/atm/mpc.c entry->next = qos_head; next 210 net/atm/mpc.c qos = qos->next; next 226 net/atm/mpc.c qos_head = qos_head->next; next 233 net/atm/mpc.c if (curr->next == entry) { next 234 net/atm/mpc.c curr->next = entry->next; next 238 net/atm/mpc.c curr = curr->next; next 266 net/atm/mpc.c qos = qos->next; next 290 net/atm/mpc.c mpc->next = mpcs; next 1300 net/atm/mpc.c entry = entry->next; next 1403 net/atm/mpc.c entry = entry->next; next 1439 net/atm/mpc.c mpc = mpc->next; next 1503 net/atm/mpc.c tmp = mpc->next; next 1525 net/atm/mpc.c nextqos = qos->next; next 16 net/atm/mpc.h struct mpoa_client *next; next 42 net/atm/mpc.h struct atm_mpoa_qos *next; next 48 net/atm/mpoa_caches.c entry = entry->next; next 69 net/atm/mpoa_caches.c entry = entry->next; next 90 net/atm/mpoa_caches.c entry = entry->next; next 112 net/atm/mpoa_caches.c entry->next = client->in_cache; next 201 net/atm/mpoa_caches.c entry->prev->next = entry->next; next 203 net/atm/mpoa_caches.c client->in_cache = entry->next; next 204 net/atm/mpoa_caches.c if (entry->next != NULL) next 205 net/atm/mpoa_caches.c entry->next->prev = entry->prev; next 237 net/atm/mpoa_caches.c next_entry = entry->next; next 266 net/atm/mpoa_caches.c entry = entry->next; /* Entry in hold down */ next 278 net/atm/mpoa_caches.c entry = entry->next; next 293 net/atm/mpoa_caches.c entry = entry->next; next 319 net/atm/mpoa_caches.c entry = entry->next; next 345 net/atm/mpoa_caches.c entry = entry->next; next 366 net/atm/mpoa_caches.c entry = entry->next; next 388 net/atm/mpoa_caches.c entry = entry->next; next 408 net/atm/mpoa_caches.c entry = entry->next; next 434 net/atm/mpoa_caches.c entry->prev->next = entry->next; next 436 net/atm/mpoa_caches.c client->eg_cache = entry->next; next 437 net/atm/mpoa_caches.c if (entry->next != NULL) next 438 net/atm/mpoa_caches.c entry->next->prev = entry->prev; next 472 net/atm/mpoa_caches.c entry->next = client->eg_cache; next 511 net/atm/mpoa_caches.c next_entry = entry->next; next 18 net/atm/mpoa_caches.h struct in_cache_entry *next; next 55 net/atm/mpoa_caches.h struct eg_cache_entry *next; next 114 net/atm/mpoa_proc.c for (mpc = mpcs; mpc; mpc = mpc->next) next 124 net/atm/mpoa_proc.c return v == SEQ_START_TOKEN ? mpcs : p->next; next 152 net/atm/mpoa_proc.c for (in_entry = mpc->in_cache; in_entry; in_entry = in_entry->next) { next 171 net/atm/mpoa_proc.c for (eg_entry = mpc->eg_cache; eg_entry; eg_entry = eg_entry->next) { next 199 net/atm/mpoa_proc.c .next = mpc_next, next 243 net/atm/proc.c .next = atm_dev_seq_next, next 266 net/atm/proc.c .next = vcc_seq_next, next 288 net/atm/proc.c .next = vcc_seq_next, next 311 net/atm/proc.c .next = vcc_seq_next, next 1923 net/ax25/af_ax25.c .next = ax25_info_next, next 37 net/ax25/ax25_dev.c for (ax25_dev = ax25_dev_list; ax25_dev != NULL; ax25_dev = ax25_dev->next) next 84 net/ax25/ax25_dev.c ax25_dev->next = ax25_dev_list; next 109 net/ax25/ax25_dev.c for (s = ax25_dev_list; s != NULL; s = s->next) next 114 net/ax25/ax25_dev.c ax25_dev_list = s->next; next 122 net/ax25/ax25_dev.c while (s != NULL && s->next != NULL) { next 123 net/ax25/ax25_dev.c if (s->next == ax25_dev) { next 124 net/ax25/ax25_dev.c s->next = ax25_dev->next; next 132 net/ax25/ax25_dev.c s = s->next; next 192 net/ax25/ax25_dev.c ax25_dev = ax25_dev->next; next 35 net/ax25/ax25_iface.c struct listen_struct *next; next 48 net/ax25/ax25_iface.c ap->next = protocol_list; next 65 net/ax25/ax25_iface.c protocol_list = protocol->next; next 69 net/ax25/ax25_iface.c while (protocol != NULL && protocol->next != NULL) { next 70 net/ax25/ax25_iface.c if (protocol->next->pid == pid) { next 71 net/ax25/ax25_iface.c protocol->next = protocol->next->next; next 75 net/ax25/ax25_iface.c protocol = protocol->next; next 115 net/ax25/ax25_iface.c listen->next = listen_list; next 136 net/ax25/ax25_iface.c listen_list = listen->next; next 142 net/ax25/ax25_iface.c while (listen != NULL && listen->next != NULL) { next 143 net/ax25/ax25_iface.c if (ax25cmp(&listen->next->callsign, callsign) == 0 && listen->next->dev == dev) { next 144 net/ax25/ax25_iface.c s = listen->next; next 145 net/ax25/ax25_iface.c listen->next = listen->next->next; next 151 net/ax25/ax25_iface.c listen = listen->next; next 164 net/ax25/ax25_iface.c for (protocol = protocol_list; protocol != NULL; protocol = protocol->next) next 179 net/ax25/ax25_iface.c for (listen = listen_list; listen != NULL; listen = listen->next) next 206 net/ax25/ax25_iface.c for (protocol = protocol_list; protocol != NULL; protocol = protocol->next) next 245 net/ax25/ax25_out.c unsigned short start, end, next; next 287 net/ax25/ax25_out.c next = (ax25->vs + 1) % ax25->modulus; next 288 net/ax25/ax25_out.c last = (next == end); next 308 net/ax25/ax25_out.c ax25->vs = next; next 50 net/ax25/ax25_route.c ax25_rt = ax25_rt->next; next 54 net/ax25/ax25_route.c ax25_route_list = s->next; next 58 net/ax25/ax25_route.c for (t = ax25_route_list; t != NULL; t = t->next) { next 59 net/ax25/ax25_route.c if (t->next == s) { next 60 net/ax25/ax25_route.c t->next = s->next; next 106 net/ax25/ax25_route.c ax25_rt = ax25_rt->next; next 132 net/ax25/ax25_route.c ax25_rt->next = ax25_route_list; next 158 net/ax25/ax25_route.c ax25_rt = ax25_rt->next; next 162 net/ax25/ax25_route.c ax25_route_list = s->next; next 165 net/ax25/ax25_route.c for (t = ax25_route_list; t != NULL; t = t->next) { next 166 net/ax25/ax25_route.c if (t->next == s) { next 167 net/ax25/ax25_route.c t->next = s->next; next 213 net/ax25/ax25_route.c ax25_rt = ax25_rt->next; next 259 net/ax25/ax25_route.c for (ax25_rt = ax25_route_list; ax25_rt != NULL; ax25_rt = ax25_rt->next) { next 272 net/ax25/ax25_route.c ((struct ax25_route *) v)->next; next 325 net/ax25/ax25_route.c .next = ax25_rt_seq_next, next 347 net/ax25/ax25_route.c for (ax25_rt = ax25_route_list; ax25_rt != NULL; ax25_rt = ax25_rt->next) { next 481 net/ax25/ax25_route.c ax25_rt = ax25_rt->next; next 183 net/ax25/ax25_uid.c .next = ax25_uid_seq_next, next 1847 net/batman-adv/bat_iv_ogm.c goto next; next 1850 net/batman-adv/bat_iv_ogm.c goto next; next 1868 net/batman-adv/bat_iv_ogm.c next: next 2431 net/batman-adv/bat_iv_ogm.c goto next; next 2434 net/batman-adv/bat_iv_ogm.c goto next; next 2479 net/batman-adv/bat_iv_ogm.c next: next 377 net/batman-adv/bat_v.c goto next; next 396 net/batman-adv/bat_v.c next: next 756 net/batman-adv/bat_v.c goto next; next 759 net/batman-adv/bat_v.c goto next; next 768 net/batman-adv/bat_v.c next: next 458 net/batman-adv/multicast.c for (pmc6 = in6_dev->mc_list; pmc6; pmc6 = pmc6->next) { next 588 net/batman-adv/routing.c goto next; next 592 net/batman-adv/routing.c goto next; next 601 net/batman-adv/routing.c goto next; next 605 net/batman-adv/routing.c goto next; next 627 net/batman-adv/routing.c next: next 673 net/bluetooth/af_bluetooth.c .next = bt_seq_next, next 3768 net/bluetooth/hci_core.c skb = list; list = list->next; next 2192 net/bluetooth/l2cap_core.c frag = &(*frag)->next; next 2978 net/bluetooth/l2cap_core.c frag = &(*frag)->next; next 5894 net/bluetooth/l2cap_core.c new_frag->next = NULL; next 5896 net/bluetooth/l2cap_core.c (*last_frag)->next = new_frag; next 7367 net/bluetooth/l2cap_core.c c = list_entry(chan_list.next, typeof(*c), global_l); next 7423 net/bluetooth/l2cap_core.c struct l2cap_chan *chan, *next; next 7427 net/bluetooth/l2cap_core.c goto next; next 7441 net/bluetooth/l2cap_core.c next: next 7442 net/bluetooth/l2cap_core.c next = l2cap_global_fixed_chan(pchan, hcon); next 7444 net/bluetooth/l2cap_core.c pchan = next; next 304 net/bridge/br_forward.c p = rcu_dereference(p->next); next 164 net/bridge/br_mdb.c pp = &p->next) { next 331 net/bridge/br_mdb.c pp = &p->next) { next 634 net/bridge/br_mdb.c pp = &p->next) { next 760 net/bridge/br_mdb.c pp = &p->next) { next 768 net/bridge/br_mdb.c rcu_assign_pointer(*pp, p->next); next 179 net/bridge/br_multicast.c pp = &p->next) { next 183 net/bridge/br_multicast.c rcu_assign_pointer(*pp, p->next); next 476 net/bridge/br_multicast.c struct net_bridge_port_group __rcu *next, next 489 net/bridge/br_multicast.c rcu_assign_pointer(p->next, next); next 563 net/bridge/br_multicast.c pp = &p->next) { next 1280 net/bridge/br_multicast.c pp = &p->next) { next 1363 net/bridge/br_multicast.c pp = &p->next) { next 1403 net/bridge/br_multicast.c pp = &p->next) { next 1410 net/bridge/br_multicast.c rcu_assign_pointer(*pp, p->next); next 1437 net/bridge/br_multicast.c p = mlock_dereference(p->next, br)) { next 1469 net/bridge/br_multicast.c p = mlock_dereference(p->next, br)) { next 206 net/bridge/br_private.h struct net_bridge_port_group __rcu *next; next 688 net/bridge/br_private.h struct net_bridge_port_group __rcu *next, next 1464 net/can/bcm.c struct bcm_op *op, *next; next 1478 net/can/bcm.c list_for_each_entry_safe(op, next, &bo->tx_ops, list) next 1481 net/can/bcm.c list_for_each_entry_safe(op, next, &bo->rx_ops, list) { next 152 net/core/datagram.c skb->prev->next = nskb; next 153 net/core/datagram.c skb->next->prev = nskb; next 155 net/core/datagram.c nskb->next = skb->next; next 364 net/core/datagram.c if (skb->next) { next 2729 net/core/dev.c skb->next = __this_cpu_read(softnet_data.completion_queue); next 3213 net/core/dev.c struct sk_buff *next = skb->next; next 3216 net/core/dev.c rc = xmit_one(skb, dev, txq, next != NULL); next 3218 net/core/dev.c skb->next = next; next 3222 net/core/dev.c skb = next; next 3311 net/core/dev.c struct sk_buff *next, *head = NULL, *tail; next 3313 net/core/dev.c for (; skb != NULL; skb = next) { next 3314 net/core/dev.c next = skb->next; next 3327 net/core/dev.c tail->next = skb; next 4494 net/core/dev.c clist = clist->next; next 4959 net/core/dev.c struct sk_buff *skb, *next; next 4969 net/core/dev.c list_for_each_entry_safe(skb, next, head, list) { next 4991 net/core/dev.c struct sk_buff *skb, *next; next 4994 net/core/dev.c list_for_each_entry_safe(skb, next, head, list) { next 5045 net/core/dev.c struct sk_buff *skb, *next; next 5048 net/core/dev.c list_for_each_entry_safe(skb, next, head, list) { next 5134 net/core/dev.c struct sk_buff *skb, *next; next 5138 net/core/dev.c list_for_each_entry_safe(skb, next, head, list) { next 5149 net/core/dev.c list_for_each_entry_safe(skb, next, head, list) { next 5810 net/core/dev.c struct softnet_data *next = remsd->rps_ipi_next; next 5814 net/core/dev.c remsd = next; next 6570 net/core/dev.c upper = list_entry_rcu((*iter)->next, struct netdev_adjacent, list); next 6587 net/core/dev.c upper = list_entry((*iter)->next, struct netdev_adjacent, list); next 6605 net/core/dev.c upper = list_entry_rcu((*iter)->next, struct netdev_adjacent, list); next 6620 net/core/dev.c struct net_device *udev, *next, *now, *dev_stack[MAX_NEST_DEV + 1]; next 6635 net/core/dev.c next = NULL; next 6643 net/core/dev.c next = udev; next 6650 net/core/dev.c if (!next) { next 6653 net/core/dev.c next = dev_stack[--cur]; next 6657 net/core/dev.c now = next; next 6669 net/core/dev.c struct net_device *udev, *next, *now, *dev_stack[MAX_NEST_DEV + 1]; next 6683 net/core/dev.c next = NULL; next 6689 net/core/dev.c next = udev; next 6696 net/core/dev.c if (!next) { next 6699 net/core/dev.c next = dev_stack[--cur]; next 6703 net/core/dev.c now = next; next 6741 net/core/dev.c *iter = lower->list.next; next 6764 net/core/dev.c lower = list_entry_rcu((*iter)->next, struct netdev_adjacent, list); next 6795 net/core/dev.c *iter = lower->list.next; next 6806 net/core/dev.c lower = list_entry((*iter)->next, struct netdev_adjacent, list); next 6822 net/core/dev.c lower = list_entry((*iter)->next, struct netdev_adjacent, list); next 6838 net/core/dev.c struct net_device *ldev, *next, *now, *dev_stack[MAX_NEST_DEV + 1]; next 6852 net/core/dev.c next = NULL; next 6858 net/core/dev.c next = ldev; next 6865 net/core/dev.c if (!next) { next 6868 net/core/dev.c next = dev_stack[--cur]; next 6872 net/core/dev.c now = next; next 6885 net/core/dev.c struct net_device *ldev, *next, *now, *dev_stack[MAX_NEST_DEV + 1]; next 6900 net/core/dev.c next = NULL; next 6908 net/core/dev.c next = ldev; next 6915 net/core/dev.c if (!next) { next 6918 net/core/dev.c next = dev_stack[--cur]; next 6922 net/core/dev.c now = next; next 6934 net/core/dev.c lower = list_entry_rcu((*iter)->next, struct netdev_adjacent, list); next 7001 net/core/dev.c struct net_device *ldev, *next, *now, *dev_stack[MAX_NEST_DEV + 1]; next 7015 net/core/dev.c next = NULL; next 7021 net/core/dev.c next = ldev; next 7028 net/core/dev.c if (!next) { next 7031 net/core/dev.c next = dev_stack[--cur]; next 7035 net/core/dev.c now = next; next 9839 net/core/dev.c list_skb = &(*list_skb)->next; next 75 net/core/fib_rules.c pos = ops->rules_list.next; next 76 net/core/fib_rules.c if (pos->next != &ops->rules_list) { next 77 net/core/fib_rules.c rule = list_entry(pos->next, struct fib_rule, list); next 186 net/core/neighbour.c neigh = rcu_dereference_protected(n->next, next 216 net/core/neighbour.c np = &n->next; next 302 net/core/neighbour.c np = &n->next; next 306 net/core/neighbour.c np = &n->next; next 310 net/core/neighbour.c rcu_dereference_protected(n->next, next 495 net/core/neighbour.c struct neighbour *n, *next; next 500 net/core/neighbour.c n = next) { next 505 net/core/neighbour.c next = rcu_dereference_protected(n->next, next 508 net/core/neighbour.c rcu_assign_pointer(n->next, next 557 net/core/neighbour.c n = rcu_dereference_bh(n->next)) { next 634 net/core/neighbour.c n1 = rcu_dereference_protected(n1->next, next 650 net/core/neighbour.c rcu_assign_pointer(n->next, next 696 net/core/neighbour.c n = n->next; next 750 net/core/neighbour.c n->next = tbl->phash_buckets[hash_val]; next 768 net/core/neighbour.c np = &n->next) { next 771 net/core/neighbour.c *np = n->next; next 795 net/core/neighbour.c *np = n->next; next 796 net/core/neighbour.c n->next = freelist; next 800 net/core/neighbour.c np = &n->next; next 805 net/core/neighbour.c freelist = n->next; next 806 net/core/neighbour.c n->next = NULL; next 936 net/core/neighbour.c *np = n->next; next 945 net/core/neighbour.c np = &n->next; next 1018 net/core/neighbour.c unsigned long now, next; next 1027 net/core/neighbour.c next = now + HZ; next 1036 net/core/neighbour.c next = neigh->confirmed + neigh->parms->reachable_time; next 1044 net/core/neighbour.c next = now + NEIGH_VAR(neigh->parms, DELAY_PROBE_TIME); next 1061 net/core/neighbour.c next = neigh->confirmed + neigh->parms->reachable_time; next 1068 net/core/neighbour.c next = now + NEIGH_VAR(neigh->parms, RETRANS_TIME); next 1072 net/core/neighbour.c next = now + NEIGH_VAR(neigh->parms, RETRANS_TIME); next 1084 net/core/neighbour.c if (time_before(next, jiffies + HZ/2)) next 1085 net/core/neighbour.c next = jiffies + HZ/2; next 1086 net/core/neighbour.c if (!mod_timer(&neigh->timer, next)) next 1120 net/core/neighbour.c unsigned long next, now = jiffies; next 1127 net/core/neighbour.c next = now + max(NEIGH_VAR(neigh->parms, RETRANS_TIME), next 1129 net/core/neighbour.c neigh_add_timer(neigh, next); next 2401 net/core/neighbour.c goto next; next 2409 net/core/neighbour.c next: next 2567 net/core/neighbour.c n = rcu_dereference_bh(n->next)) { next 2569 net/core/neighbour.c goto next; next 2572 net/core/neighbour.c goto next; next 2580 net/core/neighbour.c next: next 2610 net/core/neighbour.c for (n = tbl->phash_buckets[h], idx = 0; n; n = n->next) { next 2612 net/core/neighbour.c goto next; next 2615 net/core/neighbour.c goto next; next 2623 net/core/neighbour.c next: next 2938 net/core/neighbour.c n = rcu_dereference_bh(n->next)) next 2968 net/core/neighbour.c rcu_dereference_protected(n->next, next 2972 net/core/neighbour.c np = &n->next; next 3041 net/core/neighbour.c goto next; next 3048 net/core/neighbour.c goto next; next 3054 net/core/neighbour.c next: next 3055 net/core/neighbour.c n = rcu_dereference_bh(n->next); next 3079 net/core/neighbour.c n = rcu_dereference_bh(n->next); next 3084 net/core/neighbour.c goto next; next 3089 net/core/neighbour.c goto next; next 3096 net/core/neighbour.c next: next 3097 net/core/neighbour.c n = rcu_dereference_bh(n->next); next 3141 net/core/neighbour.c pn = pn->next; next 3159 net/core/neighbour.c pn = pn->next; next 3167 net/core/neighbour.c pn = pn->next; next 3338 net/core/neighbour.c .next = neigh_stat_seq_next, next 173 net/core/net-procfs.c .next = dev_seq_next, next 180 net/core/net-procfs.c .next = softnet_seq_next, next 225 net/core/net-procfs.c nxt = pt->list.next; next 230 net/core/net-procfs.c nxt = ptype_base[0].next; next 237 net/core/net-procfs.c nxt = ptype_base[hash].next; next 270 net/core/net-procfs.c .next = ptype_seq_next, next 338 net/core/net-procfs.c .next = dev_seq_next, next 1327 net/core/net_namespace.c first_device = first_device->next; next 256 net/core/netpoll.c clist = clist->next; next 86 net/core/rtnetlink.c tail->next = defer_kfree_skb_list; next 101 net/core/rtnetlink.c struct sk_buff *next = head->next; next 105 net/core/rtnetlink.c head = next; next 704 net/core/skbuff.c struct sk_buff *next = segs->next; next 707 net/core/skbuff.c segs = next; next 985 net/core/skbuff.c n->next = n->prev = NULL; next 1137 net/core/skbuff.c u32 bytelen, next; next 1155 net/core/skbuff.c next = (u32)atomic_read(&sk->sk_zckey); next 1156 net/core/skbuff.c if ((u32)(uarg->id + uarg->len) == next) { next 1161 net/core/skbuff.c atomic_set(&sk->sk_zckey, ++next); next 1360 net/core/skbuff.c struct page *next = (struct page *)page_private(head); next 1362 net/core/skbuff.c head = next; next 1964 net/core/skbuff.c fragp = &frag->next) { next 1974 net/core/skbuff.c nfrag->next = frag->next; next 1989 net/core/skbuff.c if (frag->next) next 1990 net/core/skbuff.c skb_drop_list(&frag->next); next 2099 net/core/skbuff.c list = list->next; next 2109 net/core/skbuff.c insp = list->next; next 2126 net/core/skbuff.c skb_shinfo(skb)->frag_list = list->next; next 2131 net/core/skbuff.c clone->next = list; next 2524 net/core/skbuff.c } else if (skb->next) { next 2525 net/core/skbuff.c skb = skb->next; next 3522 net/core/skbuff.c } else if (st->cur_skb->next) { next 3523 net/core/skbuff.c st->cur_skb = st->cur_skb->next; next 3719 net/core/skbuff.c if (frag_len != iter->len && iter->next) next 3789 net/core/skbuff.c list_skb = list_skb->next; next 3821 net/core/skbuff.c tail->next = nskb; next 3882 net/core/skbuff.c list_skb = list_skb->next; next 3956 net/core/skbuff.c for (iter = segs; iter; iter = iter->next) { next 4078 net/core/skbuff.c NAPI_GRO_CB(p)->last->next = skb; next 4351 net/core/skbuff.c if (skb1->next == NULL && tailbits) { next 4382 net/core/skbuff.c skb2->next = skb1->next; next 4389 net/core/skbuff.c skb_p = &skb1->next; next 5791 net/core/skbuff.c list = list->next; next 5799 net/core/skbuff.c insp = list->next; next 5815 net/core/skbuff.c shinfo->frag_list = list->next; next 5820 net/core/skbuff.c clone->next = list; next 2430 net/core/sock.c struct sk_buff *skb, *next; next 2438 net/core/sock.c next = skb->next; next 2439 net/core/sock.c prefetch(next); next 2446 net/core/sock.c skb = next; next 3567 net/core/sock.c .next = proto_seq_next, next 324 net/core/sock_map.c static int sock_map_get_next_key(struct bpf_map *map, void *key, void *next) next 328 net/core/sock_map.c u32 *key_next = next; next 30 net/dccp/ackvec.c struct dccp_ackvec_record *cur, *next; next 32 net/dccp/ackvec.c list_for_each_entry_safe(cur, next, &av->av_records, avr_node) next 284 net/dccp/ackvec.c struct dccp_ackvec_record *avr, *next; next 338 net/dccp/ackvec.c list_for_each_entry_safe_from(avr, next, &av->av_records, avr_node) { next 364 net/dccp/ackvec.c struct dccp_ackvec_parsed *cur, *next; next 366 net/dccp/ackvec.c list_for_each_entry_safe(cur, next, parsed_chunks, node) next 240 net/dccp/ccids/ccid2.c struct ccid2_seq *next; next 269 net/dccp/ccids/ccid2.c next = hc->tx_seqh->ccid2s_next; next 271 net/dccp/ccids/ccid2.c if (next == hc->tx_seqt) { next 277 net/dccp/ccids/ccid2.c next = hc->tx_seqh->ccid2s_next; next 278 net/dccp/ccids/ccid2.c BUG_ON(next == hc->tx_seqt); next 280 net/dccp/ccids/ccid2.c hc->tx_seqh = next; next 378 net/dccp/ccids/ccid3.c tfrc_tx_hist_purge(&acked->next); next 57 net/dccp/ccids/lib/packet_history.c entry->next = *headp; next 67 net/dccp/ccids/lib/packet_history.c struct tfrc_tx_hist_entry *next = head->next; next 70 net/dccp/ccids/lib/packet_history.c head = next; next 37 net/dccp/ccids/lib/packet_history.h struct tfrc_tx_hist_entry *next; next 46 net/dccp/ccids/lib/packet_history.h head = head->next; next 541 net/dccp/feat.c struct dccp_feat_entry *entry, *next; next 543 net/dccp/feat.c list_for_each_entry_safe(entry, next, fn_list, node) next 633 net/dccp/feat.c struct dccp_feat_entry *pos, *next; next 638 net/dccp/feat.c list_for_each_entry_safe_reverse(pos, next, fn, node) { next 1500 net/dccp/feat.c struct dccp_feat_entry *cur, *next; next 1544 net/dccp/feat.c list_for_each_entry_safe(cur, next, fn_list, node) next 2305 net/decnet/af_decnet.c .next = dn_socket_seq_next, next 1381 net/decnet/dn_dev.c .next = dn_dev_seq_next, next 588 net/decnet/dn_neigh.c .next = neigh_seq_next, next 357 net/decnet/dn_route.c struct dn_route *rt, *next; next 365 net/decnet/dn_route.c for(; rt; rt = next) { next 366 net/decnet/dn_route.c next = rcu_dereference_raw(rt->dn_next); next 1845 net/decnet/dn_route.c .next = dn_rt_cache_seq_next, next 127 net/decnet/dn_table.c struct dn_fib_node *f, **fp, *next; next 131 net/decnet/dn_table.c for(f = old_ht[i]; f; f = next) { next 132 net/decnet/dn_table.c next = f->fn_next; next 511 net/decnet/dn_table.c goto next; next 518 net/decnet/dn_table.c next: next 917 net/decnet/dn_table.c struct hlist_node *next; next 922 net/decnet/dn_table.c hlist_for_each_entry_safe(t, next, &dn_fib_table_hash[h], next 1360 net/ipv4/af_inet.c if (skb->next) next 1387 net/ipv4/af_inet.c } while ((skb = skb->next)); next 1414 net/ipv4/arp.c .next = neigh_seq_next, next 268 net/ipv4/cipso_ipv4.c __list_del(entry->list.prev, entry->list.next); next 699 net/ipv4/devinet.c unsigned long now, next, next_sec, next_sched; next 705 net/ipv4/devinet.c next = round_jiffies_up(now + ADDR_CHECK_FREQUENCY); next 729 net/ipv4/devinet.c ifa->ifa_valid_lft * HZ, next)) next 730 net/ipv4/devinet.c next = ifa->ifa_tstamp + next 737 net/ipv4/devinet.c next)) { next 738 net/ipv4/devinet.c next = ifa->ifa_tstamp + next 783 net/ipv4/devinet.c next_sec = round_jiffies_up(next); next 784 net/ipv4/devinet.c next_sched = next; next 787 net/ipv4/devinet.c if (time_before(next_sec, next + ADDRCONF_TIMER_FUZZ)) next 1026 net/ipv4/fib_frontend.c goto next; next 1038 net/ipv4/fib_frontend.c next: next 175 net/ipv4/fib_semantics.c struct fib_nh_exception *next; next 177 net/ipv4/fib_semantics.c next = rcu_dereference_protected(fnhe->fnhe_next, 1); next 184 net/ipv4/fib_semantics.c fnhe = next; next 481 net/ipv4/fib_trie.c tn_info(tn)->rcu.next = NULL; next 487 net/ipv4/fib_trie.c tn_info(n)->rcu.next = tn_info(tn)->rcu.next; next 488 net/ipv4/fib_trie.c tn_info(tn)->rcu.next = &tn_info(n)->rcu; next 496 net/ipv4/fib_trie.c head = head->next; next 1520 net/ipv4/fib_trie.c struct fib_alias *fa = hlist_entry(pprev, typeof(*fa), fa_list.next); next 2109 net/ipv4/fib_trie.c goto next; next 2114 net/ipv4/fib_trie.c goto next; next 2118 net/ipv4/fib_trie.c goto next; next 2122 net/ipv4/fib_trie.c goto next; next 2126 net/ipv4/fib_trie.c goto next; next 2153 net/ipv4/fib_trie.c next: next 2653 net/ipv4/fib_trie.c .next = fib_trie_seq_next, next 2842 net/ipv4/fib_trie.c .next = fib_route_seq_next, next 1257 net/ipv4/fou.c struct fou *fou, *next; next 1261 net/ipv4/fou.c list_for_each_entry_safe(fou, next, &fn->fou_list, list) next 102 net/ipv4/gre_offload.c } while ((skb = skb->next)); next 189 net/ipv4/igmp.c struct ip_sf_list *next; next 192 net/ipv4/igmp.c next = psf->sf_next; next 194 net/ipv4/igmp.c psf = next; next 663 net/ipv4/igmp.c pmc_next = pmc->next; next 683 net/ipv4/igmp.c pmc_prev->next = pmc_next; next 1198 net/ipv4/igmp.c pmc->next = in_dev->mc_tomb; next 1215 net/ipv4/igmp.c for (pmc = in_dev->mc_tomb; pmc; pmc = pmc->next) { next 1222 net/ipv4/igmp.c pmc_prev->next = pmc->next; next 1224 net/ipv4/igmp.c in_dev->mc_tomb = pmc->next; next 1258 net/ipv4/igmp.c nextpmc = pmc->next; next 2865 net/ipv4/igmp.c .next = igmp_mc_seq_next, next 2915 net/ipv4/igmp.c state->im = state->im->next; next 2998 net/ipv4/igmp.c .next = igmp_mcf_seq_next, next 997 net/ipv4/inet_connection_sock.c struct request_sock *next, *req; next 1030 net/ipv4/inet_connection_sock.c next = req->dl_next; next 1032 net/ipv4/inet_connection_sock.c req = next; next 243 net/ipv4/inet_fragment.c struct sk_buff *next = FRAG_CB(skb)->next_frag; next 247 net/ipv4/inet_fragment.c skb = next; next 468 net/ipv4/inet_fragment.c nextp = &clone->next; next 522 net/ipv4/inet_fragment.c nextp = &fp->next; next 640 net/ipv4/inet_hashtables.c if (sk_head(&tb->owners) == sk && !sk->sk_bind_node.next) { next 106 net/ipv4/inetpeer.c struct rb_node **pp, *parent, *next; next 114 net/ipv4/inetpeer.c next = rcu_dereference_raw(*pp); next 115 net/ipv4/inetpeer.c if (!next) next 117 net/ipv4/inetpeer.c parent = next; next 132 net/ipv4/inetpeer.c pp = &next->rb_left; next 134 net/ipv4/inetpeer.c pp = &next->rb_right; next 156 net/ipv4/ip_input.c for (ra = rcu_dereference(net->ipv4.ra_chain); ra; ra = rcu_dereference(ra->next)) { next 530 net/ipv4/ip_input.c struct sk_buff *skb, *next; next 532 net/ipv4/ip_input.c list_for_each_entry_safe(skb, next, head, list) { next 542 net/ipv4/ip_input.c struct sk_buff *skb, *next; next 546 net/ipv4/ip_input.c list_for_each_entry_safe(skb, next, head, list) { next 589 net/ipv4/ip_input.c struct sk_buff *skb, *next; next 593 net/ipv4/ip_input.c list_for_each_entry_safe(skb, next, head, list) { next 276 net/ipv4/ip_output.c struct sk_buff *nskb = segs->next; next 640 net/ipv4/ip_output.c if (frag->next) next 823 net/ipv4/ip_output.c ((frag->len & 7) && frag->next) || next 1491 net/ipv4/ip_output.c tail_skb = &(tmp_skb->next); next 353 net/ipv4/ip_sockglue.c rap = &ra->next) { next 362 net/ipv4/ip_sockglue.c RCU_INIT_POINTER(*rap, ra->next); next 384 net/ipv4/ip_sockglue.c RCU_INIT_POINTER(new_ra->next, ra); next 191 net/ipv4/ipconfig.c struct ic_device *next; next 254 net/ipv4/ipconfig.c last = &d->next; next 310 net/ipv4/ipconfig.c struct ic_device *d, *next; next 314 net/ipv4/ipconfig.c next = ic_first_dev; next 315 net/ipv4/ipconfig.c while ((d = next)) { next 316 net/ipv4/ipconfig.c next = d->next; next 527 net/ipv4/ipconfig.c d = d->next; next 1030 net/ipv4/ipconfig.c d = d->next; next 1231 net/ipv4/ipconfig.c if (!d->next) { next 1254 net/ipv4/ipconfig.c if ((d = d->next)) next 1491 net/ipv4/ipconfig.c ic_first_dev->next) { next 121 net/ipv4/ipmr.c ret = list_entry_rcu(net->ipv4.mr_tables.next, next 124 net/ipv4/ipmr.c ret = list_entry_rcu(mrt->list.next, next 270 net/ipv4/ipmr.c struct mr_table *mrt, *next; next 273 net/ipv4/ipmr.c list_for_each_entry_safe(mrt, next, &net->ipv4.mr_tables, list) { next 773 net/ipv4/ipmr.c struct mr_mfc *c, *next; next 788 net/ipv4/ipmr.c list_for_each_entry_safe(c, next, &mrt->mfc_unres_queue, list) { next 2964 net/ipv4/ipmr.c .next = mr_vif_seq_next, next 3024 net/ipv4/ipmr.c .next = mr_mfc_seq_next, next 185 net/ipv4/ipmr_base.c if (c->list.next != it->cache) next 186 net/ipv4/ipmr_base.c return list_entry(c->list.next, struct mr_mfc, list); next 349 net/ipv4/netfilter/arp_tables.c goto next; next 382 net/ipv4/netfilter/arp_tables.c next: ; next 413 net/ipv4/netfilter/ip_tables.c goto next; next 446 net/ipv4/netfilter/ip_tables.c next: ; next 740 net/ipv4/netfilter/ipt_CLUSTERIP.c .next = clusterip_seq_next, next 145 net/ipv4/nexthop.c struct rb_node **pp, *parent = NULL, *next; next 151 net/ipv4/nexthop.c next = rcu_dereference_raw(*pp); next 152 net/ipv4/nexthop.c if (!next) next 154 net/ipv4/nexthop.c parent = next; next 158 net/ipv4/nexthop.c pp = &next->rb_left; next 160 net/ipv4/nexthop.c pp = &next->rb_right; next 1010 net/ipv4/nexthop.c struct rb_node **pp, *parent = NULL, *next; next 1022 net/ipv4/nexthop.c next = rtnl_dereference(*pp); next 1023 net/ipv4/nexthop.c if (!next) next 1026 net/ipv4/nexthop.c parent = next; next 1030 net/ipv4/nexthop.c pp = &next->rb_left; next 1032 net/ipv4/nexthop.c pp = &next->rb_right; next 1141 net/ipv4/ping.c .next = ping_seq_next, next 1098 net/ipv4/raw.c .next = raw_seq_next, next 172 net/ipv4/raw_diag.c goto next; next 174 net/ipv4/raw_diag.c goto next; next 177 net/ipv4/raw_diag.c goto next; next 180 net/ipv4/raw_diag.c goto next; next 183 net/ipv4/raw_diag.c next: next 230 net/ipv4/route.c .next = rt_cache_seq_next, next 319 net/ipv4/route.c .next = rt_cpu_seq_next, next 2886 net/ipv4/route.c goto next; next 2889 net/ipv4/route.c goto next; next 2893 net/ipv4/route.c goto next; next 2899 net/ipv4/route.c goto next; next 2907 net/ipv4/route.c next: next 1790 net/ipv4/tcp.c skb = skb->next; next 2299 net/ipv4/tcp.c int next = (int)new_state[sk->sk_state]; next 2300 net/ipv4/tcp.c int ns = next & TCP_STATE_MASK; next 2304 net/ipv4/tcp.c return next & TCP_ACTION_FIN; next 3079 net/ipv4/tcp_input.c struct sk_buff *skb, *next; next 3091 net/ipv4/tcp_input.c for (skb = skb_rb_first(&sk->tcp_rtx_queue); skb; skb = next) { next 3162 net/ipv4/tcp_input.c next = skb_rb_next(skb); next 3167 net/ipv4/tcp_input.c tcp_highest_sack_replace(sk, skb, next); next 4863 net/ipv4/tcp_input.c return !skb_queue_is_last(list, skb) ? skb->next : NULL; next 4872 net/ipv4/tcp_input.c struct sk_buff *next = tcp_skb_next(skb, list); next 4882 net/ipv4/tcp_input.c return next; next 2535 net/ipv4/tcp_ipv4.c .next = tcp_seq_next, next 24 net/ipv4/tcp_offload.c skb = skb->next; next 124 net/ipv4/tcp_offload.c while (skb->next) { next 139 net/ipv4/tcp_offload.c skb = skb->next; next 2068 net/ipv4/tcp_output.c struct sk_buff *skb, *next; next 2071 net/ipv4/tcp_output.c tcp_for_write_queue_from_safe(skb, next, sk) { next 2097 net/ipv4/tcp_output.c struct sk_buff *skb, *nskb, *next; next 2180 net/ipv4/tcp_output.c tcp_for_write_queue_from_safe(skb, next, sk) { next 44 net/ipv4/tunnel4.c pprev = &t->next) { next 51 net/ipv4/tunnel4.c handler->next = *pprev; next 74 net/ipv4/tunnel4.c pprev = &t->next) { next 76 net/ipv4/tunnel4.c *pprev = handler->next; next 93 net/ipv4/tunnel4.c handler = rcu_dereference(handler->next)) \ next 2103 net/ipv4/udp.c struct sk_buff *next, *segs; next 2112 net/ipv4/udp.c for (skb = segs; skb; skb = next) { next 2113 net/ipv4/udp.c next = skb->next; next 2941 net/ipv4/udp.c .next = udp_seq_next, next 122 net/ipv4/udp_diag.c goto next; next 124 net/ipv4/udp_diag.c goto next; next 127 net/ipv4/udp_diag.c goto next; next 130 net/ipv4/udp_diag.c goto next; next 133 net/ipv4/udp_diag.c goto next; next 139 net/ipv4/udp_diag.c next: next 143 net/ipv4/udp_offload.c } while ((skb = skb->next)); next 243 net/ipv4/udp_offload.c if (!seg->next) next 255 net/ipv4/udp_offload.c seg = seg->next; next 43 net/ipv4/xfrm4_protocol.c handler = rcu_dereference(handler->next)) \ next 232 net/ipv4/xfrm4_protocol.c pprev = &t->next) { next 239 net/ipv4/xfrm4_protocol.c handler->next = *pprev; next 273 net/ipv4/xfrm4_protocol.c pprev = &t->next) { next 275 net/ipv4/xfrm4_protocol.c *pprev = handler->next; next 4350 net/ipv6/addrconf.c .next = if6_seq_next, next 4413 net/ipv6/addrconf.c unsigned long now, next, next_sec, next_sched; next 4421 net/ipv6/addrconf.c next = round_jiffies_up(now + ADDR_CHECK_FREQUENCY); next 4461 net/ipv6/addrconf.c (time_before(ifp->tstamp + ifp->valid_lft * HZ, next))) next 4462 net/ipv6/addrconf.c next = ifp->tstamp + ifp->valid_lft * HZ; next 4481 net/ipv6/addrconf.c if (time_before(ifp->tstamp + ifp->prefered_lft * HZ, next)) next 4482 net/ipv6/addrconf.c next = ifp->tstamp + ifp->prefered_lft * HZ; next 4499 net/ipv6/addrconf.c } else if (time_before(ifp->tstamp + ifp->prefered_lft * HZ - regen_advance * HZ, next)) next 4500 net/ipv6/addrconf.c next = ifp->tstamp + ifp->prefered_lft * HZ - regen_advance * HZ; next 4504 net/ipv6/addrconf.c if (time_before(ifp->tstamp + ifp->prefered_lft * HZ, next)) next 4505 net/ipv6/addrconf.c next = ifp->tstamp + ifp->prefered_lft * HZ; next 4511 net/ipv6/addrconf.c next_sec = round_jiffies_up(next); next 4512 net/ipv6/addrconf.c next_sched = next; next 4515 net/ipv6/addrconf.c if (time_before(next_sec, next + ADDRCONF_TIMER_FUZZ)) next 4523 net/ipv6/addrconf.c now, next, next_sec, next_sched); next 5053 net/ipv6/addrconf.c goto next; next 5058 net/ipv6/addrconf.c next: next 5068 net/ipv6/addrconf.c ifmca = ifmca->next, ip_idx++) { next 203 net/ipv6/anycast.c struct ipv6_ac_socklist *next = pac->acl_next; next 212 net/ipv6/anycast.c pac = next; next 573 net/ipv6/anycast.c .next = ac6_seq_next, next 230 net/ipv6/calipso.c __list_del(entry->list.prev, entry->list.next); next 21 net/ipv6/ila/ila_xlat.c struct ila_map __rcu *next; next 138 net/ipv6/ila/ila_xlat.c ila = rcu_access_pointer(ila->next); next 156 net/ipv6/ila/ila_xlat.c ila = rcu_access_pointer(ila->next); next 169 net/ipv6/ila/ila_xlat.c struct ila_map *next; next 173 net/ipv6/ila/ila_xlat.c next = rcu_access_pointer(ila->next); next 175 net/ipv6/ila/ila_xlat.c ila = next; next 255 net/ipv6/ila/ila_xlat.c tila = rcu_dereference_protected(tila->next, next 261 net/ipv6/ila/ila_xlat.c RCU_INIT_POINTER(ila->next, tila); next 262 net/ipv6/ila/ila_xlat.c rcu_assign_pointer(prev->next, ila); next 265 net/ipv6/ila/ila_xlat.c RCU_INIT_POINTER(ila->next, head); next 301 net/ipv6/ila/ila_xlat.c ila = rcu_dereference_protected(ila->next, next 310 net/ipv6/ila/ila_xlat.c rcu_assign_pointer(prev->next, ila->next); next 315 net/ipv6/ila/ila_xlat.c head = rcu_dereference_protected(ila->next, next 554 net/ipv6/ila/ila_xlat.c ila = rcu_access_pointer(ila->next); next 588 net/ipv6/ila/ila_xlat.c ila = rcu_access_pointer(ila->next); next 640 net/ipv6/ip6_fib.c goto next; next 644 net/ipv6/ip6_fib.c next: next 1492 net/ipv6/ip6_fib.c struct fib6_node *next; next 1496 net/ipv6/ip6_fib.c next = dir ? rcu_dereference(fn->right) : next 1499 net/ipv6/ip6_fib.c if (next) { next 1500 net/ipv6/ip6_fib.c fn = next; next 1601 net/ipv6/ip6_fib.c goto next; next 1619 net/ipv6/ip6_fib.c next: next 2547 net/ipv6/ip6_fib.c .next = ipv6_route_seq_next, next 63 net/ipv6/ip6_flowlabel.c fl = rcu_dereference_bh(fl->next)) next 65 net/ipv6/ip6_flowlabel.c for (fl = rcu_dereference_bh(fl->next); \ next 67 net/ipv6/ip6_flowlabel.c fl = rcu_dereference_bh(fl->next)) next 72 net/ipv6/ip6_flowlabel.c sfl = rcu_dereference_bh(sfl->next)) next 169 net/ipv6/ip6_flowlabel.c *flp = fl->next; next 177 net/ipv6/ip6_flowlabel.c flp = &fl->next; next 202 net/ipv6/ip6_flowlabel.c *flp = fl->next; next 207 net/ipv6/ip6_flowlabel.c flp = &fl->next; next 248 net/ipv6/ip6_flowlabel.c fl->next = fl_ht[FL_HASH(fl->label)]; next 292 net/ipv6/ip6_flowlabel.c np->ipv6_fl_list = sfl->next; next 494 net/ipv6/ip6_flowlabel.c sfl->next = np->ipv6_fl_list; next 569 net/ipv6/ip6_flowlabel.c sflp = &sfl->next) { next 573 net/ipv6/ip6_flowlabel.c *sflp = sfl->next; next 833 net/ipv6/ip6_flowlabel.c .next = ip6fl_seq_next, next 301 net/ipv6/ip6_gre.c rcu_assign_pointer(t->next, rtnl_dereference(*tp)); next 312 net/ipv6/ip6_gre.c tp = &iter->next) { next 314 net/ipv6/ip6_gre.c rcu_assign_pointer(*tp, t->next); next 334 net/ipv6/ip6_gre.c tp = &t->next) next 1551 net/ipv6/ip6_gre.c t = rtnl_dereference(t->next); next 81 net/ipv6/ip6_input.c struct sk_buff *skb, *next; next 83 net/ipv6/ip6_input.c list_for_each_entry_safe(skb, next, head, list) { next 93 net/ipv6/ip6_input.c struct sk_buff *skb, *next; next 97 net/ipv6/ip6_input.c list_for_each_entry_safe(skb, next, head, list) { next 303 net/ipv6/ip6_input.c struct sk_buff *skb, *next; next 307 net/ipv6/ip6_input.c list_for_each_entry_safe(skb, next, head, list) { next 121 net/ipv6/ip6_offload.c for (skb = segs; skb; skb = skb->next) { next 141 net/ipv6/ip6_offload.c if (skb->next) next 302 net/ipv6/ip6_output.c for (ra = ip6_ra_chain; ra; ra = ra->next) { next 657 net/ipv6/ip6_output.c if (frag->next) next 832 net/ipv6/ip6_output.c ((frag->len & 7) && frag->next) || next 1740 net/ipv6/ip6_output.c tail_skb = &(tmp_skb->next); next 134 net/ipv6/ip6_tunnel.c for (t = rcu_dereference(start); t; t = rcu_dereference(t->next)) next 217 net/ipv6/ip6_tunnel.c rcu_assign_pointer(t->next , rtnl_dereference(*tp)); next 237 net/ipv6/ip6_tunnel.c tp = &iter->next) { next 239 net/ipv6/ip6_tunnel.c rcu_assign_pointer(*tp, t->next); next 352 net/ipv6/ip6_tunnel.c tp = &t->next) { next 403 net/ipv6/ip6_tunnel.c u8 next, nexthdr = ipv6h->nexthdr; next 426 net/ipv6/ip6_tunnel.c next = hdr->nexthdr; next 453 net/ipv6/ip6_tunnel.c nexthdr = next; next 2209 net/ipv6/ip6_tunnel.c t = rtnl_dereference(t->next); next 75 net/ipv6/ip6_vti.c for (t = rcu_dereference(start); t; t = rcu_dereference(t->next)) next 156 net/ipv6/ip6_vti.c rcu_assign_pointer(t->next , rtnl_dereference(*tp)); next 168 net/ipv6/ip6_vti.c tp = &iter->next) { next 170 net/ipv6/ip6_vti.c rcu_assign_pointer(*tp, t->next); next 266 net/ipv6/ip6_vti.c tp = &t->next) { next 1116 net/ipv6/ip6_vti.c t = rtnl_dereference(t->next); next 108 net/ipv6/ip6mr.c ret = list_entry_rcu(net->ipv6.mr6_tables.next, next 111 net/ipv6/ip6mr.c ret = list_entry_rcu(mrt->list.next, next 257 net/ipv6/ip6mr.c struct mr_table *mrt, *next; next 260 net/ipv6/ip6mr.c list_for_each_entry_safe(mrt, next, &net->ipv6.mr6_tables, list) { next 448 net/ipv6/ip6mr.c .next = mr_vif_seq_next, next 509 net/ipv6/ip6mr.c .next = mr_mfc_seq_next, next 796 net/ipv6/ip6mr.c struct mr_mfc *c, *next; next 798 net/ipv6/ip6mr.c list_for_each_entry_safe(c, next, &mrt->mfc_unres_queue, list) { next 71 net/ipv6/ipv6_sockglue.c for (rap = &ip6_ra_chain; (ra = *rap) != NULL; rap = &ra->next) { next 79 net/ipv6/ipv6_sockglue.c *rap = ra->next; next 93 net/ipv6/ipv6_sockglue.c new_ra->next = ra; next 119 net/ipv6/mcast.c pmc = rcu_dereference(pmc->next)) next 162 net/ipv6/mcast.c mc_lst->next = NULL; next 196 net/ipv6/mcast.c mc_lst->next = np->ipv6_mc_list; next 231 net/ipv6/mcast.c lnk = &mc_lst->next) { next 236 net/ipv6/mcast.c *lnk = mc_lst->next; next 300 net/ipv6/mcast.c np->ipv6_mc_list = mc_lst->next; next 760 net/ipv6/mcast.c pmc->next = idev->mc_tomb; next 773 net/ipv6/mcast.c for (pmc = idev->mc_tomb; pmc; pmc = pmc->next) { next 780 net/ipv6/mcast.c pmc_prev->next = pmc->next; next 782 net/ipv6/mcast.c idev->mc_tomb = pmc->next; next 814 net/ipv6/mcast.c nextpmc = pmc->next; next 822 net/ipv6/mcast.c for (pmc = idev->mc_list; pmc; pmc = pmc->next) { next 904 net/ipv6/mcast.c for (mc = idev->mc_list; mc; mc = mc->next) { next 921 net/ipv6/mcast.c mc->next = idev->mc_list; next 952 net/ipv6/mcast.c for (map = &idev->mc_list; (ma = *map) != NULL; map = &ma->next) { next 955 net/ipv6/mcast.c *map = ma->next; next 1004 net/ipv6/mcast.c for (mc = idev->mc_list; mc; mc = mc->next) { next 1422 net/ipv6/mcast.c for (ma = idev->mc_list; ma; ma = ma->next) { next 1428 net/ipv6/mcast.c for (ma = idev->mc_list; ma; ma = ma->next) { next 1492 net/ipv6/mcast.c for (ma = idev->mc_list; ma; ma = ma->next) { next 1862 net/ipv6/mcast.c for (pmc = idev->mc_list; pmc; pmc = pmc->next) { next 1920 net/ipv6/mcast.c pmc_next = pmc->next; next 1941 net/ipv6/mcast.c pmc_prev->next = pmc_next; next 1952 net/ipv6/mcast.c for (pmc = idev->mc_list; pmc; pmc = pmc->next) { next 2088 net/ipv6/mcast.c for (pmc = idev->mc_list; pmc; pmc = pmc->next) { next 2176 net/ipv6/mcast.c for (pmc = idev->mc_list; pmc; pmc = pmc->next) { next 2340 net/ipv6/mcast.c for (pmc = idev->mc_list; pmc; pmc = pmc->next) { next 2520 net/ipv6/mcast.c for (i = idev->mc_list; i; i = i->next) next 2540 net/ipv6/mcast.c for (i = idev->mc_list; i; i = i->next) next 2571 net/ipv6/mcast.c for (i = idev->mc_list; i; i = i->next) { next 2618 net/ipv6/mcast.c idev->mc_list = i->next; next 2635 net/ipv6/mcast.c for (pmc = idev->mc_list; pmc; pmc = pmc->next) next 2701 net/ipv6/mcast.c im = im->next; next 2774 net/ipv6/mcast.c .next = igmp6_mc_seq_next, next 2826 net/ipv6/mcast.c state->im = state->im->next; next 2915 net/ipv6/mcast.c .next = igmp6_mcf_seq_next, next 431 net/ipv6/netfilter/ip6_tables.c goto next; next 464 net/ipv6/netfilter/ip6_tables.c next: ; next 214 net/ipv6/ping.c .next = ping_seq_next, next 1323 net/ipv6/raw.c .next = raw_seq_next, next 904 net/ipv6/route.c struct fib6_info *next = rcu_dereference(rt0->fib6_next); next 907 net/ipv6/route.c if (!next || next->fib6_metric != rt0->fib6_metric) next 908 net/ipv6/route.c next = leaf; next 910 net/ipv6/route.c if (next != rt0) { next 913 net/ipv6/route.c if (next->fib6_node) next 914 net/ipv6/route.c rcu_assign_pointer(fn->rr_ptr, next); next 5010 net/ipv6/route.c struct list_head next; next 5021 net/ipv6/route.c list_for_each_entry(nh, rt6_nh_list, next) { next 5032 net/ipv6/route.c list_add_tail(&nh->next, rt6_nh_list); next 5153 net/ipv6/route.c list_for_each_entry(nh, &rt6_nh_list, next) { next 5211 net/ipv6/route.c list_for_each_entry(nh, &rt6_nh_list, next) { next 5218 net/ipv6/route.c list_for_each_entry_safe(nh, nh_safe, &rt6_nh_list, next) { next 5221 net/ipv6/route.c list_del(&nh->next); next 160 net/ipv6/sit.c tp = &iter->next) { next 162 net/ipv6/sit.c rcu_assign_pointer(*tp, t->next); next 172 net/ipv6/sit.c rcu_assign_pointer(t->next, rtnl_dereference(*tp)); next 236 net/ipv6/sit.c tp = &t->next) { next 280 net/ipv6/sit.c prl = rcu_dereference(prl->next)) next 368 net/ipv6/sit.c for (p = rtnl_dereference(t->prl); p; p = rtnl_dereference(p->next)) { next 390 net/ipv6/sit.c p->next = t->prl; next 405 net/ipv6/sit.c n = rcu_dereference_protected(p->next, 1); next 423 net/ipv6/sit.c p = &x->next) { next 425 net/ipv6/sit.c *p = x->next; next 1834 net/ipv6/sit.c t = rtnl_dereference(t->next); next 1984 net/ipv6/tcp_ipv6.c .next = tcp_seq_next, next 38 net/ipv6/tunnel6.c pprev = &t->next) { next 45 net/ipv6/tunnel6.c handler->next = *pprev; next 68 net/ipv6/tunnel6.c pprev = &t->next) { next 70 net/ipv6/tunnel6.c *pprev = handler->next; next 87 net/ipv6/tunnel6.c handler = rcu_dereference(handler->next)) \ next 685 net/ipv6/udp.c struct sk_buff *next, *segs; next 693 net/ipv6/udp.c for (skb = segs; skb; skb = next) { next 694 net/ipv6/udp.c next = skb->next; next 1089 net/ipv6/udp.c } while ((frags = frags->next)); next 1627 net/ipv6/udp.c .next = udp_seq_next, next 43 net/ipv6/xfrm6_protocol.c handler = rcu_dereference(handler->next)) \ next 206 net/ipv6/xfrm6_protocol.c pprev = &t->next) { next 213 net/ipv6/xfrm6_protocol.c handler->next = *pprev; next 247 net/ipv6/xfrm6_protocol.c pprev = &t->next) { next 249 net/ipv6/xfrm6_protocol.c *pprev = handler->next; next 238 net/kcm/kcmproc.c .next = kcm_seq_next, next 696 net/kcm/kcmsock.c } else if (skb->next) { next 697 net/kcm/kcmsock.c skb = skb->next; next 809 net/kcm/kcmsock.c skb->next = tskb; next 976 net/kcm/kcmsock.c skb->next = tskb; next 3800 net/key/af_key.c .next = pfkey_seq_next, next 262 net/l2tp/l2tp_debugfs.c .next = l2tp_dfs_seq_next, next 1625 net/l2tp/l2tp_ppp.c .next = pppol2tp_seq_next, next 66 net/lapb/lapb_iface.c if (lapb->node.next) { next 82 net/llc/llc_proc.c struct sock* sk, *next; next 92 net/llc/llc_proc.c next = sk_nulls_next(sk); next 93 net/llc/llc_proc.c if (next) { next 94 net/llc/llc_proc.c sk = next; next 205 net/llc/llc_proc.c .next = llc_seq_next, next 212 net/llc/llc_proc.c .next = llc_seq_next, next 1485 net/mac80211/mlme.c goto next; next 1497 net/mac80211/mlme.c next: next 1066 net/mac80211/tx.c struct sk_buff *next = skb_queue_next(&tx->skbs, skb); next 1067 net/mac80211/tx.c next_len = next->len; next 3299 net/mac80211/tx.c frag_tail = &(*frag_tail)->next; next 3847 net/mac80211/tx.c struct sk_buff *next; next 3908 net/mac80211/tx.c next = skb; next 3909 net/mac80211/tx.c while (next) { next 3910 net/mac80211/tx.c skb = next; next 3911 net/mac80211/tx.c next = skb->next; next 3914 net/mac80211/tx.c skb->next = NULL; next 915 net/mac802154/llsec.c struct mac802154_llsec_device_key *next; next 917 net/mac802154/llsec.c next = kzalloc(sizeof(*devkey), GFP_ATOMIC); next 918 net/mac802154/llsec.c if (!next) next 921 net/mac802154/llsec.c next->devkey.key_id = *in_key; next 927 net/mac802154/llsec.c list_add_rcu(&next->devkey.list, &dev->dev.keys); next 929 net/mac802154/llsec.c kzfree(next); next 1514 net/mpls/af_mpls.c goto next; next 1530 net/mpls/af_mpls.c next: next 63 net/mpls/mpls_gso.c } while ((skb = skb->next)); next 331 net/netfilter/ipset/ip_set_hash_gen.h struct mtype_elem next; /* temporary storage for uadd */ next 956 net/netfilter/ipset/ip_set_hash_gen.h mtype_data_next(&h->next, d); next 67 net/netfilter/ipset/ip_set_hash_ip.c hash_ip4_data_next(struct hash_ip4_elem *next, const struct hash_ip4_elem *e) next 69 net/netfilter/ipset/ip_set_hash_ip.c next->ip = e->ip; next 147 net/netfilter/ipset/ip_set_hash_ip.c ip = ntohl(h->next.ip); next 200 net/netfilter/ipset/ip_set_hash_ip.c hash_ip6_data_next(struct hash_ip6_elem *next, const struct hash_ip6_elem *e) next 71 net/netfilter/ipset/ip_set_hash_ipmac.c hash_ipmac4_data_next(struct hash_ipmac4_elem *next, next 74 net/netfilter/ipset/ip_set_hash_ipmac.c next->ip = e->ip; next 179 net/netfilter/ipset/ip_set_hash_ipmac.c hash_ipmac6_data_next(struct hash_ipmac6_elem *next, next 68 net/netfilter/ipset/ip_set_hash_ipmark.c hash_ipmark4_data_next(struct hash_ipmark4_elem *next, next 71 net/netfilter/ipset/ip_set_hash_ipmark.c next->ip = d->ip; next 146 net/netfilter/ipset/ip_set_hash_ipmark.c ip = ntohl(h->next.ip); next 191 net/netfilter/ipset/ip_set_hash_ipmark.c hash_ipmark6_data_next(struct hash_ipmark6_elem *next, next 75 net/netfilter/ipset/ip_set_hash_ipport.c hash_ipport4_data_next(struct hash_ipport4_elem *next, next 78 net/netfilter/ipset/ip_set_hash_ipport.c next->ip = d->ip; next 79 net/netfilter/ipset/ip_set_hash_ipport.c next->port = d->port; next 176 net/netfilter/ipset/ip_set_hash_ipport.c ip = ntohl(h->next.ip); next 178 net/netfilter/ipset/ip_set_hash_ipport.c p = retried && ip == ntohl(h->next.ip) ? ntohs(h->next.port) next 230 net/netfilter/ipset/ip_set_hash_ipport.c hash_ipport6_data_next(struct hash_ipport6_elem *next, next 233 net/netfilter/ipset/ip_set_hash_ipport.c next->port = d->port; next 323 net/netfilter/ipset/ip_set_hash_ipport.c port = ntohs(h->next.port); next 76 net/netfilter/ipset/ip_set_hash_ipportip.c hash_ipportip4_data_next(struct hash_ipportip4_elem *next, next 79 net/netfilter/ipset/ip_set_hash_ipportip.c next->ip = d->ip; next 80 net/netfilter/ipset/ip_set_hash_ipportip.c next->port = d->port; next 183 net/netfilter/ipset/ip_set_hash_ipportip.c ip = ntohl(h->next.ip); next 185 net/netfilter/ipset/ip_set_hash_ipportip.c p = retried && ip == ntohl(h->next.ip) ? ntohs(h->next.port) next 240 net/netfilter/ipset/ip_set_hash_ipportip.c hash_ipportip6_data_next(struct hash_ipportip6_elem *next, next 243 net/netfilter/ipset/ip_set_hash_ipportip.c next->port = d->port; next 338 net/netfilter/ipset/ip_set_hash_ipportip.c port = ntohs(h->next.port); next 120 net/netfilter/ipset/ip_set_hash_ipportnet.c hash_ipportnet4_data_next(struct hash_ipportnet4_elem *next, next 123 net/netfilter/ipset/ip_set_hash_ipportnet.c next->ip = d->ip; next 124 net/netfilter/ipset/ip_set_hash_ipportnet.c next->port = d->port; next 125 net/netfilter/ipset/ip_set_hash_ipportnet.c next->ip2 = d->ip2; next 269 net/netfilter/ipset/ip_set_hash_ipportnet.c ip = ntohl(h->next.ip); next 270 net/netfilter/ipset/ip_set_hash_ipportnet.c p = ntohs(h->next.port); next 271 net/netfilter/ipset/ip_set_hash_ipportnet.c ip2 = ntohl(h->next.ip2); next 369 net/netfilter/ipset/ip_set_hash_ipportnet.c hash_ipportnet6_data_next(struct hash_ipportnet6_elem *next, next 372 net/netfilter/ipset/ip_set_hash_ipportnet.c next->port = d->port; next 494 net/netfilter/ipset/ip_set_hash_ipportnet.c port = ntohs(h->next.port); next 60 net/netfilter/ipset/ip_set_hash_mac.c hash_mac4_data_next(struct hash_mac4_elem *next, next 101 net/netfilter/ipset/ip_set_hash_net.c hash_net4_data_next(struct hash_net4_elem *next, next 104 net/netfilter/ipset/ip_set_hash_net.c next->ip = d->ip; next 191 net/netfilter/ipset/ip_set_hash_net.c ip = ntohl(h->next.ip); next 266 net/netfilter/ipset/ip_set_hash_net.c hash_net6_data_next(struct hash_net6_elem *next, next 123 net/netfilter/ipset/ip_set_hash_netiface.c hash_netiface4_data_next(struct hash_netiface4_elem *next, next 126 net/netfilter/ipset/ip_set_hash_netiface.c next->ip = d->ip; next 253 net/netfilter/ipset/ip_set_hash_netiface.c ip = ntohl(h->next.ip); next 346 net/netfilter/ipset/ip_set_hash_netiface.c hash_netiface6_data_next(struct hash_netiface6_elem *next, next 121 net/netfilter/ipset/ip_set_hash_netnet.c hash_netnet4_data_next(struct hash_netnet4_elem *next, next 124 net/netfilter/ipset/ip_set_hash_netnet.c next->ipcmp = d->ipcmp; next 248 net/netfilter/ipset/ip_set_hash_netnet.c ip = ntohl(h->next.ip[0]); next 249 net/netfilter/ipset/ip_set_hash_netnet.c ip2 = ntohl(h->next.ip[1]); next 352 net/netfilter/ipset/ip_set_hash_netnet.c hash_netnet6_data_next(struct hash_netnet6_elem *next, next 116 net/netfilter/ipset/ip_set_hash_netport.c hash_netport4_data_next(struct hash_netport4_elem *next, next 119 net/netfilter/ipset/ip_set_hash_netport.c next->ip = d->ip; next 120 net/netfilter/ipset/ip_set_hash_netport.c next->port = d->port; next 239 net/netfilter/ipset/ip_set_hash_netport.c ip = ntohl(h->next.ip); next 240 net/netfilter/ipset/ip_set_hash_netport.c p = ntohs(h->next.port); next 329 net/netfilter/ipset/ip_set_hash_netport.c hash_netport6_data_next(struct hash_netport6_elem *next, next 332 net/netfilter/ipset/ip_set_hash_netport.c next->port = d->port; next 442 net/netfilter/ipset/ip_set_hash_netport.c port = ntohs(h->next.port); next 130 net/netfilter/ipset/ip_set_hash_netportnet.c hash_netportnet4_data_next(struct hash_netportnet4_elem *next, next 133 net/netfilter/ipset/ip_set_hash_netportnet.c next->ipcmp = d->ipcmp; next 134 net/netfilter/ipset/ip_set_hash_netportnet.c next->port = d->port; next 288 net/netfilter/ipset/ip_set_hash_netportnet.c ip = ntohl(h->next.ip[0]); next 289 net/netfilter/ipset/ip_set_hash_netportnet.c p = ntohs(h->next.port); next 290 net/netfilter/ipset/ip_set_hash_netportnet.c ip2 = ntohl(h->next.ip[1]); next 406 net/netfilter/ipset/ip_set_hash_netportnet.c hash_netportnet6_data_next(struct hash_netportnet6_elem *next, next 409 net/netfilter/ipset/ip_set_hash_netportnet.c next->port = d->port; next 539 net/netfilter/ipset/ip_set_hash_netportnet.c port = ntohs(h->next.port); next 190 net/netfilter/ipset/ip_set_list_set.c struct set_elem *e, *next, *prev = NULL; next 205 net/netfilter/ipset/ip_set_list_set.c next = list_next_entry(e, list); next 207 net/netfilter/ipset/ip_set_list_set.c next->id == d->refid; next 237 net/netfilter/ipset/ip_set_list_set.c struct set_elem *e, *n, *prev, *next; next 241 net/netfilter/ipset/ip_set_list_set.c n = prev = next = NULL; next 251 net/netfilter/ipset/ip_set_list_set.c next = e; next 257 net/netfilter/ipset/ip_set_list_set.c if ((d->before > 0 && !next) || next 280 net/netfilter/ipset/ip_set_list_set.c if (!list_is_last(&next->list, &map->members)) next 281 net/netfilter/ipset/ip_set_list_set.c n = list_next_entry(next, list); next 302 net/netfilter/ipset/ip_set_list_set.c else if (next) next 303 net/netfilter/ipset/ip_set_list_set.c list_add_tail_rcu(&e->list, &next->list); next 319 net/netfilter/ipset/ip_set_list_set.c struct set_elem *e, *next, *prev = NULL; next 331 net/netfilter/ipset/ip_set_list_set.c next = list_next_entry(e, list); next 333 net/netfilter/ipset/ip_set_list_set.c next->id != d->refid) next 557 net/netfilter/ipvs/ip_vs_app.c if ((e = inc->a_list.next) != &app->incs_list) next 561 net/netfilter/ipvs/ip_vs_app.c for (e = app->a_list.next; e != &ipvs->app_list; e = e->next) { next 593 net/netfilter/ipvs/ip_vs_app.c .next = ip_vs_app_seq_next, next 1143 net/netfilter/ipvs/ip_vs_conn.c .next = ip_vs_conn_seq_next, next 1209 net/netfilter/ipvs/ip_vs_conn.c .next = ip_vs_conn_seq_next, next 1236 net/netfilter/ipvs/ip_vs_ctl.c struct ip_vs_dest *dest, *next; next 1240 net/netfilter/ipvs/ip_vs_ctl.c list_for_each_entry_safe(dest, next, &ipvs->dest_trash, t_list) { next 2197 net/netfilter/ipvs/ip_vs_ctl.c .next = ip_vs_info_seq_next, next 117 net/netfilter/ipvs/ip_vs_dh.c p = p->next; next 123 net/netfilter/ipvs/ip_vs_dh.c p = p->next; next 233 net/netfilter/ipvs/ip_vs_lblc.c struct hlist_node *next; next 239 net/netfilter/ipvs/ip_vs_lblc.c hlist_for_each_entry_safe(en, next, &tbl->bucket[i], list) { next 260 net/netfilter/ipvs/ip_vs_lblc.c struct hlist_node *next; next 268 net/netfilter/ipvs/ip_vs_lblc.c hlist_for_each_entry_safe(en, next, &tbl->bucket[j], list) { next 302 net/netfilter/ipvs/ip_vs_lblc.c struct hlist_node *next; next 324 net/netfilter/ipvs/ip_vs_lblc.c hlist_for_each_entry_safe(en, next, &tbl->bucket[j], list) { next 400 net/netfilter/ipvs/ip_vs_lblcr.c struct hlist_node *next; next 405 net/netfilter/ipvs/ip_vs_lblcr.c hlist_for_each_entry_safe(en, next, &tbl->bucket[i], list) { next 427 net/netfilter/ipvs/ip_vs_lblcr.c struct hlist_node *next; next 433 net/netfilter/ipvs/ip_vs_lblcr.c hlist_for_each_entry_safe(en, next, &tbl->bucket[j], list) { next 466 net/netfilter/ipvs/ip_vs_lblcr.c struct hlist_node *next; next 488 net/netfilter/ipvs/ip_vs_lblcr.c hlist_for_each_entry_safe(en, next, &tbl->bucket[j], list) { next 140 net/netfilter/ipvs/ip_vs_mh.c while ((p = p->next) != &svc->destinations) { next 187 net/netfilter/ipvs/ip_vs_mh.c p = p->next; next 193 net/netfilter/ipvs/ip_vs_mh.c p = p->next; next 223 net/netfilter/ipvs/ip_vs_mh.c p = p->next; next 53 net/netfilter/ipvs/ip_vs_proto.c pp->next = ip_vs_proto_table[hash]; next 76 net/netfilter/ipvs/ip_vs_proto.c pd->next = ipvs->proto_data_table[hash]; next 84 net/netfilter/ipvs/ip_vs_proto.c ipvs->proto_data_table[hash] = pd->next; next 102 net/netfilter/ipvs/ip_vs_proto.c for (; *pp_p; pp_p = &(*pp_p)->next) { next 104 net/netfilter/ipvs/ip_vs_proto.c *pp_p = pp->next; next 124 net/netfilter/ipvs/ip_vs_proto.c for (; *pd_p; pd_p = &(*pd_p)->next) { next 126 net/netfilter/ipvs/ip_vs_proto.c *pd_p = pd->next; next 145 net/netfilter/ipvs/ip_vs_proto.c for (pp = ip_vs_proto_table[hash]; pp; pp = pp->next) { next 163 net/netfilter/ipvs/ip_vs_proto.c for (pd = ipvs->proto_data_table[hash]; pd; pd = pd->next) { next 181 net/netfilter/ipvs/ip_vs_proto.c for (pd = ipvs->proto_data_table[i]; pd; pd = pd->next) { next 43 net/netfilter/ipvs/ip_vs_rr.c svc->sched_data = p->next->prev; next 184 net/netfilter/ipvs/ip_vs_sh.c p = p->next; next 196 net/netfilter/ipvs/ip_vs_sh.c p = p->next; next 316 net/netfilter/ipvs/ip_vs_sync.c sb = list_entry(ms->sync_queue.next, struct ip_vs_sync_buff, next 469 net/netfilter/nf_conncount.c goto next; /* do not bother */ next 488 net/netfilter/nf_conncount.c next: next 216 net/netfilter/nf_conntrack_expect.c struct hlist_node *next; next 223 net/netfilter/nf_conntrack_expect.c hlist_for_each_entry_safe(exp, next, &help->expectations, lnode) { next 422 net/netfilter/nf_conntrack_expect.c struct hlist_node *next; next 431 net/netfilter/nf_conntrack_expect.c hlist_for_each_entry_safe(i, next, &nf_ct_expect_hash[h], hnode) { next 494 net/netfilter/nf_conntrack_expect.c const struct hlist_node *next; next 500 net/netfilter/nf_conntrack_expect.c hlist_for_each_entry_safe(exp, next, next 520 net/netfilter/nf_conntrack_expect.c const struct hlist_node *next; next 526 net/netfilter/nf_conntrack_expect.c hlist_for_each_entry_safe(exp, next, next 652 net/netfilter/nf_conntrack_expect.c .next = exp_seq_next, next 808 net/netfilter/nf_conntrack_sip.c struct hlist_node *next; next 812 net/netfilter/nf_conntrack_sip.c hlist_for_each_entry_safe(exp, next, &help->expectations, lnode) { next 832 net/netfilter/nf_conntrack_sip.c struct hlist_node *next; next 835 net/netfilter/nf_conntrack_sip.c hlist_for_each_entry_safe(exp, next, &help->expectations, lnode) { next 380 net/netfilter/nf_conntrack_standalone.c .next = ct_seq_next, next 459 net/netfilter/nf_conntrack_standalone.c .next = ct_cpu_seq_next, next 394 net/netfilter/nf_log.c .next = seq_next, next 191 net/netfilter/nf_nat_sip.c goto next; next 196 net/netfilter/nf_nat_sip.c goto next; next 259 net/netfilter/nf_nat_sip.c next: next 302 net/netfilter/nf_synproxy_core.c .next = synproxy_cpu_seq_next, next 2350 net/netfilter/nf_tables_api.c const struct nft_expr *expr, *next; next 2381 net/netfilter/nf_tables_api.c nft_rule_for_each_expr(expr, next, rule) { next 2637 net/netfilter/nf_tables_api.c struct nft_expr *expr, *next; next 2645 net/netfilter/nf_tables_api.c next = nft_expr_next(expr); next 2647 net/netfilter/nf_tables_api.c expr = next; next 6590 net/netfilter/nf_tables_api.c struct nft_trans *trans, *next; next 6602 net/netfilter/nf_tables_api.c list_for_each_entry_safe(trans, next, &head, list) { next 6641 net/netfilter/nf_tables_api.c struct nft_trans *trans, *next; next 6643 net/netfilter/nf_tables_api.c list_for_each_entry_safe(trans, next, &net->nft.commit_list, list) { next 6741 net/netfilter/nf_tables_api.c struct nft_module_request *req, *next; next 6744 net/netfilter/nf_tables_api.c list_for_each_entry_safe(req, next, &net->nft.module_list, list) { next 6786 net/netfilter/nf_tables_api.c struct nft_trans *trans, *next; next 6806 net/netfilter/nf_tables_api.c list_for_each_entry_safe(trans, next, &net->nft.commit_list, list) { next 6836 net/netfilter/nf_tables_api.c list_for_each_entry_safe(trans, next, &net->nft.commit_list, list) { next 6972 net/netfilter/nf_tables_api.c struct nft_module_request *req, *next; next 6977 net/netfilter/nf_tables_api.c list_for_each_entry_safe(req, next, &module_list, list) { next 7016 net/netfilter/nf_tables_api.c struct nft_trans *trans, *next; next 7019 net/netfilter/nf_tables_api.c list_for_each_entry_safe_reverse(trans, next, &net->nft.commit_list, next 7130 net/netfilter/nf_tables_api.c list_for_each_entry_safe_reverse(trans, next, next 199 net/netfilter/nf_tables_offload.c struct flow_block_cb *block_cb, *next; next 201 net/netfilter/nf_tables_offload.c list_for_each_entry_safe(block_cb, next, &bo->cb_list, list) { next 274 net/netfilter/nfnetlink.c struct nfnl_err *nfnl_err, *next; next 276 net/netfilter/nfnetlink.c list_for_each_entry_safe(nfnl_err, next, err_list, head) next 282 net/netfilter/nfnetlink.c struct nfnl_err *nfnl_err, *next; next 284 net/netfilter/nfnetlink.c list_for_each_entry_safe(nfnl_err, next, err_list, head) { next 1097 net/netfilter/nfnetlink_log.c .next = seq_next, next 249 net/netfilter/nfnetlink_queue.c struct nf_queue_entry *entry, *next; next 252 net/netfilter/nfnetlink_queue.c list_for_each_entry_safe(entry, next, &queue->queue_list, list) { next 755 net/netfilter/nfnetlink_queue.c if (skb->next == NULL) { /* last packet, no need to copy entry */ next 819 net/netfilter/nfnetlink_queue.c struct sk_buff *nskb = segs->next; next 1431 net/netfilter/nfnetlink_queue.c h = h->next; next 1489 net/netfilter/nfnetlink_queue.c .next = seq_next, next 618 net/netfilter/nft_set_hash.c struct hlist_node *next; next 622 net/netfilter/nft_set_hash.c hlist_for_each_entry_safe(he, next, &priv->table[i], node) { next 1523 net/netfilter/x_tables.c .next = xt_table_seq_next, next 1565 net/netfilter/x_tables.c trav->curr = trav->curr->next; next 1575 net/netfilter/x_tables.c trav->curr = trav->curr->next; next 1642 net/netfilter/x_tables.c .next = xt_match_seq_next, next 1676 net/netfilter/x_tables.c .next = xt_target_seq_next, next 1214 net/netfilter/xt_hashlimit.c .next = dl_seq_next, next 1221 net/netfilter/xt_hashlimit.c .next = dl_seq_next, next 1228 net/netfilter/xt_hashlimit.c .next = dl_seq_next, next 162 net/netfilter/xt_recent.c e = list_entry(t->lru_list.next, struct recent_entry, lru_list); next 179 net/netfilter/xt_recent.c e = list_entry(t->lru_list.next, struct recent_entry, lru_list); next 224 net/netfilter/xt_recent.c struct recent_entry *e, *next; next 228 net/netfilter/xt_recent.c list_for_each_entry_safe(e, next, &t->iphash[i], list) next 493 net/netfilter/xt_recent.c const struct list_head *head = e->list.next; next 499 net/netfilter/xt_recent.c head = t->iphash[st->bucket].next; next 533 net/netfilter/xt_recent.c .next = recent_seq_next, next 64 net/netlabel/netlabel_addrlist.h i = i->next; next 83 net/netlabel/netlabel_addrlist.h for (iter = __af4list_valid((head)->next, head); \ next 85 net/netlabel/netlabel_addrlist.h iter = __af4list_valid(iter->list.next, head)) next 88 net/netlabel/netlabel_addrlist.h for (iter = __af4list_valid_rcu((head)->next, head); \ next 90 net/netlabel/netlabel_addrlist.h iter = __af4list_valid_rcu(iter->list.next, head)) next 93 net/netlabel/netlabel_addrlist.h for (iter = __af4list_valid((head)->next, head), \ next 94 net/netlabel/netlabel_addrlist.h tmp = __af4list_valid(iter->list.next, head); \ next 96 net/netlabel/netlabel_addrlist.h iter = tmp, tmp = __af4list_valid(iter->list.next, head)) next 131 net/netlabel/netlabel_addrlist.h i = i->next; next 150 net/netlabel/netlabel_addrlist.h for (iter = __af6list_valid((head)->next, head); \ next 152 net/netlabel/netlabel_addrlist.h iter = __af6list_valid(iter->list.next, head)) next 155 net/netlabel/netlabel_addrlist.h for (iter = __af6list_valid_rcu((head)->next, head); \ next 157 net/netlabel/netlabel_addrlist.h iter = __af6list_valid_rcu(iter->list.next, head)) next 160 net/netlabel/netlabel_addrlist.h for (iter = __af6list_valid((head)->next, head), \ next 161 net/netlabel/netlabel_addrlist.h tmp = __af6list_valid(iter->list.next, head); \ next 163 net/netlabel/netlabel_addrlist.h iter = tmp, tmp = __af6list_valid(iter->list.next, head)) next 568 net/netlabel/netlabel_kapi.c iter = iter->next; next 588 net/netlabel/netlabel_kapi.c iter->next = *catmap; next 591 net/netlabel/netlabel_kapi.c iter->next = prev->next; next 592 net/netlabel/netlabel_kapi.c prev->next = iter; next 638 net/netlabel/netlabel_kapi.c if (iter->next != NULL) { next 639 net/netlabel/netlabel_kapi.c iter = iter->next; next 698 net/netlabel/netlabel_kapi.c if (iter->next == NULL) next 701 net/netlabel/netlabel_kapi.c iter = iter->next; next 2647 net/netlink/af_netlink.c .next = netlink_seq_next, next 1306 net/netrom/af_netrom.c .next = nr_info_next, next 891 net/netrom/nr_route.c .next = nr_node_next, next 945 net/netrom/nr_route.c .next = nr_neigh_next, next 1389 net/nfc/llcp_core.c goto next; next 1402 net/nfc/llcp_core.c next: next 116 net/nsh/nsh.c for (skb = segs; skb; skb = skb->next) { next 332 net/openvswitch/datapath.c } while ((skb = skb->next)); next 337 net/openvswitch/datapath.c nskb = skb->next; next 45 net/openvswitch/flow_netlink.c const struct ovs_len_tbl *next; next 390 net/openvswitch/flow_netlink.c .next = ovs_vxlan_ext_key_lens }, next 426 net/openvswitch/flow_netlink.c .next = ovs_tunnel_key_lens, }, next 437 net/openvswitch/flow_netlink.c .next = ovs_nsh_key_attr_lens, }, next 1728 net/openvswitch/flow_netlink.c nlattr_set(nla, val, tbl[nla_type(nla)].next ? : tbl); next 3544 net/packet/af_packet.c *mlp = ml->next; next 3547 net/packet/af_packet.c mlp = &ml->next; next 3575 net/packet/af_packet.c for (ml = po->mclist; ml; ml = ml->next) { next 3593 net/packet/af_packet.c i->next = po->mclist; next 3597 net/packet/af_packet.c po->mclist = i->next; next 3612 net/packet/af_packet.c for (mlp = &pkt_sk(sk)->mclist; (ml = *mlp) != NULL; mlp = &ml->next) { next 3619 net/packet/af_packet.c *mlp = ml->next; next 3644 net/packet/af_packet.c po->mclist = ml->next; next 4619 net/packet/af_packet.c .next = packet_seq_next, next 48 net/packet/diag.c for (ml = po->mclist; ml; ml = ml->next) { next 205 net/packet/diag.c goto next; next 214 net/packet/diag.c next: next 8 net/packet/internal.h struct packet_mclist *next; next 107 net/phonet/pep-gprs.c skb->next = skb_shinfo(skb)->frag_list; next 1218 net/phonet/pep.c skb->next = skb_shinfo(skb)->frag_list; next 612 net/phonet/socket.c .next = pn_sock_seq_next, next 781 net/phonet/socket.c .next = pn_res_seq_next, next 665 net/rds/af_rds.c WARN_ON((&rs->rs_item != rs->rs_item.next || next 127 net/rds/bind.c rs->rs_bound_node.next = NULL; next 225 net/rds/ib_fmr.c struct rds_ib_mr *ibmr, *next; next 242 net/rds/ib_fmr.c list_for_each_entry_safe(ibmr, next, list, unmap_list) { next 360 net/rds/ib_frmr.c struct rds_ib_mr *ibmr, *next; next 378 net/rds/ib_frmr.c list_for_each_entry_safe(ibmr, next, list, unmap_list) { next 290 net/rds/ib_rdma.c struct llist_node *next; next 295 net/rds/ib_rdma.c next = node->next; next 298 net/rds/ib_rdma.c node = next; next 315 net/rds/ib_rdma.c struct llist_node **next = nodes_head; next 319 net/rds/ib_rdma.c *next = cur; next 320 net/rds/ib_rdma.c next = &cur->next; next 322 net/rds/ib_rdma.c *next = NULL; next 410 net/rds/ib_rdma.c clean_nodes = clean_nodes->next; next 58 net/rds/ib_recv.c recv->r_wr.next = NULL; next 523 net/rds/ib_recv.c cache->ready = head->next; next 543 net/rds/ib_recv.c frag = list_entry(ibinc->ii_frags.next, struct rds_page_frag, f_item); next 548 net/rds/ib_recv.c frag = list_entry(frag->f_item.next, next 797 net/rds/ib_recv.c frag = list_entry(ibinc->ii_frags.next, struct rds_page_frag, f_item); next 831 net/rds/ib_recv.c frag = list_entry(frag->f_item.next, next 630 net/rds/ib_send.c send->s_wr.next = NULL; next 674 net/rds/ib_send.c &send->s_wr, send->s_wr.num_sge, send->s_wr.next); next 687 net/rds/ib_send.c prev->s_wr.next = &send->s_wr; next 790 net/rds/ib_send.c send->s_atomic_wr.wr.next = NULL; next 915 net/rds/ib_send.c send->s_rdma_wr.wr.next = NULL; next 918 net/rds/ib_send.c prev->s_rdma_wr.wr.next = &send->s_rdma_wr.wr; next 937 net/rds/ib_send.c send->s_rdma_wr.wr.next); next 410 net/rds/recv.c *inc = list_entry(rs->rs_recv_queue.next, next 475 net/rds/recv.c notifier = list_entry(rs->rs_notify_queue.next, next 486 net/rds/recv.c notifier = list_entry(copy.next, struct rds_notifier, n_list); next 613 net/rds/recv.c info = list_entry(q->zcookie_head.next, next 255 net/rds/send.c rm = list_entry(cp->cp_send_queue.next, next 614 net/rds/send.c rm = list_entry(messages->next, struct rds_message, next 792 net/rds/send.c rm = list_entry(list.next, struct rds_message, m_sock_item); next 1423 net/rose/af_rose.c .next = rose_info_next, next 67 net/rose/rose_route.c rose_node = rose_node->next; next 81 net/rose/rose_route.c rose_neigh = rose_neigh->next; next 126 net/rose/rose_route.c rose_neigh->next = rose_neigh_list; next 144 net/rose/rose_route.c rose_tmpn = rose_tmpn->next; next 166 net/rose/rose_route.c rose_node->next = NULL; next 168 net/rose/rose_route.c rose_tmpp->next = rose_node; next 169 net/rose/rose_route.c rose_node->next = NULL; next 173 net/rose/rose_route.c rose_node->next = rose_node_list; next 176 net/rose/rose_route.c rose_tmpp->next = rose_node; next 177 net/rose/rose_route.c rose_node->next = rose_tmpn; next 207 net/rose/rose_route.c rose_node_list = rose_node->next; next 212 net/rose/rose_route.c while (s != NULL && s->next != NULL) { next 213 net/rose/rose_route.c if (s->next == rose_node) { next 214 net/rose/rose_route.c s->next = rose_node->next; next 219 net/rose/rose_route.c s = s->next; next 236 net/rose/rose_route.c rose_neigh_list = rose_neigh->next; next 244 net/rose/rose_route.c while (s != NULL && s->next != NULL) { next 245 net/rose/rose_route.c if (s->next == rose_neigh) { next 246 net/rose/rose_route.c s->next = rose_neigh->next; next 254 net/rose/rose_route.c s = s->next; next 272 net/rose/rose_route.c rose_route_list = rose_route->next; next 277 net/rose/rose_route.c while (s != NULL && s->next != NULL) { next 278 net/rose/rose_route.c if (s->next == rose_route) { next 279 net/rose/rose_route.c s->next = rose_route->next; next 284 net/rose/rose_route.c s = s->next; next 308 net/rose/rose_route.c rose_node = rose_node->next; next 322 net/rose/rose_route.c rose_neigh = rose_neigh->next; next 395 net/rose/rose_route.c sn->next = rose_neigh_list; next 416 net/rose/rose_route.c rose_node = rose_node->next; next 434 net/rose/rose_route.c rose_node->next = rose_node_list; next 460 net/rose/rose_route.c rose_node = rose_node->next; next 488 net/rose/rose_route.c rose_neigh = rose_neigh->next; next 497 net/rose/rose_route.c rose_node = rose_node->next; next 538 net/rose/rose_route.c rose_route = rose_route->next; next 565 net/rose/rose_route.c rose_node = rose_node->next; next 572 net/rose/rose_route.c rose_neigh = rose_neigh->next; next 663 net/rose/rose_route.c for (rose_route = rose_route_list; rose_route != NULL; rose_route = rose_route->next) next 683 net/rose/rose_route.c for (node = rose_node_list; node != NULL; node = node->next) { next 694 net/rose/rose_route.c for (node = rose_node_list; node != NULL; node = node->next) { next 782 net/rose/rose_route.c s = rose_route->next; next 800 net/rose/rose_route.c rose_route = rose_route->next; next 819 net/rose/rose_route.c rose_neigh = rose_neigh->next; next 840 net/rose/rose_route.c for (rose_neigh = rose_neigh_list; rose_neigh != NULL; rose_neigh = rose_neigh->next) { next 890 net/rose/rose_route.c rose_neigh = rose_neigh->next; next 1005 net/rose/rose_route.c rose_route = rose_route->next; next 1039 net/rose/rose_route.c rose_route = rose_route->next; next 1070 net/rose/rose_route.c rose_route->next = rose_route_list; next 1100 net/rose/rose_route.c rose_node = rose_node->next, ++i); next 1110 net/rose/rose_route.c : ((struct rose_node *)v)->next; next 1150 net/rose/rose_route.c .next = rose_node_next, next 1166 net/rose/rose_route.c rose_neigh = rose_neigh->next, ++i); next 1176 net/rose/rose_route.c : ((struct rose_neigh *)v)->next; next 1221 net/rose/rose_route.c .next = rose_neigh_next, next 1237 net/rose/rose_route.c rose_route = rose_route->next, ++i); next 1247 net/rose/rose_route.c : ((struct rose_route *)v)->next; next 1293 net/rose/rose_route.c .next = rose_route_next, next 1310 net/rose/rose_route.c rose_neigh = rose_neigh->next; next 1317 net/rose/rose_route.c rose_node = rose_node->next; next 1324 net/rose/rose_route.c rose_route = rose_route->next; next 501 net/rxrpc/call_accept.c call = list_entry(rx->to_be_accepted.next, next 607 net/rxrpc/call_accept.c call = list_entry(rx->to_be_accepted.next, next 301 net/rxrpc/call_event.c unsigned long now, next, t; next 425 net/rxrpc/call_event.c next = call->expect_rx_by; next 427 net/rxrpc/call_event.c #define set(T) { t = READ_ONCE(T); if (time_before(t, next)) next = t; } next 438 net/rxrpc/call_event.c if (time_after_eq(now, next)) next 441 net/rxrpc/call_event.c rxrpc_reduce_call_timer(call, next, now, rxrpc_timer_restart); next 475 net/rxrpc/call_object.c call->recvmsg_link.next = NULL; next 514 net/rxrpc/call_object.c call = list_entry(rx->to_be_accepted.next, next 522 net/rxrpc/call_object.c call = list_entry(rx->sock_calls.next, next 631 net/rxrpc/call_object.c call = list_entry(rxnet->calls.next, next 552 net/rxrpc/conn_client.c struct rxrpc_call *call = list_entry(conn->waiting_calls.next, next 930 net/rxrpc/conn_client.c if (rxnet->idle_client_conns.next == &conn->cache_link && next 947 net/rxrpc/conn_client.c struct rxrpc_connection *next = NULL; next 973 net/rxrpc/conn_client.c next = list_entry(rxnet->waiting_client_conns.next, next 975 net/rxrpc/conn_client.c rxrpc_get_connection(next); next 976 net/rxrpc/conn_client.c rxrpc_activate_conn(rxnet, next); next 983 net/rxrpc/conn_client.c if (next) next 984 net/rxrpc/conn_client.c rxrpc_activate_channels(next); next 989 net/rxrpc/conn_client.c return next; next 1035 net/rxrpc/conn_client.c conn = list_entry(rxnet->active_client_conns.next, next 1093 net/rxrpc/conn_client.c next: next 1099 net/rxrpc/conn_client.c conn = list_entry(rxnet->idle_client_conns.next, next 1136 net/rxrpc/conn_client.c goto next; next 1211 net/rxrpc/conn_client.c conn = list_entry(graveyard.next, next 441 net/rxrpc/conn_object.c conn = list_entry(graveyard.next, struct rxrpc_connection, next 223 net/rxrpc/input.c skb->next = list; next 243 net/rxrpc/input.c list = skb->next; next 157 net/rxrpc/key.c pptoken = &(*pptoken)->next) next 534 net/rxrpc/key.c pptoken = &(*pptoken)->next) next 781 net/rxrpc/key.c pp = &(*pp)->next; next 800 net/rxrpc/key.c struct rxrpc_key_token *next; next 802 net/rxrpc/key.c for (; token; token = next) { next 803 net/rxrpc/key.c next = token->next; next 1071 net/rxrpc/key.c for (token = key->payload.data[0]; token; token = token->next) { next 1159 net/rxrpc/key.c for (token = key->payload.data[0]; token; token = token->next) { next 236 net/rxrpc/local_object.c for (cursor = rxnet->local_endpoints.next; next 238 net/rxrpc/local_object.c cursor = cursor->next) { next 314 net/rxrpc/peer_event.c peer = list_entry(collector->next, next 124 net/rxrpc/proc.c .next = rxrpc_call_seq_next, next 207 net/rxrpc/proc.c .next = rxrpc_connection_seq_next, next 333 net/rxrpc/proc.c .next = rxrpc_peer_seq_next, next 501 net/rxrpc/recvmsg.c l = rx->recvmsg_q.next; next 588 net/sched/cls_api.c tp_next = rcu_dereference_protected(tp->next, 1); next 599 net/sched/cls_api.c tp_next = rcu_dereference_protected(tp->next, 1); next 1005 net/sched/cls_api.c for (; tp; tp = tcf_chain_dereference(tp->next, chain)) next 1009 net/sched/cls_api.c tp = tcf_chain_dereference(tp->next, chain); next 1480 net/sched/cls_api.c struct flow_block_cb *block_cb, *next; next 1502 net/sched/cls_api.c list_for_each_entry_safe(block_cb, next, &bo->cb_list, list) { next 1521 net/sched/cls_api.c struct flow_block_cb *block_cb, *next; next 1525 net/sched/cls_api.c list_for_each_entry_safe(block_cb, next, &bo->cb_list, list) { next 1573 net/sched/cls_api.c for (; tp; tp = rcu_dereference_bh(tp->next)) { next 1626 net/sched/cls_api.c struct tcf_proto __rcu *next; next 1645 net/sched/cls_api.c RCU_INIT_POINTER(tp->next, tcf_chain_tp_prev(chain, chain_info)); next 1655 net/sched/cls_api.c struct tcf_proto *next = tcf_chain_dereference(chain_info->next, chain); next 1659 net/sched/cls_api.c tcf_chain0_head_change(chain, next); next 1660 net/sched/cls_api.c RCU_INIT_POINTER(*chain_info->pprev, next); next 1714 net/sched/cls_api.c struct tcf_proto *next; next 1721 net/sched/cls_api.c pprev = &tp_iter->next) { next 1724 net/sched/cls_api.c chain_info.next = tp_iter->next; next 1739 net/sched/cls_api.c next = tcf_chain_dereference(chain_info.next, chain); next 1741 net/sched/cls_api.c tcf_chain0_head_change(chain, next); next 1742 net/sched/cls_api.c RCU_INIT_POINTER(*chain_info.pprev, next); next 1759 net/sched/cls_api.c pprev = &tp->next) { next 1773 net/sched/cls_api.c chain_info->next = tp->next; next 1776 net/sched/cls_api.c chain_info->next = NULL; next 595 net/sched/cls_flow.c struct flow_filter *f, *next; next 597 net/sched/cls_flow.c list_for_each_entry_safe(f, next, &head->filters, list) { next 567 net/sched/cls_flower.c struct cls_fl_filter *f, *next; next 571 net/sched/cls_flower.c list_for_each_entry_safe(f, next, &mask->filters, list) { next 34 net/sched/cls_fw.c struct fw_filter __rcu *next; next 62 net/sched/cls_fw.c f = rcu_dereference_bh(f->next)) { next 98 net/sched/cls_fw.c for (; f; f = rtnl_dereference(f->next)) { next 143 net/sched/cls_fw.c rtnl_dereference(f->next)); next 170 net/sched/cls_fw.c fp = &pfp->next, pfp = rtnl_dereference(*fp)) { next 172 net/sched/cls_fw.c RCU_INIT_POINTER(*fp, rtnl_dereference(f->next)); next 289 net/sched/cls_fw.c fp = &pfp->next, pfp = rtnl_dereference(*fp)) next 293 net/sched/cls_fw.c RCU_INIT_POINTER(fnew->next, rtnl_dereference(pfp->next)); next 333 net/sched/cls_fw.c RCU_INIT_POINTER(f->next, head->ht[fw_hash(handle)]); next 361 net/sched/cls_fw.c f = rtnl_dereference(f->next)) { next 47 net/sched/cls_route.c struct route4_filter __rcu *next; next 166 net/sched/cls_route.c f = rcu_dereference_bh(f->next)) next 172 net/sched/cls_route.c f = rcu_dereference_bh(f->next)) next 178 net/sched/cls_route.c f = rcu_dereference_bh(f->next)) next 234 net/sched/cls_route.c f = rtnl_dereference(f->next)) next 293 net/sched/cls_route.c struct route4_filter *next; next 295 net/sched/cls_route.c next = rtnl_dereference(f->next); next 296 net/sched/cls_route.c RCU_INIT_POINTER(b->ht[h2], next); next 330 net/sched/cls_route.c fp = &nf->next, nf = rtnl_dereference(*fp)) { next 333 net/sched/cls_route.c RCU_INIT_POINTER(*fp, rtnl_dereference(f->next)); next 440 net/sched/cls_route.c fp = rtnl_dereference(fp->next)) next 521 net/sched/cls_route.c fp = &f1->next) next 526 net/sched/cls_route.c rcu_assign_pointer(f->next, f1); next 536 net/sched/cls_route.c fp = &pfp->next, pfp = rtnl_dereference(*fp)) { next 538 net/sched/cls_route.c rcu_assign_pointer(*fp, fold->next); next 579 net/sched/cls_route.c f = rtnl_dereference(f->next)) { next 74 net/sched/cls_rsvp.h struct rsvp_session __rcu *next; next 86 net/sched/cls_rsvp.h struct rsvp_filter __rcu *next; next 171 net/sched/cls_rsvp.h s = rcu_dereference_bh(s->next)) { next 184 net/sched/cls_rsvp.h f = rcu_dereference_bh(f->next)) { next 209 net/sched/cls_rsvp.h f = rcu_dereference_bh(f->next)) { next 230 net/sched/cls_rsvp.h s = rtnl_dereference(s->next)) { next 232 net/sched/cls_rsvp.h ins = &pins->next, pins = rtnl_dereference(*ins)) { next 234 net/sched/cls_rsvp.h RCU_INIT_POINTER(n->next, pins->next); next 259 net/sched/cls_rsvp.h s = rtnl_dereference(s->next)) { next 261 net/sched/cls_rsvp.h f = rtnl_dereference(f->next)) { next 324 net/sched/cls_rsvp.h RCU_INIT_POINTER(data->ht[h1], s->next); next 330 net/sched/cls_rsvp.h rcu_assign_pointer(s->ht[h2], f->next); next 353 net/sched/cls_rsvp.h fp = &nfp->next, nfp = rtnl_dereference(*fp)) { next 355 net/sched/cls_rsvp.h RCU_INIT_POINTER(*fp, f->next); next 367 net/sched/cls_rsvp.h sp = &nsp->next, nsp = rtnl_dereference(*sp)) { next 369 net/sched/cls_rsvp.h RCU_INIT_POINTER(*sp, s->next); next 430 net/sched/cls_rsvp.h s = rtnl_dereference(s->next)) { next 435 net/sched/cls_rsvp.h f = rtnl_dereference(f->next)) { next 586 net/sched/cls_rsvp.h sp = &s->next) { next 610 net/sched/cls_rsvp.h fp = &nfp->next, nfp = rtnl_dereference(*fp)) { next 616 net/sched/cls_rsvp.h RCU_INIT_POINTER(f->next, nfp); next 639 net/sched/cls_rsvp.h sp = &nsp->next, nsp = rtnl_dereference(*sp)) { next 643 net/sched/cls_rsvp.h RCU_INIT_POINTER(s->next, nsp); next 669 net/sched/cls_rsvp.h s = rtnl_dereference(s->next)) { next 674 net/sched/cls_rsvp.h f = rtnl_dereference(f->next)) { next 42 net/sched/cls_tcindex.c struct tcindex_filter __rcu *next; next 93 net/sched/cls_tcindex.c fp = &f->next, f = rcu_dereference_bh_rtnl(*fp)) next 215 net/sched/cls_tcindex.c walk = &f->next, f = rtnl_dereference(*walk)) { next 223 net/sched/cls_tcindex.c rcu_assign_pointer(*walk, rtnl_dereference(f->next)); next 461 net/sched/cls_tcindex.c f->next = NULL; next 498 net/sched/cls_tcindex.c fp = &nfp->next, nfp = rtnl_dereference(*fp)) next 554 net/sched/cls_tcindex.c struct tcindex_filter *f, *next; next 574 net/sched/cls_tcindex.c for (f = rtnl_dereference(p->h[i]); f; f = next) { next 575 net/sched/cls_tcindex.c next = rtnl_dereference(f->next); next 617 net/sched/cls_tcindex.c struct tcindex_filter *f, *next; next 620 net/sched/cls_tcindex.c for (f = rtnl_dereference(p->h[i]); f; f = next) { next 621 net/sched/cls_tcindex.c next = rtnl_dereference(f->next); next 666 net/sched/cls_tcindex.c fp = &f->next, f = rtnl_dereference(*fp)) { next 44 net/sched/cls_u32.c struct tc_u_knode __rcu *next; next 70 net/sched/cls_u32.c struct tc_u_hnode __rcu *next; next 135 net/sched/cls_u32.c n = rcu_dereference_bh(n->next); next 141 net/sched/cls_u32.c n = rcu_dereference_bh(n->next); next 159 net/sched/cls_u32.c n = rcu_dereference_bh(n->next); next 175 net/sched/cls_u32.c n = rcu_dereference_bh(n->next); next 183 net/sched/cls_u32.c n = rcu_dereference_bh(n->next); next 189 net/sched/cls_u32.c n = rcu_dereference_bh(n->next); next 260 net/sched/cls_u32.c ht = rtnl_dereference(ht->next)) next 278 net/sched/cls_u32.c n = rtnl_dereference(n->next)) next 380 net/sched/cls_u32.c RCU_INIT_POINTER(root_ht->next, tp_c->hlist); next 454 net/sched/cls_u32.c kp = &pkp->next, pkp = rtnl_dereference(*kp)) { next 456 net/sched/cls_u32.c RCU_INIT_POINTER(*kp, key->next); next 578 net/sched/cls_u32.c rtnl_dereference(n->next)); next 605 net/sched/cls_u32.c hn = &phn->next, phn = rtnl_dereference(*hn)) { next 610 net/sched/cls_u32.c RCU_INIT_POINTER(*hn, ht->next); next 637 net/sched/cls_u32.c RCU_INIT_POINTER(tp_c->hlist, ht->next); next 783 net/sched/cls_u32.c ins = &pins->next, pins = rtnl_dereference(*ins)) next 788 net/sched/cls_u32.c RCU_INIT_POINTER(n->next, pins->next); next 805 net/sched/cls_u32.c RCU_INIT_POINTER(new->next, n->next); next 972 net/sched/cls_u32.c RCU_INIT_POINTER(ht->next, tp_c->hlist); next 1083 net/sched/cls_u32.c ins = &pins->next, pins = rtnl_dereference(*ins)) next 1087 net/sched/cls_u32.c RCU_INIT_POINTER(n->next, pins); next 1124 net/sched/cls_u32.c ht = rtnl_dereference(ht->next)) { next 1137 net/sched/cls_u32.c n = rtnl_dereference(n->next)) { next 1222 net/sched/cls_u32.c ht = rtnl_dereference(ht->next)) { next 1240 net/sched/cls_u32.c n = rtnl_dereference(n->next)) { next 134 net/sched/sch_api.c for (qp = &qdisc_base; (q = *qp) != NULL; qp = &q->next) next 159 net/sched/sch_api.c qops->next = NULL; next 178 net/sched/sch_api.c for (qp = &qdisc_base; (q = *qp) != NULL; qp = &q->next) next 182 net/sched/sch_api.c *qp = q->next; next 183 net/sched/sch_api.c q->next = NULL; next 203 net/sched/sch_api.c for (q = qdisc_base; q; q = q->next) { next 353 net/sched/sch_api.c for (q = qdisc_base; q; q = q->next) { next 418 net/sched/sch_api.c for (rtab = qdisc_rtab_list; rtab; rtab = rtab->next) { next 433 net/sched/sch_api.c rtab->next = qdisc_rtab_list; next 451 net/sched/sch_api.c rtabp = &rtab->next) { next 453 net/sched/sch_api.c *rtabp = rtab->next; next 660 net/sched/sch_api.c struct hlist_node *next; next 679 net/sched/sch_api.c hlist_for_each_entry_safe(cl, next, &ohash[i], hnode) { next 821 net/sched/sch_cake.c flow->head = skb->next; next 835 net/sched/sch_cake.c flow->tail->next = skb; next 837 net/sched/sch_cake.c skb->next = NULL; next 1154 net/sched/sch_cake.c skb_prev = skb_check, skb_check = skb_check->next) { next 1249 net/sched/sch_cake.c if (elig_ack && aggressive && elig_ack->next == skb && next 1258 net/sched/sch_cake.c elig_ack_prev->next = elig_ack->next; next 1260 net/sched/sch_cake.c flow->head = elig_ack->next; next 1664 net/sched/sch_cake.c u64 next = \ next 1669 net/sched/sch_cake.c qdisc_watchdog_schedule_ns(&q->watchdog, next); next 1687 net/sched/sch_cake.c nskb = segs->next; next 1904 net/sched/sch_cake.c u64 next = min(ktime_to_ns(q->time_next_packet), next 1908 net/sched/sch_cake.c qdisc_watchdog_schedule_ns(&q->watchdog, next); next 2137 net/sched/sch_cake.c u64 next = min(ktime_to_ns(q->time_next_packet), next 2140 net/sched/sch_cake.c qdisc_watchdog_schedule_ns(&q->watchdog, next); next 2146 net/sched/sch_cake.c ktime_t next = \ next 2151 net/sched/sch_cake.c ktime_to_ns(next)); next 2966 net/sched/sch_cake.c skb = skb->next; next 1450 net/sched/sch_cbq.c struct hlist_node *next; next 1469 net/sched/sch_cbq.c hlist_for_each_entry_safe(cl, next, &q->clhash.hash[h], next 1791 net/sched/sch_cbq.c .next = NULL, next 454 net/sched/sch_drr.c struct hlist_node *next; next 460 net/sched/sch_drr.c hlist_for_each_entry_safe(cl, next, &q->clhash.hash[i], next 494 net/sched/sch_dsmark.c .next = NULL, next 124 net/sched/sch_etf.c ktime_t next; next 131 net/sched/sch_etf.c next = ktime_sub_ns(skb->tstamp, q->delta); next 132 net/sched/sch_etf.c qdisc_watchdog_schedule_ns(&q->watchdog, ktime_to_ns(next)); next 216 net/sched/sch_etf.c skb->next = NULL; next 240 net/sched/sch_etf.c skb->next = NULL; next 257 net/sched/sch_etf.c ktime_t now, next; next 281 net/sched/sch_etf.c next = ktime_sub_ns(skb->tstamp, q->delta); next 284 net/sched/sch_etf.c if (ktime_after(now, next)) next 80 net/sched/sch_fq.c struct fq_flow *next; /* next pointer in RR lists, or &detached */ next 132 net/sched/sch_fq.c f->next = &detached; next 138 net/sched/sch_fq.c return f->next == &detached; next 143 net/sched/sch_fq.c return f->next == &throttled; next 149 net/sched/sch_fq.c head->last->next = flow; next 153 net/sched/sch_fq.c flow->next = NULL; next 182 net/sched/sch_fq.c f->next = &throttled; next 364 net/sched/sch_fq.c flow->head = skb->next; next 399 net/sched/sch_fq.c flow->tail->next = skb; next 401 net/sched/sch_fq.c skb->next = NULL; next 516 net/sched/sch_fq.c head->first = f->next; next 527 net/sched/sch_fq.c head->first = f->next; next 541 net/sched/sch_fq.c head->first = f->next; next 121 net/sched/sch_fq_codel.c flow->head = skb->next; next 133 net/sched/sch_fq_codel.c flow->tail->next = skb; next 135 net/sched/sch_fq_codel.c skb->next = NULL; next 648 net/sched/sch_fq_codel.c skb = skb->next; next 130 net/sched/sch_generic.c struct sk_buff *next = skb->next; next 145 net/sched/sch_generic.c skb = next; next 166 net/sched/sch_generic.c skb->next = nskb; next 192 net/sched/sch_generic.c skb->next = nskb; next 563 net/sched/sch_generic.c .next = (struct sk_buff *)&noop_qdisc.gso_skb, next 569 net/sched/sch_generic.c .next = (struct sk_buff *)&noop_qdisc.skb_bad_txq, next 1495 net/sched/sch_hfsc.c struct hlist_node *next; next 1506 net/sched/sch_hfsc.c hlist_for_each_entry_safe(cl, next, &q->clhash.hash[i], next 184 net/sched/sch_hhf.c struct hh_flow_state *flow, *next; next 190 net/sched/sch_hhf.c list_for_each_entry_safe(flow, next, head, flowchain) { next 333 net/sched/sch_hhf.c bucket->head = skb->next; next 344 net/sched/sch_hhf.c bucket->tail->next = skb; next 346 net/sched/sch_hhf.c skb->next = NULL; next 485 net/sched/sch_hhf.c struct hh_flow_state *flow, *next; next 490 net/sched/sch_hhf.c list_for_each_entry_safe(flow, next, head, flowchain) { next 835 net/sched/sch_htb.c next: next 845 net/sched/sch_htb.c struct htb_class *next; next 852 net/sched/sch_htb.c next = htb_lookup_leaf(hprio, prio); next 855 net/sched/sch_htb.c start = next; next 856 net/sched/sch_htb.c cl = next; next 857 net/sched/sch_htb.c goto next; next 1221 net/sched/sch_htb.c struct hlist_node *next; next 1241 net/sched/sch_htb.c hlist_for_each_entry_safe(cl, next, &q->clhash.hash[i], next 393 net/sched/sch_multiq.c .next = NULL, next 385 net/sched/sch_netem.c q->t_tail->next = nskb; next 499 net/sched/sch_netem.c segs = skb->next; next 522 net/sched/sch_netem.c skb->next = segs; next 603 net/sched/sch_netem.c skb2 = segs->next; next 668 net/sched/sch_netem.c q->t_head = skb->next; next 703 net/sched/sch_netem.c skb->next = NULL; next 416 net/sched/sch_prio.c .next = NULL, next 144 net/sched/sch_qfq.c struct hlist_node next; /* Link for the slot list. */ next 751 net/sched/sch_qfq.c struct qfq_group *next; next 754 net/sched/sch_qfq.c next = qfq_ffs(q, mask); next 755 net/sched/sch_qfq.c if (qfq_gt(grp->F, next->F)) next 779 net/sched/sch_qfq.c struct qfq_group *next; next 782 net/sched/sch_qfq.c next = qfq_ffs(q, mask); next 783 net/sched/sch_qfq.c if (!qfq_gt(next->F, old_F)) next 892 net/sched/sch_qfq.c hlist_add_head(&agg->next, &grp->slots[i]); next 900 net/sched/sch_qfq.c struct qfq_aggregate, next); next 911 net/sched/sch_qfq.c hlist_del(&agg->next); next 1044 net/sched/sch_qfq.c struct qfq_group *next = qfq_ffs(q, mask); next 1045 net/sched/sch_qfq.c if (qfq_gt(roundedF, next->F)) { next 1046 net/sched/sch_qfq.c if (qfq_gt(limit, next->F)) next 1047 net/sched/sch_qfq.c agg->S = next->F; next 1342 net/sched/sch_qfq.c hlist_del(&agg->next); next 1473 net/sched/sch_qfq.c struct hlist_node *next; next 1479 net/sched/sch_qfq.c hlist_for_each_entry_safe(cl, next, &q->clhash.hash[i], next 96 net/sched/sch_sfq.c sfq_index next; next 104 net/sched/sch_sfq.c sfq_index next; /* next slot in sfq RR chain */ next 210 net/sched/sch_sfq.c n = q->dep[qlen].next; next 212 net/sched/sch_sfq.c slot->dep.next = n; next 215 net/sched/sch_sfq.c q->dep[qlen].next = x; /* sfq_dep_head(q, p)->next = x */ next 221 net/sched/sch_sfq.c n = q->slots[x].dep.next; \ next 223 net/sched/sch_sfq.c sfq_dep_head(q, p)->next = n; \ next 262 net/sched/sch_sfq.c skb->prev->next = (struct sk_buff *)slot; next 263 net/sched/sch_sfq.c skb->next = skb->prev = NULL; next 272 net/sched/sch_sfq.c slot->skblist_next = skb->next; next 273 net/sched/sch_sfq.c skb->next->prev = (struct sk_buff *)slot; next 274 net/sched/sch_sfq.c skb->next = skb->prev = NULL; next 288 net/sched/sch_sfq.c skb->next = (struct sk_buff *)slot; next 289 net/sched/sch_sfq.c slot->skblist_prev->next = skb; next 303 net/sched/sch_sfq.c x = q->dep[d].next; next 318 net/sched/sch_sfq.c x = q->tail->next; next 320 net/sched/sch_sfq.c q->tail->next = slot->next; next 368 net/sched/sch_sfq.c x = q->dep[0].next; /* get a free slot */ next 448 net/sched/sch_sfq.c slot->next = x; next 450 net/sched/sch_sfq.c slot->next = q->tail->next; next 451 net/sched/sch_sfq.c q->tail->next = x; next 492 net/sched/sch_sfq.c a = q->tail->next; next 508 net/sched/sch_sfq.c next_a = slot->next; next 513 net/sched/sch_sfq.c q->tail->next = next_a; next 568 net/sched/sch_sfq.c x = q->dep[0].next; /* get a free slot */ next 592 net/sched/sch_sfq.c slot->next = x; next 594 net/sched/sch_sfq.c slot->next = q->tail->next; next 595 net/sched/sch_sfq.c q->tail->next = x; next 746 net/sched/sch_sfq.c q->dep[i].next = i + SFQ_MAX_FLOWS; next 695 net/sched/sch_taprio.c struct sched_entry *entry, *next; next 717 net/sched/sch_taprio.c next = list_first_entry(&oper->entries, struct sched_entry, next 719 net/sched/sch_taprio.c close_time = next->close_time; next 724 net/sched/sch_taprio.c next = list_first_entry(&oper->entries, struct sched_entry, next 729 net/sched/sch_taprio.c next = list_next_entry(entry, list); next 732 net/sched/sch_taprio.c close_time = ktime_add_ns(entry->close_time, next->interval); next 743 net/sched/sch_taprio.c next->close_time = close_time; next 744 net/sched/sch_taprio.c taprio_set_budget(q, next); next 747 net/sched/sch_taprio.c rcu_assign_pointer(q->current_entry, next); next 159 net/sched/sch_tbf.c nskb = segs->next; next 533 net/sched/sch_tbf.c .next = NULL, next 64 net/sched/sch_teql.c struct Qdisc *next; next 69 net/sched/sch_teql.c #define NEXT_SLAVE(q) (((struct teql_sched_data *)qdisc_priv(q))->next) next 203 net/sched/sch_teql.c q->next = NEXT_SLAVE(m->slaves); next 206 net/sched/sch_teql.c q->next = sch; next 499 net/sctp/associola.c pos = asoc->peer.transport_addr_list.next; next 281 net/sctp/bind_addr.c goto next; next 290 net/sctp/bind_addr.c next: next 340 net/sctp/bind_addr.c goto next; next 345 net/sctp/bind_addr.c next: next 543 net/sctp/bind_addr.c addr = list_entry(bp->address_list.next, next 34 net/sctp/diag.c laddr = list_entry(asoc->base.bind_addr.address_list.next, next 309 net/sctp/diag.c goto next; next 313 net/sctp/diag.c goto next; next 316 net/sctp/diag.c goto next; next 338 net/sctp/diag.c next: next 356 net/sctp/diag.c list_entry(ep->asocs.next, struct sctp_association, asocs); next 383 net/sctp/diag.c goto next; next 386 net/sctp/diag.c goto next; next 390 net/sctp/diag.c goto next; next 394 net/sctp/diag.c goto next; next 398 net/sctp/diag.c goto next; next 408 net/sctp/diag.c next: next 128 net/sctp/inqueue.c if (chunk->skb->next) { next 129 net/sctp/inqueue.c chunk->skb = chunk->skb->next; next 91 net/sctp/objcnt.c .next = sctp_objcnt_seq_next, next 73 net/sctp/offload.c for (skb = segs; skb; skb = skb->next) { next 395 net/sctp/output.c SCTP_OUTPUT_CB(head)->last->next = skb; next 559 net/sctp/output.c chunk = list_entry(packet->chunk_list.next, struct sctp_chunk, list); next 195 net/sctp/proc.c .next = sctp_eps_seq_next, next 287 net/sctp/proc.c .next = sctp_transport_seq_next, next 364 net/sctp/proc.c .next = sctp_transport_seq_next, next 3840 net/sctp/sm_statefuns.c entry = asoc->addip_chunk_list.next; next 245 net/sctp/socket.c asoc = list_entry(sctp_sk(sk)->ep->asocs.next, next 602 net/sctp/socket.c p = bp->address_list.next; next 4920 net/sctp/socket.c asoc = list_entry(ep->asocs.next, struct sctp_association, asocs); next 5204 net/sctp/socket.c asoc = list_entry(ep->asocs.next, next 6308 net/sctp/socket.c addr = list_entry(bp->address_list.next, next 8203 net/sctp/socket.c goto next; next 8205 net/sctp/socket.c next: next 68 net/sctp/stream_sched.c ch = list_entry(stream->out_curr->ext->outq.next, next 71 net/sctp/stream_sched.c entry = q->out_chunk_list.next; next 39 net/sctp/stream_sched_prio.c p->next = NULL; next 84 net/sctp/stream_sched_prio.c pos = p->next->prio_list.next; next 86 net/sctp/stream_sched_prio.c pos = pos->next; next 87 net/sctp/stream_sched_prio.c p->next = list_entry(pos, struct sctp_stream_out_ext, prio_list); next 100 net/sctp/stream_sched_prio.c if (prio_head->next == soute) next 110 net/sctp/stream_sched_prio.c prio_head->next = NULL; next 132 net/sctp/stream_sched_prio.c if (prio_head->next) { next 133 net/sctp/stream_sched_prio.c list_add(&soute->prio_list, prio_head->next->prio_list.prev); next 138 net/sctp/stream_sched_prio.c prio_head->next = soute; next 263 net/sctp/stream_sched_prio.c prio = list_entry(stream->prio_list.next, next 265 net/sctp/stream_sched_prio.c soute = prio->next; next 267 net/sctp/stream_sched_prio.c ch = list_entry(soute->outq.next, struct sctp_chunk, stream_list); next 31 net/sctp/stream_sched_rr.c pos = stream->rr_next->rr_list.next; next 33 net/sctp/stream_sched_rr.c pos = pos->next; next 126 net/sctp/stream_sched_rr.c ch = list_entry(soute->outq.next, struct sctp_chunk, stream_list); next 333 net/sctp/ulpqueue.c pos = f_frag->next; next 336 net/sctp/ulpqueue.c for (last = list; list; last = list, list = list->next) next 343 net/sctp/ulpqueue.c last->next = pos; next 373 net/sctp/ulpqueue.c pnext = pos->next; next 385 net/sctp/ulpqueue.c pos->next = pnext; next 990 net/sctp/ulpqueue.c for (last = flist; flist; flist = flist->next) { next 575 net/smc/smc_core.c lower = lower->next; next 725 net/smc/smc_pnet.c lower = lower->next; next 480 net/smc/smc_wr.c lnk->wr_tx_ibs[i].next = NULL; next 498 net/smc/smc_wr.c lnk->wr_rx_ibs[i].next = NULL; next 502 net/smc/smc_wr.c lnk->wr_reg.wr.next = NULL; next 155 net/strparser/strparser.c if (WARN_ON(head->next)) { next 167 net/strparser/strparser.c strp->skb_nextp = &head->next; next 212 net/strparser/strparser.c strp->skb_nextp = &skb->next; next 381 net/sunrpc/auth.c cred = list_entry(head->next, struct rpc_cred, cr_lru); next 481 net/sunrpc/auth.c struct rpc_cred *cred, *next; next 485 net/sunrpc/auth.c list_for_each_entry_safe(cred, next, &cred_unused, cr_lru) { next 407 net/sunrpc/cache.c struct list_head *next; next 415 net/sunrpc/cache.c next = current_detail->others.next; next 417 net/sunrpc/cache.c next = cache_list.next; next 418 net/sunrpc/cache.c if (next == &cache_list) { next 423 net/sunrpc/cache.c current_detail = list_entry(next, struct cache_detail, others); next 664 net/sunrpc/cache.c discard = list_entry(cache_defer_list.next, next 719 net/sunrpc/cache.c dreq = list_entry(pending.next, struct cache_deferred_req, recent); next 743 net/sunrpc/cache.c dreq = list_entry(pending.next, struct cache_deferred_req, recent); next 812 net/sunrpc/cache.c while (rp->q.list.next != &cd->queue && next 813 net/sunrpc/cache.c list_entry(rp->q.list.next, struct cache_queue, list) next 815 net/sunrpc/cache.c struct list_head *next = rp->q.list.next; next 816 net/sunrpc/cache.c list_move(&rp->q.list, next); next 818 net/sunrpc/cache.c if (rp->q.list.next == &cd->queue) { next 824 net/sunrpc/cache.c rq = container_of(rp->q.list.next, struct cache_request, q.list); next 972 net/sunrpc/cache.c cq = list_entry(cq->list.next, struct cache_queue, list)) next 998 net/sunrpc/cache.c cq = list_entry(cq->list.next, struct cache_queue, list)) next 1047 net/sunrpc/cache.c cq = list_entry(cq->list.next, struct cache_queue, list)) next 1094 net/sunrpc/cache.c cr = list_entry(dequeued.next, struct cache_request, q.list); next 1354 net/sunrpc/cache.c else if (ch->cache_list.next == NULL) { next 1425 net/sunrpc/cache.c .next = cache_seq_next_rcu, next 61 net/sunrpc/debugfs.c struct list_head *next = task->tk_task.next; next 66 net/sunrpc/debugfs.c if (next == &clnt->cl_tasks) next 68 net/sunrpc/debugfs.c return list_entry(next, struct rpc_task, tk_task); next 81 net/sunrpc/debugfs.c .next = tasks_next, next 72 net/sunrpc/rpc_pipe.c msg = list_entry(head->next, struct rpc_pipe_msg, list); next 298 net/sunrpc/rpc_pipe.c msg = list_entry(pipe->pipe.next, next 666 net/sunrpc/rpc_pipe.c goto next; next 676 net/sunrpc/rpc_pipe.c next: next 149 net/sunrpc/sched.c task->u.tk_wait.list.next = q; next 176 net/sunrpc/sched.c q = t->u.tk_wait.list.next; next 267 net/sunrpc/stats.c struct rpc_clnt *next = clnt; next 269 net/sunrpc/stats.c _add_rpc_iostats(&stats, &next->cl_metrics[op]); next 270 net/sunrpc/stats.c if (next == next->cl_parent) next 272 net/sunrpc/stats.c next = next->cl_parent; next 273 net/sunrpc/stats.c } while (next); next 696 net/sunrpc/svc.c rqstp = list_entry(pool->sp_all_threads.next, struct svc_rqst, rq_all); next 932 net/sunrpc/svc_xprt.c struct list_head *le, *next; next 943 net/sunrpc/svc_xprt.c list_for_each_safe(le, next, &serv->sv_tempsocks) { next 974 net/sunrpc/svc_xprt.c struct list_head *le, *next; next 978 net/sunrpc/svc_xprt.c list_for_each_safe(le, next, &serv->sv_tempsocks) { next 989 net/sunrpc/svc_xprt.c le = to_be_closed.next; next 1246 net/sunrpc/svc_xprt.c dr = list_entry(xprt->xpt_deferred.next, next 1420 net/sunrpc/svc_xprt.c .next = svc_pool_stats_next, next 1307 net/sunrpc/xprt.c struct rpc_rqst *next = list_first_entry(&req->rq_xmit2, next 1310 net/sunrpc/xprt.c list_add_tail(&next->rq_xmit, &next->rq_xprt->xmit_queue); next 1504 net/sunrpc/xprt.c struct rpc_rqst *next, *req = task->tk_rqstp; next 1510 net/sunrpc/xprt.c next = list_first_entry(&xprt->xmit_queue, next 1512 net/sunrpc/xprt.c xprt_pin_rqst(next); next 1514 net/sunrpc/xprt.c status = xprt_request_transmit(next, task); next 1515 net/sunrpc/xprt.c if (status == -EBADMSG && next != req) next 1519 net/sunrpc/xprt.c xprt_unpin_rqst(next); next 1593 net/sunrpc/xprt.c req = list_entry(xprt->free.next, struct rpc_rqst, rq_list); next 435 net/sunrpc/xprtrdma/frwr_ops.c frwr->fr_regwr.wr.next = post_wr; next 549 net/sunrpc/xprtrdma/frwr_ops.c last->next = NULL; next 558 net/sunrpc/xprtrdma/frwr_ops.c prev = &last->next; next 591 net/sunrpc/xprtrdma/frwr_ops.c bad_wr = bad_wr->next; next 652 net/sunrpc/xprtrdma/frwr_ops.c last->next = NULL; next 661 net/sunrpc/xprtrdma/frwr_ops.c prev = &last->next; next 686 net/sunrpc/xprtrdma/frwr_ops.c bad_wr = bad_wr->next; next 138 net/sunrpc/xprtrdma/svc_rdma_recvfrom.c ctxt->rc_recv_wr.next = NULL; next 148 net/sunrpc/xprtrdma/svc_rdma_sendto.c ctxt->sc_send_wr.next = NULL; next 404 net/sunrpc/xprtrdma/svc_rdma_transport.c newxprt = list_entry(listen_rdma->sc_accept_q.next, next 1084 net/sunrpc/xprtrdma/verbs.c rep->rr_recv_wr.next = NULL; next 1527 net/sunrpc/xprtrdma/verbs.c rep->rr_recv_wr.next = wr; next 1534 net/sunrpc/xprtrdma/verbs.c for (i = wr; i; i = i->next) { next 1553 net/sunrpc/xprtrdma/verbs.c wr = wr->next; next 1564 net/sunrpc/xprtrdma/verbs.c i = i->next; next 70 net/tipc/msg.c skb->next = NULL; next 137 net/tipc/msg.c frag->next = NULL; next 168 net/tipc/msg.c tail->next = frag; next 684 net/tipc/msg.c skb->next = NULL; next 724 net/tipc/msg.c frag->next = NULL; next 48 net/tipc/name_distr.c struct list_head next; next 239 net/tipc/name_distr.c list_for_each_entry_safe(e, tmp, &tn->dist_queue, next) { next 242 net/tipc/name_distr.c list_del(&e->next); next 2822 net/unix/af_unix.c .next = unix_seq_next, next 217 net/unix/diag.c goto next; next 219 net/unix/diag.c goto next; next 225 net/unix/diag.c next: next 95 net/unix/garbage.c struct sk_buff *next; next 98 net/unix/garbage.c skb_queue_walk_safe(&x->sk_receive_queue, skb, next) { next 140 net/unix/garbage.c struct sk_buff *next; next 148 net/unix/garbage.c skb_queue_walk_safe(&x->sk_receive_queue, skb, next) { next 160 net/unix/garbage.c u = list_entry(embryos.next, struct unix_sock, link); next 205 net/unix/garbage.c struct unix_sock *next; next 232 net/unix/garbage.c list_for_each_entry_safe(u, next, &gc_inflight_list, link) { next 262 net/unix/garbage.c while (cursor.next != &gc_candidates) { next 263 net/unix/garbage.c u = list_entry(cursor.next, struct unix_sock, link); next 288 net/unix/garbage.c u = list_entry(not_cycle_list.next, struct unix_sock, link); next 395 net/vmw_vsock/af_vsock.c vconnected = list_entry(vlistener->accept_queue.next, next 91 net/wireless/lib80211.c struct lib80211_crypt_data *entry, *next; next 95 net/wireless/lib80211.c list_for_each_entry_safe(entry, next, &info->crypt_deinit_list, list) { next 2255 net/wireless/scan.c const u8 *pos, *end, *next; next 2269 net/wireless/scan.c next = pos + 2 + pos[1]; next 2270 net/wireless/scan.c while (next + 2 + next[1] - pos < IW_GENERIC_IE_MAX) next 2271 net/wireless/scan.c next = next + 2 + next[1]; next 2275 net/wireless/scan.c iwe.u.data.length = next - pos; next 2281 net/wireless/scan.c pos = next; next 796 net/wireless/sme.c u8 *next; next 842 net/wireless/sme.c next = ((u8 *)ev) + sizeof(*ev); next 844 net/wireless/sme.c ev->cr.bssid = next; next 846 net/wireless/sme.c next += ETH_ALEN; next 849 net/wireless/sme.c ev->cr.req_ie = next; next 853 net/wireless/sme.c next += params->req_ie_len; next 856 net/wireless/sme.c ev->cr.resp_ie = next; next 860 net/wireless/sme.c next += params->resp_ie_len; next 863 net/wireless/sme.c ev->cr.fils.kek = next; next 867 net/wireless/sme.c next += params->fils.kek_len; next 870 net/wireless/sme.c ev->cr.fils.pmk = next; next 874 net/wireless/sme.c next += params->fils.pmk_len; next 877 net/wireless/sme.c ev->cr.fils.pmkid = next; next 880 net/wireless/sme.c next += WLAN_PMKID_LEN; next 963 net/wireless/sme.c u8 *next; next 985 net/wireless/sme.c next = ((u8 *)ev) + sizeof(*ev); next 987 net/wireless/sme.c ev->rm.req_ie = next; next 990 net/wireless/sme.c next += info->req_ie_len; next 993 net/wireless/sme.c ev->rm.resp_ie = next; next 997 net/wireless/sme.c next += info->resp_ie_len; next 1000 net/wireless/sme.c ev->rm.fils.kek = next; next 1004 net/wireless/sme.c next += info->fils.kek_len; next 1007 net/wireless/sme.c ev->rm.fils.pmk = next; next 1011 net/wireless/sme.c next += info->fils.pmk_len; next 1014 net/wireless/sme.c ev->rm.fils.pmkid = next; next 1017 net/wireless/sme.c next += WLAN_PMKID_LEN; next 124 net/wireless/wext-proc.c .next = wireless_dev_seq_next, next 283 net/x25/x25_link.c if (nb->node.next) { next 150 net/x25/x25_proc.c .next = x25_seq_route_next, next 157 net/x25/x25_proc.c .next = x25_seq_socket_next, next 164 net/x25/x25_proc.c .next = x25_seq_forward_next, next 71 net/x25/x25_route.c if (rt->node.next) { next 128 net/xfrm/xfrm_device.c if (!skb->next) { next 152 net/xfrm/xfrm_device.c struct sk_buff *nskb = skb2->next; next 164 net/xfrm/xfrm_device.c skb2->next = nskb; next 167 net/xfrm/xfrm_device.c skb2->next = nskb; next 57 net/xfrm/xfrm_interface.c for (xi = rcu_dereference(start); xi; xi = rcu_dereference(xi->next)) next 109 net/xfrm/xfrm_interface.c rcu_assign_pointer(xi->next , rtnl_dereference(*xip)); next 120 net/xfrm/xfrm_interface.c xip = &iter->next) { next 122 net/xfrm/xfrm_interface.c rcu_assign_pointer(*xip, xi->next); next 165 net/xfrm/xfrm_interface.c xip = &xi->next) next 788 net/xfrm/xfrm_interface.c xip = &xi->next) next 548 net/xfrm/xfrm_output.c struct sk_buff *nskb = segs->next; next 314 net/xfrm/xfrm_policy.c time64_t next = TIME64_MAX; next 330 net/xfrm/xfrm_policy.c if (tmo < next) next 331 net/xfrm/xfrm_policy.c next = tmo; next 338 net/xfrm/xfrm_policy.c if (tmo < next) next 339 net/xfrm/xfrm_policy.c next = tmo; next 348 net/xfrm/xfrm_policy.c if (tmo < next) next 349 net/xfrm/xfrm_policy.c next = tmo; next 358 net/xfrm/xfrm_policy.c if (tmo < next) next 359 net/xfrm/xfrm_policy.c next = tmo; next 364 net/xfrm/xfrm_policy.c if (next != TIME64_MAX && next 365 net/xfrm/xfrm_policy.c !mod_timer(&xp->timer, jiffies + make_jiffies(next))) next 526 net/xfrm/xfrm_state.c time64_t next = TIME64_MAX; next 549 net/xfrm/xfrm_state.c if (tmo < next) next 550 net/xfrm/xfrm_state.c next = tmo; next 557 net/xfrm/xfrm_state.c if (tmo < next) next 558 net/xfrm/xfrm_state.c next = tmo; next 568 net/xfrm/xfrm_state.c } else if (tmo < next) { next 569 net/xfrm/xfrm_state.c next = tmo; next 579 net/xfrm/xfrm_state.c else if (tmo < next) next 580 net/xfrm/xfrm_state.c next = tmo; next 587 net/xfrm/xfrm_state.c if (next != TIME64_MAX) { next 588 net/xfrm/xfrm_state.c hrtimer_forward_now(&x->mtimer, ktime_set(next, 0)); next 1149 net/xfrm/xfrm_state.c xfrm_hash_grow_check(net, x->bydst.next != NULL); next 1265 net/xfrm/xfrm_state.c xfrm_hash_grow_check(net, x->bydst.next != NULL); next 1376 net/xfrm/xfrm_state.c xfrm_hash_grow_check(net, x->bydst.next != NULL); next 37 samples/bpf/test_lru_dist.c struct list_head *next, *prev; next 42 samples/bpf/test_lru_dist.c list->next = list; next 48 samples/bpf/test_lru_dist.c return head->next == head; next 53 samples/bpf/test_lru_dist.c struct list_head *next) next 55 samples/bpf/test_lru_dist.c next->prev = new; next 56 samples/bpf/test_lru_dist.c new->next = next; next 58 samples/bpf/test_lru_dist.c prev->next = new; next 63 samples/bpf/test_lru_dist.c __list_add(new, head, head->next); next 66 samples/bpf/test_lru_dist.c static inline void __list_del(struct list_head *prev, struct list_head *next) next 68 samples/bpf/test_lru_dist.c next->prev = prev; next 69 samples/bpf/test_lru_dist.c prev->next = next; next 74 samples/bpf/test_lru_dist.c __list_del(entry->prev, entry->next); next 328 samples/mic/mpssd/mpssd.c unsigned int next; next 332 samples/mic/mpssd/mpssd.c next = le16toh(desc->next); next 333 samples/mic/mpssd/mpssd.c return next; next 1599 samples/mic/mpssd/mpssd.c for (mic = mic_list.next; mic != NULL; mic = mic->next) next 1713 samples/mic/mpssd/mpssd.c for (mic = mic_list.next; mic; mic = mic->next) { next 1739 samples/mic/mpssd/mpssd.c mic->next = calloc(1, sizeof(struct mic_info)); next 1740 samples/mic/mpssd/mpssd.c if (mic->next) { next 1741 samples/mic/mpssd/mpssd.c mic = mic->next; next 82 samples/mic/mpssd/mpssd.h struct mic_info *next; next 157 samples/vfio-mdev/mbochs.c struct list_head next; next 946 samples/vfio-mdev/mbochs.c list_add(&dmabuf->next, &mdev_state->dmabufs); next 968 samples/vfio-mdev/mbochs.c list_for_each_entry(dmabuf, &mdev_state->dmabufs, next) next 982 samples/vfio-mdev/mbochs.c list_for_each_entry(dmabuf, &mdev_state->dmabufs, next) next 1063 samples/vfio-mdev/mbochs.c ext->type.header.next = 0; next 1316 samples/vfio-mdev/mbochs.c list_for_each_entry_safe(dmabuf, tmp, &mdev_state->dmabufs, next) { next 1317 samples/vfio-mdev/mbochs.c list_del(&dmabuf->next); next 139 samples/vfio-mdev/mtty.c struct list_head next; next 755 samples/vfio-mdev/mtty.c list_add(&mdev_state->next, &mdev_devices_list); next 768 samples/vfio-mdev/mtty.c list_for_each_entry_safe(mds, tmp_mds, &mdev_devices_list, next) { next 770 samples/vfio-mdev/mtty.c list_del(&mdev_state->next); next 1350 samples/vfio-mdev/mtty.c list_for_each_entry(mds, &mdev_devices_list, next) next 292 scripts/asn1_compiler.c struct action *next; next 687 scripts/asn1_compiler.c struct element *next; next 1098 scripts/asn1_compiler.c ppaction = &(*ppaction)->next next 1107 scripts/asn1_compiler.c action->next = *ppaction; next 1113 scripts/asn1_compiler.c action->next = NULL; next 1182 scripts/asn1_compiler.c child_p = &element->next; next 1246 scripts/asn1_compiler.c for (c = e->children; c; c = c->next) next 1328 scripts/asn1_compiler.c for (action = action_list; action; action = action->next) { next 1338 scripts/asn1_compiler.c for (action = action_list; action; action = action->next) next 1347 scripts/asn1_compiler.c for (action = action_list; action; action = action->next) next 1406 scripts/asn1_compiler.c for (ce = e->children; ce; ce = ce->next) next 1549 scripts/asn1_compiler.c for (ec = e->children; ec; ec = ec->next) next 1597 scripts/asn1_compiler.c for (ec = e->children; ec; ec = ec->next) next 165 scripts/basic/fixdep.c struct item *next; next 191 scripts/basic/fixdep.c for (aux = hashtab[hash % HASHSZ]; aux; aux = aux->next) { next 213 scripts/basic/fixdep.c aux->next = hashtab[hash % HASHSZ]; next 107 scripts/dtc/checks.c while (pos->next) { next 108 scripts/dtc/checks.c pos = pos->next; next 287 scripts/dtc/checks.c for (prop2 = prop->next; prop2; prop2 = prop2->next) { next 550 scripts/dtc/checks.c for (pp = &node->proplist; *pp; pp = &((*pp)->next)) next 566 scripts/dtc/checks.c *pp = prop->next; next 14 scripts/dtc/data.c nm = m->next; next 124 scripts/dtc/data.c m = m->next; next 136 scripts/dtc/data.c mp = &((*mp)->next); next 235 scripts/dtc/data.c m->next = NULL; next 79 scripts/dtc/dtc.h struct marker *next; next 92 scripts/dtc/dtc.h for (; (m); (m) = (m)->next) next 132 scripts/dtc/dtc.h struct label *next; next 144 scripts/dtc/dtc.h struct property *next; next 173 scripts/dtc/dtc.h for ((l) = (l0); (l); (l) = (l)->next) next 180 scripts/dtc/dtc.h for ((p) = (n)->proplist; (p); (p) = (p)->next) next 244 scripts/dtc/dtc.h struct reserve_info *next; next 300 scripts/dtc/flattree.c for (re = reservelist; re; re = re->next) { next 516 scripts/dtc/flattree.c for (re = dti->reservelist; re; re = re->next) { next 27 scripts/dtc/livetree.c new->next = *labels; next 67 scripts/dtc/livetree.c assert(first->next == NULL); next 69 scripts/dtc/livetree.c first->next = list; next 77 scripts/dtc/livetree.c struct property *next; next 80 scripts/dtc/livetree.c next = p->next; next 81 scripts/dtc/livetree.c p->next = head; next 83 scripts/dtc/livetree.c p = next; next 159 scripts/dtc/livetree.c new_node->proplist = new_prop->next; next 160 scripts/dtc/livetree.c new_prop->next = NULL; next 270 scripts/dtc/livetree.c prop->next = NULL; next 274 scripts/dtc/livetree.c p = &((*p)->next); next 288 scripts/dtc/livetree.c prop = prop->next; next 373 scripts/dtc/livetree.c assert(first->next == NULL); next 375 scripts/dtc/livetree.c first->next = list; next 384 scripts/dtc/livetree.c new->next = NULL; next 389 scripts/dtc/livetree.c for (last = list; last->next; last = last->next) next 392 scripts/dtc/livetree.c last->next = new; next 673 scripts/dtc/livetree.c ri = ri->next) next 683 scripts/dtc/livetree.c ri = ri->next) next 690 scripts/dtc/livetree.c tbl[i]->next = tbl[i+1]; next 691 scripts/dtc/livetree.c tbl[n-1]->next = NULL; next 726 scripts/dtc/livetree.c tbl[i]->next = tbl[i+1]; next 727 scripts/dtc/livetree.c tbl[n-1]->next = NULL; next 15 scripts/dtc/srcpos.c struct search_path *next; /* next node in list, NULL for end */ next 140 scripts/dtc/srcpos.c for (node = search_path_head; !*fp && node; node = node->next) next 221 scripts/dtc/srcpos.c node->next = NULL; next 229 scripts/dtc/srcpos.c search_path_tail = &node->next; next 263 scripts/dtc/srcpos.c assert(pos->next == NULL); next 281 scripts/dtc/srcpos.c for (p = pos; p->next != NULL; p = p->next); next 282 scripts/dtc/srcpos.c p->next = newtail; next 351 scripts/dtc/srcpos.c if (pos->next != NULL) { next 352 scripts/dtc/srcpos.c rest = srcpos_string_comment(pos->next, first_line, level); next 63 scripts/dtc/srcpos.h struct srcpos *next; next 83 scripts/dtc/srcpos.h (Current).next = NULL; \ next 135 scripts/dtc/treesource.c m = m->next; next 141 scripts/dtc/treesource.c struct marker *next = next_type_marker(m->next); next 143 scripts/dtc/treesource.c if (next) next 144 scripts/dtc/treesource.c return next->offset - m->offset; next 222 scripts/dtc/treesource.c dummy_marker.next = prop->val.markers; next 229 scripts/dtc/treesource.c size_t chunk_len = (m->next ? m->next->offset : len) - m->offset; next 334 scripts/dtc/treesource.c for (re = dti->reservelist; re; re = re->next) { next 326 scripts/gcc-plugins/gcc-common.h for (node = cgraph_nodes; node; node = node->next) next 334 scripts/gcc-plugins/gcc-common.h for (node = node->next; node; node = node->next) next 347 scripts/gcc-plugins/gcc-common.h for (alias = node->same_body; alias; alias = alias->next) { next 368 scripts/gcc-plugins/gcc-common.h for (node = cgraph_nodes; node; node = node->next) next 370 scripts/gcc-plugins/gcc-common.h for (node = varpool_nodes; node; node = node->next) next 96 scripts/gcc-plugins/gcc-generate-gimple-pass.h .next = NULL, next 164 scripts/gcc-plugins/gcc-generate-ipa-pass.h .next = NULL, next 96 scripts/gcc-plugins/gcc-generate-rtl-pass.h .next = NULL, next 96 scripts/gcc-plugins/gcc-generate-simple_ipa-pass.h .next = NULL, next 193 scripts/gcc-plugins/stackleak_plugin.c rtx_insn *insn, *next; next 212 scripts/gcc-plugins/stackleak_plugin.c for (insn = get_insns(); insn; insn = next) { next 215 scripts/gcc-plugins/stackleak_plugin.c next = NEXT_INSN(insn); next 253 scripts/gcc-plugins/stackleak_plugin.c if (GET_CODE(next) == NOTE && next 254 scripts/gcc-plugins/stackleak_plugin.c NOTE_KIND(next) == NOTE_INSN_CALL_ARG_LOCATION) { next 255 scripts/gcc-plugins/stackleak_plugin.c insn = next; next 256 scripts/gcc-plugins/stackleak_plugin.c next = NEXT_INSN(insn); next 179 scripts/genksyms/genksyms.c (defn = defn->next) && defn->tag == SYM_NORMAL && next 181 scripts/genksyms/genksyms.c (defn = defn->next) && defn->tag == SYM_NORMAL && next 332 scripts/genksyms/genksyms.c struct string_list *next = s->next; next 334 scripts/genksyms/genksyms.c s = next; next 345 scripts/genksyms/genksyms.c newnode->next = NULL; next 358 scripts/genksyms/genksyms.c for (n2 = n; n2->next; n2 = n2->next) next 360 scripts/genksyms/genksyms.c n2->next = start; next 386 scripts/genksyms/genksyms.c for (start = start->next; start != end; start = start->next) { next 387 scripts/genksyms/genksyms.c n->next = copy_node(start); next 388 scripts/genksyms/genksyms.c n = n->next; next 390 scripts/genksyms/genksyms.c n->next = NULL; next 399 scripts/genksyms/genksyms.c a = a->next; next 400 scripts/genksyms/genksyms.c b = b->next; next 481 scripts/genksyms/genksyms.c def->next = defn; next 513 scripts/genksyms/genksyms.c while ((tmp = tmp->next) != NULL) next 521 scripts/genksyms/genksyms.c while ((list = list->next) != NULL) next 541 scripts/genksyms/genksyms.c while ((tmp = tmp->next) != NULL) next 549 scripts/genksyms/genksyms.c while ((list = list->next) != NULL) next 27 scripts/genksyms/genksyms.h struct string_list *next; next 30 scripts/genksyms/parse.y *p = node->next; next 57 scripts/genksyms/parse.y r->next = (*keyw)->next; *body = r; (*keyw)->next = NULL; next 140 scripts/genksyms/parse.y struct string_list *decl = (*$3)->next; next 141 scripts/genksyms/parse.y (*$3)->next = NULL; next 410 scripts/genksyms/parse.y { remove_list($2, &(*$1)->next); $$ = $2; } next 267 scripts/kconfig/conf.c for (child = menu->list; child; child = child->next) { next 322 scripts/kconfig/conf.c for (child = menu->list; child; child = child->next) { next 335 scripts/kconfig/conf.c for (child = child->list; child; child = child->next) { next 405 scripts/kconfig/conf.c for (child = menu->list; child; child = child->next) next 446 scripts/kconfig/conf.c for (child = menu->list; child; child = child->next) next 828 scripts/kconfig/confdata.c else if (menu->next != NULL) { next 829 scripts/kconfig/confdata.c menu = menu->next; next 832 scripts/kconfig/confdata.c if (menu->next != NULL) { next 833 scripts/kconfig/confdata.c menu = menu->next; next 892 scripts/kconfig/confdata.c goto next; next 903 scripts/kconfig/confdata.c goto next; next 912 scripts/kconfig/confdata.c next: next 917 scripts/kconfig/confdata.c if (menu->next) next 918 scripts/kconfig/confdata.c menu = menu->next; next 926 scripts/kconfig/confdata.c if (menu->next) { next 927 scripts/kconfig/confdata.c menu = menu->next; next 968 scripts/kconfig/confdata.c for (file = file_list; file; file = file->next) { next 969 scripts/kconfig/confdata.c if (file->next) next 21 scripts/kconfig/expr.h struct file *next; next 85 scripts/kconfig/expr.h struct symbol *next; next 134 scripts/kconfig/expr.h #define for_all_symbols(i, sym) for (i = 0; i < SYMBOL_HASHSIZE; i++) for (sym = symbol_hash[i]; sym; sym = sym->next) next 193 scripts/kconfig/expr.h struct property *next; /* next property - null if last */ next 207 scripts/kconfig/expr.h for (st = sym->prop; st; st = st->next) \ next 212 scripts/kconfig/expr.h for (st = sym->prop; st; st = st->next) \ next 223 scripts/kconfig/expr.h struct menu *next; next 1096 scripts/kconfig/gconf.c for (child = menu->list; child; child = child->next) { next 1261 scripts/kconfig/gconf.c for (child1 = src->list; child1; child1 = child1->next) { next 1345 scripts/kconfig/gconf.c for (child = menu->list; child; child = child->next) { next 1431 scripts/kconfig/gconf.c for (child = menu->list; child; child = child->next) { next 25 scripts/kconfig/list.h struct list_head *next, *prev; next 50 scripts/kconfig/list.h for (pos = list_entry((head)->next, typeof(*pos), member); \ next 52 scripts/kconfig/list.h pos = list_entry(pos->member.next, typeof(*pos), member)) next 62 scripts/kconfig/list.h for (pos = list_entry((head)->next, typeof(*pos), member), \ next 63 scripts/kconfig/list.h n = list_entry(pos->member.next, typeof(*pos), member); \ next 65 scripts/kconfig/list.h pos = n, n = list_entry(n->member.next, typeof(*n), member)) next 73 scripts/kconfig/list.h return head->next == head; next 84 scripts/kconfig/list.h struct list_head *next) next 86 scripts/kconfig/list.h next->prev = _new; next 87 scripts/kconfig/list.h _new->next = next; next 89 scripts/kconfig/list.h prev->next = _new; next 112 scripts/kconfig/list.h static inline void __list_del(struct list_head *prev, struct list_head *next) next 114 scripts/kconfig/list.h next->prev = prev; next 115 scripts/kconfig/list.h prev->next = next; next 128 scripts/kconfig/list.h __list_del(entry->prev, entry->next); next 129 scripts/kconfig/list.h entry->next = (struct list_head*)LIST_POISON1; next 91 scripts/kconfig/lxdialog/dialog.h struct subtitle_list *next; next 163 scripts/kconfig/lxdialog/dialog.h struct dialog_list *next; next 178 scripts/kconfig/lxdialog/dialog.h item_cur && (item_cur != &item_nil); item_cur = item_cur->next) next 258 scripts/kconfig/lxdialog/util.c for (pos = dlg.subtitles; pos != NULL; pos = pos->next) { next 270 scripts/kconfig/lxdialog/util.c for (pos = dlg.subtitles; pos != NULL; pos = pos->next) { next 583 scripts/kconfig/lxdialog/util.c struct dialog_list *p, *next; next 585 scripts/kconfig/lxdialog/util.c for (p = item_head; p; p = next) { next 586 scripts/kconfig/lxdialog/util.c next = p->next; next 599 scripts/kconfig/lxdialog/util.c item_cur->next = p; next 661 scripts/kconfig/lxdialog/util.c for (p = item_head; p; p = p->next) next 679 scripts/kconfig/lxdialog/util.c for (p = item_head; p; p = p->next) { next 325 scripts/kconfig/mconf.c tmp = pos->next; next 333 scripts/kconfig/mconf.c pos->next = xcalloc(1, sizeof(*pos)); next 334 scripts/kconfig/mconf.c pos = pos->next; next 350 scripts/kconfig/mconf.c tmp = pos->next; next 533 scripts/kconfig/mconf.c for (child = menu->list; child; child = child->next) { next 638 scripts/kconfig/mconf.c for (child = menu->list; child; child = child->next) next 808 scripts/kconfig/mconf.c for (child = menu->list; child; child = child->next) { next 59 scripts/kconfig/menu.c last_entry_ptr = &menu->next; next 73 scripts/kconfig/menu.c last_entry_ptr = ¤t_menu->next; next 232 scripts/kconfig/menu.c for (prop = sym->prop; prop; prop = prop->next) { next 308 scripts/kconfig/menu.c for (menu = parent->list; menu; menu = menu->next) { next 316 scripts/kconfig/menu.c for (menu = parent->list; menu; menu = menu->next) { next 337 scripts/kconfig/menu.c for (menu = parent->list; menu; menu = menu->next) { next 362 scripts/kconfig/menu.c for (; prop; prop = prop->next) { next 417 scripts/kconfig/menu.c for (menu = parent->list; menu; menu = menu->next) next 448 scripts/kconfig/menu.c for (menu = parent->next; menu; menu = menu->next) { next 455 scripts/kconfig/menu.c goto next; next 478 scripts/kconfig/menu.c next: next 485 scripts/kconfig/menu.c parent->list = parent->next; next 486 scripts/kconfig/menu.c parent->next = last_menu->next; next 487 scripts/kconfig/menu.c last_menu->next = NULL; next 492 scripts/kconfig/menu.c for (menu = parent->list; menu; menu = menu->next) { next 499 scripts/kconfig/menu.c for (prop = menu->sym->prop; prop; prop = prop->next) { next 518 scripts/kconfig/menu.c for (prop = menu->sym->prop; prop; prop = prop->next) { next 562 scripts/kconfig/menu.c for (last_menu = menu->list; ; last_menu = last_menu->next) { next 564 scripts/kconfig/menu.c if (!last_menu->next) next 567 scripts/kconfig/menu.c last_menu->next = menu->next; next 568 scripts/kconfig/menu.c menu->next = menu->list; next 617 scripts/kconfig/menu.c for (child = menu->list; child; child = child->next) { next 651 scripts/kconfig/menu.c for (child = menu->list; child; child = child->next) { next 800 scripts/kconfig/nconf.c for (child = menu->list; child; child = child->next) { next 916 scripts/kconfig/nconf.c for (child = menu->list; child; child = child->next) next 1250 scripts/kconfig/nconf.c for (i = 0, child = menu->list; child; child = child->next) { next 621 scripts/kconfig/parser.y for (prop = sym->prop; prop; prop = prop->next) { next 717 scripts/kconfig/parser.y else if (menu->next) next 718 scripts/kconfig/parser.y menu = menu->next; next 722 scripts/kconfig/parser.y if (menu->next) { next 723 scripts/kconfig/parser.y menu = menu->next; next 597 scripts/kconfig/qconf.cc for (child = menu->list; child; child = child->next) { next 662 scripts/kconfig/qconf.cc for (child = menu->list; child; child = child->next) { next 1141 scripts/kconfig/qconf.cc for (struct property *prop = sym->prop; prop; prop = prop->next) { next 1836 scripts/kconfig/qconf.cc for (child = menu->list; child; child = child->next) { next 156 scripts/kconfig/symbol.c for (prop = sym->prop; prop; prop = prop->next) { next 818 scripts/kconfig/symbol.c for (symbol = symbol_hash[hash]; symbol; symbol = symbol->next) { next 837 scripts/kconfig/symbol.c symbol->next = symbol_hash[hash]; next 860 scripts/kconfig/symbol.c for (symbol = symbol_hash[hash]; symbol; symbol = symbol->next) { next 1008 scripts/kconfig/symbol.c struct dep_stack *prev, *next; next 1018 scripts/kconfig/symbol.c check_top->next = stack; next 1028 scripts/kconfig/symbol.c check_top->next = NULL; next 1057 scripts/kconfig/symbol.c for (; stack; stack = stack->next) { next 1059 scripts/kconfig/symbol.c next_sym = stack->next ? stack->next->sym : last_sym; next 1066 scripts/kconfig/symbol.c for (prop = sym->prop; prop; prop = prop->next) { next 1185 scripts/kconfig/symbol.c for (prop = sym->prop; prop; prop = prop->next) { next 1290 scripts/kconfig/symbol.c for (propp = &sym->prop; *propp; propp = &(*propp)->next) next 17 scripts/kconfig/util.c for (file = file_list; file; file = file->next) { next 26 scripts/kconfig/util.c file->next = file_list; next 129 scripts/mod/modpost.c for (mod = modules; mod; mod = mod->next) next 153 scripts/mod/modpost.c mod->next = modules; next 165 scripts/mod/modpost.c struct symbol *next; next 200 scripts/mod/modpost.c struct symbol *next) next 207 scripts/mod/modpost.c s->next = next; next 234 scripts/mod/modpost.c for (s = symbolhash[tdb_hash(name) % SYMBOL_HASH_SIZE]; s; s = s->next) { next 246 scripts/mod/modpost.c for (ns_entry = list; ns_entry != NULL; ns_entry = ns_entry->next) next 261 scripts/mod/modpost.c ns_entry->next = *list; next 2198 scripts/mod/modpost.c for (s = mod->unres; s; s = s->next) { next 2312 scripts/mod/modpost.c for (s = mod->unres; s; s = s->next) { next 2328 scripts/mod/modpost.c for (s = mod->unres; s; s = s->next) { next 2357 scripts/mod/modpost.c for (s = mod->unres; s; s = s->next) next 2363 scripts/mod/modpost.c for (s = mod->unres; s; s = s->next) { next 2523 scripts/mod/modpost.c symbol = symbol->next; next 2536 scripts/mod/modpost.c for (mod = modules; mod; mod = mod->next) { next 2544 scripts/mod/modpost.c for (ns = mod->required_namespaces; ns; ns = ns->next) next 2556 scripts/mod/modpost.c struct ext_sym_list *next; next 2585 scripts/mod/modpost.c extsym_iter->next = extsym_start; next 2627 scripts/mod/modpost.c extsym_iter = extsym_start->next; next 2640 scripts/mod/modpost.c for (mod = modules; mod; mod = mod->next) { next 2679 scripts/mod/modpost.c for (s = symbolhash[n]; s; s = s->next) { next 113 scripts/mod/modpost.h struct namespace_list *next; next 118 scripts/mod/modpost.h struct module *next; next 1980 security/apparmor/apparmorfs.c struct aa_ns *parent, *next; next 1988 security/apparmor/apparmorfs.c next = list_first_entry(&ns->sub_ns, typeof(*ns), base.list); next 1989 security/apparmor/apparmorfs.c mutex_lock_nested(&next->lock, next->level); next 1990 security/apparmor/apparmorfs.c return next; next 1997 security/apparmor/apparmorfs.c next = list_next_entry(ns, base.list); next 1998 security/apparmor/apparmorfs.c if (!list_entry_is_head(next, &parent->sub_ns, base.list)) { next 1999 security/apparmor/apparmorfs.c mutex_lock_nested(&next->lock, next->level); next 2000 security/apparmor/apparmorfs.c return next; next 2082 security/apparmor/apparmorfs.c struct aa_profile *next = __next_profile(profile); next 2083 security/apparmor/apparmorfs.c if (next) next 2084 security/apparmor/apparmorfs.c return next; next 2177 security/apparmor/apparmorfs.c .next = p_next, next 148 security/apparmor/domain.c goto next; next 155 security/apparmor/domain.c next: next 210 security/apparmor/domain.c goto next; next 216 security/apparmor/domain.c next: next 1025 security/apparmor/label.c struct aa_profile *next; next 1038 security/apparmor/label.c label_for_each_in_merge(i, a, b, next) { next 1039 security/apparmor/label.c AA_BUG(!next); next 1040 security/apparmor/label.c if (profile_is_stale(next)) { next 1041 security/apparmor/label.c new->vec[k] = aa_get_newest_profile(next); next 1044 security/apparmor/label.c if (next->label.proxy != new->vec[k]->label.proxy) next 1049 security/apparmor/label.c new->vec[k++] = aa_get_profile(next); next 1288 security/apparmor/label.c goto next; next 1295 security/apparmor/label.c next: next 1348 security/apparmor/label.c goto next; next 1354 security/apparmor/label.c next: next 370 security/apparmor/match.c #define match_char(state, def, base, next, check, C) \ next 380 security/apparmor/match.c (state) = (next)[pos]; \ next 405 security/apparmor/match.c u16 *next = NEXT_TABLE(dfa); next 417 security/apparmor/match.c match_char(state, def, base, next, check, next 422 security/apparmor/match.c match_char(state, def, base, next, check, (u8) *str++); next 445 security/apparmor/match.c u16 *next = NEXT_TABLE(dfa); next 458 security/apparmor/match.c match_char(state, def, base, next, check, next 463 security/apparmor/match.c match_char(state, def, base, next, check, (u8) *str++); next 484 security/apparmor/match.c u16 *next = NEXT_TABLE(dfa); next 491 security/apparmor/match.c match_char(state, def, base, next, check, equiv[(u8) c]); next 493 security/apparmor/match.c match_char(state, def, base, next, check, (u8) c); next 516 security/apparmor/match.c u16 *next = NEXT_TABLE(dfa); next 532 security/apparmor/match.c state = next[pos]; next 543 security/apparmor/match.c state = next[pos]; next 577 security/apparmor/match.c u16 *next = NEXT_TABLE(dfa); next 594 security/apparmor/match.c state = next[pos]; next 605 security/apparmor/match.c state = next[pos]; next 653 security/apparmor/match.c u16 *next = NEXT_TABLE(dfa); next 677 security/apparmor/match.c state = next[pos]; next 696 security/apparmor/match.c state = next[pos]; next 127 security/apparmor/procattr.c char *next = hat + strlen(hat) + 1; next 131 security/apparmor/procattr.c hat = next; next 103 security/integrity/ima/ima_fs.c qe = list_entry_rcu(qe->later.next, struct ima_queue_entry, later); next 192 security/integrity/ima/ima_fs.c .next = ima_measurements_next, next 258 security/integrity/ima/ima_fs.c .next = ima_measurements_next, next 374 security/integrity/ima/ima_fs.c .next = ima_policy_next, next 1349 security/integrity/ima/ima_policy.c entry = list_entry_rcu(entry->list.next, struct ima_rule_entry, list); next 126 security/keys/gc.c list_entry(keys->next, struct key, graveyard_link); next 417 security/keys/keyring.c if (keyring->name_link.next != NULL && next 23 security/keys/proc.c .next = proc_keys_next, next 35 security/keys/proc.c .next = proc_key_users_next, next 36 security/safesetid/lsm.c hash_for_each_possible(policy->rules, rule, next, __kuid_val(src)) { next 35 security/safesetid/lsm.h struct hlist_node next; next 68 security/safesetid/securityfs.c hash_for_each_safe(pol->rules, bucket, tmp, rule, next) next 81 security/safesetid/securityfs.c hash_add(pol->rules, &rule->next, __kuid_val(rule->src_uid)); next 90 security/safesetid/securityfs.c hash_for_each(pol->rules, bucket, rule, next) { next 210 security/security.c char *sep, *name, *next; next 240 security/security.c next = sep; next 242 security/security.c while ((name = strsep(&next, ",")) != NULL) { next 79 security/selinux/avc.c struct avc_callback_node *next; next 808 security/selinux/avc.c c->next = avc_callbacks; next 971 security/selinux/avc.c for (c = avc_callbacks; c; c = c->next) { next 1552 security/selinux/selinuxfs.c .next = sel_avc_stats_seq_next, next 93 security/selinux/ss/avtab.c newnode->next = prev->next; next 94 security/selinux/ss/avtab.c prev->next = newnode; next 98 security/selinux/ss/avtab.c newnode->next = *n; next 118 security/selinux/ss/avtab.c prev = cur, cur = cur->next) { next 162 security/selinux/ss/avtab.c prev = cur, cur = cur->next) { next 192 security/selinux/ss/avtab.c cur = cur->next) { next 228 security/selinux/ss/avtab.c cur = cur->next) { next 257 security/selinux/ss/avtab.c for (cur = node->next; cur; cur = cur->next) { next 289 security/selinux/ss/avtab.c cur = cur->next; next 360 security/selinux/ss/avtab.c cur = cur->next; next 648 security/selinux/ss/avtab.c cur = cur->next) { next 80 security/selinux/ss/avtab.h struct avtab_node *next; next 31 security/selinux/ss/conditional.c for (cur = expr; cur; cur = cur->next) { next 99 security/selinux/ss/conditional.c for (cur = node->true_list; cur; cur = cur->next) { next 106 security/selinux/ss/conditional.c for (cur = node->false_list; cur; cur = cur->next) { next 133 security/selinux/ss/conditional.c struct cond_av_list *cur, *next; next 134 security/selinux/ss/conditional.c for (cur = list; cur; cur = next) { next 135 security/selinux/ss/conditional.c next = cur->next; next 146 security/selinux/ss/conditional.c next_expr = cur_expr->next; next 156 security/selinux/ss/conditional.c struct cond_node *next, *cur; next 161 security/selinux/ss/conditional.c for (cur = list; cur; cur = next) { next 162 security/selinux/ss/conditional.c next = cur->next; next 303 security/selinux/ss/conditional.c for (cur = other; cur; cur = cur->next) { next 339 security/selinux/ss/conditional.c data->tail->next = list; next 433 security/selinux/ss/conditional.c last->next = expr; next 479 security/selinux/ss/conditional.c last->next = node; next 530 security/selinux/ss/conditional.c for (cur_list = list; cur_list != NULL; cur_list = cur_list->next) next 541 security/selinux/ss/conditional.c for (cur_list = list; cur_list != NULL; cur_list = cur_list->next) { next 563 security/selinux/ss/conditional.c for (cur_expr = node->expr; cur_expr != NULL; cur_expr = cur_expr->next) next 571 security/selinux/ss/conditional.c for (cur_expr = node->expr; cur_expr != NULL; cur_expr = cur_expr->next) { next 597 security/selinux/ss/conditional.c for (cur = list; cur != NULL; cur = cur->next) next 604 security/selinux/ss/conditional.c for (cur = list; cur != NULL; cur = cur->next) { next 33 security/selinux/ss/conditional.h struct cond_expr *next; next 43 security/selinux/ss/conditional.h struct cond_av_list *next; next 58 security/selinux/ss/conditional.h struct cond_node *next; next 54 security/selinux/ss/constraint.h struct constraint_expr *next; /* next expression */ next 60 security/selinux/ss/constraint.h struct constraint_node *next; /* next constraint */ next 42 security/selinux/ss/ebitmap.c n1 = n1->next; next 43 security/selinux/ss/ebitmap.c n2 = n2->next; next 67 security/selinux/ss/ebitmap.c new->next = NULL; next 69 security/selinux/ss/ebitmap.c prev->next = new; next 73 security/selinux/ss/ebitmap.c n = n->next; next 123 security/selinux/ss/ebitmap.c e_iter = e_iter->next; next 175 security/selinux/ss/ebitmap.c e_prev->next = e_iter; next 214 security/selinux/ss/ebitmap.c n1 = n1->next; next 232 security/selinux/ss/ebitmap.c n1 = n1->next; next 233 security/selinux/ss/ebitmap.c n2 = n2->next; next 253 security/selinux/ss/ebitmap.c n = n->next; next 279 security/selinux/ss/ebitmap.c if (!n->next) { next 291 security/selinux/ss/ebitmap.c prev->next = n->next; next 293 security/selinux/ss/ebitmap.c e->node = n->next; next 299 security/selinux/ss/ebitmap.c n = n->next; next 317 security/selinux/ss/ebitmap.c new->next = prev->next; next 318 security/selinux/ss/ebitmap.c prev->next = new; next 320 security/selinux/ss/ebitmap.c new->next = e->node; next 337 security/selinux/ss/ebitmap.c n = n->next; next 417 security/selinux/ss/ebitmap.c n->next = tmp; next 35 security/selinux/ss/ebitmap.h struct ebitmap_node *next; next 52 security/selinux/ss/ebitmap.h for (*n = e->node; *n; *n = (*n)->next) { next 75 security/selinux/ss/ebitmap.h for (*n = (*n)->next; *n; *n = (*n)->next) { next 57 security/selinux/ss/hashtab.c cur = cur->next; next 69 security/selinux/ss/hashtab.c newnode->next = prev->next; next 70 security/selinux/ss/hashtab.c prev->next = newnode; next 72 security/selinux/ss/hashtab.c newnode->next = h->htable[hvalue]; next 91 security/selinux/ss/hashtab.c cur = cur->next; next 111 security/selinux/ss/hashtab.c cur = cur->next; next 140 security/selinux/ss/hashtab.c cur = cur->next; next 161 security/selinux/ss/hashtab.c cur = cur->next; next 19 security/selinux/ss/hashtab.h struct hashtab_node *next; next 235 security/selinux/ss/policydb.c e = e->next; next 239 security/selinux/ss/policydb.c constraint = constraint->next; next 248 security/selinux/ss/policydb.c e = e->next; next 252 security/selinux/ss/policydb.c constraint = constraint->next; next 791 security/selinux/ss/policydb.c c = c->next; next 804 security/selinux/ss/policydb.c c = c->next; next 808 security/selinux/ss/policydb.c g = g->next; next 815 security/selinux/ss/policydb.c for (tr = p->role_tr; tr; tr = tr->next) { next 822 security/selinux/ss/policydb.c for (ra = p->role_allow; ra; ra = ra->next) { next 862 security/selinux/ss/policydb.c for (c = head; c; c = c->next) { next 1201 security/selinux/ss/policydb.c lc->next = c; next 1218 security/selinux/ss/policydb.c le->next = e; next 1986 security/selinux/ss/policydb.c genfs_p = genfs, genfs = genfs->next) { next 1996 security/selinux/ss/policydb.c newgenfs->next = genfs; next 1998 security/selinux/ss/policydb.c genfs_p->next = newgenfs; next 2034 security/selinux/ss/policydb.c l = c, c = c->next) { next 2049 security/selinux/ss/policydb.c newc->next = c; next 2051 security/selinux/ss/policydb.c l->next = newc; next 2091 security/selinux/ss/policydb.c l->next = c; next 2418 security/selinux/ss/policydb.c ltr->next = tr; next 2457 security/selinux/ss/policydb.c lra->next = ra; next 2657 security/selinux/ss/policydb.c for (tr = r; tr; tr = tr->next) next 2663 security/selinux/ss/policydb.c for (tr = r; tr; tr = tr->next) { next 2689 security/selinux/ss/policydb.c for (ra = r; ra; ra = ra->next) next 2695 security/selinux/ss/policydb.c for (ra = r; ra; ra = ra->next) { next 2815 security/selinux/ss/policydb.c for (c = node; c; c = c->next) { next 2817 security/selinux/ss/policydb.c for (e = c->expr; e; e = e->next) next 2824 security/selinux/ss/policydb.c for (e = c->expr; e; e = e->next) { next 2873 security/selinux/ss/policydb.c for (c = cladatum->constraints; c; c = c->next) next 2909 security/selinux/ss/policydb.c for (c = cladatum->validatetrans; c; c = c->next) next 3086 security/selinux/ss/policydb.c for (c = p->ocontexts[i]; c; c = c->next) next 3092 security/selinux/ss/policydb.c for (c = p->ocontexts[i]; c; c = c->next) { next 3214 security/selinux/ss/policydb.c for (genfs = p->genfs; genfs; genfs = genfs->next) next 3220 security/selinux/ss/policydb.c for (genfs = p->genfs; genfs; genfs = genfs->next) { next 3230 security/selinux/ss/policydb.c for (c = genfs->head; c; c = c->next) next 3236 security/selinux/ss/policydb.c for (c = genfs->head; c; c = c->next) { next 88 security/selinux/ss/policydb.h struct role_trans *next; next 105 security/selinux/ss/policydb.h struct role_allow *next; next 202 security/selinux/ss/policydb.h struct ocontext *next; next 208 security/selinux/ss/policydb.h struct genfs *next; next 276 security/selinux/ss/services.c for (e = cexpr; e; e = e->next) { next 689 security/selinux/ss/services.c constraint = constraint->next; next 700 security/selinux/ss/services.c for (ra = policydb->role_allow; ra; ra = ra->next) { next 820 security/selinux/ss/services.c constraint = constraint->next; next 1780 security/selinux/ss/services.c roletr = roletr->next) { next 2017 security/selinux/ss/services.c oc = oc->next; next 2278 security/selinux/ss/services.c c = c->next; next 2325 security/selinux/ss/services.c c = c->next; next 2372 security/selinux/ss/services.c c = c->next; next 2414 security/selinux/ss/services.c c = c->next; next 2489 security/selinux/ss/services.c c = c->next; next 2503 security/selinux/ss/services.c c = c->next; next 2688 security/selinux/ss/services.c for (genfs = policydb->genfs; genfs; genfs = genfs->next) { next 2698 security/selinux/ss/services.c for (c = genfs->head; c; c = c->next) { next 2767 security/selinux/ss/services.c c = c->next; next 2886 security/selinux/ss/services.c for (cur = policydb->cond_list; cur; cur = cur->next) { next 2943 security/selinux/ss/services.c for (cur = policydb->cond_list; cur; cur = cur->next) { next 725 security/smack/smack_lsm.c char *next = strchr(from, ','); next 729 security/smack/smack_lsm.c if (next) next 730 security/smack/smack_lsm.c len = next - from; next 632 security/smack/smackfs.c .next = load2_seq_next, next 802 security/smack/smackfs.c .next = cipso_seq_next, next 985 security/smack/smackfs.c .next = cipso_seq_next, next 1061 security/smack/smackfs.c .next = net4addr_seq_next, next 1098 security/smack/smackfs.c m = list_entry_rcu(smk_net4addr_list.next, next 1112 security/smack/smackfs.c m_next = list_entry_rcu(m->list.next, next 1325 security/smack/smackfs.c .next = net6addr_seq_next, next 1362 security/smack/smackfs.c m = list_entry_rcu(smk_net6addr_list.next, next 1375 security/smack/smackfs.c m_next = list_entry_rcu(m->list.next, next 1885 security/smack/smackfs.c .next = onlycap_seq_next, next 1913 security/smack/smackfs.c first = public->next; next 1917 security/smack/smackfs.c private->prev->next = public; next 1918 security/smack/smackfs.c private->next->prev = public; next 1919 security/smack/smackfs.c rcu_assign_pointer(public->next, private->next); next 1926 security/smack/smackfs.c private->next = first; next 1929 security/smack/smackfs.c last->next = private; next 2218 security/smack/smackfs.c .next = load_self_seq_next, next 2349 security/smack/smackfs.c .next = load2_seq_next, next 2426 security/smack/smackfs.c .next = load_self2_seq_next, next 2694 security/smack/smackfs.c .next = relabel_self_seq_next, next 449 security/tomoyo/audit.c ptr = list_entry(tomoyo_log.next, typeof(*ptr), list); next 753 security/tomoyo/common.c next: next 833 security/tomoyo/common.c goto next; next 2552 security/tomoyo/common.c if (!ns || (head->r.eof && ns->next != &tomoyo_namespace_list)) { next 2555 security/tomoyo/common.c head->r.ns = ns ? ns->next : tomoyo_namespace_list.next; next 2570 security/tomoyo/common.c head->r.ns->next != &tomoyo_namespace_list; next 1331 security/tomoyo/common.h pos = srcu_dereference((head)->next, &tomoyo_ss); \ next 1332 security/tomoyo/common.h for ( ; pos != (head); pos = srcu_dereference(pos->next, &tomoyo_ss)) next 96 sound/core/control.c cread = snd_kctl_event(ctl->events.next); next 1621 sound/core/control.c kev = snd_kctl_event(ctl->events.next); next 1844 sound/core/control.c control = snd_kcontrol(card->controls.next); next 224 sound/core/device.c struct snd_device *dev, *next; next 228 sound/core/device.c list_for_each_entry_safe_reverse(dev, next, &card->devices, list) { next 237 sound/core/device.c list_for_each_entry_safe_reverse(dev, next, &card->devices, list) next 310 sound/core/info.c size_t next; next 318 sound/core/info.c next = pos + count; next 320 sound/core/info.c if (next > 16 * 1024) next 331 sound/core/info.c if (next > buf->len) { next 332 sound/core/info.c char *nbuf = kvzalloc(PAGE_ALIGN(next), GFP_KERNEL); next 339 sound/core/info.c buf->len = PAGE_ALIGN(next); next 345 sound/core/info.c buf->size = next; next 350 sound/core/info.c *offset = next; next 542 sound/core/oss/pcm_oss.c struct snd_pcm_plugin *plugin, *next; next 546 sound/core/oss/pcm_oss.c next = plugin->next; next 548 sound/core/oss/pcm_oss.c plugin = next; next 557 sound/core/oss/pcm_oss.c plugin->next = runtime->oss.plugin_first; next 572 sound/core/oss/pcm_oss.c plugin->next = NULL; next 575 sound/core/oss/pcm_oss.c runtime->oss.plugin_last->next = plugin; next 2302 sound/core/oss/pcm_oss.c setup = setup->next) { next 2927 sound/core/oss/pcm_oss.c setup = setup->next; next 2938 sound/core/oss/pcm_oss.c setupn = setup->next; next 2963 sound/core/oss/pcm_oss.c for (setup = pstr->oss.setup_list; setup; setup = setup->next) { next 3007 sound/core/oss/pcm_oss.c setup1->next; setup1 = setup1->next); next 3008 sound/core/oss/pcm_oss.c setup1->next = setup; next 111 sound/core/oss/pcm_plugin.c while (plugin->next) { next 116 sound/core/oss/pcm_plugin.c plugin = plugin->next; next 225 sound/core/oss/pcm_plugin.c plugin_next = plugin->next; next 255 sound/core/oss/pcm_plugin.c plugin_next = plugin->next; next 618 sound/core/oss/pcm_plugin.c struct snd_pcm_plugin *plugin, *next; next 627 sound/core/oss/pcm_plugin.c if ((next = plugin->next) != NULL) { next 634 sound/core/oss/pcm_plugin.c if ((err = next->client_channels(next, frames1, &dst_channels)) < 0) { next 651 sound/core/oss/pcm_plugin.c plugin = next; next 658 sound/core/oss/pcm_plugin.c struct snd_pcm_plugin *plugin, *next; next 670 sound/core/oss/pcm_plugin.c if ((next = plugin->next) != NULL) { next 681 sound/core/oss/pcm_plugin.c plugin = next; next 60 sound/core/oss/pcm_plugin.h struct snd_pcm_plugin *next; next 137 sound/core/pcm.c substream = substream->next) next 678 sound/core/pcm.c prev->next = substream; next 688 sound/core/pcm.c prev->next = NULL; next 830 sound/core/pcm.c substream_next = substream->next; next 837 sound/core/pcm.c setupn = setup->next; next 907 sound/core/pcm.c substream = substream->next) { next 920 sound/core/pcm.c substream = substream->next) next 933 sound/core/pcm.c for (substream = pstr->substream; substream; substream = substream->next) { next 1075 sound/core/pcm.c for (substream = pcm->streams[cidx].substream; substream; substream = substream->next) next 1097 sound/core/pcm.c for (substream = pcm->streams[cidx].substream; substream; substream = substream->next) { next 477 sound/core/pcm_lib.c for (substream = stream->substream; substream != NULL; substream = substream->next) next 93 sound/core/pcm_memory.c for (substream = pcm->streams[stream].substream; substream; substream = substream->next) next 249 sound/core/pcm_memory.c for (substream = pcm->streams[stream].substream; substream; substream = substream->next) next 1524 sound/core/pcm_native.c substream; substream = substream->next) { next 1631 sound/core/rawmidi.c substream = list_entry(stream->substreams.next, struct snd_rawmidi_substream, list); next 123 sound/core/seq/seq_fifo.c f->tail->next = cell; next 127 sound/core/seq/seq_fifo.c cell->next = NULL; next 147 sound/core/seq/seq_fifo.c f->head = cell->next; next 153 sound/core/seq/seq_fifo.c cell->next = NULL; next 205 sound/core/seq/seq_fifo.c cell->next = f->head; next 227 sound/core/seq/seq_fifo.c struct snd_seq_event_cell *cell, *next, *oldhead; next 258 sound/core/seq/seq_fifo.c for (cell = oldhead; cell; cell = next) { next 259 sound/core/seq/seq_fifo.c next = cell->next; next 96 sound/core/seq/seq_memory.c for (; len > 0 && cell; cell = cell->next) { next 166 sound/core/seq/seq_memory.c cell->next = pool->free; next 189 sound/core/seq/seq_memory.c nextptr = curp->next; next 190 sound/core/seq/seq_memory.c curp->next = pool->free; next 255 sound/core/seq/seq_memory.c pool->free = cell->next; next 262 sound/core/seq/seq_memory.c cell->next = NULL; next 331 sound/core/seq/seq_memory.c tail->next = tmp; next 336 sound/core/seq/seq_memory.c src = src->next; next 396 sound/core/seq/seq_memory.c cellptr->next = pool->free; next 18 sound/core/seq/seq_memory.h struct snd_seq_event_cell *next; /* next cell */ next 153 sound/core/seq/seq_prioq.c f->tail->next = cell; next 155 sound/core/seq/seq_prioq.c cell->next = NULL; next 180 sound/core/seq/seq_prioq.c cur = cur->next; next 190 sound/core/seq/seq_prioq.c prev->next = cell; next 191 sound/core/seq/seq_prioq.c cell->next = cur; next 228 sound/core/seq/seq_prioq.c f->head = cell->next; next 234 sound/core/seq/seq_prioq.c cell->next = NULL; next 277 sound/core/seq/seq_prioq.c register struct snd_seq_event_cell *cell, *next; next 286 sound/core/seq/seq_prioq.c next = cell->next; next 290 sound/core/seq/seq_prioq.c f->head = cell->next; next 292 sound/core/seq/seq_prioq.c prev->next = cell->next; next 295 sound/core/seq/seq_prioq.c f->tail = cell->next; next 298 sound/core/seq/seq_prioq.c cell->next = NULL; next 302 sound/core/seq/seq_prioq.c freeprev->next = cell; next 316 sound/core/seq/seq_prioq.c cell = next; next 322 sound/core/seq/seq_prioq.c freenext = freefirst->next; next 387 sound/core/seq/seq_prioq.c struct snd_seq_event_cell *cell, *next; next 397 sound/core/seq/seq_prioq.c next = cell->next; next 403 sound/core/seq/seq_prioq.c f->head = cell->next; next 405 sound/core/seq/seq_prioq.c prev->next = cell->next; next 409 sound/core/seq/seq_prioq.c f->tail = cell->next; next 413 sound/core/seq/seq_prioq.c cell->next = NULL; next 417 sound/core/seq/seq_prioq.c freeprev->next = cell; next 424 sound/core/seq/seq_prioq.c cell = next; next 430 sound/core/seq/seq_prioq.c freenext = freefirst->next; next 296 sound/core/timer.c list_entry(timer->open_list_head.next, next 782 sound/core/timer.c list_del_init(head->next); next 1528 sound/core/timer.c timer = list_entry(snd_timer_list.next, next 309 sound/drivers/opl3/opl3_synth.c for (patch = opl3->patch_table[key]; patch; patch = patch->next) { next 321 sound/drivers/opl3/opl3_synth.c patch->next = opl3->patch_table[key]; next 334 sound/drivers/opl3/opl3_synth.c struct fm_patch *patch, *next; next 335 sound/drivers/opl3/opl3_synth.c for (patch = opl3->patch_table[i]; patch; patch = next) { next 336 sound/drivers/opl3/opl3_synth.c next = patch->next; next 468 sound/drivers/opl4/opl4_synth.c return list_entry(opl4->off_voices.next, struct opl4_voice, list); next 471 sound/drivers/opl4/opl4_synth.c return list_entry(opl4->on_voices.next, struct opl4_voice, list); next 1217 sound/firewire/amdtp-stream.c struct amdtp_stream *s, *next; next 1219 sound/firewire/amdtp-stream.c list_for_each_entry_safe(s, next, &d->streams, list) { next 42 sound/i2c/i2c.c device = snd_i2c_device(bus->devices.next); next 49 sound/i2c/i2c.c slave = snd_i2c_slave_bus(bus->buses.next); next 91 sound/isa/gus/gus_dma.c gus->gf1.dma_data_pcm = block->next; next 99 sound/isa/gus/gus_dma.c gus->gf1.dma_data_synth = block->next; next 166 sound/isa/gus/gus_dma.c gus->gf1.dma_data_pcm = block->next; next 170 sound/isa/gus/gus_dma.c gus->gf1.dma_data_synth = block->next; next 193 sound/isa/gus/gus_dma.c block->next = NULL; next 207 sound/isa/gus/gus_dma.c gus->gf1.dma_data_synth_last->next = block; next 215 sound/isa/gus/gus_dma.c gus->gf1.dma_data_pcm_last->next = block; next 40 sound/isa/gus/gus_mem.c nblock->next = pblock; next 45 sound/isa/gus/gus_mem.c nblock->prev->next = nblock; next 49 sound/isa/gus/gus_mem.c pblock = pblock->next; next 51 sound/isa/gus/gus_mem.c nblock->next = NULL; next 57 sound/isa/gus/gus_mem.c alloc->last->next = nblock; next 71 sound/isa/gus/gus_mem.c alloc->first = block->next; next 72 sound/isa/gus/gus_mem.c if (block->next) next 73 sound/isa/gus/gus_mem.c block->next->prev = NULL; next 75 sound/isa/gus/gus_mem.c block->prev->next = block->next; next 76 sound/isa/gus/gus_mem.c if (block->next) next 77 sound/isa/gus/gus_mem.c block->next->prev = block->prev; next 82 sound/isa/gus/gus_mem.c block->prev->next = NULL; next 84 sound/isa/gus/gus_mem.c block->next->prev = block->prev; next 86 sound/isa/gus/gus_mem.c block->prev->next = block->next; next 98 sound/isa/gus/gus_mem.c for (block = alloc->first; block; block = block->next) { next 114 sound/isa/gus/gus_mem.c for (block = alloc->first; block; block = block->next) next 139 sound/isa/gus/gus_mem.c block->prev = block->next = NULL; next 140 sound/isa/gus/gus_mem.c for (pblock = alloc->first, idx = 0; pblock; pblock = pblock->next) { next 146 sound/isa/gus/gus_mem.c if (pblock->next) { next 147 sound/isa/gus/gus_mem.c if (pblock->ptr + pblock->size == pblock->next->ptr) next 149 sound/isa/gus/gus_mem.c if (pblock->next->ptr < boundary) next 150 sound/isa/gus/gus_mem.c ptr2 = pblock->next->ptr; next 261 sound/isa/gus/gus_mem.c nblock = block->next; next 292 sound/isa/gus/gus_mem.c for (block = alloc->first, i = 0; block; block = block->next, i++) { next 877 sound/isa/gus/gus_pcm.c for (substream = pcm->streams[SNDRV_PCM_STREAM_PLAYBACK].substream; substream; substream = substream->next) next 199 sound/pci/atiixp.c __le32 next; /* address of the next packet descriptor */ next 384 sound/pci/atiixp.c desc->next = cpu_to_le32((u32)dma->desc_buf.addr); next 386 sound/pci/atiixp.c desc->next = cpu_to_le32(desc_addr); next 175 sound/pci/atiixp_modem.c __le32 next; /* address of the next packet descriptor */ next 351 sound/pci/atiixp_modem.c desc->next = cpu_to_le32((u32)dma->desc_buf.addr); next 353 sound/pci/atiixp_modem.c desc->next = cpu_to_le32(desc_addr); next 1390 sound/pci/ca0106/ca0106_main.c substream = substream->next) { next 1398 sound/pci/ca0106/ca0106_main.c substream = substream->next) { next 638 sound/pci/ctxfi/ctdaio.c hw->daio_mgr_set_imapnxt(mgr->ctrl_blk, entry->next); next 710 sound/pci/ctxfi/ctdaio.c entry->slot = entry->addr = entry->next = entry->user = 0; next 126 sound/pci/ctxfi/cthardware.h int (*srcimp_mgr_set_imapnxt)(void *blk, unsigned int next); next 172 sound/pci/ctxfi/cthardware.h int (*daio_mgr_set_imapnxt)(void *blk, unsigned int next); next 541 sound/pci/ctxfi/cthw20k1.c static int srcimp_mgr_set_imapnxt(void *blk, unsigned int next) next 545 sound/pci/ctxfi/cthw20k1.c set_field(&ctl->srcimap.srcaim, SRCAIM_NXT, next); next 1080 sound/pci/ctxfi/cthw20k1.c static int daio_mgr_set_imapnxt(void *blk, unsigned int next) next 1084 sound/pci/ctxfi/cthw20k1.c set_field(&ctl->daoimap.aim, AIM_NXT, next); next 541 sound/pci/ctxfi/cthw20k2.c static int srcimp_mgr_set_imapnxt(void *blk, unsigned int next) next 545 sound/pci/ctxfi/cthw20k2.c set_field(&ctl->srcimap.srcaim, SRCAIM_NXT, next); next 1035 sound/pci/ctxfi/cthw20k2.c static int daio_mgr_set_imapnxt(void *blk, unsigned int next) next 1039 sound/pci/ctxfi/cthw20k2.c set_field(&ctl->daoimap.aim, AIM_NXT, next); next 27 sound/pci/ctxfi/ctimap.c entry->next = entry->addr; next 49 sound/pci/ctxfi/ctimap.c pos = head->next; next 56 sound/pci/ctxfi/ctimap.c entry->next = pos_ent->addr; next 58 sound/pci/ctxfi/ctimap.c pre_ent->next = entry->addr; next 67 sound/pci/ctxfi/ctimap.c struct list_head *next, *pre, *head; next 76 sound/pci/ctxfi/ctimap.c next = (entry->list.next == head) ? head->next : entry->list.next; next 80 sound/pci/ctxfi/ctimap.c entry->next = entry->addr = entry->user = entry->slot = 0; next 87 sound/pci/ctxfi/ctimap.c next_ent = list_entry(next, struct imapper, list); next 89 sound/pci/ctxfi/ctimap.c pre_ent->next = next_ent->addr; next 102 sound/pci/ctxfi/ctimap.c pos = head->next; next 24 sound/pci/ctxfi/ctimap.h unsigned short next; /* the next input mapper ram id */ next 784 sound/pci/ctxfi/ctsrc.c hw->srcimp_mgr_set_imapnxt(mgr->ctrl_blk, entry->next); next 848 sound/pci/ctxfi/ctsrc.c entry->slot = entry->addr = entry->next = entry->user = 0; next 223 sound/pci/ctxfi/ctvmem.c pos = vm->used.next; next 229 sound/pci/ctxfi/ctvmem.c pos = vm->unused.next; next 878 sound/pci/echoaudio/echoaudio.c for (ss = pcm->streams[stream].substream; ss; ss = ss->next) next 755 sound/pci/emu10k1/emu10k1_main.c goto next; next 770 sound/pci/emu10k1/emu10k1_main.c goto next; next 789 sound/pci/emu10k1/emu10k1_main.c next: next 397 sound/pci/emu10k1/emufx.c nirq = irq->next; /* irq ptr can be removed from list */ next 418 sound/pci/emu10k1/emufx.c irq->next = NULL; next 425 sound/pci/emu10k1/emufx.c irq->next = emu->fx8010.irq_handlers; next 440 sound/pci/emu10k1/emufx.c emu->fx8010.irq_handlers = tmp->next; next 446 sound/pci/emu10k1/emufx.c while (tmp && tmp->next != irq) next 447 sound/pci/emu10k1/emufx.c tmp = tmp->next; next 449 sound/pci/emu10k1/emufx.c tmp->next = tmp->next->next; next 1415 sound/pci/emu10k1/emupcm.c for (substream = pcm->streams[SNDRV_PCM_STREAM_PLAYBACK].substream; substream; substream = substream->next) next 1420 sound/pci/emu10k1/emupcm.c for (substream = pcm->streams[SNDRV_PCM_STREAM_CAPTURE].substream; substream; substream = substream->next) next 1446 sound/pci/emu10k1/emupcm.c for (substream = pcm->streams[SNDRV_PCM_STREAM_PLAYBACK].substream; substream; substream = substream->next) next 137 sound/pci/emu10k1/memory.c struct list_head *next; next 139 sound/pci/emu10k1/memory.c page = search_empty_map_area(emu, blk->pages, &next); next 147 sound/pci/emu10k1/memory.c list_add_tail(&blk->mapped_link, next); next 177 sound/pci/emu10k1/memory.c if ((p = blk->mapped_link.next) != &emu->mapped_link_head) { next 273 sound/pci/emu10k1/memory.c p = emu->mapped_order_link_head.next; next 275 sound/pci/emu10k1/memory.c nextp = p->next; next 463 sound/pci/emu10k1/memory.c if ((p = blk->mem.list.next) != &hdr->block) { next 644 sound/pci/emu10k1/p16v.c substream = substream->next) { next 657 sound/pci/emu10k1/p16v.c substream = substream->next) { next 1394 sound/pci/es1968.c if (buf->list.next != &chip->buf_list) { next 1395 sound/pci/es1968.c chunk = list_entry(buf->list.next, struct esm_memory, list); next 1412 sound/pci/es1968.c while ((p = chip->buf_list.next) != &chip->buf_list) { next 1317 sound/pci/hda/hda_controller.c struct hda_codec *codec, *next; next 1322 sound/pci/hda/hda_controller.c list_for_each_codec_safe(codec, next, &chip->bus) { next 130 sound/pci/hda/hda_jack.c struct hda_jack_callback *cb, *next; next 136 sound/pci/hda/hda_jack.c for (cb = jack->callback; cb; cb = next) { next 137 sound/pci/hda/hda_jack.c next = cb->next; next 259 sound/pci/hda/hda_jack.c callback->next = jack->callback; next 536 sound/pci/hda/hda_jack.c for (cb = jack->callback; cb; cb = cb->next) { next 545 sound/pci/hda/hda_jack.c for (cb = gated->callback; cb; cb = cb->next) { next 26 sound/pci/hda/hda_jack.h struct hda_jack_callback *next; next 943 sound/pci/mixart/mixart.c for (subs = pcm->streams[stream].substream; subs; subs = subs->next, idx++) next 2187 sound/pci/trident/trident_main.c for (substream = pcm->streams[SNDRV_PCM_STREAM_PLAYBACK].substream; substream; substream = substream->next) next 2234 sound/pci/trident/trident_main.c substream = substream->next; next 2236 sound/pci/trident/trident_main.c substream = substream->next; next 2239 sound/pci/trident/trident_main.c substream = substream->next; next 1871 sound/pci/ymfpci/ymfpci_main.c substream = substream->next; next 100 sound/soc/au1x/ac97c.c goto next; next 116 sound/soc/au1x/ac97c.c next: next 139 sound/soc/au1x/ac97c.c goto next; next 148 sound/soc/au1x/ac97c.c next: next 30 sound/soc/au1x/dma.c struct pcm_period *next; next 56 sound/soc/au1x/dma.c pointer_next = pointer->next; next 92 sound/soc/au1x/dma.c pointer->next = kmalloc(sizeof(struct pcm_period), next 94 sound/soc/au1x/dma.c if (!pointer->next) { next 98 sound/soc/au1x/dma.c pointer = pointer->next; next 101 sound/soc/au1x/dma.c pointer->next = stream->buffer; next 121 sound/soc/au1x/dma.c set_dma_addr1(stream->dma, stream->buffer->next->start); next 127 sound/soc/au1x/dma.c set_dma_addr0(stream->dma, stream->buffer->next->start); next 141 sound/soc/au1x/dma.c stream->buffer = stream->buffer->next; next 143 sound/soc/au1x/dma.c set_dma_addr0(stream->dma, stream->buffer->next->start); next 148 sound/soc/au1x/dma.c stream->buffer = stream->buffer->next; next 150 sound/soc/au1x/dma.c set_dma_addr1(stream->dma, stream->buffer->next->start); next 461 sound/soc/fsl/fsl_dma.c dma_private->link[i].next = cpu_to_be64(temp_link); next 466 sound/soc/fsl/fsl_dma.c dma_private->link[i - 1].next = cpu_to_be64(dma_private->ld_buf_phys); next 103 sound/soc/fsl/fsl_dma.h __be64 next; /* Address of next list descriptor */ next 121 sound/soc/fsl/fsl_dma.h __be64 next; /* Address of next link descriptor */ next 156 sound/soc/intel/atom/sst/sst_ipc.c msg = list_entry(sst_drv_ctx->ipc_dispatch_list.next, next 1218 sound/soc/sh/rcar/core.c substream = substream->next) { next 1321 sound/soc/soc-dapm.c list_del(widgets.next); next 113 sound/sound_core.c struct sound_unit *next; next 169 sound/sound_core.c list=&((*list)->next); next 176 sound/sound_core.c list=&((*list)->next); next 189 sound/sound_core.c list=&((*list)->next); next 204 sound/sound_core.c s->next=*list; next 222 sound/sound_core.c *list=p->next; next 225 sound/sound_core.c list=&(p->next); next 531 sound/sound_core.c s=s->next; next 342 sound/sparc/amd7930.c struct snd_amd7930 *next; next 1050 sound/sparc/amd7930.c amd->next = amd7930_list; next 1088 sound/sparc/amd7930.c struct snd_amd7930 *next = p->next; next 1092 sound/sparc/amd7930.c p = next; next 118 sound/sparc/cs4231.c struct snd_cs4231 *next; next 263 sound/synth/emux/soundfont.c for (sf = sflist->fonts; sf; sf = sf->next) { next 278 sound/synth/emux/soundfont.c sf->next = sflist->fonts; next 354 sound/synth/emux/soundfont.c zp->next = sf->zones; next 387 sound/synth/emux/soundfont.c sp->next = sf->samples; next 404 sound/synth/emux/soundfont.c sf->samples = sp->next; next 432 sound/synth/emux/soundfont.c for (zp = sf->zones; zp; prevp = zp, zp = zp->next) { next 443 sound/synth/emux/soundfont.c prevp->next = zp->next; next 444 sound/synth/emux/soundfont.c zp->next = sf->zones; next 480 sound/synth/emux/soundfont.c struct snd_sf_zone *prev, *next, *p; next 484 sound/synth/emux/soundfont.c for (p = sf->zones; p; p = next) { next 485 sound/synth/emux/soundfont.c next = p->next; next 490 sound/synth/emux/soundfont.c prev->next = next; next 492 sound/synth/emux/soundfont.c sf->zones = next; next 550 sound/synth/emux/soundfont.c for (zone = sf->zones; zone; zone = zone->next) { next 680 sound/synth/emux/soundfont.c for (p = sf->samples; p; p = p->next) { next 1144 sound/synth/emux/soundfont.c for (sf = sflist->fonts; sf; sf = sf->next) { next 1145 sound/synth/emux/soundfont.c for (cur = sf->zones; cur; cur = cur->next) { next 1362 sound/synth/emux/soundfont.c nextsf = sf->next; next 1364 sound/synth/emux/soundfont.c nextzp = zp->next; next 1368 sound/synth/emux/soundfont.c nextsp = sp->next; next 1457 sound/synth/emux/soundfont.c for (sf = sflist->fonts; sf; sf = sf->next) { next 1461 sound/synth/emux/soundfont.c nextzp = zp->next; next 1469 sound/synth/emux/soundfont.c nextsp = sp->next; next 49 sound/synth/util_mem.c while ((p = hdr->block.next) != &hdr->block) { next 33 sound/usb/helper.c u8 *p, *end, *next; next 40 sound/usb/helper.c next = p + p[0]; next 41 sound/usb/helper.c if (next > end) next 46 sound/usb/helper.c p = next; next 888 tools/bpf/bpf_dbg.c uint32_t pkt_len, int next) next 896 tools/bpf/bpf_dbg.c if (i++ == next) next 3 tools/firewire/list.h struct list *next, *prev; next 9 tools/firewire/list.h list->next = list; next 16 tools/firewire/list.h return list->next == list; next 23 tools/firewire/list.h new_link->next = link; next 24 tools/firewire/list.h new_link->prev->next = new_link; next 25 tools/firewire/list.h new_link->next->prev = new_link; next 37 tools/firewire/list.h list_insert(list->next, new_link); next 43 tools/firewire/list.h link->prev->next = link->next; next 44 tools/firewire/list.h link->next->prev = link->prev; next 51 tools/firewire/list.h list_entry((list)->next, type, member) next 57 tools/firewire/list.h list_entry((elm)->member.next, typeof(*elm), member) next 29 tools/firmware/ihex2fw.c struct ihex_binrec *next; /* not part of the real data structure */ next 259 tools/firmware/ihex2fw.c p = &((*p)->next); next 261 tools/firmware/ihex2fw.c record->next = *p; next 282 tools/firmware/ihex2fw.c p = p->next; next 27 tools/include/linux/list.h list->next = list; next 40 tools/include/linux/list.h struct list_head *next) next 42 tools/include/linux/list.h next->prev = new; next 43 tools/include/linux/list.h new->next = next; next 45 tools/include/linux/list.h prev->next = new; next 50 tools/include/linux/list.h struct list_head *next); next 63 tools/include/linux/list.h __list_add(new, head, head->next); next 87 tools/include/linux/list.h static inline void __list_del(struct list_head * prev, struct list_head * next) next 89 tools/include/linux/list.h next->prev = prev; next 90 tools/include/linux/list.h WRITE_ONCE(prev->next, next); next 102 tools/include/linux/list.h __list_del(entry->prev, entry->next); next 107 tools/include/linux/list.h __list_del(entry->prev, entry->next); next 108 tools/include/linux/list.h entry->next = LIST_POISON1; next 126 tools/include/linux/list.h new->next = old->next; next 127 tools/include/linux/list.h new->next->prev = new; next 129 tools/include/linux/list.h new->prev->next = new; next 180 tools/include/linux/list.h return list->next == head; next 189 tools/include/linux/list.h return head->next == head; next 207 tools/include/linux/list.h struct list_head *next = head->next; next 208 tools/include/linux/list.h return (next == head) && (next == head->prev); next 220 tools/include/linux/list.h first = head->next; next 231 tools/include/linux/list.h return !list_empty(head) && (head->next == head->prev); next 237 tools/include/linux/list.h struct list_head *new_first = entry->next; next 238 tools/include/linux/list.h list->next = head->next; next 239 tools/include/linux/list.h list->next->prev = list; next 241 tools/include/linux/list.h entry->next = list; next 242 tools/include/linux/list.h head->next = new_first; next 266 tools/include/linux/list.h (head->next != entry && head != entry)) next 276 tools/include/linux/list.h struct list_head *next) next 278 tools/include/linux/list.h struct list_head *first = list->next; next 282 tools/include/linux/list.h prev->next = first; next 284 tools/include/linux/list.h last->next = next; next 285 tools/include/linux/list.h next->prev = last; next 297 tools/include/linux/list.h __list_splice(list, head, head->next); next 323 tools/include/linux/list.h __list_splice(list, head, head->next); next 363 tools/include/linux/list.h list_entry((ptr)->next, type, member) next 393 tools/include/linux/list.h list_entry((pos)->member.next, typeof(*(pos)), member) next 409 tools/include/linux/list.h for (pos = (head)->next; pos != (head); pos = pos->next) next 426 tools/include/linux/list.h for (pos = (head)->next, n = pos->next; pos != (head); \ next 427 tools/include/linux/list.h pos = n, n = pos->next) next 600 tools/include/linux/list.h h->next = NULL; next 616 tools/include/linux/list.h struct hlist_node *next = n->next; next 619 tools/include/linux/list.h WRITE_ONCE(*pprev, next); next 620 tools/include/linux/list.h if (next) next 621 tools/include/linux/list.h next->pprev = pprev; next 627 tools/include/linux/list.h n->next = LIST_POISON1; next 642 tools/include/linux/list.h n->next = first; next 644 tools/include/linux/list.h first->pprev = &n->next; next 651 tools/include/linux/list.h struct hlist_node *next) next 653 tools/include/linux/list.h n->pprev = next->pprev; next 654 tools/include/linux/list.h n->next = next; next 655 tools/include/linux/list.h next->pprev = &n->next; next 662 tools/include/linux/list.h n->next = prev->next; next 663 tools/include/linux/list.h prev->next = n; next 664 tools/include/linux/list.h n->pprev = &prev->next; next 666 tools/include/linux/list.h if (n->next) next 667 tools/include/linux/list.h n->next->pprev = &n->next; next 673 tools/include/linux/list.h n->pprev = &n->next; next 678 tools/include/linux/list.h return h->pprev == &h->next; next 697 tools/include/linux/list.h for (pos = (head)->first; pos ; pos = pos->next) next 700 tools/include/linux/list.h for (pos = (head)->first; pos && ({ n = pos->next; 1; }); \ next 717 tools/include/linux/list.h pos = hlist_entry_safe((pos)->member.next, typeof(*(pos)), member)) next 725 tools/include/linux/list.h for (pos = hlist_entry_safe((pos)->member.next, typeof(*(pos)), member);\ next 727 tools/include/linux/list.h pos = hlist_entry_safe((pos)->member.next, typeof(*(pos)), member)) next 736 tools/include/linux/list.h pos = hlist_entry_safe((pos)->member.next, typeof(*(pos)), member)) next 747 tools/include/linux/list.h pos && ({ n = pos->member.next; 1; }); \ next 760 tools/include/linux/list.h begin->prev->next = end->next; next 761 tools/include/linux/list.h end->next->prev = begin->prev; next 770 tools/include/linux/list.h for (; pos != (head); pos = pos->next) next 70 tools/include/linux/types.h struct list_head *next, *prev; next 78 tools/include/linux/types.h struct hlist_node *next, **pprev; next 116 tools/include/uapi/drm/drm.h unsigned char next; next 144 tools/io_uring/queue.c unsigned next = sq->sqe_tail + 1; next 150 tools/io_uring/queue.c if (next - sq->sqe_head > *sq->kring_entries) next 154 tools/io_uring/queue.c sq->sqe_tail = next; next 21 tools/lib/bpf/hashmap.c entry->next = *pprev; next 28 tools/lib/bpf/hashmap.c *pprev = entry->next; next 29 tools/lib/bpf/hashmap.c entry->next = NULL; next 130 tools/lib/bpf/hashmap.c prev_ptr = &cur->next, cur = cur->next) { next 32 tools/lib/bpf/hashmap.h struct hashmap_entry *next; next 140 tools/lib/bpf/hashmap.h for (cur = map->buckets[bkt]; cur; cur = cur->next) next 153 tools/lib/bpf/hashmap.h cur && ({tmp = cur->next; true; }); \ next 167 tools/lib/bpf/hashmap.h cur = cur->next) \ next 174 tools/lib/bpf/hashmap.h cur && ({ tmp = cur->next; true; }); \ next 4229 tools/lib/bpf/libbpf.c struct bpf_object *next; next 4232 tools/lib/bpf/libbpf.c next = list_first_entry(&bpf_objects_list, next 4236 tools/lib/bpf/libbpf.c next = list_next_entry(prev, list); next 4239 tools/lib/bpf/libbpf.c if (&next->list == &bpf_objects_list) next 4242 tools/lib/bpf/libbpf.c return next; next 4320 tools/lib/bpf/libbpf.c bpf_program__prev(struct bpf_program *next, const struct bpf_object *obj) next 4322 tools/lib/bpf/libbpf.c struct bpf_program *prog = next; next 4759 tools/lib/bpf/libbpf.c bpf_map__prev(const struct bpf_map *next, const struct bpf_object *obj) next 4761 tools/lib/bpf/libbpf.c if (next == NULL) { next 4767 tools/lib/bpf/libbpf.c return __bpf_map__iter(next, obj, -1); next 75 tools/lib/traceevent/event-parse.c struct event_handler *next; next 84 tools/lib/traceevent/event-parse.c struct func_params *next; next 89 tools/lib/traceevent/event-parse.c struct tep_function_handler *next; next 165 tools/lib/traceevent/event-parse.c struct cmdline_list *next; next 187 tools/lib/traceevent/event-parse.c cmdlist = cmdlist->next; next 354 tools/lib/traceevent/event-parse.c item->next = tep->cmdlist; next 403 tools/lib/traceevent/event-parse.c struct func_list *next; next 463 tools/lib/traceevent/event-parse.c funclist = funclist->next; next 622 tools/lib/traceevent/event-parse.c item->next = tep->funclist; next 679 tools/lib/traceevent/event-parse.c struct printk_list *next; next 716 tools/lib/traceevent/event-parse.c printklist = printklist->next; next 763 tools/lib/traceevent/event-parse.c item->next = tep->printklist; next 857 tools/lib/traceevent/event-parse.c struct tep_print_flag_sym *next; next 860 tools/lib/traceevent/event-parse.c next = fsym->next; next 864 tools/lib/traceevent/event-parse.c fsym = next; next 924 tools/lib/traceevent/event-parse.c arg->func.args = farg->next; next 1698 tools/lib/traceevent/event-parse.c fields = &field->next; next 2579 tools/lib/traceevent/event-parse.c list = &field->next; next 2983 tools/lib/traceevent/event-parse.c for (func = tep->func_handlers; func; func = func->next) { next 2994 tools/lib/traceevent/event-parse.c struct tep_function_handler **next; next 2996 tools/lib/traceevent/event-parse.c next = &tep->func_handlers; next 2997 tools/lib/traceevent/event-parse.c while ((func = *next)) { next 2999 tools/lib/traceevent/event-parse.c *next = func->next; next 3003 tools/lib/traceevent/event-parse.c next = &func->next; next 3050 tools/lib/traceevent/event-parse.c next_arg = &(farg->next); next 3249 tools/lib/traceevent/event-parse.c list = &arg->next; next 3256 tools/lib/traceevent/event-parse.c list = &arg->next; next 3339 tools/lib/traceevent/event-parse.c format; format = format->next) { next 3361 tools/lib/traceevent/event-parse.c format; format = format->next) { next 4020 tools/lib/traceevent/event-parse.c for (flag = arg->flags.flags; flag; flag = flag->next) { next 4042 tools/lib/traceevent/event-parse.c for (flag = arg->symbol.symbols; flag; flag = flag->next) { next 4203 tools/lib/traceevent/event-parse.c struct save_str *next; next 4238 tools/lib/traceevent/event-parse.c string->next = strings; next 4258 tools/lib/traceevent/event-parse.c farg = farg->next; next 4259 tools/lib/traceevent/event-parse.c param = param->next; next 4267 tools/lib/traceevent/event-parse.c strings = string->next; next 4279 tools/lib/traceevent/event-parse.c struct tep_print_arg *next; next 4282 tools/lib/traceevent/event-parse.c next = args->next; next 4285 tools/lib/traceevent/event-parse.c args = next; next 4293 tools/lib/traceevent/event-parse.c struct tep_print_arg *args, *arg, **next; next 4329 tools/lib/traceevent/event-parse.c arg->next = NULL; next 4330 tools/lib/traceevent/event-parse.c next = &arg->next; next 4430 tools/lib/traceevent/event-parse.c arg->next = NULL; next 4436 tools/lib/traceevent/event-parse.c *next = arg; next 4437 tools/lib/traceevent/event-parse.c next = &arg->next; next 4454 tools/lib/traceevent/event-parse.c arg->next = NULL; next 4460 tools/lib/traceevent/event-parse.c *next = arg; next 4461 tools/lib/traceevent/event-parse.c next = &arg->next; next 4949 tools/lib/traceevent/event-parse.c field = field->next; next 5038 tools/lib/traceevent/event-parse.c arg = arg->next; next 5057 tools/lib/traceevent/event-parse.c arg = arg->next; next 5066 tools/lib/traceevent/event-parse.c arg = arg->next; next 5074 tools/lib/traceevent/event-parse.c arg = arg->next; next 5105 tools/lib/traceevent/event-parse.c arg = arg->next; next 5194 tools/lib/traceevent/event-parse.c arg = arg->next; next 5370 tools/lib/traceevent/event-parse.c pid_from_cmdlist(struct tep_handle *tep, const char *comm, struct tep_cmdline *next) next 5372 tools/lib/traceevent/event-parse.c struct cmdline_list *cmdlist = (struct cmdline_list *)next; next 5375 tools/lib/traceevent/event-parse.c cmdlist = cmdlist->next; next 5380 tools/lib/traceevent/event-parse.c cmdlist = cmdlist->next; next 5399 tools/lib/traceevent/event-parse.c struct tep_cmdline *next) next 5408 tools/lib/traceevent/event-parse.c return pid_from_cmdlist(tep, comm, next); next 5410 tools/lib/traceevent/event-parse.c if (next) { next 5415 tools/lib/traceevent/event-parse.c if (next < tep->cmdlines || next 5416 tools/lib/traceevent/event-parse.c next >= tep->cmdlines + tep->cmdline_count) next 5417 tools/lib/traceevent/event-parse.c next = NULL; next 5419 tools/lib/traceevent/event-parse.c cmdline = next++; next 5422 tools/lib/traceevent/event-parse.c if (!next) next 5872 tools/lib/traceevent/event-parse.c for (field = list; field; field = field->next) { next 5922 tools/lib/traceevent/event-parse.c if (field->next) { next 5924 tools/lib/traceevent/event-parse.c print_fields(s, field->next); next 6013 tools/lib/traceevent/event-parse.c if (args->next) { next 6015 tools/lib/traceevent/event-parse.c print_args(args->next); next 6179 tools/lib/traceevent/event-parse.c struct event_handler *handle, **next; next 6181 tools/lib/traceevent/event-parse.c for (next = &tep->handlers; *next; next 6182 tools/lib/traceevent/event-parse.c next = &(*next)->next) { next 6183 tools/lib/traceevent/event-parse.c handle = *next; next 6190 tools/lib/traceevent/event-parse.c if (!(*next)) next 6199 tools/lib/traceevent/event-parse.c *next = handle->next; next 6291 tools/lib/traceevent/event-parse.c for (field = event->format.fields; field; field = field->next) { next 6306 tools/lib/traceevent/event-parse.c list = &arg->next; next 6625 tools/lib/traceevent/event-parse.c func->params = params->next; next 6705 tools/lib/traceevent/event-parse.c param->next = NULL; next 6708 tools/lib/traceevent/event-parse.c next_param = &(param->next); next 6714 tools/lib/traceevent/event-parse.c func_handle->next = tep->func_handlers; next 6835 tools/lib/traceevent/event-parse.c handle->next = tep->handlers; next 6883 tools/lib/traceevent/event-parse.c struct event_handler **next; next 6899 tools/lib/traceevent/event-parse.c for (next = &tep->handlers; *next; next = &(*next)->next) { next 6900 tools/lib/traceevent/event-parse.c handle = *next; next 6906 tools/lib/traceevent/event-parse.c if (!(*next)) next 6909 tools/lib/traceevent/event-parse.c *next = handle->next; next 6953 tools/lib/traceevent/event-parse.c struct tep_format_field *next; next 6956 tools/lib/traceevent/event-parse.c next = field->next; next 6958 tools/lib/traceevent/event-parse.c field = next; next 7012 tools/lib/traceevent/event-parse.c cmdnext = cmdlist->next; next 7027 tools/lib/traceevent/event-parse.c funcnext = funclist->next; next 7036 tools/lib/traceevent/event-parse.c tep->func_handlers = func_handler->next; next 7047 tools/lib/traceevent/event-parse.c printknext = printklist->next; next 7058 tools/lib/traceevent/event-parse.c tep->handlers = handle->next; next 52 tools/lib/traceevent/event-parse.h struct tep_record *next; next 71 tools/lib/traceevent/event-parse.h struct tep_plugin_option *next; next 145 tools/lib/traceevent/event-parse.h struct tep_format_field *next; next 184 tools/lib/traceevent/event-parse.h struct tep_print_flag_sym *next; next 257 tools/lib/traceevent/event-parse.h struct tep_print_arg *next; next 536 tools/lib/traceevent/event-parse.h struct tep_cmdline *next); next 24 tools/lib/traceevent/event-plugin.c struct registered_plugin_options *next; next 29 tools/lib/traceevent/event-plugin.c struct trace_plugin_options *next; next 36 tools/lib/traceevent/event-plugin.c struct tep_plugin_list *next; next 107 tools/lib/traceevent/event-plugin.c for (reg = registered_options; reg; reg = reg->next) { next 175 tools/lib/traceevent/event-plugin.c for (op = trace_plugin_options; op; op = op->next) { next 190 tools/lib/traceevent/event-plugin.c for (op = trace_plugin_options; op; op = op->next) { next 220 tools/lib/traceevent/event-plugin.c reg->next = registered_options; next 240 tools/lib/traceevent/event-plugin.c for (last = ®istered_options; *last; last = &(*last)->next) { next 243 tools/lib/traceevent/event-plugin.c *last = reg->next; next 267 tools/lib/traceevent/event-plugin.c list = list->next; next 313 tools/lib/traceevent/event-plugin.c list->next = *plugin_list; next 438 tools/lib/traceevent/event-plugin.c plugin_list = list->next; next 51 tools/lib/traceevent/kbuffer-parse.c unsigned int next; next 294 tools/lib/traceevent/kbuffer-parse.c kbuf->next = kbuf->size; next 309 tools/lib/traceevent/kbuffer-parse.c kbuf->next = kbuf->size; next 325 tools/lib/traceevent/kbuffer-parse.c kbuf->next = kbuf->index + length; next 335 tools/lib/traceevent/kbuffer-parse.c kbuf->curr = kbuf->next; next 336 tools/lib/traceevent/kbuffer-parse.c if (kbuf->next >= kbuf->size) next 402 tools/lib/traceevent/kbuffer-parse.c kbuf->next = kbuf->index + length; next 453 tools/lib/traceevent/kbuffer-parse.c kbuf->curr = kbuf->next; next 454 tools/lib/traceevent/kbuffer-parse.c if (kbuf->next >= kbuf->size) next 542 tools/lib/traceevent/kbuffer-parse.c kbuf->next = 0; next 671 tools/lib/traceevent/kbuffer-parse.c return kbuf->next - kbuf->index; next 683 tools/lib/traceevent/kbuffer-parse.c return kbuf->next - kbuf->curr; next 753 tools/lib/traceevent/kbuffer-parse.c void *ptr = info->next; next 771 tools/lib/traceevent/kbuffer-parse.c info->next = ptr + length; next 74 tools/lib/traceevent/kbuffer.h void *next; next 29 tools/lib/traceevent/parse-filter.c struct event_list *next; next 240 tools/lib/traceevent/parse-filter.c list->next = *events; next 331 tools/lib/traceevent/parse-filter.c events = events->next; next 1332 tools/lib/traceevent/parse-filter.c for (event = events; event; event = event->next) { next 47 tools/objtool/check.c struct instruction *next = list_next_entry(insn, list); next 49 tools/objtool/check.c if (!next || &next->list == &file->insn_list || next->sec != insn->sec) next 52 tools/objtool/check.c return next; next 58 tools/objtool/check.c struct instruction *next = list_next_entry(insn, list); next 64 tools/objtool/check.c if (&next->list != &file->insn_list && next->func == func) next 65 tools/objtool/check.c return next; next 313 tools/perf/builtin-annotate.c struct rb_node *nd = rb_first_cached(&hists->entries), *next; next 365 tools/perf/builtin-annotate.c next = rb_next(nd); next 368 tools/perf/builtin-annotate.c next = rb_prev(nd); next 374 tools/perf/builtin-annotate.c if (next != NULL) next 375 tools/perf/builtin-annotate.c nd = next; next 2106 tools/perf/builtin-c2c.c struct rb_node *next = rb_first_cached(&hists->entries); next 2109 tools/perf/builtin-c2c.c while (next) { next 2112 tools/perf/builtin-c2c.c he = rb_entry(next, struct hist_entry, rb_node); next 2116 tools/perf/builtin-c2c.c next = rb_next(&he->rb_node); next 514 tools/perf/builtin-diff.c struct rb_node *next; next 521 tools/perf/builtin-diff.c next = rb_first_cached(root); next 522 tools/perf/builtin-diff.c while (next != NULL) { next 523 tools/perf/builtin-diff.c struct hist_entry *he = rb_entry(next, struct hist_entry, rb_node_in); next 525 tools/perf/builtin-diff.c next = rb_next(&he->rb_node_in); next 675 tools/perf/builtin-diff.c struct rb_node *next = rb_first_cached(root); next 678 tools/perf/builtin-diff.c while (next != NULL) { next 679 tools/perf/builtin-diff.c struct hist_entry *he_pair = rb_entry(next, struct hist_entry, next 682 tools/perf/builtin-diff.c next = rb_next(&he_pair->rb_node_in); next 707 tools/perf/builtin-diff.c struct rb_node *next = rb_first_cached(root); next 709 tools/perf/builtin-diff.c while (next != NULL) { next 710 tools/perf/builtin-diff.c struct hist_entry *he = rb_entry(next, struct hist_entry, next 714 tools/perf/builtin-diff.c next = rb_next(&he->rb_node_in); next 733 tools/perf/builtin-diff.c struct rb_node *next; next 740 tools/perf/builtin-diff.c next = rb_first_cached(root); next 741 tools/perf/builtin-diff.c while (next != NULL) { next 747 tools/perf/builtin-diff.c he = rb_entry(next, struct hist_entry, rb_node_in); next 748 tools/perf/builtin-diff.c next = rb_next(&he->rb_node_in); next 29 tools/perf/builtin-help.c struct man_viewer_list *next; next 34 tools/perf/builtin-help.c struct man_viewer_info_list *next; next 63 tools/perf/builtin-help.c for (viewer = man_viewer_info_list; viewer; viewer = viewer->next) { next 195 tools/perf/builtin-help.c p = &((*p)->next); next 215 tools/perf/builtin-help.c new->next = man_viewer_info_list; next 368 tools/perf/builtin-help.c for (viewer = man_viewer_list; viewer; viewer = viewer->next) next 993 tools/perf/builtin-kmem.c struct rb_node *next; next 1001 tools/perf/builtin-kmem.c next = rb_first(root); next 1003 tools/perf/builtin-kmem.c while (next && n_lines--) { next 1004 tools/perf/builtin-kmem.c struct alloc_stat *data = rb_entry(next, struct alloc_stat, next 1034 tools/perf/builtin-kmem.c next = rb_next(next); next 1054 tools/perf/builtin-kmem.c struct rb_node *next = rb_first(&page_alloc_sorted); next 1070 tools/perf/builtin-kmem.c while (next && n_lines--) { next 1077 tools/perf/builtin-kmem.c data = rb_entry(next, struct page_stat, node); next 1090 tools/perf/builtin-kmem.c next = rb_next(next); next 1103 tools/perf/builtin-kmem.c struct rb_node *next = rb_first(&page_caller_sorted); next 1112 tools/perf/builtin-kmem.c while (next && n_lines--) { next 1119 tools/perf/builtin-kmem.c data = rb_entry(next, struct page_stat, node); next 1132 tools/perf/builtin-kmem.c next = rb_next(next); next 838 tools/perf/builtin-sched.c struct task_desc *prev, __maybe_unused *next; next 864 tools/perf/builtin-sched.c next = register_pid(sched, next_pid, next_comm); next 3140 tools/perf/builtin-sched.c struct rb_node *next; next 3154 tools/perf/builtin-sched.c next = rb_first_cached(&sched->sorted_atom_root); next 3156 tools/perf/builtin-sched.c while (next) { next 3159 tools/perf/builtin-sched.c work_list = rb_entry(next, struct work_atoms, node); next 3161 tools/perf/builtin-sched.c next = rb_next(next); next 90 tools/perf/builtin-timechart.c struct per_pid *next; next 107 tools/perf/builtin-timechart.c struct per_pidcomm *next; next 128 tools/perf/builtin-timechart.c struct sample_wrapper *next; next 140 tools/perf/builtin-timechart.c struct cpu_sample *next; next 159 tools/perf/builtin-timechart.c struct io_sample *next; next 174 tools/perf/builtin-timechart.c struct power_event *next; next 183 tools/perf/builtin-timechart.c struct wake_event *next; next 193 tools/perf/builtin-timechart.c struct process_filter *next; next 206 tools/perf/builtin-timechart.c cursor = cursor->next; next 211 tools/perf/builtin-timechart.c cursor->next = tchart->all_data; next 232 tools/perf/builtin-timechart.c c = c->next; next 238 tools/perf/builtin-timechart.c c->next = p->all; next 281 tools/perf/builtin-timechart.c c->next = p->all; next 290 tools/perf/builtin-timechart.c sample->next = c->samples; next 365 tools/perf/builtin-timechart.c pwr->next = tchart->power_events; next 386 tools/perf/builtin-timechart.c pwr->next = tchart->power_events; next 423 tools/perf/builtin-timechart.c we->next = tchart->wake_events; next 693 tools/perf/builtin-timechart.c pwr->next = tchart->power_events; next 708 tools/perf/builtin-timechart.c pwr->next = tchart->power_events; next 731 tools/perf/builtin-timechart.c c->next = p->all; next 743 tools/perf/builtin-timechart.c c->io_samples = prev->next; next 754 tools/perf/builtin-timechart.c sample->next = c->io_samples; next 792 tools/perf/builtin-timechart.c prev = sample->next; next 829 tools/perf/builtin-timechart.c sample->next = prev->next; next 973 tools/perf/builtin-timechart.c tchart->all_data = p->next; next 974 tools/perf/builtin-timechart.c p->next = NULL; next 978 tools/perf/builtin-timechart.c p->next = NULL; next 988 tools/perf/builtin-timechart.c p->next = prev->next; next 989 tools/perf/builtin-timechart.c prev->next = p; next 993 tools/perf/builtin-timechart.c p->next = new_list; next 1001 tools/perf/builtin-timechart.c cursor = cursor->next; next 1003 tools/perf/builtin-timechart.c prev->next = p; next 1021 tools/perf/builtin-timechart.c pwr = pwr->next; next 1031 tools/perf/builtin-timechart.c pwr = pwr->next; next 1062 tools/perf/builtin-timechart.c c = c->next; next 1074 tools/perf/builtin-timechart.c c = c->next; next 1077 tools/perf/builtin-timechart.c p = p->next; next 1096 tools/perf/builtin-timechart.c we = we->next; next 1123 tools/perf/builtin-timechart.c sample = sample->next; next 1125 tools/perf/builtin-timechart.c c = c->next; next 1127 tools/perf/builtin-timechart.c p = p->next; next 1147 tools/perf/builtin-timechart.c c = c->next; next 1153 tools/perf/builtin-timechart.c for (sample = c->io_samples; sample; sample = sample->next) { next 1240 tools/perf/builtin-timechart.c c = c->next; next 1242 tools/perf/builtin-timechart.c p = p->next; next 1261 tools/perf/builtin-timechart.c c = c->next; next 1283 tools/perf/builtin-timechart.c sample = sample->next; next 1297 tools/perf/builtin-timechart.c c = c->next; next 1299 tools/perf/builtin-timechart.c p = p->next; next 1313 tools/perf/builtin-timechart.c filt->next = process_filter; next 1330 tools/perf/builtin-timechart.c filt = filt->next; next 1368 tools/perf/builtin-timechart.c c = c->next; next 1370 tools/perf/builtin-timechart.c p = p->next; next 1409 tools/perf/builtin-timechart.c c = c->next; next 1411 tools/perf/builtin-timechart.c p = p->next; next 1441 tools/perf/builtin-timechart.c c = c->next; next 1443 tools/perf/builtin-timechart.c p = p->next; next 392 tools/perf/builtin-top.c struct rb_node *next; next 409 tools/perf/builtin-top.c next = rb_first_cached(&hists->entries); next 410 tools/perf/builtin-top.c while (next) { next 411 tools/perf/builtin-top.c n = rb_entry(next, struct hist_entry, rb_node); next 416 tools/perf/builtin-top.c next = rb_next(&n->rb_node); next 1461 tools/perf/builtin-trace.c for (field = sc->args; field; field = field->next, ++idx) { next 1545 tools/perf/builtin-trace.c sc->args = sc->args->next; next 1735 tools/perf/builtin-trace.c field = field->next, ++arg.idx, bit <<= 1) { next 2935 tools/perf/builtin-trace.c for (field = sc->args; field; field = field->next) { next 2953 tools/perf/builtin-trace.c field && candidate_field; field = field->next, candidate_field = candidate_field->next) { next 2985 tools/perf/builtin-trace.c for (candidate_field = candidate_field->next; candidate_field; candidate_field = candidate_field->next) next 82 tools/perf/jvmti/libjvmti.c for (hdr = compile_info; hdr != NULL; hdr = hdr->next) { next 113 tools/perf/jvmti/libjvmti.c for (hdr = compile_info; hdr != NULL; hdr = hdr->next) { next 87 tools/perf/lib/evlist.c struct perf_evsel *next; next 90 tools/perf/lib/evlist.c next = list_first_entry(&evlist->entries, next 94 tools/perf/lib/evlist.c next = list_next_entry(prev, node); next 98 tools/perf/lib/evlist.c if (&next->node == &evlist->entries) next 101 tools/perf/lib/evlist.c return next; next 65 tools/perf/lib/include/internal/evlist.h return list_entry(evlist->entries.next, struct perf_evsel, node); next 407 tools/perf/pmu-events/jevents.c struct event_struct *es, *next; next 409 tools/perf/pmu-events/jevents.c list_for_each_entry_safe(es, next, &arch_std_events, list) { next 599 tools/perf/tests/parse-events.c term = list_entry(terms->next, struct parse_events_term, list); next 608 tools/perf/tests/parse-events.c term = list_entry(term->list.next, struct parse_events_term, list); next 617 tools/perf/tests/parse-events.c term = list_entry(term->list.next, struct parse_events_term, list); next 626 tools/perf/tests/parse-events.c term = list_entry(term->list.next, struct parse_events_term, list); next 245 tools/perf/tests/switch-tracking.c node = list_entry(events->next, struct event_node, list); next 84 tools/perf/ui/browser.c pos = pos->next; next 113 tools/perf/ui/browser.c pos = ui_browser__list_head_filter_entries(browser, head->next); next 129 tools/perf/ui/browser.c pos = ui_browser__list_head_filter_entries(browser, pos->next); next 505 tools/perf/ui/browser.c browser->top = ui_browser__list_head_filter_entries(browser, head->next); next 337 tools/perf/ui/browsers/hists.c chain->has_children = chain->list.next != &child->val || next 340 tools/perf/ui/browsers/hists.c chain->has_children = chain->list.next == &child->val && next 353 tools/perf/ui/browsers/hists.c chain = list_entry(node->val.next, struct callchain_list, list); next 874 tools/perf/ui/browsers/hists.c struct rb_node *next = rb_next(node); next 900 tools/perf/ui/browsers/hists.c goto next; next 926 tools/perf/ui/browsers/hists.c next: next 929 tools/perf/ui/browsers/hists.c node = next; next 978 tools/perf/ui/browsers/hists.c struct rb_node *next = rb_next(node); next 986 tools/perf/ui/browsers/hists.c goto next; next 1037 tools/perf/ui/browsers/hists.c next: next 1040 tools/perf/ui/browsers/hists.c node = next; next 1067 tools/perf/ui/browsers/hists.c struct rb_node *next = rb_next(node); next 1106 tools/perf/ui/browsers/hists.c node = next; next 2800 tools/perf/ui/browsers/hists.c goto next; next 2814 tools/perf/ui/browsers/hists.c next: next 3322 tools/perf/ui/browsers/hists.c if (pos->core.node.next == &evlist->core.entries) next 486 tools/perf/ui/hist.c if (fmt->sort_list.next == NULL) next 583 tools/perf/ui/hist.c goto next; next 587 tools/perf/ui/hist.c next: next 602 tools/perf/ui/hist.c goto next; next 606 tools/perf/ui/hist.c next: next 22 tools/perf/ui/progress.c if (p->curr >= p->next) { next 25 tools/perf/ui/progress.c p->next += nr * p->step; next 34 tools/perf/ui/progress.c p->next = p->step = total / 16 ?: 1; next 11 tools/perf/ui/progress.h u64 curr, next, step, total; next 112 tools/perf/ui/stdio/hist.c struct rb_node *node, *next; next 141 tools/perf/ui/stdio/hist.c next = rb_next(node); next 142 tools/perf/ui/stdio/hist.c if (!next && (callchain_param.mode != CHAIN_GRAPH_REL || !remaining)) next 168 tools/perf/ui/stdio/hist.c node = next; next 12 tools/perf/ui/tui/progress.c p->next = p->step = p->total / (SLtt_Screen_Cols - 2) ?: 1; next 2039 tools/perf/util/annotate.c struct annotation_line *al, *next; next 2049 tools/perf/util/annotate.c next = annotation_line__next(al, ¬es->src->source); next 2050 tools/perf/util/annotate.c end = next ? next->offset : len; next 2191 tools/perf/util/annotate.c struct rb_node *next; next 2194 tools/perf/util/annotate.c next = rb_next(node); next 2198 tools/perf/util/annotate.c node = next; next 2352 tools/perf/util/annotate.c queue = list_entry(queue->node.next, typeof(*queue), node); next 412 tools/perf/util/auxtrace.c buffer = list_entry(queues->queue_array[i].head.next, next 804 tools/perf/util/auxtrace.c return list_entry(buffer->list.next, struct auxtrace_buffer, next 809 tools/perf/util/auxtrace.c return list_entry(queue->head.next, struct auxtrace_buffer, next 85 tools/perf/util/block-range.c struct block_range *next, *entry = NULL; next 117 tools/perf/util/block-range.c next = rb_entry(n, struct block_range, node); next 119 tools/perf/util/block-range.c if (next->start <= end) { /* add head: [start...][n->start...] */ next 126 tools/perf/util/block-range.c .end = next->start - 1, next 131 tools/perf/util/block-range.c rb_link_left_of_node(&head->node, &next->node); next 242 tools/perf/util/block-range.c next = block_range__next(entry); next 243 tools/perf/util/block-range.c if (!next) next 249 tools/perf/util/block-range.c if (end < next->start) { /* add tail: [...e->end][...end] */ next 274 tools/perf/util/block-range.c if (entry->end + 1 != next->start) { next 281 tools/perf/util/block-range.c .end = next->start - 1, next 286 tools/perf/util/block-range.c rb_link_left_of_node(&hole->node, &next->node); next 291 tools/perf/util/block-range.c entry = next; next 514 tools/perf/util/bpf-loader.c ref1 = ref1->next; next 515 tools/perf/util/bpf-loader.c ref2 = ref2->next; next 289 tools/perf/util/bpf-prologue.c ref = ref->next; next 427 tools/perf/util/bpf-prologue.c ref = ref->next; next 58 tools/perf/util/call-path.c if (cpr->next < cpr->sz) { next 69 tools/perf/util/call-path.c n = cpr->next++ & CALL_PATH_BLOCK_MASK; next 57 tools/perf/util/call-path.h size_t next; next 196 tools/perf/util/callchain.c goto next; next 206 tools/perf/util/callchain.c goto next; next 229 tools/perf/util/callchain.c next: next 806 tools/perf/util/callchain.c new->val.next = &to_split->list; next 809 tools/perf/util/callchain.c old_tail->next = &new->val; next 1078 tools/perf/util/callchain.c cursor->last = &node->next; next 1599 tools/perf/util/callchain.c for (node = cursor->first; node != NULL; node = node->next) next 152 tools/perf/util/callchain.h struct callchain_cursor_node *next; next 222 tools/perf/util/callchain.h cursor->curr = cursor->curr->next; next 405 tools/perf/util/data-convert-bt.c for (field = fields; field; field = field->next) { next 1018 tools/perf/util/data-convert-bt.c for (field = fields; field; field = field->next) { next 757 tools/perf/util/dso.c struct rb_node *next = rb_first(root); next 760 tools/perf/util/dso.c while (next) { next 763 tools/perf/util/dso.c cache = rb_entry(next, struct dso_cache, rb_node); next 764 tools/perf/util/dso.c next = rb_next(&cache->rb_node); next 130 tools/perf/util/env.c struct rb_node *next; next 135 tools/perf/util/env.c next = rb_first(root); next 137 tools/perf/util/env.c while (next) { next 140 tools/perf/util/env.c node = rb_entry(next, struct bpf_prog_info_node, rb_node); next 141 tools/perf/util/env.c next = rb_next(&node->rb_node); next 149 tools/perf/util/env.c next = rb_first(root); next 151 tools/perf/util/env.c while (next) { next 154 tools/perf/util/env.c node = rb_entry(next, struct btf_node, rb_node); next 155 tools/perf/util/env.c next = rb_next(&node->rb_node); next 193 tools/perf/util/evlist.c leader = list_entry(list->next, struct evsel, core.node); next 309 tools/perf/util/evsel.h return list_entry(evsel->core.node.next, struct evsel, core.node); next 370 tools/perf/util/evsel.h for ((_evsel) = list_entry((_leader)->core.node.next, struct evsel, core.node); \ next 372 tools/perf/util/evsel.h (_evsel) = list_entry((_evsel)->core.node.next, struct evsel, core.node)) next 378 tools/perf/util/evsel.h (_evsel) = list_entry((_evsel)->core.node.next, struct evsel, core.node)) next 93 tools/perf/util/evsel_fprintf.c field = field->next; next 96 tools/perf/util/evsel_fprintf.c field = field->next; next 135 tools/perf/util/evsel_fprintf.c goto next; next 185 tools/perf/util/evsel_fprintf.c next: next 920 tools/perf/util/header.c struct rb_node *next; next 931 tools/perf/util/header.c next = rb_first(root); next 932 tools/perf/util/header.c while (next) { next 936 tools/perf/util/header.c node = rb_entry(next, struct bpf_prog_info_node, rb_node); next 937 tools/perf/util/header.c next = rb_next(&node->rb_node); next 969 tools/perf/util/header.c struct rb_node *next; next 981 tools/perf/util/header.c next = rb_first(root); next 982 tools/perf/util/header.c while (next) { next 985 tools/perf/util/header.c node = rb_entry(next, struct btf_node, rb_node); next 986 tools/perf/util/header.c next = rb_next(&node->rb_node); next 1536 tools/perf/util/header.c struct rb_node *next; next 1541 tools/perf/util/header.c next = rb_first(root); next 1543 tools/perf/util/header.c while (next) { next 1546 tools/perf/util/header.c node = rb_entry(next, struct bpf_prog_info_node, rb_node); next 1547 tools/perf/util/header.c next = rb_next(&node->rb_node); next 1560 tools/perf/util/header.c struct rb_node *next; next 1565 tools/perf/util/header.c next = rb_first(root); next 1567 tools/perf/util/header.c while (next) { next 1570 tools/perf/util/header.c node = rb_entry(next, struct btf_node, rb_node); next 1571 tools/perf/util/header.c next = rb_next(&node->rb_node); next 226 tools/perf/util/hist.c struct rb_node *next = rb_first_cached(&hists->entries); next 232 tools/perf/util/hist.c while (next && row++ < max_rows) { next 233 tools/perf/util/hist.c n = rb_entry(next, struct hist_entry, rb_node); next 236 tools/perf/util/hist.c next = rb_next(&n->rb_node); next 362 tools/perf/util/hist.c struct rb_node *next = rb_first_cached(&hists->entries); next 365 tools/perf/util/hist.c while (next) { next 366 tools/perf/util/hist.c n = rb_entry(next, struct hist_entry, rb_node); next 367 tools/perf/util/hist.c next = rb_next(&n->rb_node); next 378 tools/perf/util/hist.c struct rb_node *next = rb_first_cached(&hists->entries); next 381 tools/perf/util/hist.c while (next) { next 382 tools/perf/util/hist.c n = rb_entry(next, struct hist_entry, rb_node); next 383 tools/perf/util/hist.c next = rb_next(&n->rb_node); next 391 tools/perf/util/hist.c struct rb_node *next = rb_first_cached(&hists->entries); next 395 tools/perf/util/hist.c while (next) { next 396 tools/perf/util/hist.c n = rb_entry(next, struct hist_entry, rb_node); next 400 tools/perf/util/hist.c next = rb_next(&n->rb_node); next 1590 tools/perf/util/hist.c struct rb_node *next; next 1601 tools/perf/util/hist.c next = rb_first_cached(root); next 1603 tools/perf/util/hist.c while (next) { next 1606 tools/perf/util/hist.c n = rb_entry(next, struct hist_entry, rb_node_in); next 1607 tools/perf/util/hist.c next = rb_next(&n->rb_node_in); next 1836 tools/perf/util/hist.c struct rb_node *next; next 1865 tools/perf/util/hist.c next = rb_first_cached(root); next 1868 tools/perf/util/hist.c while (next) { next 1869 tools/perf/util/hist.c n = rb_entry(next, struct hist_entry, rb_node_in); next 1870 tools/perf/util/hist.c next = rb_next(&n->rb_node_in); next 2012 tools/perf/util/hist.c goto next; next 2019 tools/perf/util/hist.c next: next 1886 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c next: next 1975 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c goto next; next 2042 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c goto next; next 2065 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c goto next; next 2103 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c goto next; next 2143 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c goto next; next 2490 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c unsigned char *next; next 2501 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c next = memmem(decoder->buf, decoder->len, INTEL_PT_PSB_STR, next 2503 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c if (!next) { next 2518 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c decoder->pkt_step = next - decoder->buf; next 2667 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c unsigned char *next; next 2669 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c next = memmem(*buf, *len, INTEL_PT_PSB_STR, INTEL_PT_PSB_LEN); next 2670 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c if (next) { next 2671 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c *len -= next - *buf; next 2672 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c *buf = next; next 2692 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c unsigned char *next; next 2697 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c next = memmem(*buf + 1, *len - 1, INTEL_PT_PSB_STR, INTEL_PT_PSB_LEN); next 2698 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c if (next) { next 2699 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c *len -= next - *buf; next 2700 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c *buf = next; next 1178 tools/perf/util/machine.c struct rb_node *next = rb_first_cached(&machines->guests); next 1182 tools/perf/util/machine.c while (next) { next 1183 tools/perf/util/machine.c struct machine *pos = rb_entry(next, struct machine, rb_node); next 1185 tools/perf/util/machine.c next = rb_next(&pos->rb_node); next 598 tools/perf/util/map.c struct rb_node *next = rb_first(root); next 600 tools/perf/util/map.c while (next) { next 601 tools/perf/util/map.c struct map *pos = rb_entry(next, struct map, rb_node); next 603 tools/perf/util/map.c next = rb_next(&pos->rb_node); next 612 tools/perf/util/map.c struct rb_node *next = rb_first(root); next 614 tools/perf/util/map.c while (next) { next 615 tools/perf/util/map.c struct map *pos = rb_entry(next, struct map, rb_node_name); next 617 tools/perf/util/map.c next = rb_next(&pos->rb_node_name); next 776 tools/perf/util/map.c struct rb_node *next, *first; next 787 tools/perf/util/map.c next = root->rb_node; next 789 tools/perf/util/map.c while (next) { next 790 tools/perf/util/map.c struct map *pos = rb_entry(next, struct map, rb_node); next 793 tools/perf/util/map.c first = next; next 796 tools/perf/util/map.c next = next->rb_left; next 798 tools/perf/util/map.c next = next->rb_right; next 801 tools/perf/util/map.c next = first; next 802 tools/perf/util/map.c while (next) { next 803 tools/perf/util/map.c struct map *pos = rb_entry(next, struct map, rb_node); next 804 tools/perf/util/map.c next = rb_next(&pos->rb_node); next 1012 tools/perf/util/map.c struct rb_node *next = rb_next(&map->rb_node); next 1014 tools/perf/util/map.c if (next) next 1015 tools/perf/util/map.c return rb_entry(next, struct map, rb_node); next 306 tools/perf/util/metricgroup.c struct rb_node *node, *next; next 387 tools/perf/util/metricgroup.c for (node = rb_first_cached(&groups.entries); node; node = next) { next 394 tools/perf/util/metricgroup.c next = rb_next(node); next 42 tools/perf/util/ordered-events.c p = last->list.next; next 141 tools/perf/util/ordered-events.c new = list_entry(cache->next, struct ordered_event, list); next 296 tools/perf/util/ordered-events.c first = list_entry(head->next, struct ordered_event, list); next 207 tools/perf/util/parse-events.c goto next; next 248 tools/perf/util/parse-events.c next: next 1575 tools/perf/util/parse-events.c leader = list_entry(list->next, struct evsel, core.node); next 2225 tools/perf/util/parse-events.c goto next; next 2248 tools/perf/util/parse-events.c next: next 2309 tools/perf/util/parse-events.c goto next; next 2321 tools/perf/util/parse-events.c next: next 24 tools/perf/util/parse-events.h struct tracepoint_path *next; next 39 tools/perf/util/parse-events.y if (!list_is_last(list->next, list)) next 1642 tools/perf/util/probe-event.c fieldp = &(*fieldp)->next; next 1884 tools/perf/util/probe-event.c field = field->next; next 1974 tools/perf/util/probe-event.c if (ref->next) { next 1975 tools/perf/util/probe-event.c depth = __synthesize_probe_trace_arg_ref(ref->next, buf, next 2000 tools/perf/util/probe-event.c ref = ref->next; next 2214 tools/perf/util/probe-event.c struct perf_probe_arg_field *field, *next; next 2228 tools/perf/util/probe-event.c next = field->next; next 2231 tools/perf/util/probe-event.c field = next; next 2275 tools/perf/util/probe-event.c field = field->next; next 2276 tools/perf/util/probe-event.c ppfield = &((*ppfield)->next); next 2313 tools/perf/util/probe-event.c struct probe_trace_arg_ref *ref, *next; next 2327 tools/perf/util/probe-event.c next = ref->next; next 2329 tools/perf/util/probe-event.c ref = next; next 2542 tools/perf/util/probe-event.c goto next; next 2546 tools/perf/util/probe-event.c goto next; next 2551 tools/perf/util/probe-event.c next: next 38 tools/perf/util/probe-event.h struct probe_trace_arg_ref *next; /* Next reference */ next 74 tools/perf/util/probe-event.h struct perf_probe_arg_field *next; /* Next field */ next 341 tools/perf/util/probe-finder.c ref_ptr = &(*ref_ptr)->next; next 435 tools/perf/util/probe-finder.c (*ref_ptr)->next = ref; next 441 tools/perf/util/probe-finder.c goto next; next 466 tools/perf/util/probe-finder.c (*ref_ptr)->next = ref; next 519 tools/perf/util/probe-finder.c next: next 521 tools/perf/util/probe-finder.c if (field->next) next 523 tools/perf/util/probe-finder.c field->next, &ref, die_mem, user_access); next 112 tools/perf/util/rblist.c struct rb_node *pos, *next = rb_first_cached(&rblist->entries); next 114 tools/perf/util/rblist.c while (next) { next 115 tools/perf/util/rblist.c pos = next; next 116 tools/perf/util/rblist.c next = rb_next(pos); next 109 tools/perf/util/scripting-engines/trace-event-perl.c if (field->next) next 110 tools/perf/util/scripting-engines/trace-event-perl.c define_symbolic_values(field->next, ev_name, field_name); next 167 tools/perf/util/scripting-engines/trace-event-perl.c if (field->next) next 168 tools/perf/util/scripting-engines/trace-event-perl.c define_flag_values(field->next, ev_name, field_name); next 256 tools/perf/util/scripting-engines/trace-event-perl.c if (args->next) next 257 tools/perf/util/scripting-engines/trace-event-perl.c define_event_symbols(event, ev_name, args->next); next 392 tools/perf/util/scripting-engines/trace-event-perl.c for (field = event->format.fields; field; field = field->next) { next 627 tools/perf/util/scripting-engines/trace-event-perl.c for (f = event->format.fields; f; f = f->next) { next 646 tools/perf/util/scripting-engines/trace-event-perl.c for (f = event->format.fields; f; f = f->next) { next 670 tools/perf/util/scripting-engines/trace-event-perl.c for (f = event->format.fields; f; f = f->next) { next 238 tools/perf/util/scripting-engines/trace-event-python.c if (field->next) next 239 tools/perf/util/scripting-engines/trace-event-python.c define_values(field_type, field->next, ev_name, field_name); next 335 tools/perf/util/scripting-engines/trace-event-python.c if (args->next) next 336 tools/perf/util/scripting-engines/trace-event-python.c define_event_symbols(event, ev_name, args->next); next 869 tools/perf/util/scripting-engines/trace-event-python.c for (field = event->format.fields; field; field = field->next) { next 1757 tools/perf/util/scripting-engines/trace-event-python.c for (f = event->format.fields; f; f = f->next) { next 1782 tools/perf/util/scripting-engines/trace-event-python.c for (f = event->format.fields; f; f = f->next) { next 1807 tools/perf/util/scripting-engines/trace-event-python.c for (f = event->format.fields; f; f = f->next) { next 87 tools/perf/util/session.c session->decomp_last->next = decomp; next 275 tools/perf/util/session.c struct decomp *next, *decomp; next 277 tools/perf/util/session.c next = session->decomp; next 279 tools/perf/util/session.c decomp = next; next 282 tools/perf/util/session.c next = decomp->next; next 47 tools/perf/util/session.h struct decomp *next; next 2385 tools/perf/util/sort.c field = field->next; next 171 tools/perf/util/sort.h return list_entry(he->pairs.node.next, struct hist_entry, pairs.node); next 115 tools/perf/util/srccode.c srcfile_list.next != &srcfile_list) { next 666 tools/perf/util/srcline.c struct rb_node *next = rb_first_cached(tree); next 668 tools/perf/util/srcline.c while (next) { next 669 tools/perf/util/srcline.c pos = rb_entry(next, struct srcline_node, rb_node); next 670 tools/perf/util/srcline.c next = rb_next(&pos->rb_node); next 750 tools/perf/util/srcline.c struct rb_node *next = rb_first_cached(tree); next 752 tools/perf/util/srcline.c while (next) { next 753 tools/perf/util/srcline.c pos = rb_entry(next, struct inline_node, rb_node); next 754 tools/perf/util/srcline.c next = rb_next(&pos->rb_node); next 170 tools/perf/util/stat-shadow.c struct rb_node *pos, *next; next 173 tools/perf/util/stat-shadow.c next = rb_first_cached(&rblist->entries); next 174 tools/perf/util/stat-shadow.c while (next) { next 175 tools/perf/util/stat-shadow.c pos = next; next 176 tools/perf/util/stat-shadow.c next = rb_next(pos); next 2172 tools/perf/util/symbol-elf.c size_t shstrndx, next; next 2201 tools/perf/util/symbol-elf.c for (offset = 0; (next = gelf_getnote(data, offset, &nhdr, &name_off, next 2202 tools/perf/util/symbol-elf.c &desc_off)) > 0; offset = next) { next 188 tools/perf/util/symbol.c struct symbol *curr, *next; next 199 tools/perf/util/symbol.c next = rb_entry(nd, struct symbol, rb_node); next 204 tools/perf/util/symbol.c if (curr->start != next->start) next 207 tools/perf/util/symbol.c if (choose_best_symbol(curr, next) == SYMBOL_A) { next 208 tools/perf/util/symbol.c rb_erase_cached(&next->rb_node, symbols); next 209 tools/perf/util/symbol.c symbol__delete(next); next 245 tools/perf/util/symbol.c struct map *next, *curr; next 253 tools/perf/util/symbol.c for (next = map__next(curr); next; next = map__next(curr)) { next 255 tools/perf/util/symbol.c curr->end = next->start; next 256 tools/perf/util/symbol.c curr = next; next 307 tools/perf/util/symbol.c struct rb_node *next = rb_first_cached(symbols); next 309 tools/perf/util/symbol.c while (next) { next 310 tools/perf/util/symbol.c pos = rb_entry(next, struct symbol, rb_node); next 311 tools/perf/util/symbol.c next = rb_next(&pos->rb_node); next 712 tools/perf/util/symbol.c struct rb_node *next = rb_first_cached(root); next 719 tools/perf/util/symbol.c while (next) { next 722 tools/perf/util/symbol.c pos = rb_entry(next, struct symbol, rb_node); next 723 tools/perf/util/symbol.c next = rb_next(&pos->rb_node); next 766 tools/perf/util/symbol.c struct rb_node *next = rb_first_cached(root); next 777 tools/perf/util/symbol.c while (next) { next 780 tools/perf/util/symbol.c pos = rb_entry(next, struct symbol, rb_node); next 781 tools/perf/util/symbol.c next = rb_next(&pos->rb_node); next 946 tools/perf/util/symbol.c struct rb_node *next = rb_first(modules); next 948 tools/perf/util/symbol.c while (next) { next 949 tools/perf/util/symbol.c mi = rb_entry(next, struct module_info, rb_node); next 950 tools/perf/util/symbol.c next = rb_next(&mi->rb_node); next 1074 tools/perf/util/symbol.c struct map *next = map_groups__next(old_map); next 1078 tools/perf/util/symbol.c old_map = next; next 1089 tools/perf/util/symbol.c old_map = next; next 1249 tools/perf/util/symbol.c old_map = list_entry(merged.next, struct map, node); next 1316 tools/perf/util/symbol.c struct map *next = map_groups__next(old_map); next 1325 tools/perf/util/symbol.c old_map = next; next 1340 tools/perf/util/symbol.c replacement_map = list_entry(md.maps.next, struct map, node); next 1344 tools/perf/util/symbol.c new_map = list_entry(md.maps.next, struct map, node); next 1403 tools/perf/util/symbol.c map = list_entry(md.maps.next, struct map, node); next 149 tools/perf/util/trace-event-info.c tps = tps->next; next 247 tools/perf/util/trace-event-info.c tps = tps->next; next 396 tools/perf/util/trace-event-info.c tps = tps->next; next 416 tools/perf/util/trace-event-info.c ppath->next = tracepoint_name_to_path(pos->name); next 417 tools/perf/util/trace-event-info.c if (ppath->next) next 418 tools/perf/util/trace-event-info.c goto next; next 427 tools/perf/util/trace-event-info.c ppath->next = tracepoint_id_to_path(pos->core.attr.config); next 428 tools/perf/util/trace-event-info.c if (!ppath->next) { next 434 tools/perf/util/trace-event-info.c next: next 435 tools/perf/util/trace-event-info.c ppath = ppath->next; next 438 tools/perf/util/trace-event-info.c return nr_tracepoints > 0 ? path.next : NULL; next 129 tools/perf/util/trace-event-parse.c char *next = NULL; next 133 tools/perf/util/trace-event-parse.c line = strtok_r(file, "\n", &next); next 143 tools/perf/util/trace-event-parse.c line = strtok_r(NULL, "\n", &next); next 154 tools/perf/util/trace-event-parse.c char *next = NULL; next 157 tools/perf/util/trace-event-parse.c line = strtok_r(file, "\n", &next); next 161 tools/perf/util/trace-event-parse.c line = strtok_r(NULL, "\n", &next); next 20 tools/power/acpi/os_specific/service_layers/oslinuxtbl.c struct osl_table_info *next; next 280 tools/power/acpi/os_specific/service_layers/oslinuxtbl.c struct osl_table_info *next; next 294 tools/power/acpi/os_specific/service_layers/oslinuxtbl.c next = gbl_table_list_head; next 296 tools/power/acpi/os_specific/service_layers/oslinuxtbl.c if (ACPI_COMPARE_NAMESEG(next->signature, signature)) { next 297 tools/power/acpi/os_specific/service_layers/oslinuxtbl.c if (next->instance == instance) { next 300 tools/power/acpi/os_specific/service_layers/oslinuxtbl.c if (next->instance >= next_instance) { next 301 tools/power/acpi/os_specific/service_layers/oslinuxtbl.c next_instance = next->instance + 1; next 305 tools/power/acpi/os_specific/service_layers/oslinuxtbl.c if (!next->next) { next 308 tools/power/acpi/os_specific/service_layers/oslinuxtbl.c next = next->next; next 310 tools/power/acpi/os_specific/service_layers/oslinuxtbl.c next->next = new_info; next 373 tools/power/acpi/os_specific/service_layers/oslinuxtbl.c info = info->next; next 281 tools/power/cpupower/lib/cpufreq.c current->next = malloc(sizeof(*current)); next 282 tools/power/cpupower/lib/cpufreq.c if (!current->next) next 284 tools/power/cpupower/lib/cpufreq.c current = current->next; next 292 tools/power/cpupower/lib/cpufreq.c current->next = NULL; next 308 tools/power/cpupower/lib/cpufreq.c current = first->next; next 319 tools/power/cpupower/lib/cpufreq.c struct cpufreq_available_governors *tmp, *next; next 326 tools/power/cpupower/lib/cpufreq.c next = tmp->next; next 330 tools/power/cpupower/lib/cpufreq.c tmp = next; next 358 tools/power/cpupower/lib/cpufreq.c current->next = malloc(sizeof(*current)); next 359 tools/power/cpupower/lib/cpufreq.c if (!current->next) next 361 tools/power/cpupower/lib/cpufreq.c current = current->next; next 369 tools/power/cpupower/lib/cpufreq.c current->next = NULL; next 384 tools/power/cpupower/lib/cpufreq.c current = first->next; next 414 tools/power/cpupower/lib/cpufreq.c current->next = malloc(sizeof(*current)); next 415 tools/power/cpupower/lib/cpufreq.c if (!current->next) next 417 tools/power/cpupower/lib/cpufreq.c current = current->next; next 425 tools/power/cpupower/lib/cpufreq.c current->next = NULL; next 440 tools/power/cpupower/lib/cpufreq.c current = first->next; next 449 tools/power/cpupower/lib/cpufreq.c struct cpufreq_available_frequencies *tmp, *next; next 456 tools/power/cpupower/lib/cpufreq.c next = tmp->next; next 458 tools/power/cpupower/lib/cpufreq.c tmp = next; next 489 tools/power/cpupower/lib/cpufreq.c current->next = malloc(sizeof(*current)); next 490 tools/power/cpupower/lib/cpufreq.c if (!current->next) next 492 tools/power/cpupower/lib/cpufreq.c current = current->next; next 500 tools/power/cpupower/lib/cpufreq.c current->next = NULL; next 516 tools/power/cpupower/lib/cpufreq.c current = first->next; next 530 tools/power/cpupower/lib/cpufreq.c struct cpufreq_affected_cpus *tmp, *next; next 537 tools/power/cpupower/lib/cpufreq.c next = tmp->next; next 539 tools/power/cpupower/lib/cpufreq.c tmp = next; next 722 tools/power/cpupower/lib/cpufreq.c current->next = malloc(sizeof(*current)); next 723 tools/power/cpupower/lib/cpufreq.c if (!current->next) next 725 tools/power/cpupower/lib/cpufreq.c current = current->next; next 733 tools/power/cpupower/lib/cpufreq.c current->next = NULL; next 751 tools/power/cpupower/lib/cpufreq.c current = first->next; next 760 tools/power/cpupower/lib/cpufreq.c struct cpufreq_stats *tmp, *next; next 767 tools/power/cpupower/lib/cpufreq.c next = tmp->next; next 769 tools/power/cpupower/lib/cpufreq.c tmp = next; next 19 tools/power/cpupower/lib/cpufreq.h struct cpufreq_available_governors *next; next 25 tools/power/cpupower/lib/cpufreq.h struct cpufreq_available_frequencies *next; next 32 tools/power/cpupower/lib/cpufreq.h struct cpufreq_affected_cpus *next; next 39 tools/power/cpupower/lib/cpufreq.h struct cpufreq_stats *next; next 257 tools/power/cpupower/utils/cpufreq-info.c while (freqs->next) { next 260 tools/power/cpupower/utils/cpufreq-info.c freqs = freqs->next; next 379 tools/power/cpupower/utils/cpufreq-info.c while (governors->next) { next 381 tools/power/cpupower/utils/cpufreq-info.c governors = governors->next; next 401 tools/power/cpupower/utils/cpufreq-info.c while (cpus->next) { next 403 tools/power/cpupower/utils/cpufreq-info.c cpus = cpus->next; next 422 tools/power/cpupower/utils/cpufreq-info.c while (cpus->next) { next 424 tools/power/cpupower/utils/cpufreq-info.c cpus = cpus->next; next 446 tools/power/cpupower/utils/cpufreq-info.c stats = stats->next; next 489 tools/power/cpupower/utils/cpufreq-info.c while (freqs->next) { next 492 tools/power/cpupower/utils/cpufreq-info.c freqs = freqs->next; next 304 tools/power/cpupower/utils/cpufreq-set.c while (cpus->next) { next 306 tools/power/cpupower/utils/cpufreq-set.c cpus = cpus->next; next 47 tools/power/cpupower/utils/helpers/pci.c for (device = (*pacc)->devices; device; device = device->next) { next 1343 tools/power/x86/intel-speed-select/isst-config.c char *next; next 1345 tools/power/x86/intel-speed-select/isst-config.c next = optarg; next 1347 tools/power/x86/intel-speed-select/isst-config.c while (next && *next) { next 1348 tools/power/x86/intel-speed-select/isst-config.c if (*next == '-') /* no negative cpu numbers */ next 1351 tools/power/x86/intel-speed-select/isst-config.c start = strtoul(next, &next, 10); next 1356 tools/power/x86/intel-speed-select/isst-config.c if (*next == '\0') next 1359 tools/power/x86/intel-speed-select/isst-config.c if (*next == ',') { next 1360 tools/power/x86/intel-speed-select/isst-config.c next += 1; next 1364 tools/power/x86/intel-speed-select/isst-config.c if (*next == '-') { next 1365 tools/power/x86/intel-speed-select/isst-config.c next += 1; /* start range */ next 1366 tools/power/x86/intel-speed-select/isst-config.c } else if (*next == '.') { next 1367 tools/power/x86/intel-speed-select/isst-config.c next += 1; next 1368 tools/power/x86/intel-speed-select/isst-config.c if (*next == '.') next 1369 tools/power/x86/intel-speed-select/isst-config.c next += 1; /* start range */ next 1374 tools/power/x86/intel-speed-select/isst-config.c end = strtoul(next, &next, 10); next 1383 tools/power/x86/intel-speed-select/isst-config.c if (*next == ',') next 1384 tools/power/x86/intel-speed-select/isst-config.c next += 1; next 1385 tools/power/x86/intel-speed-select/isst-config.c else if (*next != '\0') next 255 tools/power/x86/turbostat/turbostat.c struct msr_counter *next; next 666 tools/power/x86/turbostat/turbostat.c for (mp = sys.tp; mp; mp = mp->next) { next 704 tools/power/x86/turbostat/turbostat.c for (mp = sys.cp; mp; mp = mp->next) { next 782 tools/power/x86/turbostat/turbostat.c for (mp = sys.pp; mp; mp = mp->next) { next 820 tools/power/x86/turbostat/turbostat.c for (i = 0, mp = sys.tp; mp; i++, mp = mp->next) { next 834 tools/power/x86/turbostat/turbostat.c for (i = 0, mp = sys.cp; mp; i++, mp = mp->next) { next 871 tools/power/x86/turbostat/turbostat.c for (i = 0, mp = sys.pp; mp; i++, mp = mp->next) { next 1007 tools/power/x86/turbostat/turbostat.c for (i = 0, mp = sys.tp; mp; i++, mp = mp->next) { next 1049 tools/power/x86/turbostat/turbostat.c for (i = 0, mp = sys.cp; mp; i++, mp = mp->next) { next 1152 tools/power/x86/turbostat/turbostat.c for (i = 0, mp = sys.pp; mp; i++, mp = mp->next) { next 1264 tools/power/x86/turbostat/turbostat.c for (i = 0, mp = sys.pp; mp; i++, mp = mp->next) { next 1288 tools/power/x86/turbostat/turbostat.c for (i = 0, mp = sys.cp; mp; i++, mp = mp->next) { next 1383 tools/power/x86/turbostat/turbostat.c for (i = 0, mp = sys.tp; mp; i++, mp = mp->next) { next 1472 tools/power/x86/turbostat/turbostat.c for (i = 0, mp = sys.tp; mp; i++, mp = mp->next) next 1475 tools/power/x86/turbostat/turbostat.c for (i = 0, mp = sys.cp; mp; i++, mp = mp->next) next 1478 tools/power/x86/turbostat/turbostat.c for (i = 0, mp = sys.pp; mp; i++, mp = mp->next) next 1508 tools/power/x86/turbostat/turbostat.c for (i = 0, mp = sys.tp; mp; i++, mp = mp->next) { next 1527 tools/power/x86/turbostat/turbostat.c for (i = 0, mp = sys.cp; mp; i++, mp = mp->next) { next 1573 tools/power/x86/turbostat/turbostat.c for (i = 0, mp = sys.pp; mp; i++, mp = mp->next) { next 1631 tools/power/x86/turbostat/turbostat.c for (i = 0, mp = sys.tp; mp; i++, mp = mp->next) { next 1641 tools/power/x86/turbostat/turbostat.c for (i = 0, mp = sys.cp; mp; i++, mp = mp->next) { next 1650 tools/power/x86/turbostat/turbostat.c for (i = 0, mp = sys.pp; mp; i++, mp = mp->next) { next 1867 tools/power/x86/turbostat/turbostat.c for (i = 0, mp = sys.tp; mp; i++, mp = mp->next) { next 1909 tools/power/x86/turbostat/turbostat.c for (i = 0, mp = sys.cp; mp; i++, mp = mp->next) { next 2015 tools/power/x86/turbostat/turbostat.c for (i = 0, mp = sys.pp; mp; i++, mp = mp->next) { next 5356 tools/power/x86/turbostat/turbostat.c msrp->next = sys.tp; next 5367 tools/power/x86/turbostat/turbostat.c msrp->next = sys.cp; next 5378 tools/power/x86/turbostat/turbostat.c msrp->next = sys.pp; next 5406 tools/power/x86/turbostat/turbostat.c goto next; next 5409 tools/power/x86/turbostat/turbostat.c goto next; next 5413 tools/power/x86/turbostat/turbostat.c goto next; next 5418 tools/power/x86/turbostat/turbostat.c goto next; next 5423 tools/power/x86/turbostat/turbostat.c goto next; next 5427 tools/power/x86/turbostat/turbostat.c goto next; next 5431 tools/power/x86/turbostat/turbostat.c goto next; next 5435 tools/power/x86/turbostat/turbostat.c goto next; next 5439 tools/power/x86/turbostat/turbostat.c goto next; next 5443 tools/power/x86/turbostat/turbostat.c goto next; next 5447 tools/power/x86/turbostat/turbostat.c goto next; next 5451 tools/power/x86/turbostat/turbostat.c goto next; next 5455 tools/power/x86/turbostat/turbostat.c goto next; next 5464 tools/power/x86/turbostat/turbostat.c goto next; next 5467 tools/power/x86/turbostat/turbostat.c next: next 5581 tools/power/x86/turbostat/turbostat.c char *next; next 5605 tools/power/x86/turbostat/turbostat.c next = optarg; next 5607 tools/power/x86/turbostat/turbostat.c while (next && *next) { next 5609 tools/power/x86/turbostat/turbostat.c if (*next == '-') /* no negative cpu numbers */ next 5612 tools/power/x86/turbostat/turbostat.c start = strtoul(next, &next, 10); next 5618 tools/power/x86/turbostat/turbostat.c if (*next == '\0') next 5621 tools/power/x86/turbostat/turbostat.c if (*next == ',') { next 5622 tools/power/x86/turbostat/turbostat.c next += 1; next 5626 tools/power/x86/turbostat/turbostat.c if (*next == '-') { next 5627 tools/power/x86/turbostat/turbostat.c next += 1; /* start range */ next 5628 tools/power/x86/turbostat/turbostat.c } else if (*next == '.') { next 5629 tools/power/x86/turbostat/turbostat.c next += 1; next 5630 tools/power/x86/turbostat/turbostat.c if (*next == '.') next 5631 tools/power/x86/turbostat/turbostat.c next += 1; /* start range */ next 5636 tools/power/x86/turbostat/turbostat.c end = strtoul(next, &next, 10); next 5646 tools/power/x86/turbostat/turbostat.c if (*next == ',') next 5647 tools/power/x86/turbostat/turbostat.c next += 1; next 5648 tools/power/x86/turbostat/turbostat.c else if (*next != '\0') next 32 tools/testing/selftests/bpf/progs/btf_dump_test_case_ordering.c struct list_head *next; next 37 tools/testing/selftests/bpf/progs/btf_dump_test_case_ordering.c struct hlist_node *next; next 46 tools/testing/selftests/bpf/progs/btf_dump_test_case_ordering.c struct callback_head *next; next 32 tools/testing/selftests/bpf/test_lpm_map.c struct tlpm_node *next; next 62 tools/testing/selftests/bpf/test_lpm_map.c node->next = list; next 76 tools/testing/selftests/bpf/test_lpm_map.c list = list->next; next 94 tools/testing/selftests/bpf/test_lpm_map.c for ( ; list; list = list->next) { next 121 tools/testing/selftests/bpf/test_lpm_map.c node = best->next; next 126 tools/testing/selftests/bpf/test_lpm_map.c for (node = list; node; node = node->next) { next 127 tools/testing/selftests/bpf/test_lpm_map.c if (node->next == best) { next 128 tools/testing/selftests/bpf/test_lpm_map.c node->next = best->next; next 187 tools/testing/selftests/bpf/test_lpm_map.c for (t1 = l1; t1; t1 = t1->next) next 283 tools/testing/selftests/bpf/test_lpm_map.c for (i = 0, t = list; t; i++, t = t->next) next 135 tools/testing/selftests/bpf/test_lru_map.c int next = *next_to_try; next 138 tools/testing/selftests/bpf/test_lru_map.c while (next < nr_cpus) { next 140 tools/testing/selftests/bpf/test_lru_map.c CPU_SET(next++, &cpuset); next 147 tools/testing/selftests/bpf/test_lru_map.c *next_to_try = next; next 366 tools/testing/selftests/bpf/test_progs.c char *next; next 370 tools/testing/selftests/bpf/test_progs.c num = strtol(s, &next, 10); next 379 tools/testing/selftests/bpf/test_progs.c if (!parsing_end && *next == '-') { next 380 tools/testing/selftests/bpf/test_progs.c s = next + 1; next 383 tools/testing/selftests/bpf/test_progs.c } else if (*next == ',') { next 385 tools/testing/selftests/bpf/test_progs.c s = next + 1; next 387 tools/testing/selftests/bpf/test_progs.c } else if (*next == '\0') { next 389 tools/testing/selftests/bpf/test_progs.c s = next; next 644 tools/testing/selftests/kselftest_harness.h struct __test_metadata *prev, *next; next 671 tools/testing/selftests/kselftest_harness.h t->next = NULL; next 676 tools/testing/selftests/kselftest_harness.h t->next = NULL; next 678 tools/testing/selftests/kselftest_harness.h t->prev->next = t; next 681 tools/testing/selftests/kselftest_harness.h t->next = __test_list; next 682 tools/testing/selftests/kselftest_harness.h t->next->prev = t; next 769 tools/testing/selftests/kselftest_harness.h for (t = __test_list; t; t = t->next) { next 264 tools/testing/selftests/kvm/lib/kvm_util.c region = region->next) { next 325 tools/testing/selftests/kvm/lib/kvm_util.c region = region->next) { next 385 tools/testing/selftests/kvm/lib/kvm_util.c for (vcpup = vm->vcpu_head; vcpup; vcpup = vcpup->next) { next 418 tools/testing/selftests/kvm/lib/kvm_util.c if (vcpu->next) next 419 tools/testing/selftests/kvm/lib/kvm_util.c vcpu->next->prev = vcpu->prev; next 421 tools/testing/selftests/kvm/lib/kvm_util.c vcpu->prev->next = vcpu->next; next 423 tools/testing/selftests/kvm/lib/kvm_util.c vm->vcpu_head = vcpu->next; next 464 tools/testing/selftests/kvm/lib/kvm_util.c vmp->userspace_mem_region_head = region->next; next 614 tools/testing/selftests/kvm/lib/kvm_util.c region = region->next) { next 689 tools/testing/selftests/kvm/lib/kvm_util.c region->next = vm->userspace_mem_region_head; next 714 tools/testing/selftests/kvm/lib/kvm_util.c region = region->next) { next 836 tools/testing/selftests/kvm/lib/kvm_util.c vcpu->next = vm->vcpu_head; next 1032 tools/testing/selftests/kvm/lib/kvm_util.c region = region->next) { next 1065 tools/testing/selftests/kvm/lib/kvm_util.c region = region->next) { next 1458 tools/testing/selftests/kvm/lib/kvm_util.c region = region->next) { next 1477 tools/testing/selftests/kvm/lib/kvm_util.c for (vcpu = vm->vcpu_head; vcpu; vcpu = vcpu->next) next 27 tools/testing/selftests/kvm/lib/kvm_util_internal.h struct userspace_mem_region *next, *prev; next 38 tools/testing/selftests/kvm/lib/kvm_util_internal.h struct vcpu *next, *prev; next 605 tools/testing/selftests/kvm/lib/sparsebit.c struct node *prev, *next, *tmp; next 746 tools/testing/selftests/kvm/lib/sparsebit.c next = node_next(s, nodep); next 747 tools/testing/selftests/kvm/lib/sparsebit.c if (next) { next 749 tools/testing/selftests/kvm/lib/sparsebit.c if (next->mask == 0 && next->num_after == 0) { next 750 tools/testing/selftests/kvm/lib/sparsebit.c node_rm(s, next); next 759 tools/testing/selftests/kvm/lib/sparsebit.c if (next->idx == nodep->idx + MASK_BITS + nodep->num_after && next 760 tools/testing/selftests/kvm/lib/sparsebit.c next->mask == ~(mask_t) 0) { next 762 tools/testing/selftests/kvm/lib/sparsebit.c next->mask = 0; next 763 tools/testing/selftests/kvm/lib/sparsebit.c nodep->num_after += next->num_after; next 764 tools/testing/selftests/kvm/lib/sparsebit.c next->num_after = 0; next 766 tools/testing/selftests/kvm/lib/sparsebit.c node_rm(s, next); next 767 tools/testing/selftests/kvm/lib/sparsebit.c next = NULL; next 1374 tools/testing/selftests/kvm/lib/sparsebit.c struct node *nodep, *next; next 1421 tools/testing/selftests/kvm/lib/sparsebit.c for (next = node_next(s, nodep); next 1422 tools/testing/selftests/kvm/lib/sparsebit.c next && (next->idx < middle_end); next 1423 tools/testing/selftests/kvm/lib/sparsebit.c next = node_next(s, nodep)) { next 1424 tools/testing/selftests/kvm/lib/sparsebit.c assert(next->idx + MASK_BITS + next->num_after - 1 <= middle_end); next 1425 tools/testing/selftests/kvm/lib/sparsebit.c node_rm(s, next); next 1426 tools/testing/selftests/kvm/lib/sparsebit.c next = NULL; next 1456 tools/testing/selftests/kvm/lib/sparsebit.c struct node *nodep, *next; next 1484 tools/testing/selftests/kvm/lib/sparsebit.c for (next = node_next(s, nodep); next 1485 tools/testing/selftests/kvm/lib/sparsebit.c next && (next->idx < middle_end); next 1486 tools/testing/selftests/kvm/lib/sparsebit.c next = node_next(s, nodep)) { next 1487 tools/testing/selftests/kvm/lib/sparsebit.c assert(next->idx + MASK_BITS + next->num_after - 1 <= middle_end); next 1488 tools/testing/selftests/kvm/lib/sparsebit.c node_rm(s, next); next 1489 tools/testing/selftests/kvm/lib/sparsebit.c next = NULL; next 1918 tools/testing/selftests/kvm/lib/sparsebit.c sparsebit_idx_t next; next 1982 tools/testing/selftests/kvm/lib/sparsebit.c next = sparsebit_next_set(s, first); next 1983 tools/testing/selftests/kvm/lib/sparsebit.c assert(next == 0 || next > first); next 1984 tools/testing/selftests/kvm/lib/sparsebit.c assert(next == 0 || get_value(next)); next 1987 tools/testing/selftests/kvm/lib/sparsebit.c next = sparsebit_next_clear(s, first); next 1988 tools/testing/selftests/kvm/lib/sparsebit.c assert(next == 0 || next > first); next 1989 tools/testing/selftests/kvm/lib/sparsebit.c assert(next == 0 || !get_value(next)); next 1992 tools/testing/selftests/kvm/lib/sparsebit.c next = sparsebit_next_clear(s, first); next 1994 tools/testing/selftests/kvm/lib/sparsebit.c assert(next == 0 || next > last); next 1996 tools/testing/selftests/kvm/lib/sparsebit.c next = sparsebit_next_set(s, first - 1); next 1998 tools/testing/selftests/kvm/lib/sparsebit.c next = sparsebit_first_set(s); next 2001 tools/testing/selftests/kvm/lib/sparsebit.c assert(next == first); next 2003 tools/testing/selftests/kvm/lib/sparsebit.c assert(sparsebit_is_clear(s, first) || next <= last); next 2007 tools/testing/selftests/kvm/lib/sparsebit.c next = sparsebit_next_set(s, first); next 2009 tools/testing/selftests/kvm/lib/sparsebit.c assert(next == 0 || next > last); next 2011 tools/testing/selftests/kvm/lib/sparsebit.c next = sparsebit_next_clear(s, first - 1); next 2013 tools/testing/selftests/kvm/lib/sparsebit.c next = sparsebit_first_clear(s); next 2016 tools/testing/selftests/kvm/lib/sparsebit.c assert(next == first); next 2018 tools/testing/selftests/kvm/lib/sparsebit.c assert(sparsebit_is_set(s, first) || next <= last); next 419 tools/testing/selftests/net/psock_tpacket.c void *next = get_next_frame(ring, frame_num); next 421 tools/testing/selftests/net/psock_tpacket.c while (__tx_kernel_ready(next, ring->version) && next 423 tools/testing/selftests/net/psock_tpacket.c ppd.raw = next; next 446 tools/testing/selftests/net/psock_tpacket.c struct tpacket3_hdr *tx = next; next 463 tools/testing/selftests/net/psock_tpacket.c __tx_user_ready(next, ring->version); next 315 tools/testing/selftests/networking/timestamping/timestamping.c struct timeval next; next 455 tools/testing/selftests/networking/timestamping/timestamping.c gettimeofday(&next, 0); next 456 tools/testing/selftests/networking/timestamping/timestamping.c next.tv_sec = (next.tv_sec + 1) / 5 * 5; next 457 tools/testing/selftests/networking/timestamping/timestamping.c next.tv_usec = 0; next 466 tools/testing/selftests/networking/timestamping/timestamping.c delta_us = (long)(next.tv_sec - now.tv_sec) * 1000000 + next 467 tools/testing/selftests/networking/timestamping/timestamping.c (long)(next.tv_usec - now.tv_usec); next 503 tools/testing/selftests/networking/timestamping/timestamping.c next.tv_sec += 5; next 122 tools/testing/selftests/proc/fd-001-lookup.c next: next 137 tools/testing/selftests/proc/fd-001-lookup.c goto next; next 109 tools/testing/selftests/rcutorture/formal/srcu-cbmc/include/linux/types.h struct list_head *next, *prev; next 117 tools/testing/selftests/rcutorture/formal/srcu-cbmc/include/linux/types.h struct hlist_node *next, **pprev; next 140 tools/testing/selftests/rcutorture/formal/srcu-cbmc/include/linux/types.h struct callback_head *next; next 36 tools/testing/selftests/rseq/basic_percpu_ops_test.c struct percpu_list_node *next; next 150 tools/testing/selftests/rseq/basic_percpu_ops_test.c node->next = (struct percpu_list_node *)expect; next 177 tools/testing/selftests/rseq/basic_percpu_ops_test.c offset = offsetof(struct percpu_list_node, next); next 203 tools/testing/selftests/rseq/basic_percpu_ops_test.c list->c[cpu].head = node->next; next 260 tools/testing/selftests/rseq/basic_percpu_ops_test.c node->next = list.c[i].head; next 281 tools/testing/selftests/rseq/param_test.c struct percpu_list_node *next; next 528 tools/testing/selftests/rseq/param_test.c node->next = (struct percpu_list_node *)expect; next 558 tools/testing/selftests/rseq/param_test.c offset = offsetof(struct percpu_list_node, next); next 586 tools/testing/selftests/rseq/param_test.c list->c[cpu].head = node->next; next 642 tools/testing/selftests/rseq/param_test.c node->next = list.c[i].head; next 152 tools/testing/selftests/timers/nanosleep.c goto next; next 161 tools/testing/selftests/timers/nanosleep.c next: next 25 tools/testing/selftests/vm/compaction_test.c struct map_list *next; next 207 tools/testing/selftests/vm/compaction_test.c entry->next = list; next 219 tools/testing/selftests/vm/compaction_test.c for (entry = list; entry != NULL; entry = entry->next) { next 221 tools/testing/selftests/vm/compaction_test.c if (!entry->next) next 223 tools/testing/selftests/vm/compaction_test.c entry = entry->next; next 46 tools/testing/selftests/vm/mlock2.h goto next; next 51 tools/testing/selftests/vm/mlock2.h next: next 61 tools/testing/vsock/control.c goto next; next 75 tools/testing/vsock/control.c goto next; next 77 tools/testing/vsock/control.c goto next; next 93 tools/testing/vsock/control.c next: next 332 tools/testing/vsock/vsock_diag_test.c struct vsock_stat *next; next 334 tools/testing/vsock/vsock_diag_test.c list_for_each_entry_safe(st, next, sockets, list) next 116 tools/usb/testusb.c struct testdev *next; next 274 tools/usb/testusb.c entry->next = testdevs; next 483 tools/usb/testusb.c if (testdevs && testdevs->next == 0 && !device) next 485 tools/usb/testusb.c for (entry = testdevs; entry; entry = entry->next) { next 517 tools/usb/testusb.c for (entry = testdevs; entry; entry = entry->next) { next 20 tools/usb/usbip/libsrc/list.h struct list_head *next, *prev; next 30 tools/usb/usbip/libsrc/list.h list->next = list; next 42 tools/usb/usbip/libsrc/list.h struct list_head *next) next 44 tools/usb/usbip/libsrc/list.h next->prev = new; next 45 tools/usb/usbip/libsrc/list.h new->next = next; next 47 tools/usb/usbip/libsrc/list.h prev->next = new; next 60 tools/usb/usbip/libsrc/list.h __list_add(new, head, head->next); next 70 tools/usb/usbip/libsrc/list.h static inline void __list_del(struct list_head * prev, struct list_head * next) next 72 tools/usb/usbip/libsrc/list.h next->prev = prev; next 73 tools/usb/usbip/libsrc/list.h prev->next = next; next 88 tools/usb/usbip/libsrc/list.h __list_del(entry->prev, entry->next); next 93 tools/usb/usbip/libsrc/list.h __list_del(entry->prev, entry->next); next 94 tools/usb/usbip/libsrc/list.h entry->next = LIST_POISON1; next 112 tools/usb/usbip/libsrc/list.h for (pos = (head)->next; pos != (head); pos = pos->next) next 121 tools/usb/usbip/libsrc/list.h for (pos = (head)->next, n = pos->next; pos != (head); \ next 122 tools/usb/usbip/libsrc/list.h pos = n, n = pos->next) next 26 tools/usb/usbip/libsrc/names.c struct vendor *next; next 32 tools/usb/usbip/libsrc/names.c struct product *next; next 38 tools/usb/usbip/libsrc/names.c struct class *next; next 44 tools/usb/usbip/libsrc/names.c struct subclass *next; next 50 tools/usb/usbip/libsrc/names.c struct protocol *next; next 56 tools/usb/usbip/libsrc/names.c struct genericstrtable *next; next 88 tools/usb/usbip/libsrc/names.c for (; v; v = v->next) next 99 tools/usb/usbip/libsrc/names.c for (; p; p = p->next) next 110 tools/usb/usbip/libsrc/names.c for (; c; c = c->next) next 121 tools/usb/usbip/libsrc/names.c for (; s; s = s->next) next 134 tools/usb/usbip/libsrc/names.c for (; p; p = p->next) next 143 tools/usb/usbip/libsrc/names.c struct pool *next; next 163 tools/usb/usbip/libsrc/names.c p->next = pool_head; next 183 tools/usb/usbip/libsrc/names.c pool = pool->next; next 194 tools/usb/usbip/libsrc/names.c for (; v; v = v->next) next 202 tools/usb/usbip/libsrc/names.c v->next = vendors[h]; next 214 tools/usb/usbip/libsrc/names.c for (; p; p = p->next) next 223 tools/usb/usbip/libsrc/names.c p->next = products[h]; next 234 tools/usb/usbip/libsrc/names.c for (; c; c = c->next) next 242 tools/usb/usbip/libsrc/names.c c->next = classes[h]; next 253 tools/usb/usbip/libsrc/names.c for (; s; s = s->next) next 262 tools/usb/usbip/libsrc/names.c s->next = subclasses[h]; next 275 tools/usb/usbip/libsrc/names.c for (; p; p = p->next) next 286 tools/usb/usbip/libsrc/names.c p->next = protocols[h]; next 20 tools/virtio/ringtest/ring.c unsigned short next, next 23 tools/virtio/ringtest/ring.c return (unsigned short)(next - event - 1) < (unsigned short)(next - prev); next 86 tools/virtio/ringtest/virtio_ring_0_9.c ring.desc[i].next = i + 1; next 127 tools/virtio/ringtest/virtio_ring_0_9.c guest.free_head = desc[head].next; next 189 tools/virtio/ringtest/virtio_ring_0_9.c ring.desc[index].next = guest.free_head; next 698 tools/virtio/vringh_test.c d[0].next = 1; next 712 tools/virtio/vringh_test.c d[3].next = 1; next 716 tools/virtio/vringh_test.c d[4].next = 2; next 271 virt/kvm/arm/mmu.c phys_addr_t next, start_addr = addr; next 276 virt/kvm/arm/mmu.c next = stage2_pmd_addr_end(kvm, addr, end); next 288 virt/kvm/arm/mmu.c unmap_stage2_ptes(kvm, pmd, addr, next); next 291 virt/kvm/arm/mmu.c } while (pmd++, addr = next, addr != end); next 300 virt/kvm/arm/mmu.c phys_addr_t next, start_addr = addr; next 305 virt/kvm/arm/mmu.c next = stage2_pud_addr_end(kvm, addr, end); next 315 virt/kvm/arm/mmu.c unmap_stage2_pmds(kvm, pud, addr, next); next 318 virt/kvm/arm/mmu.c } while (pud++, addr = next, addr != end); next 339 virt/kvm/arm/mmu.c phys_addr_t next; next 353 virt/kvm/arm/mmu.c next = stage2_pgd_addr_end(kvm, addr, end); next 355 virt/kvm/arm/mmu.c unmap_stage2_puds(kvm, pgd, addr, next); next 360 virt/kvm/arm/mmu.c if (next != end) next 362 virt/kvm/arm/mmu.c } while (pgd++, addr = next, addr != end); next 381 virt/kvm/arm/mmu.c phys_addr_t next; next 385 virt/kvm/arm/mmu.c next = stage2_pmd_addr_end(kvm, addr, end); next 390 virt/kvm/arm/mmu.c stage2_flush_ptes(kvm, pmd, addr, next); next 392 virt/kvm/arm/mmu.c } while (pmd++, addr = next, addr != end); next 399 virt/kvm/arm/mmu.c phys_addr_t next; next 403 virt/kvm/arm/mmu.c next = stage2_pud_addr_end(kvm, addr, end); next 408 virt/kvm/arm/mmu.c stage2_flush_pmds(kvm, pud, addr, next); next 410 virt/kvm/arm/mmu.c } while (pud++, addr = next, addr != end); next 418 virt/kvm/arm/mmu.c phys_addr_t next; next 423 virt/kvm/arm/mmu.c next = stage2_pgd_addr_end(kvm, addr, end); next 425 virt/kvm/arm/mmu.c stage2_flush_puds(kvm, pgd, addr, next); next 426 virt/kvm/arm/mmu.c } while (pgd++, addr = next, addr != end); next 497 virt/kvm/arm/mmu.c phys_addr_t next; next 502 virt/kvm/arm/mmu.c next = pmd_addr_end(addr, end); next 505 virt/kvm/arm/mmu.c unmap_hyp_ptes(pmd, addr, next); next 506 virt/kvm/arm/mmu.c } while (pmd++, addr = next, addr != end); next 514 virt/kvm/arm/mmu.c phys_addr_t next; next 519 virt/kvm/arm/mmu.c next = pud_addr_end(addr, end); next 522 virt/kvm/arm/mmu.c unmap_hyp_pmds(pud, addr, next); next 523 virt/kvm/arm/mmu.c } while (pud++, addr = next, addr != end); next 539 virt/kvm/arm/mmu.c phys_addr_t next; next 547 virt/kvm/arm/mmu.c next = pgd_addr_end(addr, end); next 549 virt/kvm/arm/mmu.c unmap_hyp_puds(pgd, addr, next); next 550 virt/kvm/arm/mmu.c } while (pgd++, addr = next, addr != end); next 632 virt/kvm/arm/mmu.c unsigned long addr, next; next 650 virt/kvm/arm/mmu.c next = pmd_addr_end(addr, end); next 652 virt/kvm/arm/mmu.c create_hyp_pte_mappings(pmd, addr, next, pfn, prot); next 653 virt/kvm/arm/mmu.c pfn += (next - addr) >> PAGE_SHIFT; next 654 virt/kvm/arm/mmu.c } while (addr = next, addr != end); next 665 virt/kvm/arm/mmu.c unsigned long addr, next; next 682 virt/kvm/arm/mmu.c next = pud_addr_end(addr, end); next 683 virt/kvm/arm/mmu.c ret = create_hyp_pmd_mappings(pud, addr, next, pfn, prot); next 686 virt/kvm/arm/mmu.c pfn += (next - addr) >> PAGE_SHIFT; next 687 virt/kvm/arm/mmu.c } while (addr = next, addr != end); next 698 virt/kvm/arm/mmu.c unsigned long addr, next; next 718 virt/kvm/arm/mmu.c next = pgd_addr_end(addr, end); next 719 virt/kvm/arm/mmu.c err = create_hyp_pud_mappings(pgd, addr, next, pfn, prot); next 722 virt/kvm/arm/mmu.c pfn += (next - addr) >> PAGE_SHIFT; next 723 virt/kvm/arm/mmu.c } while (addr = next, addr != end); next 1453 virt/kvm/arm/mmu.c phys_addr_t next; next 1458 virt/kvm/arm/mmu.c next = stage2_pmd_addr_end(kvm, addr, end); next 1464 virt/kvm/arm/mmu.c stage2_wp_ptes(pmd, addr, next); next 1467 virt/kvm/arm/mmu.c } while (pmd++, addr = next, addr != end); next 1480 virt/kvm/arm/mmu.c phys_addr_t next; next 1484 virt/kvm/arm/mmu.c next = stage2_pud_addr_end(kvm, addr, end); next 1490 virt/kvm/arm/mmu.c stage2_wp_pmds(kvm, pud, addr, next); next 1493 virt/kvm/arm/mmu.c } while (pud++, addr = next, addr != end); next 1505 virt/kvm/arm/mmu.c phys_addr_t next; next 1521 virt/kvm/arm/mmu.c next = stage2_pgd_addr_end(kvm, addr, end); next 1523 virt/kvm/arm/mmu.c stage2_wp_puds(kvm, pgd, addr, next); next 1524 virt/kvm/arm/mmu.c } while (pgd++, addr = next, addr != end); next 253 virt/kvm/arm/vgic/vgic-debug.c .next = vgic_debug_next, next 332 virt/kvm/arm/vgic/vgic-init.c struct vgic_redist_region *rdreg, *next; next 342 virt/kvm/arm/vgic/vgic-init.c list_for_each_entry_safe(rdreg, next, &dist->rd_regions, list) { next 2027 virt/kvm/arm/vgic/vgic-its.c struct its_device *next; next 2032 virt/kvm/arm/vgic/vgic-its.c next = list_next_entry(dev, dev_list); next 2033 virt/kvm/arm/vgic/vgic-its.c next_offset = next->device_id - dev->device_id; next 2040 virt/kvm/arm/vgic/vgic-its.c struct its_ite *next; next 2045 virt/kvm/arm/vgic/vgic-its.c next = list_next_entry(ite, ite_list); next 2046 virt/kvm/arm/vgic/vgic-its.c next_offset = next->event_id - ite->event_id; next 4385 virt/kvm/kvm_main.c struct task_struct *next)