prev 69 arch/alpha/include/asm/futex.h u32 prev; prev 87 arch/alpha/include/asm/futex.h : "+r"(ret), "=&r"(prev), "=&r"(cmp) prev 91 arch/alpha/include/asm/futex.h *uval = prev; prev 25 arch/alpha/include/asm/gct.h gct6_handle prev; prev 131 arch/alpha/include/asm/xchg.h unsigned long prev, tmp, cmp, addr64; prev 148 arch/alpha/include/asm/xchg.h : "=&r" (prev), "=&r" (new), "=&r" (tmp), "=&r" (cmp), "=&r" (addr64) prev 151 arch/alpha/include/asm/xchg.h return prev; prev 157 arch/alpha/include/asm/xchg.h unsigned long prev, tmp, cmp, addr64; prev 174 arch/alpha/include/asm/xchg.h : "=&r" (prev), "=&r" (new), "=&r" (tmp), "=&r" (cmp), "=&r" (addr64) prev 177 arch/alpha/include/asm/xchg.h return prev; prev 183 arch/alpha/include/asm/xchg.h unsigned long prev, cmp; prev 196 arch/alpha/include/asm/xchg.h : "=&r"(prev), "=&r"(cmp), "=m"(*m) prev 199 arch/alpha/include/asm/xchg.h return prev; prev 205 arch/alpha/include/asm/xchg.h unsigned long prev, cmp; prev 218 arch/alpha/include/asm/xchg.h : "=&r"(prev), "=&r"(cmp), "=m"(*m) prev 221 arch/alpha/include/asm/xchg.h return prev; prev 95 arch/alpha/kernel/core_marvel.c marvel_next_io7(struct io7 *prev) prev 97 arch/alpha/kernel/core_marvel.c return (prev ? prev->next : io7_head); prev 60 arch/alpha/kernel/proto.h struct io7 *marvel_next_io7(struct io7 *prev); prev 447 arch/arc/include/asm/atomic.h s64 prev; prev 458 arch/arc/include/asm/atomic.h : "=&r"(prev) prev 464 arch/arc/include/asm/atomic.h return prev; prev 469 arch/arc/include/asm/atomic.h s64 prev; prev 478 arch/arc/include/asm/atomic.h : "=&r"(prev) prev 484 arch/arc/include/asm/atomic.h return prev; prev 19 arch/arc/include/asm/cmpxchg.h unsigned long prev; prev 33 arch/arc/include/asm/cmpxchg.h : "=&r"(prev) /* Early clobber, to prevent reg reuse */ prev 41 arch/arc/include/asm/cmpxchg.h return prev; prev 50 arch/arc/include/asm/cmpxchg.h int prev; prev 57 arch/arc/include/asm/cmpxchg.h prev = *p; prev 58 arch/arc/include/asm/cmpxchg.h if (prev == expected) prev 61 arch/arc/include/asm/cmpxchg.h return prev; prev 129 arch/arc/include/asm/mmu_context.h static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, prev 162 arch/arc/include/asm/mmu_context.h #define activate_mm(prev, next) switch_mm(prev, next, NULL) prev 36 arch/arc/include/asm/switch_to.h #define switch_to(prev, next, last) \ prev 38 arch/arc/include/asm/switch_to.h ARC_EZNPS_DP_PREV(prev, next); \ prev 39 arch/arc/include/asm/switch_to.h ARC_FPU_PREV(prev, next); \ prev 40 arch/arc/include/asm/switch_to.h last = __switch_to(prev, next);\ prev 27 arch/arc/kernel/ctx_sw.c unsigned int prev = (unsigned int)prev_task; prev 117 arch/arc/kernel/ctx_sw.c : "n"(KSP_WORD_OFF), "r"(next), "r"(prev) prev 30 arch/arc/kernel/fpu.c void fpu_save_restore(struct task_struct *prev, struct task_struct *next) prev 32 arch/arc/kernel/fpu.c unsigned int *saveto = &prev->thread.fpu.aux_dpfp[0].l; prev 413 arch/arc/kernel/unwind.c struct unwind_table *table = info->table, *prev; prev 415 arch/arc/kernel/unwind.c for (prev = &root_table; prev->link && prev->link != table; prev 416 arch/arc/kernel/unwind.c prev = prev->link) prev 419 arch/arc/kernel/unwind.c if (prev->link) { prev 425 arch/arc/kernel/unwind.c prev->link = table->link; prev 426 arch/arc/kernel/unwind.c if (!prev->link) prev 427 arch/arc/kernel/unwind.c last_table = prev; prev 10 arch/arc/plat-eznps/ctop.c void dp_save_restore(struct task_struct *prev, struct task_struct *next) prev 12 arch/arc/plat-eznps/ctop.c struct eznps_dp *prev_task_dp = &prev->thread.dp; prev 105 arch/arm/include/asm/mmu_context.h #define activate_mm(prev,next) switch_mm(prev, next, NULL) prev 128 arch/arm/include/asm/mmu_context.h switch_mm(struct mm_struct *prev, struct mm_struct *next, prev 144 arch/arm/include/asm/mmu_context.h if (!cpumask_test_and_set_cpu(cpu, mm_cpumask(next)) || prev != next) { prev 147 arch/arm/include/asm/mmu_context.h cpumask_clear_cpu(cpu, mm_cpumask(prev)); prev 26 arch/arm/include/asm/switch_to.h #define switch_to(prev,next,last) \ prev 29 arch/arm/include/asm/switch_to.h last = __switch_to(prev,task_thread_info(prev), task_thread_info(next)); \ prev 120 arch/arm/kernel/module-plts.c const Elf32_Rel *prev; prev 130 arch/arm/kernel/module-plts.c prev = rel + num - 1; prev 131 arch/arm/kernel/module-plts.c return cmp_rel(rel + num, prev) == 0 && prev 132 arch/arm/kernel/module-plts.c is_zero_addend_relocation(base, prev); prev 800 arch/arm/mach-omap1/clock.c if (clk->node.next || clk->node.prev) prev 54 arch/arm/mach-omap2/pm-debug.c void pm_dbg_update_time(struct powerdomain *pwrdm, int prev) prev 64 arch/arm/mach-omap2/pm-debug.c pwrdm->state_timer[prev] += t - pwrdm->timer; prev 59 arch/arm/mach-omap2/pm.h extern void pm_dbg_update_time(struct powerdomain *pwrdm, int prev); prev 61 arch/arm/mach-omap2/pm.h #define pm_dbg_update_time(pwrdm, prev) do {} while (0); prev 164 arch/arm/mach-omap2/powerdomain.c int prev, next, state, trace_state = 0; prev 173 arch/arm/mach-omap2/powerdomain.c prev = pwrdm->state; prev 176 arch/arm/mach-omap2/powerdomain.c prev = pwrdm_read_prev_pwrst(pwrdm); prev 177 arch/arm/mach-omap2/powerdomain.c if (pwrdm->state != prev) prev 178 arch/arm/mach-omap2/powerdomain.c pwrdm->state_counter[prev]++; prev 179 arch/arm/mach-omap2/powerdomain.c if (prev == PWRDM_POWER_RET) prev 186 arch/arm/mach-omap2/powerdomain.c if (next != prev) { prev 189 arch/arm/mach-omap2/powerdomain.c ((prev & OMAP_POWERSTATE_MASK) << 0)); prev 199 arch/arm/mach-omap2/powerdomain.c if (state != prev) prev 202 arch/arm/mach-omap2/powerdomain.c pm_dbg_update_time(pwrdm, prev); prev 82 arch/arm/mach-omap2/sdrc2xxx.c u32 prev = curr_perf_level; prev 86 arch/arm/mach-omap2/sdrc2xxx.c return prev; prev 93 arch/arm/mach-omap2/sdrc2xxx.c return prev; prev 110 arch/arm/mach-omap2/sdrc2xxx.c return prev; prev 230 arch/arm64/include/asm/mmu_context.h switch_mm(struct mm_struct *prev, struct mm_struct *next, prev 233 arch/arm64/include/asm/mmu_context.h if (prev != next) prev 246 arch/arm64/include/asm/mmu_context.h #define activate_mm(prev,next) switch_mm(prev, next, current) prev 251 arch/arm64/include/asm/processor.h extern struct task_struct *cpu_switch_to(struct task_struct *prev, prev 139 arch/arm64/kernel/armv8_deprecated.c enum insn_emulation_mode prev) prev 143 arch/arm64/kernel/armv8_deprecated.c switch (prev) { prev 504 arch/arm64/kernel/process.c __notrace_funcgraph struct task_struct *__switch_to(struct task_struct *prev, prev 527 arch/arm64/kernel/process.c last = cpu_switch_to(prev, next); prev 17 arch/c6x/include/asm/switch_to.h asmlinkage void *__switch_to(struct thread_struct *prev, prev 21 arch/c6x/include/asm/switch_to.h #define switch_to(prev, next, last) \ prev 24 arch/c6x/include/asm/switch_to.h (last) = __switch_to(&(prev)->thread, \ prev 25 arch/c6x/include/asm/switch_to.h &(next)->thread, (prev)); \ prev 27 arch/csky/include/asm/mmu_context.h #define activate_mm(prev,next) switch_mm(prev, next, current) prev 36 arch/csky/include/asm/mmu_context.h switch_mm(struct mm_struct *prev, struct mm_struct *next, prev 41 arch/csky/include/asm/mmu_context.h if (prev != next) prev 10 arch/csky/include/asm/switch_to.h static inline void __switch_to_fpu(struct task_struct *prev, prev 13 arch/csky/include/asm/switch_to.h save_to_user_fp(&prev->thread.user_fp); prev 17 arch/csky/include/asm/switch_to.h static inline void __switch_to_fpu(struct task_struct *prev, prev 28 arch/csky/include/asm/switch_to.h #define switch_to(prev, next, last) \ prev 30 arch/csky/include/asm/switch_to.h struct task_struct *__prev = (prev); \ prev 33 arch/csky/include/asm/switch_to.h ((last) = __switch_to((prev), (next))); \ prev 35 arch/h8300/include/asm/switch_to.h #define switch_to(prev, next, last) \ prev 45 arch/h8300/include/asm/switch_to.h : "r" (&(prev->thread)), \ prev 47 arch/h8300/include/asm/switch_to.h "g" (prev) \ prev 77 arch/hexagon/include/asm/futex.h int prev; prev 99 arch/hexagon/include/asm/futex.h : "+r" (ret), "=&r" (prev), "+m" (*uaddr) prev 103 arch/hexagon/include/asm/futex.h *uval = prev; prev 55 arch/hexagon/include/asm/mmu_context.h static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, prev 64 arch/hexagon/include/asm/mmu_context.h if (next->context.generation < prev->context.generation) { prev 68 arch/hexagon/include/asm/mmu_context.h next->context.generation = prev->context.generation; prev 77 arch/hexagon/include/asm/mmu_context.h static inline void activate_mm(struct mm_struct *prev, struct mm_struct *next) prev 82 arch/hexagon/include/asm/mmu_context.h switch_mm(prev, next, current_thread_info()->task); prev 94 arch/ia64/include/asm/futex.h unsigned long prev; prev 101 arch/ia64/include/asm/futex.h : "+r" (r8), "=&r" (prev) prev 105 arch/ia64/include/asm/futex.h *uval = prev; prev 187 arch/ia64/include/asm/mmu_context.h activate_mm (struct mm_struct *prev, struct mm_struct *next) prev 45 arch/ia64/include/asm/switch_to.h #define __switch_to(prev,next,last) do { \ prev 46 arch/ia64/include/asm/switch_to.h if (IA64_HAS_EXTRA_STATE(prev)) \ prev 47 arch/ia64/include/asm/switch_to.h ia64_save_extra(prev); \ prev 61 arch/ia64/include/asm/switch_to.h # define switch_to(prev,next,last) do { \ prev 62 arch/ia64/include/asm/switch_to.h if (ia64_psr(task_pt_regs(prev))->mfh && ia64_is_local_fpu_owner(prev)) { \ prev 63 arch/ia64/include/asm/switch_to.h ia64_psr(task_pt_regs(prev))->mfh = 0; \ prev 64 arch/ia64/include/asm/switch_to.h (prev)->thread.flags |= IA64_THREAD_FPH_VALID; \ prev 65 arch/ia64/include/asm/switch_to.h __ia64_save_fpu((prev)->thread.fph); \ prev 67 arch/ia64/include/asm/switch_to.h __switch_to(prev, next, last); \ prev 76 arch/ia64/include/asm/switch_to.h # define switch_to(prev,next,last) __switch_to(prev, next, last) prev 1051 arch/ia64/kernel/efi.c struct kern_memdesc *k, *prev = NULL; prev 1152 arch/ia64/kernel/efi.c if (prev && kmd_end(prev) == md->phys_addr) { prev 1153 arch/ia64/kernel/efi.c prev->num_pages += (ae - as) >> EFI_PAGE_SHIFT; prev 1161 arch/ia64/kernel/efi.c prev = k++; prev 108 arch/ia64/kernel/time.c void arch_vtime_task_switch(struct task_struct *prev) prev 110 arch/ia64/kernel/time.c struct thread_info *pi = task_thread_info(prev); prev 646 arch/ia64/kernel/unwind.c struct unw_reg_state *copy, *prev = NULL, *first = NULL; prev 656 arch/ia64/kernel/unwind.c prev->next = copy; prev 660 arch/ia64/kernel/unwind.c prev = copy; prev 1269 arch/ia64/kernel/unwind.c struct unw_script *script, *prev, *tmp; prev 1300 arch/ia64/kernel/unwind.c prev = NULL; prev 1303 arch/ia64/kernel/unwind.c if (prev) prev 1304 arch/ia64/kernel/unwind.c prev->coll_chain = tmp->coll_chain; prev 1309 arch/ia64/kernel/unwind.c prev = tmp; prev 1534 arch/ia64/kernel/unwind.c struct unw_table *table, *prev; prev 1563 arch/ia64/kernel/unwind.c prev = NULL; prev 1573 arch/ia64/kernel/unwind.c if (prev && prev != unw.tables) { prev 1575 arch/ia64/kernel/unwind.c prev->next = table->next; prev 1582 arch/ia64/kernel/unwind.c prev = table; prev 2130 arch/ia64/kernel/unwind.c struct unw_table *table, *prev; prev 2152 arch/ia64/kernel/unwind.c for (prev = (struct unw_table *) &unw.tables; prev; prev = prev->next) prev 2153 arch/ia64/kernel/unwind.c if (prev->next == table) prev 2155 arch/ia64/kernel/unwind.c if (!prev) { prev 2161 arch/ia64/kernel/unwind.c prev->next = table->next; prev 173 arch/m68k/include/asm/atomic.h int prev; prev 176 arch/m68k/include/asm/atomic.h prev = atomic_read(v); prev 177 arch/m68k/include/asm/atomic.h if (prev == old) prev 180 arch/m68k/include/asm/atomic.h return prev; prev 186 arch/m68k/include/asm/atomic.h int prev; prev 189 arch/m68k/include/asm/atomic.h prev = atomic_read(v); prev 192 arch/m68k/include/asm/atomic.h return prev; prev 75 arch/m68k/include/asm/mmu_context.h static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, prev 197 arch/m68k/include/asm/mmu_context.h static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, prev 283 arch/m68k/include/asm/mmu_context.h static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, struct task_struct *tsk) prev 285 arch/m68k/include/asm/mmu_context.h if (prev != next) { prev 316 arch/m68k/include/asm/mmu_context.h static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, struct task_struct *tsk) prev 31 arch/m68k/include/asm/switch_to.h #define switch_to(prev,next,last) do { \ prev 32 arch/m68k/include/asm/switch_to.h register void *_prev __asm__ ("a0") = (prev); \ prev 110 arch/m68k/sun3/sun3dvma.c struct hole *prev = NULL; prev 117 arch/m68k/sun3/sun3dvma.c if(!prev) { prev 118 arch/m68k/sun3/sun3dvma.c prev = hole; prev 122 arch/m68k/sun3/sun3dvma.c if(hole->end == prev->start) { prev 123 arch/m68k/sun3/sun3dvma.c hole->size += prev->size; prev 124 arch/m68k/sun3/sun3dvma.c hole->end = prev->end; prev 125 arch/m68k/sun3/sun3dvma.c list_move(&(prev->list), &hole_cache); prev 72 arch/microblaze/include/asm/futex.h u32 prev; prev 91 arch/microblaze/include/asm/futex.h : "+r" (ret), "=&r" (prev), "=&r"(cmp) \ prev 94 arch/microblaze/include/asm/futex.h *uval = prev; prev 120 arch/microblaze/include/asm/mmu_context_mm.h static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, prev 15 arch/microblaze/include/asm/switch_to.h extern struct task_struct *_switch_to(struct thread_info *prev, prev 18 arch/microblaze/include/asm/switch_to.h #define switch_to(prev, next, last) \ prev 20 arch/microblaze/include/asm/switch_to.h (last) = _switch_to(task_thread_info(prev), \ prev 158 arch/mips/include/asm/mmu_context.h static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, prev 172 arch/mips/include/asm/mmu_context.h cpumask_clear_cpu(cpu, mm_cpumask(prev)); prev 188 arch/mips/include/asm/mmu_context.h #define activate_mm(prev, next) switch_mm(prev, next, current) prev 32 arch/mips/include/asm/switch_to.h extern asmlinkage struct task_struct *resume(struct task_struct *prev, prev 52 arch/mips/include/asm/switch_to.h #define __mips_mt_fpaff_switch_to(prev) \ prev 54 arch/mips/include/asm/switch_to.h struct thread_info *__prev_ti = task_thread_info(prev); \ prev 58 arch/mips/include/asm/switch_to.h (!(KSTK_STATUS(prev) & ST0_CU1))) { \ prev 60 arch/mips/include/asm/switch_to.h prev->cpus_mask = prev->thread.user_cpus_allowed; \ prev 66 arch/mips/include/asm/switch_to.h #define __mips_mt_fpaff_switch_to(prev) do { (void) (prev); } while (0) prev 109 arch/mips/include/asm/switch_to.h #define switch_to(prev, next, last) \ prev 111 arch/mips/include/asm/switch_to.h __mips_mt_fpaff_switch_to(prev); \ prev 112 arch/mips/include/asm/switch_to.h lose_fpu_inatomic(1, prev); \ prev 116 arch/mips/include/asm/switch_to.h __save_dsp(prev); \ prev 121 arch/mips/include/asm/switch_to.h if ((KSTK_STATUS(prev) & ST0_CU2)) { \ prev 123 arch/mips/include/asm/switch_to.h KSTK_STATUS(prev) &= ~ST0_CU2; \ prev 124 arch/mips/include/asm/switch_to.h cop2_save(prev); \ prev 137 arch/mips/include/asm/switch_to.h (last) = resume(prev, next, task_thread_info(next)); \ prev 47 arch/mips/kernel/csrc-r4k.c unsigned int prev, curr, i; prev 55 arch/mips/kernel/csrc-r4k.c for (i = 0, prev = rdhwr_count(); i < 100; i++) { prev 58 arch/mips/kernel/csrc-r4k.c if (curr != prev) prev 61 arch/mips/kernel/csrc-r4k.c prev = curr; prev 54 arch/nds32/include/asm/mmu_context.h static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, prev 59 arch/nds32/include/asm/mmu_context.h if (!cpumask_test_and_set_cpu(cpu, mm_cpumask(next)) || prev != next) { prev 66 arch/nds32/include/asm/mmu_context.h #define activate_mm(prev,next) switch_mm(prev, next, NULL) prev 205 arch/nds32/kernel/process.c struct task_struct *_switch_fpu(struct task_struct *prev, struct task_struct *next) prev 208 arch/nds32/kernel/process.c unlazy_fpu(prev); prev 212 arch/nds32/kernel/process.c return prev; prev 54 arch/nios2/include/asm/mmu_context.h void switch_mm(struct mm_struct *prev, struct mm_struct *next, prev 66 arch/nios2/include/asm/mmu_context.h void activate_mm(struct mm_struct *prev, struct mm_struct *next); prev 17 arch/nios2/include/asm/switch_to.h #define switch_to(prev, next, last) \ prev 26 arch/nios2/include/asm/switch_to.h : "r" (prev), "r" (next) \ prev 80 arch/nios2/mm/mmu_context.c void switch_mm(struct mm_struct *prev, struct mm_struct *next, prev 106 arch/nios2/mm/mmu_context.c void activate_mm(struct mm_struct *prev, struct mm_struct *next) prev 73 arch/openrisc/include/asm/futex.h u32 prev; prev 94 arch/openrisc/include/asm/futex.h : "+r" (ret), "=&r" (prev), "+m" (*uaddr) \ prev 99 arch/openrisc/include/asm/futex.h *uval = prev; prev 22 arch/openrisc/include/asm/mmu_context.h extern void switch_mm(struct mm_struct *prev, struct mm_struct *next, prev 27 arch/openrisc/include/asm/mmu_context.h #define activate_mm(prev, next) switch_mm((prev), (next), NULL) prev 138 arch/openrisc/mm/tlb.c void switch_mm(struct mm_struct *prev, struct mm_struct *next, prev 153 arch/openrisc/mm/tlb.c if (prev != next) prev 154 arch/openrisc/mm/tlb.c local_flush_tlb_mm(prev); prev 53 arch/parisc/include/asm/mmu_context.h static inline void switch_mm_irqs_off(struct mm_struct *prev, prev 56 arch/parisc/include/asm/mmu_context.h if (prev != next) { prev 62 arch/parisc/include/asm/mmu_context.h static inline void switch_mm(struct mm_struct *prev, prev 67 arch/parisc/include/asm/mmu_context.h if (prev == next) prev 71 arch/parisc/include/asm/mmu_context.h switch_mm_irqs_off(prev, next, tsk); prev 78 arch/parisc/include/asm/mmu_context.h static inline void activate_mm(struct mm_struct *prev, struct mm_struct *next) prev 93 arch/parisc/include/asm/mmu_context.h switch_mm(prev,next,current); prev 9 arch/parisc/include/asm/switch_to.h #define switch_to(prev, next, last) do { \ prev 10 arch/parisc/include/asm/switch_to.h (last) = _switch_to(prev, next); \ prev 86 arch/parisc/kernel/sys_parisc.c struct vm_area_struct *vma, *prev; prev 113 arch/parisc/kernel/sys_parisc.c vma = find_vma_prev(mm, addr, &prev); prev 116 arch/parisc/kernel/sys_parisc.c (!prev || addr >= vm_end_gap(prev))) prev 140 arch/parisc/kernel/sys_parisc.c struct vm_area_struct *vma, *prev; prev 170 arch/parisc/kernel/sys_parisc.c vma = find_vma_prev(mm, addr, &prev); prev 173 arch/parisc/kernel/sys_parisc.c (!prev || addr >= vm_end_gap(prev))) prev 62 arch/parisc/lib/bitops.c u64 prev; prev 65 arch/parisc/lib/bitops.c if ((prev = *ptr) == old) prev 68 arch/parisc/lib/bitops.c return prev; prev 74 arch/parisc/lib/bitops.c unsigned int prev; prev 77 arch/parisc/lib/bitops.c if ((prev = *ptr) == old) prev 80 arch/parisc/lib/bitops.c return (unsigned long)prev; prev 130 arch/parisc/mm/fault.c prev = tree; prev 131 arch/parisc/mm/fault.c if (prev->vm_next == NULL) prev 133 arch/parisc/mm/fault.c if (prev->vm_next->vm_start > addr) prev 117 arch/powerpc/boot/libfdt-wrapper.c static void *fdt_wrapper_find_node_by_prop_value(const void *prev, prev 122 arch/powerpc/boot/libfdt-wrapper.c int offset = fdt_node_offset_by_prop_value(fdt, devp_offset_find(prev), prev 127 arch/powerpc/boot/libfdt-wrapper.c static void *fdt_wrapper_find_node_by_compatible(const void *prev, prev 130 arch/powerpc/boot/libfdt-wrapper.c int offset = fdt_node_offset_by_compatible(fdt, devp_offset_find(prev), prev 48 arch/powerpc/boot/ops.h void *(*find_node_by_prop_value)(const void *prev, prev 51 arch/powerpc/boot/ops.h void *(*find_node_by_compatible)(const void *prev, prev 147 arch/powerpc/boot/ops.h static inline void *find_node_by_prop_value(const void *prev, prev 152 arch/powerpc/boot/ops.h return dt_ops.find_node_by_prop_value(prev, propname, prev 158 arch/powerpc/boot/ops.h static inline void *find_node_by_prop_value_str(const void *prev, prev 162 arch/powerpc/boot/ops.h return find_node_by_prop_value(prev, propname, propval, prev 166 arch/powerpc/boot/ops.h static inline void *find_node_by_devtype(const void *prev, prev 169 arch/powerpc/boot/ops.h return find_node_by_prop_value_str(prev, "device_type", type); prev 185 arch/powerpc/boot/ops.h static inline void *find_node_by_compatible(const void *prev, prev 189 arch/powerpc/boot/ops.h return dt_ops.find_node_by_compatible(prev, compat); prev 20 arch/powerpc/include/asm/cmpxchg.h unsigned int prev, prev_mask, tmp, bitoff, off; \ prev 35 arch/powerpc/include/asm/cmpxchg.h : "=&r" (prev), "=&r" (tmp), "+m" (*(u32*)p) \ prev 39 arch/powerpc/include/asm/cmpxchg.h return prev >> bitoff; \ prev 46 arch/powerpc/include/asm/cmpxchg.h unsigned int prev, prev_mask, tmp, bitoff, off; \ prev 69 arch/powerpc/include/asm/cmpxchg.h : "=&r" (prev), "=&r" (tmp), "+m" (*(u32*)p) \ prev 73 arch/powerpc/include/asm/cmpxchg.h return prev >> bitoff; \ prev 91 arch/powerpc/include/asm/cmpxchg.h unsigned long prev; prev 98 arch/powerpc/include/asm/cmpxchg.h : "=&r" (prev), "+m" (*(volatile unsigned int *)p) prev 102 arch/powerpc/include/asm/cmpxchg.h return prev; prev 108 arch/powerpc/include/asm/cmpxchg.h unsigned long prev; prev 115 arch/powerpc/include/asm/cmpxchg.h : "=&r" (prev), "+m" (*p) prev 119 arch/powerpc/include/asm/cmpxchg.h return prev; prev 126 arch/powerpc/include/asm/cmpxchg.h unsigned long prev; prev 133 arch/powerpc/include/asm/cmpxchg.h : "=&r" (prev), "+m" (*(volatile unsigned long *)p) prev 137 arch/powerpc/include/asm/cmpxchg.h return prev; prev 143 arch/powerpc/include/asm/cmpxchg.h unsigned long prev; prev 150 arch/powerpc/include/asm/cmpxchg.h : "=&r" (prev), "+m" (*p) prev 154 arch/powerpc/include/asm/cmpxchg.h return prev; prev 225 arch/powerpc/include/asm/cmpxchg.h unsigned int prev; prev 238 arch/powerpc/include/asm/cmpxchg.h : "=&r" (prev), "+m" (*p) prev 242 arch/powerpc/include/asm/cmpxchg.h return prev; prev 249 arch/powerpc/include/asm/cmpxchg.h unsigned int prev; prev 260 arch/powerpc/include/asm/cmpxchg.h : "=&r" (prev), "+m" (*p) prev 264 arch/powerpc/include/asm/cmpxchg.h return prev; prev 270 arch/powerpc/include/asm/cmpxchg.h unsigned long prev; prev 280 arch/powerpc/include/asm/cmpxchg.h : "=&r" (prev), "+m" (*p) prev 284 arch/powerpc/include/asm/cmpxchg.h return prev; prev 298 arch/powerpc/include/asm/cmpxchg.h unsigned long prev; prev 310 arch/powerpc/include/asm/cmpxchg.h : "=&r" (prev), "+m" (*p) prev 314 arch/powerpc/include/asm/cmpxchg.h return prev; prev 321 arch/powerpc/include/asm/cmpxchg.h unsigned long prev; prev 333 arch/powerpc/include/asm/cmpxchg.h : "=&r" (prev), "+m" (*p) prev 337 arch/powerpc/include/asm/cmpxchg.h return prev; prev 344 arch/powerpc/include/asm/cmpxchg.h unsigned long prev; prev 354 arch/powerpc/include/asm/cmpxchg.h : "=&r" (prev), "+m" (*p) prev 358 arch/powerpc/include/asm/cmpxchg.h return prev; prev 364 arch/powerpc/include/asm/cmpxchg.h unsigned long prev; prev 373 arch/powerpc/include/asm/cmpxchg.h : "=&r" (prev), "+m" (*p) prev 377 arch/powerpc/include/asm/cmpxchg.h return prev; prev 383 arch/powerpc/include/asm/cmpxchg.h unsigned long prev; prev 394 arch/powerpc/include/asm/cmpxchg.h : "=&r" (prev), "+m" (*p) prev 398 arch/powerpc/include/asm/cmpxchg.h return prev; prev 54 arch/powerpc/include/asm/cputime.h static inline void arch_vtime_task_switch(struct task_struct *prev) prev 57 arch/powerpc/include/asm/cputime.h struct cpu_accounting_data *acct0 = get_accounting(prev); prev 74 arch/powerpc/include/asm/futex.h u32 prev; prev 96 arch/powerpc/include/asm/futex.h : "+r" (ret), "=&r" (prev), "+m" (*uaddr) prev 100 arch/powerpc/include/asm/futex.h *uval = prev; prev 4 arch/powerpc/include/asm/membarrier.h static inline void membarrier_arch_switch_mm(struct mm_struct *prev, prev 17 arch/powerpc/include/asm/membarrier.h MEMBARRIER_STATE_GLOBAL_EXPEDITED)) || !prev)) prev 61 arch/powerpc/include/asm/mmu_context.h extern void radix__switch_mmu_context(struct mm_struct *prev, prev 63 arch/powerpc/include/asm/mmu_context.h static inline void switch_mmu_context(struct mm_struct *prev, prev 68 arch/powerpc/include/asm/mmu_context.h return radix__switch_mmu_context(prev, next); prev 104 arch/powerpc/include/asm/mmu_context.h extern void switch_mmu_context(struct mm_struct *prev, struct mm_struct *next, prev 196 arch/powerpc/include/asm/mmu_context.h extern void switch_mm_irqs_off(struct mm_struct *prev, struct mm_struct *next, prev 199 arch/powerpc/include/asm/mmu_context.h static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, prev 205 arch/powerpc/include/asm/mmu_context.h switch_mm_irqs_off(prev, next, tsk); prev 217 arch/powerpc/include/asm/mmu_context.h static inline void activate_mm(struct mm_struct *prev, struct mm_struct *next) prev 219 arch/powerpc/include/asm/mmu_context.h switch_mm(prev, next, current); prev 84 arch/powerpc/include/asm/pgtable-be-types.h __be64 prev; prev 87 arch/powerpc/include/asm/pgtable-be-types.h prev = (__force __be64)__cmpxchg_u64(p, (__force unsigned long)pte_raw(old), prev 90 arch/powerpc/include/asm/pgtable-be-types.h return pte_raw(old) == prev; prev 96 arch/powerpc/include/asm/pgtable-be-types.h __be64 prev; prev 98 arch/powerpc/include/asm/pgtable-be-types.h prev = (__force __be64)__cmpxchg_u64(p, (__force unsigned long)pmd_raw(old), prev 101 arch/powerpc/include/asm/pgtable-be-types.h return pmd_raw(old) == prev; prev 16 arch/powerpc/include/asm/switch_to.h #define switch_to(prev, next, last) ((last) = __switch_to((prev), (next))) prev 18 arch/powerpc/include/asm/switch_to.h extern struct task_struct *_switch(struct thread_struct *prev, prev 788 arch/powerpc/kernel/nvram_64.c struct nvram_partition *part, *prev, *tmp; prev 807 arch/powerpc/kernel/nvram_64.c prev = NULL; prev 810 arch/powerpc/kernel/nvram_64.c prev = NULL; prev 813 arch/powerpc/kernel/nvram_64.c if (prev) { prev 814 arch/powerpc/kernel/nvram_64.c prev->header.length += part->header.length; prev 815 arch/powerpc/kernel/nvram_64.c prev->header.checksum = nvram_checksum(&prev->header); prev 816 arch/powerpc/kernel/nvram_64.c rc = nvram_write_header(prev); prev 824 arch/powerpc/kernel/nvram_64.c prev = part; prev 985 arch/powerpc/kernel/process.c static inline void __switch_to_tm(struct task_struct *prev, prev 989 arch/powerpc/kernel/process.c if (tm_enabled(prev) || tm_enabled(new)) prev 992 arch/powerpc/kernel/process.c if (tm_enabled(prev)) { prev 993 arch/powerpc/kernel/process.c prev->thread.load_tm++; prev 994 arch/powerpc/kernel/process.c tm_reclaim_task(prev); prev 995 arch/powerpc/kernel/process.c if (!MSR_TM_ACTIVE(prev->thread.regs->msr) && prev->thread.load_tm == 0) prev 996 arch/powerpc/kernel/process.c prev->thread.regs->msr &= ~MSR_TM; prev 1048 arch/powerpc/kernel/process.c #define __switch_to_tm(prev, new) prev 1122 arch/powerpc/kernel/process.c struct task_struct *__switch_to(struct task_struct *prev, prev 1163 arch/powerpc/kernel/process.c save_sprs(&prev->thread); prev 1166 arch/powerpc/kernel/process.c giveup_all(prev); prev 1168 arch/powerpc/kernel/process.c __switch_to_tm(prev, new); prev 164 arch/powerpc/kvm/book3s_hv_rm_mmu.c struct revmap_entry *next, *prev; prev 180 arch/powerpc/kvm/book3s_hv_rm_mmu.c prev = real_vmalloc_addr(&kvm->arch.hpt.rev[rev->back]); prev 182 arch/powerpc/kvm/book3s_hv_rm_mmu.c prev->forw = rev->forw; prev 39 arch/powerpc/lib/rheap.c pp = (unsigned long *)&l->prev; prev 489 arch/powerpc/lib/rheap.c list_add(&spblk->list, blk->list.prev); prev 304 arch/powerpc/mm/book3s64/mmu_context.c void radix__switch_mmu_context(struct mm_struct *prev, struct mm_struct *next) prev 201 arch/powerpc/mm/mem.c struct memblock_region *reg, *prev = NULL; prev 204 arch/powerpc/mm/mem.c if (prev && prev 205 arch/powerpc/mm/mem.c memblock_region_memory_end_pfn(prev) < memblock_region_memory_base_pfn(reg)) prev 206 arch/powerpc/mm/mem.c register_nosave_region(memblock_region_memory_end_pfn(prev), prev 208 arch/powerpc/mm/mem.c prev = reg; prev 34 arch/powerpc/mm/mmu_context.c void switch_mm_irqs_off(struct mm_struct *prev, struct mm_struct *next, prev 74 arch/powerpc/mm/mmu_context.c if (prev == next) prev 87 arch/powerpc/mm/mmu_context.c membarrier_arch_switch_mm(prev, next, tsk); prev 93 arch/powerpc/mm/mmu_context.c switch_mmu_context(prev, next, tsk); prev 261 arch/powerpc/mm/nohash/mmu_context.c void switch_mmu_context(struct mm_struct *prev, struct mm_struct *next, prev 279 arch/powerpc/mm/nohash/mmu_context.c if (prev) { prev 280 arch/powerpc/mm/nohash/mmu_context.c pr_hardcont(" (old=0x%p a=%d)", prev, prev->context.active); prev 281 arch/powerpc/mm/nohash/mmu_context.c WARN_ON(prev->context.active < 1); prev 282 arch/powerpc/mm/nohash/mmu_context.c prev->context.active--; prev 330 arch/powerpc/mm/slice.c unsigned long addr, found, prev; prev 363 arch/powerpc/mm/slice.c else if (slice_scan_available(addr - 1, available, 0, &prev)) { prev 364 arch/powerpc/mm/slice.c addr = prev; prev 126 arch/powerpc/perf/8xx-pmu.c s64 prev, val = 0, delta = 0; prev 132 arch/powerpc/perf/8xx-pmu.c prev = local64_read(&event->hw.prev_count); prev 136 arch/powerpc/perf/8xx-pmu.c delta = 16 * (val - prev); prev 140 arch/powerpc/perf/8xx-pmu.c delta = prev - val; prev 146 arch/powerpc/perf/8xx-pmu.c delta = (s64)((s32)val - (s32)prev); prev 150 arch/powerpc/perf/8xx-pmu.c delta = (s64)((s32)val - (s32)prev); prev 153 arch/powerpc/perf/8xx-pmu.c } while (local64_cmpxchg(&event->hw.prev_count, prev, val) != prev); prev 1034 arch/powerpc/perf/core-book3s.c static u64 check_and_compute_delta(u64 prev, u64 val) prev 1036 arch/powerpc/perf/core-book3s.c u64 delta = (val - prev) & 0xfffffffful; prev 1047 arch/powerpc/perf/core-book3s.c if (prev > val && (prev - val) < 256) prev 1055 arch/powerpc/perf/core-book3s.c s64 val, delta, prev; prev 1075 arch/powerpc/perf/core-book3s.c prev = local64_read(&event->hw.prev_count); prev 1078 arch/powerpc/perf/core-book3s.c delta = check_and_compute_delta(prev, val); prev 1081 arch/powerpc/perf/core-book3s.c } while (local64_cmpxchg(&event->hw.prev_count, prev, val) != prev); prev 1095 arch/powerpc/perf/core-book3s.c prev = local64_read(&event->hw.period_left); prev 1096 arch/powerpc/perf/core-book3s.c val = prev - delta; prev 1099 arch/powerpc/perf/core-book3s.c } while (local64_cmpxchg(&event->hw.period_left, prev, val) != prev); prev 1117 arch/powerpc/perf/core-book3s.c u64 val, prev, delta; prev 1125 arch/powerpc/perf/core-book3s.c prev = local64_read(&event->hw.prev_count); prev 1127 arch/powerpc/perf/core-book3s.c delta = check_and_compute_delta(prev, val); prev 1137 arch/powerpc/perf/core-book3s.c u64 val, prev; prev 1144 arch/powerpc/perf/core-book3s.c prev = local64_read(&event->hw.prev_count); prev 1145 arch/powerpc/perf/core-book3s.c if (check_and_compute_delta(prev, val)) prev 2042 arch/powerpc/perf/core-book3s.c s64 prev, delta, left; prev 2051 arch/powerpc/perf/core-book3s.c prev = local64_read(&event->hw.prev_count); prev 2052 arch/powerpc/perf/core-book3s.c delta = check_and_compute_delta(prev, val); prev 177 arch/powerpc/perf/core-fsl-emb.c s64 val, delta, prev; prev 188 arch/powerpc/perf/core-fsl-emb.c prev = local64_read(&event->hw.prev_count); prev 191 arch/powerpc/perf/core-fsl-emb.c } while (local64_cmpxchg(&event->hw.prev_count, prev, val) != prev); prev 194 arch/powerpc/perf/core-fsl-emb.c delta = (val - prev) & 0xfffffffful; prev 606 arch/powerpc/perf/core-fsl-emb.c s64 prev, delta, left; prev 615 arch/powerpc/perf/core-fsl-emb.c prev = local64_read(&event->hw.prev_count); prev 616 arch/powerpc/perf/core-fsl-emb.c delta = (val - prev) & 0xfffffffful; prev 1359 arch/powerpc/perf/hv-24x7.c s64 prev; prev 1361 arch/powerpc/perf/hv-24x7.c prev = local64_xchg(&event->hw.prev_count, now); prev 1362 arch/powerpc/perf/hv-24x7.c local64_add(now - prev, &event->count); prev 192 arch/powerpc/perf/hv-gpci.c s64 prev; prev 194 arch/powerpc/perf/hv-gpci.c prev = local64_xchg(&event->hw.prev_count, now); prev 195 arch/powerpc/perf/hv-gpci.c local64_add(now - prev, &event->count); prev 72 arch/powerpc/platforms/85xx/smp.c u64 prev; prev 78 arch/powerpc/platforms/85xx/smp.c prev = timebase; prev 81 arch/powerpc/platforms/85xx/smp.c } while (prev != timebase); prev 283 arch/powerpc/platforms/cell/spufs/sched.c list_for_each_entry(ctx, gang->aff_ref_ctx->aff_list.prev, aff_list) { prev 374 arch/powerpc/platforms/cell/spufs/sched.c list_for_each_entry(spu, ref->aff_list.prev, aff_list) { prev 350 arch/powerpc/platforms/cell/spufs/spufs.h extern int spu_save(struct spu_state *prev, struct spu *spu); prev 352 arch/powerpc/platforms/cell/spufs/spufs.h extern int spu_switch(struct spu_state *prev, struct spu_state *new, prev 1777 arch/powerpc/platforms/cell/spufs/switch.c static int quiece_spu(struct spu_state *prev, struct spu *spu) prev 1789 arch/powerpc/platforms/cell/spufs/switch.c if (check_spu_isolate(prev, spu)) { /* Step 2. */ prev 1792 arch/powerpc/platforms/cell/spufs/switch.c disable_interrupts(prev, spu); /* Step 3. */ prev 1793 arch/powerpc/platforms/cell/spufs/switch.c set_watchdog_timer(prev, spu); /* Step 4. */ prev 1794 arch/powerpc/platforms/cell/spufs/switch.c inhibit_user_access(prev, spu); /* Step 5. */ prev 1795 arch/powerpc/platforms/cell/spufs/switch.c if (check_spu_isolate(prev, spu)) { /* Step 6. */ prev 1798 arch/powerpc/platforms/cell/spufs/switch.c set_switch_pending(prev, spu); /* Step 7. */ prev 1799 arch/powerpc/platforms/cell/spufs/switch.c save_mfc_cntl(prev, spu); /* Step 8. */ prev 1800 arch/powerpc/platforms/cell/spufs/switch.c save_spu_runcntl(prev, spu); /* Step 9. */ prev 1801 arch/powerpc/platforms/cell/spufs/switch.c save_mfc_sr1(prev, spu); /* Step 10. */ prev 1802 arch/powerpc/platforms/cell/spufs/switch.c save_spu_status(prev, spu); /* Step 11. */ prev 1803 arch/powerpc/platforms/cell/spufs/switch.c save_mfc_stopped_status(prev, spu); /* Step 12. */ prev 1804 arch/powerpc/platforms/cell/spufs/switch.c halt_mfc_decr(prev, spu); /* Step 13. */ prev 1805 arch/powerpc/platforms/cell/spufs/switch.c save_timebase(prev, spu); /* Step 14. */ prev 1806 arch/powerpc/platforms/cell/spufs/switch.c remove_other_spu_access(prev, spu); /* Step 15. */ prev 1807 arch/powerpc/platforms/cell/spufs/switch.c do_mfc_mssync(prev, spu); /* Step 16. */ prev 1808 arch/powerpc/platforms/cell/spufs/switch.c issue_mfc_tlbie(prev, spu); /* Step 17. */ prev 1809 arch/powerpc/platforms/cell/spufs/switch.c handle_pending_interrupts(prev, spu); /* Step 18. */ prev 1814 arch/powerpc/platforms/cell/spufs/switch.c static void save_csa(struct spu_state *prev, struct spu *spu) prev 1821 arch/powerpc/platforms/cell/spufs/switch.c save_mfc_queues(prev, spu); /* Step 19. */ prev 1822 arch/powerpc/platforms/cell/spufs/switch.c save_ppu_querymask(prev, spu); /* Step 20. */ prev 1823 arch/powerpc/platforms/cell/spufs/switch.c save_ppu_querytype(prev, spu); /* Step 21. */ prev 1824 arch/powerpc/platforms/cell/spufs/switch.c save_ppu_tagstatus(prev, spu); /* NEW. */ prev 1825 arch/powerpc/platforms/cell/spufs/switch.c save_mfc_csr_tsq(prev, spu); /* Step 22. */ prev 1826 arch/powerpc/platforms/cell/spufs/switch.c save_mfc_csr_cmd(prev, spu); /* Step 23. */ prev 1827 arch/powerpc/platforms/cell/spufs/switch.c save_mfc_csr_ato(prev, spu); /* Step 24. */ prev 1828 arch/powerpc/platforms/cell/spufs/switch.c save_mfc_tclass_id(prev, spu); /* Step 25. */ prev 1829 arch/powerpc/platforms/cell/spufs/switch.c set_mfc_tclass_id(prev, spu); /* Step 26. */ prev 1830 arch/powerpc/platforms/cell/spufs/switch.c save_mfc_cmd(prev, spu); /* Step 26a - moved from 44. */ prev 1831 arch/powerpc/platforms/cell/spufs/switch.c purge_mfc_queue(prev, spu); /* Step 27. */ prev 1832 arch/powerpc/platforms/cell/spufs/switch.c wait_purge_complete(prev, spu); /* Step 28. */ prev 1833 arch/powerpc/platforms/cell/spufs/switch.c setup_mfc_sr1(prev, spu); /* Step 30. */ prev 1834 arch/powerpc/platforms/cell/spufs/switch.c save_spu_npc(prev, spu); /* Step 31. */ prev 1835 arch/powerpc/platforms/cell/spufs/switch.c save_spu_privcntl(prev, spu); /* Step 32. */ prev 1836 arch/powerpc/platforms/cell/spufs/switch.c reset_spu_privcntl(prev, spu); /* Step 33. */ prev 1837 arch/powerpc/platforms/cell/spufs/switch.c save_spu_lslr(prev, spu); /* Step 34. */ prev 1838 arch/powerpc/platforms/cell/spufs/switch.c reset_spu_lslr(prev, spu); /* Step 35. */ prev 1839 arch/powerpc/platforms/cell/spufs/switch.c save_spu_cfg(prev, spu); /* Step 36. */ prev 1840 arch/powerpc/platforms/cell/spufs/switch.c save_pm_trace(prev, spu); /* Step 37. */ prev 1841 arch/powerpc/platforms/cell/spufs/switch.c save_mfc_rag(prev, spu); /* Step 38. */ prev 1842 arch/powerpc/platforms/cell/spufs/switch.c save_ppu_mb_stat(prev, spu); /* Step 39. */ prev 1843 arch/powerpc/platforms/cell/spufs/switch.c save_ppu_mb(prev, spu); /* Step 40. */ prev 1844 arch/powerpc/platforms/cell/spufs/switch.c save_ppuint_mb(prev, spu); /* Step 41. */ prev 1845 arch/powerpc/platforms/cell/spufs/switch.c save_ch_part1(prev, spu); /* Step 42. */ prev 1846 arch/powerpc/platforms/cell/spufs/switch.c save_spu_mb(prev, spu); /* Step 43. */ prev 1847 arch/powerpc/platforms/cell/spufs/switch.c reset_ch(prev, spu); /* Step 45. */ prev 1850 arch/powerpc/platforms/cell/spufs/switch.c static void save_lscsa(struct spu_state *prev, struct spu *spu) prev 1858 arch/powerpc/platforms/cell/spufs/switch.c resume_mfc_queue(prev, spu); /* Step 46. */ prev 1860 arch/powerpc/platforms/cell/spufs/switch.c setup_mfc_slbs(prev, spu, spu_save_code, sizeof(spu_save_code)); prev 1861 arch/powerpc/platforms/cell/spufs/switch.c set_switch_active(prev, spu); /* Step 48. */ prev 1862 arch/powerpc/platforms/cell/spufs/switch.c enable_interrupts(prev, spu); /* Step 49. */ prev 1863 arch/powerpc/platforms/cell/spufs/switch.c save_ls_16kb(prev, spu); /* Step 50. */ prev 1864 arch/powerpc/platforms/cell/spufs/switch.c set_spu_npc(prev, spu); /* Step 51. */ prev 1865 arch/powerpc/platforms/cell/spufs/switch.c set_signot1(prev, spu); /* Step 52. */ prev 1866 arch/powerpc/platforms/cell/spufs/switch.c set_signot2(prev, spu); /* Step 53. */ prev 1867 arch/powerpc/platforms/cell/spufs/switch.c send_save_code(prev, spu); /* Step 54. */ prev 1868 arch/powerpc/platforms/cell/spufs/switch.c set_ppu_querymask(prev, spu); /* Step 55. */ prev 1869 arch/powerpc/platforms/cell/spufs/switch.c wait_tag_complete(prev, spu); /* Step 56. */ prev 1870 arch/powerpc/platforms/cell/spufs/switch.c wait_spu_stopped(prev, spu); /* Step 57. */ prev 1918 arch/powerpc/platforms/cell/spufs/switch.c static void harvest(struct spu_state *prev, struct spu *spu) prev 1926 arch/powerpc/platforms/cell/spufs/switch.c disable_interrupts(prev, spu); /* Step 2. */ prev 1927 arch/powerpc/platforms/cell/spufs/switch.c inhibit_user_access(prev, spu); /* Step 3. */ prev 1928 arch/powerpc/platforms/cell/spufs/switch.c terminate_spu_app(prev, spu); /* Step 4. */ prev 1929 arch/powerpc/platforms/cell/spufs/switch.c set_switch_pending(prev, spu); /* Step 5. */ prev 1931 arch/powerpc/platforms/cell/spufs/switch.c remove_other_spu_access(prev, spu); /* Step 6. */ prev 1932 arch/powerpc/platforms/cell/spufs/switch.c suspend_mfc_and_halt_decr(prev, spu); /* Step 7. */ prev 1933 arch/powerpc/platforms/cell/spufs/switch.c wait_suspend_mfc_complete(prev, spu); /* Step 8. */ prev 1934 arch/powerpc/platforms/cell/spufs/switch.c if (!suspend_spe(prev, spu)) /* Step 9. */ prev 1935 arch/powerpc/platforms/cell/spufs/switch.c clear_spu_status(prev, spu); /* Step 10. */ prev 1936 arch/powerpc/platforms/cell/spufs/switch.c do_mfc_mssync(prev, spu); /* Step 11. */ prev 1937 arch/powerpc/platforms/cell/spufs/switch.c issue_mfc_tlbie(prev, spu); /* Step 12. */ prev 1938 arch/powerpc/platforms/cell/spufs/switch.c handle_pending_interrupts(prev, spu); /* Step 13. */ prev 1939 arch/powerpc/platforms/cell/spufs/switch.c purge_mfc_queue(prev, spu); /* Step 14. */ prev 1940 arch/powerpc/platforms/cell/spufs/switch.c wait_purge_complete(prev, spu); /* Step 15. */ prev 1941 arch/powerpc/platforms/cell/spufs/switch.c reset_spu_privcntl(prev, spu); /* Step 16. */ prev 1942 arch/powerpc/platforms/cell/spufs/switch.c reset_spu_lslr(prev, spu); /* Step 17. */ prev 1943 arch/powerpc/platforms/cell/spufs/switch.c setup_mfc_sr1(prev, spu); /* Step 18. */ prev 1945 arch/powerpc/platforms/cell/spufs/switch.c reset_ch_part1(prev, spu); /* Step 20. */ prev 1946 arch/powerpc/platforms/cell/spufs/switch.c reset_ch_part2(prev, spu); /* Step 21. */ prev 1947 arch/powerpc/platforms/cell/spufs/switch.c enable_interrupts(prev, spu); /* Step 22. */ prev 1948 arch/powerpc/platforms/cell/spufs/switch.c set_switch_active(prev, spu); /* Step 23. */ prev 1949 arch/powerpc/platforms/cell/spufs/switch.c set_mfc_tclass_id(prev, spu); /* Step 24. */ prev 1950 arch/powerpc/platforms/cell/spufs/switch.c resume_mfc_queue(prev, spu); /* Step 25. */ prev 2024 arch/powerpc/platforms/cell/spufs/switch.c static int __do_spu_save(struct spu_state *prev, struct spu *spu) prev 2040 arch/powerpc/platforms/cell/spufs/switch.c rc = quiece_spu(prev, spu); /* Steps 2-16. */ prev 2045 arch/powerpc/platforms/cell/spufs/switch.c harvest(prev, spu); prev 2051 arch/powerpc/platforms/cell/spufs/switch.c save_csa(prev, spu); /* Steps 17-43. */ prev 2052 arch/powerpc/platforms/cell/spufs/switch.c save_lscsa(prev, spu); /* Steps 44-53. */ prev 2053 arch/powerpc/platforms/cell/spufs/switch.c return check_save_status(prev, spu); /* Step 54. */ prev 2094 arch/powerpc/platforms/cell/spufs/switch.c int spu_save(struct spu_state *prev, struct spu *spu) prev 2099 arch/powerpc/platforms/cell/spufs/switch.c rc = __do_spu_save(prev, spu); /* Steps 2-53. */ prev 958 arch/powerpc/platforms/powermac/low_i2c.c struct device_node *prev = NULL; prev 964 arch/powerpc/platforms/powermac/low_i2c.c if (prev && bus->flags & pmac_i2c_multibus) { prev 966 arch/powerpc/platforms/powermac/low_i2c.c reg = of_get_property(prev, "reg", prev 974 arch/powerpc/platforms/powermac/low_i2c.c of_node_put(prev); prev 978 arch/powerpc/platforms/powermac/low_i2c.c of_node_put(prev); prev 979 arch/powerpc/platforms/powermac/low_i2c.c prev = p; prev 329 arch/powerpc/sysdev/fsl_msi.c if (msi->list.prev != NULL) prev 417 arch/powerpc/xmon/xmon.c int last_speaker = 0, prev; prev 440 arch/powerpc/xmon/xmon.c prev = cmpxchg(&xmon_speaker, last_speaker, me); prev 441 arch/powerpc/xmon/xmon.c if (prev == last_speaker) prev 204 arch/riscv/include/asm/atomic.h int prev, rc; prev 214 arch/riscv/include/asm/atomic.h : [p]"=&r" (prev), [rc]"=&r" (rc), [c]"+A" (v->counter) prev 217 arch/riscv/include/asm/atomic.h return prev; prev 224 arch/riscv/include/asm/atomic.h s64 prev; prev 235 arch/riscv/include/asm/atomic.h : [p]"=&r" (prev), [rc]"=&r" (rc), [c]"+A" (v->counter) prev 238 arch/riscv/include/asm/atomic.h return prev; prev 317 arch/riscv/include/asm/atomic.h int prev, rc; prev 327 arch/riscv/include/asm/atomic.h : [p]"=&r" (prev), [rc]"=&r" (rc), [c]"+A" (v->counter) prev 330 arch/riscv/include/asm/atomic.h return prev - offset; prev 338 arch/riscv/include/asm/atomic.h s64 prev; prev 349 arch/riscv/include/asm/atomic.h : [p]"=&r" (prev), [rc]"=&r" (rc), [c]"+A" (v->counter) prev 352 arch/riscv/include/asm/atomic.h return prev - offset; prev 32 arch/riscv/include/asm/mmu_context.h void switch_mm(struct mm_struct *prev, struct mm_struct *next, prev 35 arch/riscv/include/asm/mmu_context.h static inline void activate_mm(struct mm_struct *prev, prev 38 arch/riscv/include/asm/mmu_context.h switch_mm(prev, next, NULL); prev 47 arch/riscv/include/asm/switch_to.h static inline void __switch_to_aux(struct task_struct *prev, prev 52 arch/riscv/include/asm/switch_to.h regs = task_pt_regs(prev); prev 54 arch/riscv/include/asm/switch_to.h fstate_save(prev, regs); prev 69 arch/riscv/include/asm/switch_to.h #define switch_to(prev, next, last) \ prev 71 arch/riscv/include/asm/switch_to.h struct task_struct *__prev = (prev); \ prev 23 arch/riscv/kernel/cacheinfo.c struct device_node *prev = NULL; prev 35 arch/riscv/kernel/cacheinfo.c prev = np; prev 37 arch/riscv/kernel/cacheinfo.c of_node_put(prev); prev 38 arch/riscv/kernel/cacheinfo.c prev = np; prev 66 arch/riscv/kernel/cacheinfo.c struct device_node *prev = NULL; prev 76 arch/riscv/kernel/cacheinfo.c prev = np; prev 78 arch/riscv/kernel/cacheinfo.c of_node_put(prev); prev 79 arch/riscv/kernel/cacheinfo.c prev = np; prev 43 arch/riscv/mm/context.c void switch_mm(struct mm_struct *prev, struct mm_struct *next, prev 48 arch/riscv/mm/context.c if (unlikely(prev == next)) prev 58 arch/riscv/mm/context.c cpumask_clear_cpu(cpu, mm_cpumask(prev)); prev 102 arch/s390/crypto/prng.c u8 *prev; prev 428 arch/s390/crypto/prng.c prng_data->prev = prng_data->buf + prng_chunk_size; prev 431 arch/s390/crypto/prng.c prng_data->prev, prng_chunk_size, NULL, 0); prev 493 arch/s390/crypto/prng.c if (!memcmp(prng_data->prev, buf, nbytes)) { prev 497 arch/s390/crypto/prng.c memcpy(prng_data->prev, buf, nbytes); prev 35 arch/s390/include/asm/debug.h struct debug_info *prev; prev 88 arch/s390/include/asm/mmu_context.h static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, prev 104 arch/s390/include/asm/mmu_context.h if (prev != next) prev 105 arch/s390/include/asm/mmu_context.h cpumask_clear_cpu(cpu, &prev->context.cpu_attach_mask); prev 128 arch/s390/include/asm/mmu_context.h static inline void activate_mm(struct mm_struct *prev, prev 131 arch/s390/include/asm/mmu_context.h switch_mm(prev, next, current); prev 99 arch/s390/include/asm/stacktrace.h unsigned long prev; \ prev 107 arch/s390/include/asm/stacktrace.h : [_prev] "=&a" (prev), CALL_FMT_##nr \ prev 33 arch/s390/include/asm/switch_to.h #define switch_to(prev, next, last) do { \ prev 39 arch/s390/include/asm/switch_to.h save_access_regs(&prev->thread.acrs[0]); \ prev 40 arch/s390/include/asm/switch_to.h save_ri_cb(prev->thread.ri_cb); \ prev 41 arch/s390/include/asm/switch_to.h save_gs_cb(prev->thread.gs_cb); \ prev 44 arch/s390/include/asm/switch_to.h restore_ri_cb(next->thread.ri_cb, prev->thread.ri_cb); \ prev 46 arch/s390/include/asm/switch_to.h prev = __switch_to(prev, next); \ prev 338 arch/s390/kernel/debug.c rc->prev = NULL; prev 342 arch/s390/kernel/debug.c rc->prev = debug_area_last; prev 420 arch/s390/kernel/debug.c debug_area_last = db_info->prev; prev 421 arch/s390/kernel/debug.c if (db_info->prev) prev 422 arch/s390/kernel/debug.c db_info->prev->next = db_info->next; prev 424 arch/s390/kernel/debug.c db_info->next->prev = db_info->prev; prev 316 arch/s390/kernel/perf_cpum_cf.c u64 prev, new; prev 320 arch/s390/kernel/perf_cpum_cf.c prev = local64_read(&event->hw.prev_count); prev 332 arch/s390/kernel/perf_cpum_cf.c } while (local64_cmpxchg(&event->hw.prev_count, prev, new) != prev); prev 339 arch/s390/kernel/perf_cpum_cf.c u64 prev, new, delta; prev 343 arch/s390/kernel/perf_cpum_cf.c prev = local64_read(&event->hw.prev_count); prev 347 arch/s390/kernel/perf_cpum_cf.c } while (local64_cmpxchg(&event->hw.prev_count, prev, new) != prev); prev 349 arch/s390/kernel/perf_cpum_cf.c delta = (prev <= new) ? new - prev prev 350 arch/s390/kernel/perf_cpum_cf.c : (-1ULL - prev) + new + 1; /* overflow */ prev 188 arch/s390/kernel/vtime.c void vtime_task_switch(struct task_struct *prev) prev 190 arch/s390/kernel/vtime.c do_account_vtime(prev); prev 191 arch/s390/kernel/vtime.c prev->thread.user_timer = S390_lowcore.user_timer; prev 192 arch/s390/kernel/vtime.c prev->thread.guest_timer = S390_lowcore.guest_timer; prev 193 arch/s390/kernel/vtime.c prev->thread.system_timer = S390_lowcore.system_timer; prev 194 arch/s390/kernel/vtime.c prev->thread.hardirq_timer = S390_lowcore.hardirq_timer; prev 195 arch/s390/kernel/vtime.c prev->thread.softirq_timer = S390_lowcore.softirq_timer; prev 40 arch/s390/lib/spinlock.c struct spin_wait *next, *prev; prev 110 arch/s390/lib/spinlock.c if (node == NULL || node->prev == NULL) prev 112 arch/s390/lib/spinlock.c while (node->prev) prev 113 arch/s390/lib/spinlock.c node = node->prev; prev 126 arch/s390/lib/spinlock.c node->prev = node->next = NULL; prev 156 arch/s390/lib/spinlock.c node->prev = arch_spin_decode_tail(tail_id); prev 157 arch/s390/lib/spinlock.c WRITE_ONCE(node->prev->next, node); prev 168 arch/s390/lib/spinlock.c while (READ_ONCE(node->prev) != NULL) { prev 204 arch/s390/lib/spinlock.c next->prev = NULL; prev 44 arch/s390/numa/toptree.h (ptree->parent->children.prev == &ptree->sibling)) prev 273 arch/sh/include/asm/dwarf.h struct dwarf_frame *prev, *next; prev 11 arch/sh/include/asm/futex-irq.h u32 prev = 0; prev 15 arch/sh/include/asm/futex-irq.h ret = get_user(prev, uaddr); prev 16 arch/sh/include/asm/futex-irq.h if (!ret && oldval == prev) prev 21 arch/sh/include/asm/futex-irq.h *uval = prev; prev 34 arch/sh/include/asm/futex.h u32 oldval, newval, prev; prev 67 arch/sh/include/asm/futex.h ret = futex_atomic_cmpxchg_inatomic(&prev, uaddr, oldval, newval); prev 68 arch/sh/include/asm/futex.h } while (!ret && prev != oldval); prev 121 arch/sh/include/asm/mmu_context.h static inline void switch_mm(struct mm_struct *prev, prev 127 arch/sh/include/asm/mmu_context.h if (likely(prev != next)) { prev 136 arch/sh/include/asm/mmu_context.h #define activate_mm(prev, next) switch_mm((prev),(next),NULL) prev 64 arch/sh/include/asm/switch_to_32.h struct task_struct *__switch_to(struct task_struct *prev, prev 70 arch/sh/include/asm/switch_to_32.h #define switch_to(prev, next, last) \ prev 80 arch/sh/include/asm/switch_to_32.h if (is_dsp_enabled(prev)) \ prev 81 arch/sh/include/asm/switch_to_32.h __save_dsp(prev); \ prev 85 arch/sh/include/asm/switch_to_32.h __ts1 = (u32 *)&prev->thread.sp; \ prev 86 arch/sh/include/asm/switch_to_32.h __ts2 = (u32 *)&prev->thread.pc; \ prev 87 arch/sh/include/asm/switch_to_32.h __ts4 = (u32 *)prev; \ prev 16 arch/sh/include/asm/switch_to_64.h struct task_struct *sh64_switch_to(struct task_struct *prev, prev 21 arch/sh/include/asm/switch_to_64.h #define switch_to(prev,next,last) \ prev 27 arch/sh/include/asm/switch_to_64.h last = sh64_switch_to(prev, &prev->thread, next, \ prev 575 arch/sh/kernel/dwarf.c struct dwarf_frame *prev) prev 598 arch/sh/kernel/dwarf.c if (!pc || !prev) prev 631 arch/sh/kernel/dwarf.c frame->prev = prev; prev 670 arch/sh/kernel/dwarf.c if (prev) { prev 671 arch/sh/kernel/dwarf.c reg = dwarf_frame_reg(prev, frame->cfa_register); prev 675 arch/sh/kernel/dwarf.c addr = prev->cfa + reg->addr; prev 725 arch/sh/kernel/dwarf.c if (prev && prev->pc == (unsigned long)ret_from_irq) prev 173 arch/sh/kernel/process_32.c __switch_to(struct task_struct *prev, struct task_struct *next) prev 181 arch/sh/kernel/process_32.c unlazy_fpu(prev, task_pt_regs(prev)); prev 205 arch/sh/kernel/process_32.c return prev; prev 35 arch/sh/kernel/unwinder.c .prev = &unwinder_list, prev 53 arch/sh/kernel/unwinder.c .prev = &stack_reader.list, prev 143 arch/sparc/include/asm/mmu_context_64.h static inline void arch_start_context_switch(struct task_struct *prev) prev 158 arch/sparc/include/asm/mmu_context_64.h set_tsk_thread_flag(prev, TIF_MCDPER); prev 160 arch/sparc/include/asm/mmu_context_64.h clear_tsk_thread_flag(prev, TIF_MCDPER); prev 58 arch/sparc/include/asm/switch_to_32.h #define switch_to(prev, next, last) do { \ prev 59 arch/sparc/include/asm/switch_to_32.h SWITCH_ENTER(prev); \ prev 21 arch/sparc/include/asm/switch_to_64.h #define switch_to(prev, next, last) \ prev 234 arch/sparc/kernel/leon_smp.c int *prev; prev 238 arch/sparc/kernel/leon_smp.c prev = &first; prev 241 arch/sparc/kernel/leon_smp.c *prev = i; prev 242 arch/sparc/kernel/leon_smp.c prev = &cpu_data(i).next; prev 245 arch/sparc/kernel/leon_smp.c *prev = first; prev 132 arch/sparc/kernel/prom_common.c static int __init prom_common_nextprop(phandle node, char *prev, char *buf) prev 137 arch/sparc/kernel/prom_common.c name = prom_nextprop(node, prev, buf); prev 166 arch/sparc/kernel/sun4d_smp.c int *prev; prev 170 arch/sparc/kernel/sun4d_smp.c prev = &first; prev 172 arch/sparc/kernel/sun4d_smp.c *prev = i; prev 173 arch/sparc/kernel/sun4d_smp.c prev = &cpu_data(i).next; prev 175 arch/sparc/kernel/sun4d_smp.c *prev = first; prev 124 arch/sparc/kernel/sun4m_smp.c int *prev; prev 128 arch/sparc/kernel/sun4m_smp.c prev = &first; prev 130 arch/sparc/kernel/sun4m_smp.c *prev = i; prev 131 arch/sparc/kernel/sun4m_smp.c prev = &cpu_data(i).next; prev 133 arch/sparc/kernel/sun4m_smp.c *prev = first; prev 165 arch/sparc/lib/atomic32.c u32 prev; prev 168 arch/sparc/lib/atomic32.c if ((prev = *ptr) == old) prev 172 arch/sparc/lib/atomic32.c return (unsigned long)prev; prev 179 arch/sparc/lib/atomic32.c u64 prev; prev 182 arch/sparc/lib/atomic32.c if ((prev = *ptr) == old) prev 186 arch/sparc/lib/atomic32.c return prev; prev 193 arch/sparc/lib/atomic32.c u32 prev; prev 196 arch/sparc/lib/atomic32.c prev = *ptr; prev 200 arch/sparc/lib/atomic32.c return (unsigned long)prev; prev 407 arch/sparc/mm/srmmu.c struct ctx_list *prev; prev 421 arch/sparc/mm/srmmu.c entry->next->prev = entry->prev; prev 422 arch/sparc/mm/srmmu.c entry->prev->next = entry->next; prev 428 arch/sparc/mm/srmmu.c (entry->prev = head->prev)->next = entry; prev 429 arch/sparc/mm/srmmu.c head->prev = entry; prev 487 arch/sparc/mm/srmmu.c ctx_free.next = ctx_free.prev = &ctx_free; prev 488 arch/sparc/mm/srmmu.c ctx_used.next = ctx_used.prev = &ctx_used; prev 60 arch/um/include/asm/mmu_context.h static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, prev 65 arch/um/include/asm/mmu_context.h if(prev != next){ prev 66 arch/um/include/asm/mmu_context.h cpumask_clear_cpu(cpu, mm_cpumask(prev)); prev 46 arch/unicore32/include/asm/mmu_context.h switch_mm(struct mm_struct *prev, struct mm_struct *next, prev 51 arch/unicore32/include/asm/mmu_context.h if (!cpumask_test_and_set_cpu(cpu, mm_cpumask(next)) || prev != next) prev 56 arch/unicore32/include/asm/mmu_context.h #define activate_mm(prev, next) switch_mm(prev, next, NULL) prev 21 arch/unicore32/include/asm/switch_to.h #define switch_to(prev, next, last) \ prev 23 arch/unicore32/include/asm/switch_to.h last = __switch_to(prev, task_thread_info(prev), \ prev 516 arch/x86/boot/compressed/eboot.c struct boot_e820_entry *prev = NULL; prev 573 arch/x86/boot/compressed/eboot.c if (prev && prev->type == e820_type && prev 574 arch/x86/boot/compressed/eboot.c (prev->addr + prev->size) == d->phys_addr) { prev 575 arch/x86/boot/compressed/eboot.c prev->size += d->num_pages << 12; prev 593 arch/x86/boot/compressed/eboot.c prev = entry++; prev 303 arch/x86/events/amd/iommu.c u64 count, prev, delta; prev 314 arch/x86/events/amd/iommu.c prev = local64_read(&hwc->prev_count); prev 315 arch/x86/events/amd/iommu.c if (local64_cmpxchg(&hwc->prev_count, prev, count) != prev) prev 319 arch/x86/events/amd/iommu.c delta = (count << COUNTER_SHIFT) - (prev << COUNTER_SHIFT); prev 85 arch/x86/events/amd/uncore.c u64 prev, new; prev 93 arch/x86/events/amd/uncore.c prev = local64_read(&hwc->prev_count); prev 96 arch/x86/events/amd/uncore.c delta = (new << COUNTER_SHIFT) - (prev << COUNTER_SHIFT); prev 229 arch/x86/events/msr.c u64 prev, now; prev 234 arch/x86/events/msr.c prev = local64_read(&event->hw.prev_count); prev 237 arch/x86/events/msr.c if (local64_cmpxchg(&event->hw.prev_count, prev, now) != prev) prev 240 arch/x86/events/msr.c delta = now - prev; prev 28 arch/x86/include/asm/cmpxchg_32.h u64 prev = *ptr; prev 33 arch/x86/include/asm/cmpxchg_32.h : "=m" (*ptr), "+A" (prev) prev 49 arch/x86/include/asm/cmpxchg_32.h u64 prev; prev 51 arch/x86/include/asm/cmpxchg_32.h : "=A" (prev), prev 57 arch/x86/include/asm/cmpxchg_32.h return prev; prev 62 arch/x86/include/asm/cmpxchg_32.h u64 prev; prev 64 arch/x86/include/asm/cmpxchg_32.h : "=A" (prev), prev 70 arch/x86/include/asm/cmpxchg_32.h return prev; prev 21 arch/x86/include/asm/mmu_context.h static inline void paravirt_activate_mm(struct mm_struct *prev, prev 152 arch/x86/include/asm/mmu_context.h static inline void switch_ldt(struct mm_struct *prev, struct mm_struct *next) prev 172 arch/x86/include/asm/mmu_context.h if (unlikely((unsigned long)prev->context.ldt | prev 210 arch/x86/include/asm/mmu_context.h extern void switch_mm(struct mm_struct *prev, struct mm_struct *next, prev 213 arch/x86/include/asm/mmu_context.h extern void switch_mm_irqs_off(struct mm_struct *prev, struct mm_struct *next, prev 217 arch/x86/include/asm/mmu_context.h #define activate_mm(prev, next) \ prev 219 arch/x86/include/asm/mmu_context.h paravirt_activate_mm((prev), (next)); \ prev 220 arch/x86/include/asm/mmu_context.h switch_mm((prev), (next), NULL); \ prev 302 arch/x86/include/asm/paravirt.h static inline void paravirt_activate_mm(struct mm_struct *prev, prev 305 arch/x86/include/asm/paravirt.h PVOP_VCALL2(mmu.activate_mm, prev, next); prev 604 arch/x86/include/asm/paravirt.h static inline void arch_start_context_switch(struct task_struct *prev) prev 606 arch/x86/include/asm/paravirt.h PVOP_VCALL1(cpu.start_context_switch, prev); prev 178 arch/x86/include/asm/paravirt_types.h void (*start_context_switch)(struct task_struct *prev); prev 225 arch/x86/include/asm/paravirt_types.h void (*activate_mm)(struct mm_struct *prev, prev 656 arch/x86/include/asm/paravirt_types.h void paravirt_start_context_switch(struct task_struct *prev); prev 116 arch/x86/include/asm/pgtable.h #define arch_end_context_switch(prev) do {} while(0) prev 9 arch/x86/include/asm/switch_to.h struct task_struct *__switch_to_asm(struct task_struct *prev, prev 12 arch/x86/include/asm/switch_to.h __visible struct task_struct *__switch_to(struct task_struct *prev, prev 68 arch/x86/include/asm/switch_to.h #define switch_to(prev, next, last) \ prev 72 arch/x86/include/asm/switch_to.h ((last) = __switch_to_asm((prev), (next))); \ prev 792 arch/x86/kernel/alternative.c temp_mm_state_t prev; prev 850 arch/x86/kernel/alternative.c prev = use_temporary_mm(poking_mm); prev 871 arch/x86/kernel/alternative.c unuse_temporary_mm(prev); prev 630 arch/x86/kernel/cpu/microcode/amd.c __list_del(p->plist.prev, p->plist.next); prev 764 arch/x86/kernel/cpu/microcode/intel.c static struct cpu_signature prev; prev 781 arch/x86/kernel/cpu/microcode/intel.c if (csig->sig != prev.sig || csig->pf != prev.pf || csig->rev != prev.rev) { prev 784 arch/x86/kernel/cpu/microcode/intel.c prev = *csig; prev 83 arch/x86/kernel/cpu/mtrr/generic.c static int check_type_overlap(u8 *prev, u8 *curr) prev 85 arch/x86/kernel/cpu/mtrr/generic.c if (*prev == MTRR_TYPE_UNCACHABLE || *curr == MTRR_TYPE_UNCACHABLE) { prev 86 arch/x86/kernel/cpu/mtrr/generic.c *prev = MTRR_TYPE_UNCACHABLE; prev 91 arch/x86/kernel/cpu/mtrr/generic.c if ((*prev == MTRR_TYPE_WRBACK && *curr == MTRR_TYPE_WRTHROUGH) || prev 92 arch/x86/kernel/cpu/mtrr/generic.c (*prev == MTRR_TYPE_WRTHROUGH && *curr == MTRR_TYPE_WRBACK)) { prev 93 arch/x86/kernel/cpu/mtrr/generic.c *prev = MTRR_TYPE_WRTHROUGH; prev 97 arch/x86/kernel/cpu/mtrr/generic.c if (*prev != *curr) { prev 98 arch/x86/kernel/cpu/mtrr/generic.c *prev = MTRR_TYPE_UNCACHABLE; prev 142 arch/x86/kernel/jump_label.c int prev = tp_vec_nr - 1; prev 143 arch/x86/kernel/jump_label.c struct text_poke_loc *prev_tp = &tp_vec[prev]; prev 251 arch/x86/kernel/paravirt.c void paravirt_start_context_switch(struct task_struct *prev) prev 257 arch/x86/kernel/paravirt.c set_ti_thread_flag(task_thread_info(prev), TIF_LAZY_MMU_UPDATES); prev 272 arch/x86/kernel/process.c static inline void switch_to_bitmap(struct thread_struct *prev, prev 284 arch/x86/kernel/process.c max(prev->io_bitmap_max, next->io_bitmap_max)); prev 294 arch/x86/kernel/process.c memset(tss->io_bitmap, 0xff, prev->io_bitmap_max); prev 492 arch/x86/kernel/process.c struct thread_struct *prev, *next; prev 495 arch/x86/kernel/process.c prev = &prev_p->thread; prev 500 arch/x86/kernel/process.c switch_to_bitmap(prev, next, tifp, tifn); prev 13 arch/x86/kernel/process.h static inline void switch_to_extra(struct task_struct *prev, prev 17 arch/x86/kernel/process.h unsigned long prev_tif = task_thread_info(prev)->flags; prev 38 arch/x86/kernel/process.h __switch_to_xtra(prev, next); prev 230 arch/x86/kernel/process_32.c struct thread_struct *prev = &prev_p->thread, prev 232 arch/x86/kernel/process_32.c struct fpu *prev_fpu = &prev->fpu; prev 251 arch/x86/kernel/process_32.c lazy_save_gs(prev->gs); prev 264 arch/x86/kernel/process_32.c if (get_kernel_rpl() && unlikely(prev->iopl != next->iopl)) prev 290 arch/x86/kernel/process_32.c if (prev->gs | next->gs) prev 280 arch/x86/kernel/process_64.c static __always_inline void x86_fsgsbase_load(struct thread_struct *prev, prev 283 arch/x86/kernel/process_64.c load_seg_legacy(prev->fsindex, prev->fsbase, prev 285 arch/x86/kernel/process_64.c load_seg_legacy(prev->gsindex, prev->gsbase, prev 506 arch/x86/kernel/process_64.c struct thread_struct *prev = &prev_p->thread; prev 508 arch/x86/kernel/process_64.c struct fpu *prev_fpu = &prev->fpu; prev 552 arch/x86/kernel/process_64.c savesegment(es, prev->es); prev 553 arch/x86/kernel/process_64.c if (unlikely(next->es | prev->es)) prev 556 arch/x86/kernel/process_64.c savesegment(ds, prev->ds); prev 557 arch/x86/kernel/process_64.c if (unlikely(next->ds | prev->ds)) prev 560 arch/x86/kernel/process_64.c x86_fsgsbase_load(prev, next); prev 582 arch/x86/kernel/process_64.c prev->iopl != next->iopl)) prev 228 arch/x86/kernel/tsc_sync.c cycles_t start, now, prev, end, cur_max_warp = 0; prev 245 arch/x86/kernel/tsc_sync.c prev = last_tsc; prev 266 arch/x86/kernel/tsc_sync.c if (unlikely(prev > now)) { prev 268 arch/x86/kernel/tsc_sync.c max_warp = max(max_warp, prev - now); prev 275 arch/x86/kvm/vmx/nested.c struct loaded_vmcs *prev) prev 282 arch/x86/kvm/vmx/nested.c src = &prev->host_state; prev 296 arch/x86/kvm/vmx/nested.c struct loaded_vmcs *prev; prev 303 arch/x86/kvm/vmx/nested.c prev = vmx->loaded_vmcs; prev 305 arch/x86/kvm/vmx/nested.c vmx_vcpu_load_vmcs(vcpu, cpu, prev); prev 306 arch/x86/kvm/vmx/nested.c vmx_sync_vmcs_host_state(vmx, prev); prev 1294 arch/x86/kvm/vmx/vmx.c struct vmcs *prev; prev 1313 arch/x86/kvm/vmx/vmx.c prev = per_cpu(current_vmcs, cpu); prev 1314 arch/x86/kvm/vmx/vmx.c if (prev != vmx->loaded_vmcs->vmcs) { prev 1323 arch/x86/kvm/vmx/vmx.c if (!buddy || WARN_ON_ONCE(buddy->vmcs != prev)) prev 778 arch/x86/mm/mpx.c struct vm_area_struct *prev; prev 794 arch/x86/mm/mpx.c next = find_vma_prev(mm, start, &prev); prev 805 arch/x86/mm/mpx.c while (prev && (prev->vm_flags & VM_MPX)) prev 806 arch/x86/mm/mpx.c prev = prev->vm_prev; prev 814 arch/x86/mm/mpx.c next = find_vma_prev(mm, start, &prev); prev 815 arch/x86/mm/mpx.c if ((!prev || prev->vm_end <= bta_start_vaddr) && prev 154 arch/x86/mm/tlb.c void switch_mm(struct mm_struct *prev, struct mm_struct *next, prev 160 arch/x86/mm/tlb.c switch_mm_irqs_off(prev, next, tsk); prev 275 arch/x86/mm/tlb.c void switch_mm_irqs_off(struct mm_struct *prev, struct mm_struct *next, prev 90 arch/x86/platform/olpc/olpc_dt.c static int __init olpc_dt_nextprop(phandle node, char *prev, char *buf) prev 92 arch/x86/platform/olpc/olpc_dt.c const void *args[] = { (void *)node, prev, buf }; prev 977 arch/x86/xen/mmu_pv.c static void xen_activate_mm(struct mm_struct *prev, struct mm_struct *next) prev 125 arch/xtensa/include/asm/mmu_context.h static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, prev 135 arch/xtensa/include/asm/mmu_context.h if (migrated || prev != next) prev 139 arch/xtensa/include/asm/mmu_context.h #define activate_mm(prev, next) switch_mm((prev), (next), NULL) prev 23 arch/xtensa/include/asm/nommu_context.h static inline void activate_mm(struct mm_struct *prev, struct mm_struct *next) prev 27 arch/xtensa/include/asm/nommu_context.h static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, prev 17 arch/xtensa/include/asm/switch_to.h #define switch_to(prev,next,last) \ prev 19 arch/xtensa/include/asm/switch_to.h (last) = _switch_to(prev, next); \ prev 896 block/bfq-iosched.c struct request *next, *prev = NULL; prev 904 block/bfq-iosched.c prev = rb_entry_rq(rbprev); prev 914 block/bfq-iosched.c return bfq_choose_req(bfqd, next, prev, blk_rq_pos(last)); prev 1868 block/bfq-iosched.c struct request *next_rq, *prev; prev 2055 block/bfq-iosched.c prev = bfqq->next_rq; prev 2063 block/bfq-iosched.c if (unlikely(!bfqd->nonrot_with_queueing && prev != bfqq->next_rq)) prev 2080 block/bfq-iosched.c if (prev != bfqq->next_rq) prev 2164 block/bfq-iosched.c if (rq->queuelist.prev != &rq->queuelist) prev 2273 block/bfq-iosched.c struct request *prev, *next_rq; prev 2285 block/bfq-iosched.c prev = bfqq->next_rq; prev 2294 block/bfq-iosched.c if (prev != bfqq->next_rq) { prev 33 block/blk-integrity.c int prev = 0; prev 37 block/blk-integrity.c if (prev) { prev 50 block/blk-integrity.c prev = 1; prev 75 block/blk-integrity.c int prev = 0; prev 79 block/blk-integrity.c if (prev) { prev 99 block/blk-integrity.c prev = 1; prev 16 block/blk-merge.c struct request *prev_rq, struct bio *prev, struct bio *next) prev 20 block/blk-merge.c if (!bio_has_data(prev) || !queue_virt_boundary(q)) prev 31 block/blk-merge.c bio_get_first_bvec(prev, &pb); prev 44 block/blk-merge.c bio_get_last_bvec(prev, &pb); prev 826 block/blk-merge.c struct request *prev = elv_former_request(q, rq); prev 828 block/blk-merge.c if (prev) prev 829 block/blk-merge.c return attempt_merge(q, prev, rq); prev 2016 block/blk-mq.c last = list_entry_rq(plug->mq_list.prev); prev 493 crypto/af_alg.c sgl = list_entry(ctx->tsgl_list.prev, struct af_alg_tsgl, list); prev 870 crypto/af_alg.c sgl = list_entry(ctx->tsgl_list.prev, prev 905 crypto/af_alg.c sgl = list_entry(ctx->tsgl_list.prev, struct af_alg_tsgl, prev 991 crypto/af_alg.c sgl = list_entry(ctx->tsgl_list.prev, struct af_alg_tsgl, list); prev 109 crypto/cmac.c u8 *prev = PTR_ALIGN((void *)ctx->ctx, alignmask + 1) + bs; prev 112 crypto/cmac.c memset(prev, 0, bs); prev 127 crypto/cmac.c u8 *prev = odds + bs; prev 141 crypto/cmac.c crypto_xor(prev, odds, bs); prev 142 crypto/cmac.c crypto_cipher_encrypt_one(tfm, prev, prev); prev 149 crypto/cmac.c crypto_xor(prev, p, bs); prev 150 crypto/cmac.c crypto_cipher_encrypt_one(tfm, prev, prev); prev 175 crypto/cmac.c u8 *prev = odds + bs; prev 192 crypto/cmac.c crypto_xor(prev, odds, bs); prev 193 crypto/cmac.c crypto_xor(prev, consts + offset, bs); prev 195 crypto/cmac.c crypto_cipher_encrypt_one(tfm, out, prev); prev 259 crypto/drbg.c memcpy(drbg->prev, entropy, entropylen); prev 264 crypto/drbg.c ret = memcmp(drbg->prev, entropy, entropylen); prev 267 crypto/drbg.c memcpy(drbg->prev, entropy, entropylen); prev 1221 crypto/drbg.c kzfree(drbg->prev); prev 1222 crypto/drbg.c drbg->prev = NULL; prev 1295 crypto/drbg.c drbg->prev = kzalloc(drbg_sec_strength(drbg->core->flags), prev 1297 crypto/drbg.c if (!drbg->prev) { prev 76 crypto/xcbc.c u8 *prev = PTR_ALIGN(&ctx->ctx[0], alignmask + 1) + bs; prev 79 crypto/xcbc.c memset(prev, 0, bs); prev 94 crypto/xcbc.c u8 *prev = odds + bs; prev 108 crypto/xcbc.c crypto_xor(prev, odds, bs); prev 109 crypto/xcbc.c crypto_cipher_encrypt_one(tfm, prev, prev); prev 116 crypto/xcbc.c crypto_xor(prev, p, bs); prev 117 crypto/xcbc.c crypto_cipher_encrypt_one(tfm, prev, prev); prev 141 crypto/xcbc.c u8 *prev = odds + bs; prev 158 crypto/xcbc.c crypto_xor(prev, odds, bs); prev 159 crypto/xcbc.c crypto_xor(prev, consts + offset, bs); prev 161 crypto/xcbc.c crypto_cipher_encrypt_one(tfm, out, prev); prev 128 drivers/acpi/acpica/acobject.h union acpi_operand_object *prev; /* Link for list of acquired mutexes */ prev 46 drivers/acpi/acpica/exmutex.c (obj_desc->mutex.next)->mutex.prev = obj_desc->mutex.prev; prev 49 drivers/acpi/acpica/exmutex.c if (obj_desc->mutex.prev) { prev 50 drivers/acpi/acpica/exmutex.c (obj_desc->mutex.prev)->mutex.next = obj_desc->mutex.next; prev 58 drivers/acpi/acpica/exmutex.c (obj_desc->mutex.prev)->mutex.original_sync_level = prev 88 drivers/acpi/acpica/exmutex.c obj_desc->mutex.prev = NULL; prev 94 drivers/acpi/acpica/exmutex.c list_head->mutex.prev = obj_desc; prev 498 drivers/acpi/acpica/exmutex.c obj_desc->mutex.prev = NULL; prev 708 drivers/acpi/acpica/psargs.c union acpi_parse_object *prev = NULL; prev 757 drivers/acpi/acpica/psargs.c if (prev) { prev 758 drivers/acpi/acpica/psargs.c prev->common.next = field; prev 762 drivers/acpi/acpica/psargs.c prev = field; prev 103 drivers/acpi/acpica/psparse.c union acpi_parse_object *prev; prev 130 drivers/acpi/acpica/psparse.c prev = op->common.parent->common.value.arg; prev 131 drivers/acpi/acpica/psparse.c if (!prev) { prev 224 drivers/acpi/acpica/psparse.c if (prev == op) { prev 245 drivers/acpi/acpica/psparse.c while (prev) { prev 249 drivers/acpi/acpica/psparse.c next = prev->common.next; prev 258 drivers/acpi/acpica/psparse.c prev->common.next = prev 264 drivers/acpi/acpica/psparse.c prev->common.next = prev 269 drivers/acpi/acpica/psparse.c prev = next; prev 99 drivers/acpi/acpica/utaddress.c struct acpi_address_range *prev; prev 110 drivers/acpi/acpica/utaddress.c range_info = prev = acpi_gbl_address_range_list[space_id]; prev 113 drivers/acpi/acpica/utaddress.c if (range_info == prev) { /* Found at list head */ prev 117 drivers/acpi/acpica/utaddress.c prev->next = range_info->next; prev 133 drivers/acpi/acpica/utaddress.c prev = range_info; prev 679 drivers/acpi/nfit/core.c struct nfit_table_prev *prev, prev 688 drivers/acpi/nfit/core.c list_for_each_entry(nfit_spa, &prev->spas, list) { prev 709 drivers/acpi/nfit/core.c struct nfit_table_prev *prev, prev 718 drivers/acpi/nfit/core.c list_for_each_entry(nfit_memdev, &prev->memdevs, list) prev 780 drivers/acpi/nfit/core.c struct nfit_table_prev *prev, prev 789 drivers/acpi/nfit/core.c list_for_each_entry(nfit_dcr, &prev->dcrs, list) prev 808 drivers/acpi/nfit/core.c struct nfit_table_prev *prev, prev 816 drivers/acpi/nfit/core.c list_for_each_entry(nfit_bdw, &prev->bdws, list) prev 842 drivers/acpi/nfit/core.c struct nfit_table_prev *prev, prev 851 drivers/acpi/nfit/core.c list_for_each_entry(nfit_idt, &prev->idts, list) { prev 881 drivers/acpi/nfit/core.c struct nfit_table_prev *prev, prev 890 drivers/acpi/nfit/core.c list_for_each_entry(nfit_flush, &prev->flushes, list) { prev 926 drivers/acpi/nfit/core.c struct nfit_table_prev *prev, void *table, const void *end) prev 944 drivers/acpi/nfit/core.c if (!add_spa(acpi_desc, prev, table)) prev 948 drivers/acpi/nfit/core.c if (!add_memdev(acpi_desc, prev, table)) prev 952 drivers/acpi/nfit/core.c if (!add_dcr(acpi_desc, prev, table)) prev 956 drivers/acpi/nfit/core.c if (!add_bdw(acpi_desc, prev, table)) prev 960 drivers/acpi/nfit/core.c if (!add_idt(acpi_desc, prev, table)) prev 964 drivers/acpi/nfit/core.c if (!add_flush(acpi_desc, prev, table)) prev 3325 drivers/acpi/nfit/core.c struct nfit_table_prev *prev) prev 3329 drivers/acpi/nfit/core.c if (!list_empty(&prev->spas) || prev 3330 drivers/acpi/nfit/core.c !list_empty(&prev->memdevs) || prev 3331 drivers/acpi/nfit/core.c !list_empty(&prev->dcrs) || prev 3332 drivers/acpi/nfit/core.c !list_empty(&prev->bdws) || prev 3333 drivers/acpi/nfit/core.c !list_empty(&prev->idts) || prev 3334 drivers/acpi/nfit/core.c !list_empty(&prev->flushes)) { prev 3376 drivers/acpi/nfit/core.c struct nfit_table_prev prev; prev 3405 drivers/acpi/nfit/core.c INIT_LIST_HEAD(&prev.spas); prev 3406 drivers/acpi/nfit/core.c INIT_LIST_HEAD(&prev.memdevs); prev 3407 drivers/acpi/nfit/core.c INIT_LIST_HEAD(&prev.dcrs); prev 3408 drivers/acpi/nfit/core.c INIT_LIST_HEAD(&prev.bdws); prev 3409 drivers/acpi/nfit/core.c INIT_LIST_HEAD(&prev.idts); prev 3410 drivers/acpi/nfit/core.c INIT_LIST_HEAD(&prev.flushes); prev 3412 drivers/acpi/nfit/core.c list_cut_position(&prev.spas, &acpi_desc->spas, prev 3413 drivers/acpi/nfit/core.c acpi_desc->spas.prev); prev 3414 drivers/acpi/nfit/core.c list_cut_position(&prev.memdevs, &acpi_desc->memdevs, prev 3415 drivers/acpi/nfit/core.c acpi_desc->memdevs.prev); prev 3416 drivers/acpi/nfit/core.c list_cut_position(&prev.dcrs, &acpi_desc->dcrs, prev 3417 drivers/acpi/nfit/core.c acpi_desc->dcrs.prev); prev 3418 drivers/acpi/nfit/core.c list_cut_position(&prev.bdws, &acpi_desc->bdws, prev 3419 drivers/acpi/nfit/core.c acpi_desc->bdws.prev); prev 3420 drivers/acpi/nfit/core.c list_cut_position(&prev.idts, &acpi_desc->idts, prev 3421 drivers/acpi/nfit/core.c acpi_desc->idts.prev); prev 3422 drivers/acpi/nfit/core.c list_cut_position(&prev.flushes, &acpi_desc->flushes, prev 3423 drivers/acpi/nfit/core.c acpi_desc->flushes.prev); prev 3427 drivers/acpi/nfit/core.c data = add_table(acpi_desc, &prev, data, end); prev 3435 drivers/acpi/nfit/core.c rc = acpi_nfit_check_deletions(acpi_desc, &prev); prev 1170 drivers/acpi/processor_idle.c struct acpi_lpi_states_array info[2], *tmp, *prev, *curr; prev 1179 drivers/acpi/processor_idle.c prev = &info[0]; prev 1182 drivers/acpi/processor_idle.c ret = acpi_processor_evaluate_lpi(handle, prev); prev 1185 drivers/acpi/processor_idle.c flatten_lpi_states(pr, prev, NULL); prev 1204 drivers/acpi/processor_idle.c flatten_lpi_states(pr, curr, prev); prev 1206 drivers/acpi/processor_idle.c tmp = prev, prev = curr, curr = tmp; prev 1142 drivers/acpi/property.c const struct fwnode_handle *fwnode, struct fwnode_handle *prev) prev 1147 drivers/acpi/property.c if (!prev) { prev 1161 drivers/acpi/property.c port = fwnode_get_parent(prev); prev 1167 drivers/acpi/property.c endpoint = fwnode_get_next_child_node(port, prev); prev 57 drivers/android/binder_alloc.c return list_entry(buffer->entry.prev, struct binder_buffer, entry); prev 549 drivers/android/binder_alloc.c struct binder_buffer *prev, *next = NULL; prev 552 drivers/android/binder_alloc.c prev = binder_buffer_prev(buffer); prev 553 drivers/android/binder_alloc.c BUG_ON(!prev->free); prev 554 drivers/android/binder_alloc.c if (prev_buffer_end_page(prev) == buffer_start_page(buffer)) { prev 559 drivers/android/binder_alloc.c prev->user_data); prev 585 drivers/android/binder_alloc.c prev->user_data, prev 639 drivers/android/binder_alloc.c struct binder_buffer *prev = binder_buffer_prev(buffer); prev 641 drivers/android/binder_alloc.c if (prev->free) { prev 643 drivers/android/binder_alloc.c rb_erase(&prev->rb_node, &alloc->free_buffers); prev 644 drivers/android/binder_alloc.c buffer = prev; prev 253 drivers/android/binder_alloc_selftest.c size_t end, prev; prev 259 drivers/android/binder_alloc_selftest.c prev = index == 0 ? 0 : end_offset[index - 1]; prev 260 drivers/android/binder_alloc_selftest.c end = prev; prev 135 drivers/ata/pata_cs5530.c struct ata_device *prev = ap->private_data; prev 138 drivers/ata/pata_cs5530.c if (ata_dma_enabled(adev) && adev != prev && prev != NULL) { prev 140 drivers/ata/pata_cs5530.c if ((ata_using_udma(adev) && !ata_using_udma(prev)) || prev 141 drivers/ata/pata_cs5530.c (ata_using_udma(prev) && !ata_using_udma(adev))) prev 156 drivers/ata/pata_sc1200.c struct ata_device *prev = ap->private_data; prev 159 drivers/ata/pata_sc1200.c if (ata_dma_enabled(adev) && adev != prev && prev != NULL) { prev 161 drivers/ata/pata_sc1200.c if ((ata_using_udma(adev) && !ata_using_udma(prev)) || prev 162 drivers/ata/pata_sc1200.c (ata_using_udma(prev) && !ata_using_udma(adev))) prev 323 drivers/atm/idt77105.c struct idt77105_priv *walk, *prev; prev 331 drivers/atm/idt77105.c for (prev = NULL, walk = idt77105_all ; prev 333 drivers/atm/idt77105.c prev = walk, walk = walk->next) { prev 335 drivers/atm/idt77105.c if (prev != NULL) prev 336 drivers/atm/idt77105.c prev->next = walk->next; prev 332 drivers/atm/suni.c printk(KERN_DEBUG "[u] p=0x%lx,n=0x%lx\n",(unsigned long) poll_timer.list.prev, prev 3192 drivers/base/core.c dev = list_entry(devices_kset->list.prev, struct device, prev 1225 drivers/base/dd.c dev_prv = list_entry(drv->p->klist_devices.k_list.prev, prev 380 drivers/base/power/domain.c unsigned int prev; prev 397 drivers/base/power/domain.c prev = gpd_data->performance_state; prev 403 drivers/base/power/domain.c gpd_data->performance_state = prev; prev 1168 drivers/base/power/main.c struct device *dev = to_device(dpm_prepared_list.prev); prev 1400 drivers/base/power/main.c struct device *dev = to_device(dpm_late_early_list.prev); prev 1600 drivers/base/power/main.c struct device *dev = to_device(dpm_suspended_list.prev); prev 1872 drivers/base/power/main.c struct device *dev = to_device(dpm_prepared_list.prev); prev 200 drivers/base/power/trace.c entry = dpm_list.prev; prev 208 drivers/base/power/trace.c entry = entry->prev; prev 227 drivers/base/power/trace.c entry = dpm_list.prev; prev 241 drivers/base/power/trace.c entry = entry->prev; prev 900 drivers/base/property.c struct fwnode_handle *prev) prev 902 drivers/base/property.c return fwnode_call_ptr_op(fwnode, graph_get_next_endpoint, prev); prev 939 drivers/block/drbd/drbd_nl.c } prev; prev 963 drivers/block/drbd/drbd_nl.c prev.last_agreed_sect = md->la_size_sect; prev 964 drivers/block/drbd/drbd_nl.c prev.md_offset = md->md_offset; prev 965 drivers/block/drbd/drbd_nl.c prev.al_offset = md->al_offset; prev 966 drivers/block/drbd/drbd_nl.c prev.bm_offset = md->bm_offset; prev 967 drivers/block/drbd/drbd_nl.c prev.md_size_sect = md->md_size_sect; prev 968 drivers/block/drbd/drbd_nl.c prev.al_stripes = md->al_stripes; prev 969 drivers/block/drbd/drbd_nl.c prev.al_stripe_size_4k = md->al_stripe_size_4k; prev 985 drivers/block/drbd/drbd_nl.c if (size < prev.last_agreed_sect) { prev 1023 drivers/block/drbd/drbd_nl.c la_size_changed = (prev.last_agreed_sect != md->la_size_sect); prev 1025 drivers/block/drbd/drbd_nl.c md_moved = prev.md_offset != md->md_offset prev 1026 drivers/block/drbd/drbd_nl.c || prev.md_size_sect != md->md_size_sect; prev 1066 drivers/block/drbd/drbd_nl.c if (size > prev.last_agreed_sect) prev 1067 drivers/block/drbd/drbd_nl.c rv = prev.last_agreed_sect ? DS_GREW : DS_GREW_FROM_ZERO; prev 1068 drivers/block/drbd/drbd_nl.c if (size < prev.last_agreed_sect) prev 1074 drivers/block/drbd/drbd_nl.c md->la_size_sect = prev.last_agreed_sect; prev 1075 drivers/block/drbd/drbd_nl.c md->md_offset = prev.md_offset; prev 1076 drivers/block/drbd/drbd_nl.c md->al_offset = prev.al_offset; prev 1077 drivers/block/drbd/drbd_nl.c md->bm_offset = prev.bm_offset; prev 1078 drivers/block/drbd/drbd_nl.c md->md_size_sect = prev.md_size_sect; prev 1079 drivers/block/drbd/drbd_nl.c md->al_stripes = prev.al_stripes; prev 1080 drivers/block/drbd/drbd_nl.c md->al_stripe_size_4k = prev.al_stripe_size_4k; prev 1081 drivers/block/drbd/drbd_nl.c md->al_size_4k = (u64)prev.al_stripes * prev.al_stripe_size_4k; prev 68 drivers/char/agp/frontend.c struct agp_memory *prev; prev 76 drivers/char/agp/frontend.c prev = temp->prev; prev 78 drivers/char/agp/frontend.c if (prev != NULL) { prev 79 drivers/char/agp/frontend.c prev->next = next; prev 81 drivers/char/agp/frontend.c next->prev = prev; prev 86 drivers/char/agp/frontend.c next->prev = NULL; prev 201 drivers/char/agp/frontend.c struct agp_memory *prev; prev 203 drivers/char/agp/frontend.c prev = agp_fe.current_controller->pool; prev 205 drivers/char/agp/frontend.c if (prev != NULL) { prev 206 drivers/char/agp/frontend.c prev->prev = temp; prev 207 drivers/char/agp/frontend.c temp->next = prev; prev 232 drivers/char/agp/frontend.c struct agp_file_private *prev; prev 234 drivers/char/agp/frontend.c prev = agp_fe.file_priv_list; prev 236 drivers/char/agp/frontend.c if (prev != NULL) prev 237 drivers/char/agp/frontend.c prev->prev = priv; prev 238 drivers/char/agp/frontend.c priv->next = prev; prev 245 drivers/char/agp/frontend.c struct agp_file_private *prev; prev 248 drivers/char/agp/frontend.c prev = priv->prev; prev 250 drivers/char/agp/frontend.c if (prev != NULL) { prev 251 drivers/char/agp/frontend.c prev->next = next; prev 254 drivers/char/agp/frontend.c next->prev = prev; prev 258 drivers/char/agp/frontend.c next->prev = NULL; prev 328 drivers/char/agp/frontend.c prev_controller->prev = controller; prev 377 drivers/char/agp/frontend.c prev_controller = controller->prev; prev 383 drivers/char/agp/frontend.c next_controller->prev = prev_controller; prev 387 drivers/char/agp/frontend.c next_controller->prev = NULL; prev 507 drivers/char/agp/frontend.c prev_client->prev = client; prev 541 drivers/char/agp/frontend.c prev_client = client->prev; prev 547 drivers/char/agp/frontend.c next_client->prev = prev_client; prev 551 drivers/char/agp/frontend.c next_client->prev = NULL; prev 402 drivers/char/virtio_console.c pending_free_dma_bufs.prev); prev 165 drivers/clk/clk-max9485.c const struct max9485_rate *curr, *prev = NULL; prev 183 drivers/clk/clk-max9485.c if (!prev) prev 190 drivers/clk/clk-max9485.c mid = prev->out + ((curr->out - prev->out) / 2); prev 192 drivers/clk/clk-max9485.c return (mid > rate) ? prev->out : curr->out; prev 195 drivers/clk/clk-max9485.c prev = curr; prev 199 drivers/clk/clk-max9485.c return prev->out; prev 508 drivers/clk/keystone/sci-clk.c struct sci_clk *sci_clk, *prev; prev 607 drivers/clk/keystone/sci-clk.c prev = NULL; prev 610 drivers/clk/keystone/sci-clk.c if (prev && prev->dev_id == sci_clk->dev_id && prev 611 drivers/clk/keystone/sci-clk.c prev->clk_id == sci_clk->clk_id) prev 615 drivers/clk/keystone/sci-clk.c prev = sci_clk; prev 296 drivers/cpufreq/freq_table.c struct cpufreq_frequency_table *prev = NULL; prev 302 drivers/cpufreq/freq_table.c if (!prev) { prev 303 drivers/cpufreq/freq_table.c prev = pos; prev 307 drivers/cpufreq/freq_table.c if (pos->frequency == prev->frequency) { prev 314 drivers/cpufreq/freq_table.c if (pos->frequency > prev->frequency) { prev 334 drivers/cpufreq/freq_table.c prev = pos; prev 127 drivers/cpufreq/qcom-cpufreq-hw.c struct cpufreq_frequency_table *prev = &table[i - 1]; prev 133 drivers/cpufreq/qcom-cpufreq-hw.c if (prev->frequency == CPUFREQ_ENTRY_INVALID) { prev 134 drivers/cpufreq/qcom-cpufreq-hw.c prev->frequency = prev_freq; prev 135 drivers/cpufreq/qcom-cpufreq-hw.c prev->flags = CPUFREQ_BOOST_FREQ; prev 993 drivers/crypto/chelsio/chcr_algo.c u32 c, prev; prev 997 drivers/crypto/chelsio/chcr_algo.c prev = be32_to_cpu(*--b); prev 998 drivers/crypto/chelsio/chcr_algo.c c = prev + add; prev 1000 drivers/crypto/chelsio/chcr_algo.c if (prev < c) prev 227 drivers/crypto/chelsio/chtls/chtls.h struct sk_buff *prev; prev 541 drivers/crypto/chelsio/chtls/chtls_cm.c struct listen_info *p, **prev; prev 546 drivers/crypto/chelsio/chtls/chtls_cm.c prev = &cdev->listen_hash_tab[key]; prev 549 drivers/crypto/chelsio/chtls/chtls_cm.c for (p = *prev; p; prev = &p->next, p = p->next) prev 552 drivers/crypto/chelsio/chtls/chtls_cm.c *prev = p->next; prev 23 drivers/dma-buf/dma-fence-chain.c struct dma_fence *prev; prev 26 drivers/dma-buf/dma-fence-chain.c prev = dma_fence_get_rcu_safe(&chain->prev); prev 28 drivers/dma-buf/dma-fence-chain.c return prev; prev 42 drivers/dma-buf/dma-fence-chain.c struct dma_fence *prev, *replacement, *tmp; prev 50 drivers/dma-buf/dma-fence-chain.c while ((prev = dma_fence_chain_get_prev(chain))) { prev 52 drivers/dma-buf/dma-fence-chain.c prev_chain = to_dma_fence_chain(prev); prev 59 drivers/dma-buf/dma-fence-chain.c if (!dma_fence_is_signaled(prev)) prev 65 drivers/dma-buf/dma-fence-chain.c tmp = cmpxchg((void **)&chain->prev, (void *)prev, (void *)replacement); prev 66 drivers/dma-buf/dma-fence-chain.c if (tmp == prev) prev 70 drivers/dma-buf/dma-fence-chain.c dma_fence_put(prev); prev 74 drivers/dma-buf/dma-fence-chain.c return prev; prev 181 drivers/dma-buf/dma-fence-chain.c struct dma_fence *prev; prev 186 drivers/dma-buf/dma-fence-chain.c while ((prev = rcu_dereference_protected(chain->prev, true))) { prev 189 drivers/dma-buf/dma-fence-chain.c if (kref_read(&prev->refcount) > 1) prev 192 drivers/dma-buf/dma-fence-chain.c prev_chain = to_dma_fence_chain(prev); prev 199 drivers/dma-buf/dma-fence-chain.c chain->prev = prev_chain->prev; prev 200 drivers/dma-buf/dma-fence-chain.c RCU_INIT_POINTER(prev_chain->prev, NULL); prev 201 drivers/dma-buf/dma-fence-chain.c dma_fence_put(prev); prev 203 drivers/dma-buf/dma-fence-chain.c dma_fence_put(prev); prev 229 drivers/dma-buf/dma-fence-chain.c struct dma_fence *prev, prev 233 drivers/dma-buf/dma-fence-chain.c struct dma_fence_chain *prev_chain = to_dma_fence_chain(prev); prev 237 drivers/dma-buf/dma-fence-chain.c rcu_assign_pointer(chain->prev, prev); prev 243 drivers/dma-buf/dma-fence-chain.c if (prev_chain && __dma_fence_is_later(seqno, prev->seqno, prev->ops)) { prev 244 drivers/dma-buf/dma-fence-chain.c context = prev->context; prev 245 drivers/dma-buf/dma-fence-chain.c chain->prev_seqno = prev->seqno; prev 250 drivers/dma-buf/dma-fence-chain.c seqno = max(prev->seqno, seqno); prev 196 drivers/dma/at_hdmac.c static void atc_desc_chain(struct at_desc **first, struct at_desc **prev, prev 203 drivers/dma/at_hdmac.c (*prev)->lli.dscr = desc->txd.phys; prev 208 drivers/dma/at_hdmac.c *prev = desc; prev 555 drivers/dma/at_hdmac.c list_splice_init(&atchan->queue, atchan->active_list.prev); prev 807 drivers/dma/at_hdmac.c struct at_desc *prev = NULL; prev 853 drivers/dma/at_hdmac.c atc_desc_chain(&first, &prev, desc); prev 985 drivers/dma/at_hdmac.c struct at_desc *desc = NULL, *first = NULL, *prev = NULL; prev 1026 drivers/dma/at_hdmac.c atc_desc_chain(&first, &prev, desc); prev 1072 drivers/dma/at_hdmac.c struct at_desc *prev = NULL; prev 1133 drivers/dma/at_hdmac.c atc_desc_chain(&first, &prev, desc); prev 1174 drivers/dma/at_hdmac.c atc_desc_chain(&first, &prev, desc); prev 1183 drivers/dma/at_hdmac.c set_desc_eol(prev); prev 1292 drivers/dma/at_hdmac.c struct at_desc *prev = NULL; prev 1338 drivers/dma/at_hdmac.c atc_desc_chain(&first, &prev, desc); prev 1342 drivers/dma/at_hdmac.c prev->lli.dscr = first->txd.phys; prev 485 drivers/dma/at_xdmac.c struct at_xdmac_desc *prev, prev 488 drivers/dma/at_xdmac.c if (!prev || !desc) prev 491 drivers/dma/at_xdmac.c prev->lld.mbr_nda = desc->tx_dma_desc.phys; prev 492 drivers/dma/at_xdmac.c prev->lld.mbr_ubc |= AT_XDMAC_MBR_UBC_NDE; prev 495 drivers/dma/at_xdmac.c __func__, prev, &prev->lld.mbr_nda); prev 638 drivers/dma/at_xdmac.c struct at_xdmac_desc *first = NULL, *prev = NULL; prev 709 drivers/dma/at_xdmac.c if (prev) prev 710 drivers/dma/at_xdmac.c at_xdmac_queue_desc(chan, prev, desc); prev 712 drivers/dma/at_xdmac.c prev = desc; prev 740 drivers/dma/at_xdmac.c struct at_xdmac_desc *first = NULL, *prev = NULL; prev 797 drivers/dma/at_xdmac.c if (prev) prev 798 drivers/dma/at_xdmac.c at_xdmac_queue_desc(chan, prev, desc); prev 800 drivers/dma/at_xdmac.c prev = desc; prev 809 drivers/dma/at_xdmac.c at_xdmac_queue_desc(chan, prev, first); prev 849 drivers/dma/at_xdmac.c struct at_xdmac_desc *prev, prev 885 drivers/dma/at_xdmac.c if (prev) prev 887 drivers/dma/at_xdmac.c "Adding items at the end of desc 0x%p\n", prev); prev 932 drivers/dma/at_xdmac.c if (prev) prev 933 drivers/dma/at_xdmac.c at_xdmac_queue_desc(chan, prev, desc); prev 944 drivers/dma/at_xdmac.c struct at_xdmac_desc *prev = NULL, *first = NULL; prev 998 drivers/dma/at_xdmac.c prev, prev 1021 drivers/dma/at_xdmac.c prev = desc; prev 1037 drivers/dma/at_xdmac.c struct at_xdmac_desc *first = NULL, *prev = NULL; prev 1118 drivers/dma/at_xdmac.c if (prev) prev 1119 drivers/dma/at_xdmac.c at_xdmac_queue_desc(chan, prev, desc); prev 1121 drivers/dma/at_xdmac.c prev = desc; prev 428 drivers/dma/dw-axi-dmac/dw-axi-dmac-platform.c struct axi_dma_desc *first = NULL, *desc = NULL, *prev = NULL; prev 495 drivers/dma/dw-axi-dmac/dw-axi-dmac-platform.c write_desc_llp(prev, desc->vd.tx.phys | lms); prev 497 drivers/dma/dw-axi-dmac/dw-axi-dmac-platform.c prev = desc; prev 336 drivers/dma/dw/core.c desc->residue -= to_dw_desc(active->prev)->len; prev 445 drivers/dma/dw/core.c list_move(dwc->queue.next, dwc->active_list.prev); prev 552 drivers/dma/dw/core.c struct dw_desc *prev; prev 581 drivers/dma/dw/core.c prev = first = NULL; prev 599 drivers/dma/dw/core.c lli_write(prev, llp, desc->txd.phys | lms); prev 602 drivers/dma/dw/core.c prev = desc; prev 607 drivers/dma/dw/core.c lli_set(prev, ctllo, DWC_CTLL_INT_EN); prev 609 drivers/dma/dw/core.c prev->lli.llp = 0; prev 610 drivers/dma/dw/core.c lli_clear(prev, ctllo, DWC_CTLL_LLP_D_EN | DWC_CTLL_LLP_S_EN); prev 629 drivers/dma/dw/core.c struct dw_desc *prev; prev 649 drivers/dma/dw/core.c prev = first = NULL; prev 689 drivers/dma/dw/core.c lli_write(prev, llp, desc->txd.phys | lms); prev 692 drivers/dma/dw/core.c prev = desc; prev 738 drivers/dma/dw/core.c lli_write(prev, llp, desc->txd.phys | lms); prev 741 drivers/dma/dw/core.c prev = desc; prev 757 drivers/dma/dw/core.c lli_set(prev, ctllo, DWC_CTLL_INT_EN); prev 759 drivers/dma/dw/core.c prev->lli.llp = 0; prev 760 drivers/dma/dw/core.c lli_clear(prev, ctllo, DWC_CTLL_LLP_D_EN | DWC_CTLL_LLP_S_EN); prev 384 drivers/dma/fsldma.c struct fsl_desc_sw *tail = to_fsl_desc(chan->ld_pending.prev); prev 764 drivers/dma/fsldma.c struct fsl_desc_sw *first = NULL, *prev = NULL, *new; prev 793 drivers/dma/fsldma.c set_desc_next(chan, &prev->hw, new->async_tx.phys); prev 798 drivers/dma/fsldma.c prev = new; prev 292 drivers/dma/img-mdc-dma.c struct mdc_hw_list_desc *curr, *prev = NULL; prev 311 drivers/dma/img-mdc-dma.c if (prev) { prev 312 drivers/dma/img-mdc-dma.c prev->node_addr = curr_phys; prev 313 drivers/dma/img-mdc-dma.c prev->next_desc = curr; prev 324 drivers/dma/img-mdc-dma.c prev = curr; prev 374 drivers/dma/img-mdc-dma.c struct mdc_hw_list_desc *curr, *prev = NULL; prev 406 drivers/dma/img-mdc-dma.c if (!prev) { prev 410 drivers/dma/img-mdc-dma.c prev->node_addr = curr_phys; prev 411 drivers/dma/img-mdc-dma.c prev->next_desc = curr; prev 429 drivers/dma/img-mdc-dma.c prev = curr; prev 437 drivers/dma/img-mdc-dma.c prev->node_addr = mdesc->list_phys; prev 456 drivers/dma/img-mdc-dma.c struct mdc_hw_list_desc *curr, *prev = NULL; prev 486 drivers/dma/img-mdc-dma.c if (!prev) { prev 490 drivers/dma/img-mdc-dma.c prev->node_addr = curr_phys; prev 491 drivers/dma/img-mdc-dma.c prev->next_desc = curr; prev 507 drivers/dma/img-mdc-dma.c prev = curr; prev 376 drivers/dma/iop-adma.c old_chain_tail = list_entry(iop_chan->chain.prev, prev 418 drivers/dma/mediatek/mtk-hsdma.c u16 reserved, prev, tlen, num_sgs; prev 477 drivers/dma/mediatek/mtk-hsdma.c prev = MTK_HSDMA_LAST_DESP_IDX(ring->cur_tptr, MTK_DMA_SIZE); prev 478 drivers/dma/mediatek/mtk-hsdma.c ring->cb[prev].flag = MTK_HSDMA_VDESC_FINISHED; prev 450 drivers/dma/mmp_pdma.c struct mmp_pdma_desc_sw *first = NULL, *prev = NULL, *new; prev 487 drivers/dma/mmp_pdma.c prev->desc.ddadr = new->async_tx.phys; prev 492 drivers/dma/mmp_pdma.c prev = new; prev 531 drivers/dma/mmp_pdma.c struct mmp_pdma_desc_sw *first = NULL, *prev = NULL, *new = NULL; prev 572 drivers/dma/mmp_pdma.c prev->desc.ddadr = new->async_tx.phys; prev 576 drivers/dma/mmp_pdma.c prev = new; prev 612 drivers/dma/mmp_pdma.c struct mmp_pdma_desc_sw *first = NULL, *prev = NULL, *new; prev 660 drivers/dma/mmp_pdma.c prev->desc.ddadr = new->async_tx.phys; prev 665 drivers/dma/mmp_pdma.c prev = new; prev 258 drivers/dma/mpc512x_dma.c struct mpc_dma_desc *prev = NULL; prev 285 drivers/dma/mpc512x_dma.c if (!prev) { prev 286 drivers/dma/mpc512x_dma.c prev = mdesc; prev 290 drivers/dma/mpc512x_dma.c prev->tcd->dlast_sga = mdesc->tcd_paddr; prev 291 drivers/dma/mpc512x_dma.c prev->tcd->e_sg = 1; prev 294 drivers/dma/mpc512x_dma.c prev = mdesc; prev 297 drivers/dma/mpc512x_dma.c prev->tcd->int_maj = 1; prev 302 drivers/dma/mpc512x_dma.c if (first != prev) prev 402 drivers/dma/mv_xor.c old_chain_tail = list_entry(mv_chan->chain.prev, prev 779 drivers/dma/nbpfaxi.c struct nbpf_link_desc *ldesc, *prev = NULL; prev 816 drivers/dma/nbpfaxi.c if (prev) prev 817 drivers/dma/nbpfaxi.c prev->hwdesc->next = (u32)ldesc->hwdesc_dma_addr; prev 819 drivers/dma/nbpfaxi.c prev = ldesc; prev 826 drivers/dma/nbpfaxi.c prev->hwdesc->next = 0; prev 344 drivers/dma/owl-dma.c struct owl_dma_lli *prev, prev 351 drivers/dma/owl-dma.c if (prev) { prev 352 drivers/dma/owl-dma.c prev->hw.next_lli = next->phys; prev 353 drivers/dma/owl-dma.c prev->hw.ctrla |= llc_hw_ctrla(OWL_DMA_MODE_LME, 0); prev 833 drivers/dma/owl-dma.c struct owl_dma_lli *lli, *prev = NULL; prev 864 drivers/dma/owl-dma.c prev = owl_dma_add_lli(txd, prev, lli, false); prev 885 drivers/dma/owl-dma.c struct owl_dma_lli *lli, *prev = NULL; prev 928 drivers/dma/owl-dma.c prev = owl_dma_add_lli(txd, prev, lli, false); prev 950 drivers/dma/owl-dma.c struct owl_dma_lli *lli, *prev = NULL, *first = NULL; prev 987 drivers/dma/owl-dma.c prev = owl_dma_add_lli(txd, prev, lli, false); prev 991 drivers/dma/owl-dma.c owl_dma_add_lli(txd, prev, first, true); prev 385 drivers/dma/pch_dma.c list_splice_init(&pd_chan->queue, pd_chan->active_list.prev); prev 574 drivers/dma/pch_dma.c struct pch_dma_desc *prev = NULL; prev 629 drivers/dma/pch_dma.c prev->regs.next |= desc->txd.phys; prev 633 drivers/dma/pch_dma.c prev = desc; prev 1921 drivers/dma/ppc4xx/adma.c old_chain_tail = list_entry(chan->chain.prev, prev 574 drivers/dma/pxa_dma.c vd_last_issued = list_entry(vc->desc_issued.prev, prev 802 drivers/dma/pxa_dma.c vd_chained = list_entry(vc->desc_submitted.prev, prev 86 drivers/dma/sh/shdma-base.c list_for_each_entry_safe(chunk, c, desc->node.prev, node) { prev 365 drivers/dma/sun6i-dma.c static void *sun6i_dma_lli_add(struct sun6i_dma_lli *prev, prev 370 drivers/dma/sun6i-dma.c if ((!prev && !txd) || !next) prev 373 drivers/dma/sun6i-dma.c if (!prev) { prev 377 drivers/dma/sun6i-dma.c prev->p_lli_next = next_phy; prev 378 drivers/dma/sun6i-dma.c prev->v_lli_next = next; prev 690 drivers/dma/sun6i-dma.c struct sun6i_dma_lli *v_lli, *prev = NULL; prev 745 drivers/dma/sun6i-dma.c prev = sun6i_dma_lli_add(prev, v_lli, p_lli, txd); prev 749 drivers/dma/sun6i-dma.c for (prev = txd->v_lli; prev; prev = prev->v_lli_next) prev 750 drivers/dma/sun6i-dma.c sun6i_dma_dump_lli(vchan, prev); prev 755 drivers/dma/sun6i-dma.c for (prev = txd->v_lli; prev; prev = prev->v_lli_next) prev 756 drivers/dma/sun6i-dma.c dma_pool_free(sdev->pool, prev, virt_to_phys(prev)); prev 772 drivers/dma/sun6i-dma.c struct sun6i_dma_lli *v_lli, *prev = NULL; prev 813 drivers/dma/sun6i-dma.c prev = sun6i_dma_lli_add(prev, v_lli, p_lli, txd); prev 816 drivers/dma/sun6i-dma.c prev->p_lli_next = txd->p_lli; /* cyclic list */ prev 823 drivers/dma/sun6i-dma.c for (prev = txd->v_lli; prev; prev = prev->v_lli_next) prev 824 drivers/dma/sun6i-dma.c dma_pool_free(sdev->pool, prev, virt_to_phys(prev)); prev 171 drivers/dma/txx9dmac.c return list_entry(dc->active_list.prev, prev 183 drivers/dma/txx9dmac.c desc = list_entry(desc->tx_list.prev, typeof(*desc), desc_node); prev 428 drivers/dma/txx9dmac.c struct txx9dmac_desc *prev = NULL; prev 433 drivers/dma/txx9dmac.c if (prev) { prev 434 drivers/dma/txx9dmac.c desc_write_CHAR(dc, prev, desc->txd.phys); prev 436 drivers/dma/txx9dmac.c prev->txd.phys, ddev->descsize, prev 439 drivers/dma/txx9dmac.c prev = txx9dmac_last_child(desc); prev 716 drivers/dma/txx9dmac.c struct txx9dmac_desc *prev; prev 728 drivers/dma/txx9dmac.c prev = first = NULL; prev 778 drivers/dma/txx9dmac.c desc_write_CHAR(dc, prev, desc->txd.phys); prev 780 drivers/dma/txx9dmac.c prev->txd.phys, ddev->descsize, prev 784 drivers/dma/txx9dmac.c prev = desc; prev 789 drivers/dma/txx9dmac.c txx9dmac_desc_set_INTENT(ddev, prev); prev 791 drivers/dma/txx9dmac.c desc_write_CHAR(dc, prev, 0); prev 793 drivers/dma/txx9dmac.c prev->txd.phys, ddev->descsize, prev 810 drivers/dma/txx9dmac.c struct txx9dmac_desc *prev; prev 825 drivers/dma/txx9dmac.c prev = first = NULL; prev 872 drivers/dma/txx9dmac.c desc_write_CHAR(dc, prev, desc->txd.phys); prev 874 drivers/dma/txx9dmac.c prev->txd.phys, prev 879 drivers/dma/txx9dmac.c prev = desc; prev 884 drivers/dma/txx9dmac.c txx9dmac_desc_set_INTENT(ddev, prev); prev 886 drivers/dma/txx9dmac.c desc_write_CHAR(dc, prev, 0); prev 888 drivers/dma/txx9dmac.c prev->txd.phys, ddev->descsize, prev 940 drivers/dma/txx9dmac.c struct txx9dmac_desc *prev) prev 946 drivers/dma/txx9dmac.c prev = txx9dmac_last_child(prev); prev 949 drivers/dma/txx9dmac.c desc_write_CHAR(dc, prev, desc->txd.phys); prev 951 drivers/dma/txx9dmac.c prev->txd.phys, ddev->descsize, prev 954 drivers/dma/txx9dmac.c channel_read_CHAR(dc) == prev->txd.phys) prev 973 drivers/dma/txx9dmac.c struct txx9dmac_desc *prev = txx9dmac_last_active(dc); prev 975 drivers/dma/txx9dmac.c if (!(prev->txd.flags & DMA_PREP_INTERRUPT) || prev 977 drivers/dma/txx9dmac.c txx9dmac_chain_dynamic(dc, prev); prev 1881 drivers/dma/xilinx/xilinx_dma.c struct xilinx_axidma_tx_segment *segment, *head_segment, *prev = NULL; prev 1929 drivers/dma/xilinx/xilinx_dma.c if (prev) prev 1930 drivers/dma/xilinx/xilinx_dma.c prev->hw.next_desc = segment->phys; prev 1932 drivers/dma/xilinx/xilinx_dma.c prev = segment; prev 305 drivers/dma/xilinx/zynqmp_dma.c struct zynqmp_dma_desc_ll *prev) prev 319 drivers/dma/xilinx/zynqmp_dma.c if (prev) { prev 322 drivers/dma/xilinx/zynqmp_dma.c ddesc = prev + 1; prev 323 drivers/dma/xilinx/zynqmp_dma.c prev->nxtdscraddr = addr; prev 806 drivers/dma/xilinx/zynqmp_dma.c void *desc = NULL, *prev = NULL; prev 831 drivers/dma/xilinx/zynqmp_dma.c dma_dst, copy, prev); prev 832 drivers/dma/xilinx/zynqmp_dma.c prev = desc; prev 1305 drivers/edac/i7core_edac.c static int i7core_get_onedevice(struct pci_dev **prev, prev 1318 drivers/edac/i7core_edac.c dev_descr->dev_id, *prev); prev 1326 drivers/edac/i7core_edac.c pci_dev_get(*prev); /* pci_get_device will put it */ prev 1328 drivers/edac/i7core_edac.c PCI_DEVICE_ID_INTEL_I7_NONCORE_ALT, *prev); prev 1333 drivers/edac/i7core_edac.c pci_dev_get(*prev); /* pci_get_device will put it */ prev 1336 drivers/edac/i7core_edac.c *prev); prev 1340 drivers/edac/i7core_edac.c if (*prev) { prev 1341 drivers/edac/i7core_edac.c *prev = pdev; prev 1418 drivers/edac/i7core_edac.c *prev = pdev; prev 734 drivers/edac/sb_edac.c struct sbridge_dev *prev) prev 747 drivers/edac/sb_edac.c sbridge_dev = list_entry(prev ? prev->list.next prev 2329 drivers/edac/sb_edac.c static int sbridge_get_onedevice(struct pci_dev **prev, prev 2347 drivers/edac/sb_edac.c dev_descr->dev_id, *prev); prev 2350 drivers/edac/sb_edac.c if (*prev) { prev 2351 drivers/edac/sb_edac.c *prev = pdev; prev 2431 drivers/edac/sb_edac.c *prev = pdev; prev 77 drivers/edac/skx_base.c struct pci_dev *pdev, *prev; prev 82 drivers/edac/skx_base.c prev = NULL; prev 84 drivers/edac/skx_base.c pdev = pci_get_device(PCI_VENDOR_ID_INTEL, m->did, prev); prev 139 drivers/edac/skx_base.c prev = pdev; prev 186 drivers/edac/skx_common.c struct pci_dev *pdev, *prev; prev 191 drivers/edac/skx_common.c prev = NULL; prev 193 drivers/edac/skx_common.c pdev = pci_get_device(PCI_VENDOR_ID_INTEL, did, prev); prev 223 drivers/edac/skx_common.c prev = pdev; prev 215 drivers/firewire/core-topology.c h = h->prev; prev 279 drivers/firewire/core-topology.c __list_del(h->prev, &stack); prev 325 drivers/firewire/net.c fi2 = list_entry(fi->fi_link.prev, prev 345 drivers/firewire/net.c list = fi->fi_link.prev; prev 148 drivers/firewire/ohci.c struct descriptor *prev; prev 1161 drivers/firewire/ohci.c ctx->prev = ctx->buffer_tail->buffer; prev 1235 drivers/firewire/ohci.c d_branch = find_branch_descriptor(ctx->prev, ctx->prev_z); prev 1248 drivers/firewire/ohci.c d_branch != ctx->prev && prev 1249 drivers/firewire/ohci.c (ctx->prev->control & cpu_to_le16(DESCRIPTOR_CMD)) == prev 1251 drivers/firewire/ohci.c ctx->prev->branch_address = cpu_to_le32(d_bus | z); prev 1254 drivers/firewire/ohci.c ctx->prev = d; prev 341 drivers/firmware/efi/libstub/arm-stub.c efi_memory_desc_t *in, *prev = NULL, *out = runtime_map; prev 357 drivers/firmware/efi/libstub/arm-stub.c for (l = 0; l < map_size; l += desc_size, prev = in) { prev 378 drivers/firmware/efi/libstub/arm-stub.c !regions_are_adjacent(prev, in)) || prev 379 drivers/firmware/efi/libstub/arm-stub.c !regions_have_compatible_memory_type_attrs(prev, in)) { prev 1087 drivers/firmware/efi/vars.c struct efivar_entry **prev) prev 1092 drivers/firmware/efi/vars.c if (!prev || !*prev) { prev 1099 drivers/firmware/efi/vars.c if (prev) prev 1100 drivers/firmware/efi/vars.c *prev = entry; prev 1106 drivers/firmware/efi/vars.c list_for_each_entry_safe_continue((*prev), n, head, list) { prev 1107 drivers/firmware/efi/vars.c err = func(*prev, data); prev 257 drivers/gpio/gpiolib.c struct gpio_device *prev, *next; prev 272 drivers/gpio/gpiolib.c prev = list_entry(gpio_devices.prev, struct gpio_device, list); prev 273 drivers/gpio/gpiolib.c if (prev->base + prev->ngpio <= gdev->base) { prev 279 drivers/gpio/gpiolib.c list_for_each_entry_safe(prev, next, &gpio_devices, list) { prev 285 drivers/gpio/gpiolib.c if (prev->base + prev->ngpio <= gdev->base prev 287 drivers/gpio/gpiolib.c list_add(&gdev->list, &prev->list); prev 97 drivers/gpu/drm/amd/amdgpu/amdgpu_pmu.c u64 count, prev; prev 100 drivers/gpu/drm/amd/amdgpu/amdgpu_pmu.c prev = local64_read(&hwc->prev_count); prev 111 drivers/gpu/drm/amd/amdgpu/amdgpu_pmu.c } while (local64_cmpxchg(&hwc->prev_count, prev, count) != prev); prev 113 drivers/gpu/drm/amd/amdgpu/amdgpu_pmu.c local64_add(count - prev, &event->count); prev 106 drivers/gpu/drm/amd/amdgpu/amdgpu_sa.c sa_manager->hole = sa_bo->olist.prev; prev 265 drivers/gpu/drm/amd/amdgpu/amdgpu_sa.c sa_manager->hole = best_bo->olist.prev; prev 42 drivers/gpu/drm/amd/powerplay/inc/power_state.h struct pp_power_state *prev; prev 62 drivers/gpu/drm/drm_lock.c unsigned int old, new, prev; prev 75 drivers/gpu/drm/drm_lock.c prev = cmpxchg(lock, old, new); prev 76 drivers/gpu/drm/drm_lock.c } while (prev != old); prev 111 drivers/gpu/drm/drm_lock.c unsigned int old, new, prev; prev 118 drivers/gpu/drm/drm_lock.c prev = cmpxchg(lock, old, new); prev 119 drivers/gpu/drm/drm_lock.c } while (prev != old); prev 126 drivers/gpu/drm/drm_lock.c unsigned int old, new, prev; prev 141 drivers/gpu/drm/drm_lock.c prev = cmpxchg(lock, old, new); prev 142 drivers/gpu/drm/drm_lock.c } while (prev != old); prev 312 drivers/gpu/drm/drm_lock.c unsigned int old, prev; prev 320 drivers/gpu/drm/drm_lock.c prev = cmpxchg(lock, old, DRM_KERNEL_CONTEXT); prev 321 drivers/gpu/drm/drm_lock.c } while (prev != old); prev 229 drivers/gpu/drm/drm_syncobj.c struct dma_fence *prev; prev 235 drivers/gpu/drm/drm_syncobj.c prev = drm_syncobj_fence_get(syncobj); prev 237 drivers/gpu/drm/drm_syncobj.c if (prev && prev->seqno >= point) prev 239 drivers/gpu/drm/drm_syncobj.c dma_fence_chain_init(chain, prev, fence, point); prev 247 drivers/gpu/drm/drm_syncobj.c dma_fence_chain_for_each(fence, prev); prev 248 drivers/gpu/drm/drm_syncobj.c dma_fence_put(prev); prev 365 drivers/gpu/drm/exynos/exynos_drm_g2d.c lnode = list_entry(file_priv->inuse_cmdlist.prev, prev 254 drivers/gpu/drm/i915/gt/intel_engine.h u32 next, u32 prev) prev 257 drivers/gpu/drm/i915/gt/intel_engine.h typecheck(typeof(ring->size), prev); prev 258 drivers/gpu/drm/i915/gt/intel_engine.h return (next - prev) << ring->wrap; prev 1170 drivers/gpu/drm/i915/gt/intel_engine_cs.c const void *prev = NULL; prev 1177 drivers/gpu/drm/i915/gt/intel_engine_cs.c if (prev && !memcmp(prev, buf + pos, rowsize)) { prev 1191 drivers/gpu/drm/i915/gt/intel_engine_cs.c prev = buf + pos; prev 198 drivers/gpu/drm/i915/gt/intel_engine_user.c struct rb_node **p, *prev; prev 203 drivers/gpu/drm/i915/gt/intel_engine_user.c prev = NULL; prev 224 drivers/gpu/drm/i915/gt/intel_engine_user.c rb_link_node(&engine->uabi_node, prev, p); prev 234 drivers/gpu/drm/i915/gt/intel_engine_user.c prev = &engine->uabi_node; prev 235 drivers/gpu/drm/i915/gt/intel_engine_user.c p = &prev->rb_right; prev 387 drivers/gpu/drm/i915/gt/intel_lrc.c assert_priority_queue(const struct i915_request *prev, prev 397 drivers/gpu/drm/i915/gt/intel_lrc.c if (i915_request_is_active(prev)) prev 400 drivers/gpu/drm/i915/gt/intel_lrc.c return rq_prio(prev) >= rq_prio(next); prev 650 drivers/gpu/drm/i915/gt/intel_lrc.c u32 tail, prev; prev 670 drivers/gpu/drm/i915/gt/intel_lrc.c prev = ce->lrc_reg_state[CTX_RING_TAIL + 1]; prev 671 drivers/gpu/drm/i915/gt/intel_lrc.c if (unlikely(intel_ring_direction(rq->ring, tail, prev) <= 0)) prev 803 drivers/gpu/drm/i915/gt/intel_lrc.c static bool can_merge_ctx(const struct intel_context *prev, prev 806 drivers/gpu/drm/i915/gt/intel_lrc.c if (prev != next) prev 809 drivers/gpu/drm/i915/gt/intel_lrc.c if (ctx_single_port_submission(prev)) prev 815 drivers/gpu/drm/i915/gt/intel_lrc.c static bool can_merge_rq(const struct i915_request *prev, prev 818 drivers/gpu/drm/i915/gt/intel_lrc.c GEM_BUG_ON(prev == next); prev 819 drivers/gpu/drm/i915/gt/intel_lrc.c GEM_BUG_ON(!assert_priority_queue(prev, next)); prev 832 drivers/gpu/drm/i915/gt/intel_lrc.c if (!can_merge_ctx(prev->hw_context, next->hw_context)) prev 15 drivers/gpu/drm/i915/gt/selftest_engine_cs.c u8 prev = U8_MAX; prev 21 drivers/gpu/drm/i915/gt/selftest_engine_cs.c if (gen >= prev) { prev 26 drivers/gpu/drm/i915/gt/selftest_engine_cs.c prev, gen); prev 42 drivers/gpu/drm/i915/gt/selftest_engine_cs.c prev = gen; prev 49 drivers/gpu/drm/i915/gt/selftest_engine_cs.c prev); prev 1412 drivers/gpu/drm/i915/gt/selftest_hangcheck.c struct i915_request *prev; prev 1419 drivers/gpu/drm/i915/gt/selftest_hangcheck.c prev = hang_create_request(&h, engine); prev 1420 drivers/gpu/drm/i915/gt/selftest_hangcheck.c if (IS_ERR(prev)) { prev 1421 drivers/gpu/drm/i915/gt/selftest_hangcheck.c err = PTR_ERR(prev); prev 1425 drivers/gpu/drm/i915/gt/selftest_hangcheck.c i915_request_get(prev); prev 1426 drivers/gpu/drm/i915/gt/selftest_hangcheck.c i915_request_add(prev); prev 1457 drivers/gpu/drm/i915/gt/selftest_hangcheck.c i915_request_put(prev); prev 1464 drivers/gpu/drm/i915/gt/selftest_hangcheck.c if (!wait_until_running(&h, prev)) { prev 1469 drivers/gpu/drm/i915/gt/selftest_hangcheck.c prev->fence.seqno, hws_seqno(&h, prev)); prev 1474 drivers/gpu/drm/i915/gt/selftest_hangcheck.c i915_request_put(prev); prev 1484 drivers/gpu/drm/i915/gt/selftest_hangcheck.c if (prev->fence.error != -EIO) { prev 1486 drivers/gpu/drm/i915/gt/selftest_hangcheck.c prev->fence.error); prev 1488 drivers/gpu/drm/i915/gt/selftest_hangcheck.c i915_request_put(prev); prev 1497 drivers/gpu/drm/i915/gt/selftest_hangcheck.c i915_request_put(prev); prev 1505 drivers/gpu/drm/i915/gt/selftest_hangcheck.c i915_request_put(prev); prev 1510 drivers/gpu/drm/i915/gt/selftest_hangcheck.c i915_request_put(prev); prev 1511 drivers/gpu/drm/i915/gt/selftest_hangcheck.c prev = rq; prev 1519 drivers/gpu/drm/i915/gt/selftest_hangcheck.c i915_request_put(prev); prev 58 drivers/gpu/drm/i915/i915_active.c return (struct intel_engine_cs *)READ_ONCE(node->base.link.prev); prev 502 drivers/gpu/drm/i915/i915_active.c struct rb_node *prev, *p; prev 522 drivers/gpu/drm/i915/i915_active.c prev = NULL; prev 531 drivers/gpu/drm/i915/i915_active.c prev = p; prev 544 drivers/gpu/drm/i915/i915_active.c for (p = prev; p; p = rb_next(p)) { prev 634 drivers/gpu/drm/i915/i915_active.c node->base.link.prev = (void *)engine; prev 1132 drivers/gpu/drm/i915/i915_irq.c const struct intel_rps_ei *prev = &rps->ei; prev 1141 drivers/gpu/drm/i915/i915_irq.c if (prev->ktime) { prev 1145 drivers/gpu/drm/i915/i915_irq.c time = ktime_us_delta(now.ktime, prev->ktime); prev 1154 drivers/gpu/drm/i915/i915_irq.c render = now.render_c0 - prev->render_c0; prev 1155 drivers/gpu/drm/i915/i915_irq.c media = now.media_c0 - prev->media_c0; prev 577 drivers/gpu/drm/i915/i915_pmu.c u64 prev, new; prev 580 drivers/gpu/drm/i915/i915_pmu.c prev = local64_read(&hwc->prev_count); prev 583 drivers/gpu/drm/i915/i915_pmu.c if (local64_cmpxchg(&hwc->prev_count, prev, new) != prev) prev 586 drivers/gpu/drm/i915/i915_pmu.c local64_add(new - prev, &event->count); prev 1129 drivers/gpu/drm/i915/i915_request.c struct i915_request *prev; prev 1151 drivers/gpu/drm/i915/i915_request.c prev = rcu_dereference_protected(timeline->last_request.request, prev 1153 drivers/gpu/drm/i915/i915_request.c if (prev && !i915_request_completed(prev)) { prev 1154 drivers/gpu/drm/i915/i915_request.c if (is_power_of_2(prev->engine->mask | rq->engine->mask)) prev 1156 drivers/gpu/drm/i915/i915_request.c &prev->submit, prev 1160 drivers/gpu/drm/i915/i915_request.c &prev->fence, prev 1164 drivers/gpu/drm/i915/i915_request.c &prev->sched, prev 1179 drivers/gpu/drm/i915/i915_request.c return prev; prev 1242 drivers/gpu/drm/i915/i915_request.c struct i915_request *prev; prev 1249 drivers/gpu/drm/i915/i915_request.c prev = __i915_request_commit(rq); prev 1296 drivers/gpu/drm/i915/i915_request.c if (prev && i915_request_completed(prev) && prev->timeline == tl) prev 1297 drivers/gpu/drm/i915/i915_request.c i915_request_retire_upto(prev); prev 305 drivers/gpu/drm/i915/i915_scheduler.c if (stack.dfs_link.next == stack.dfs_link.prev) prev 244 drivers/gpu/drm/i915/i915_utils.h first->prev = head; prev 119 drivers/gpu/drm/i915/selftests/i915_buddy.c struct i915_buddy_block *prev; prev 124 drivers/gpu/drm/i915/selftests/i915_buddy.c prev = NULL; prev 135 drivers/gpu/drm/i915/selftests/i915_buddy.c if (is_contiguous && prev) { prev 140 drivers/gpu/drm/i915/selftests/i915_buddy.c prev_offset = i915_buddy_block_offset(prev); prev 141 drivers/gpu/drm/i915/selftests/i915_buddy.c prev_block_size = i915_buddy_block_size(mm, prev); prev 154 drivers/gpu/drm/i915/selftests/i915_buddy.c prev = block; prev 166 drivers/gpu/drm/i915/selftests/i915_buddy.c if (prev) { prev 168 drivers/gpu/drm/i915/selftests/i915_buddy.c igt_dump_block(mm, prev); prev 182 drivers/gpu/drm/i915/selftests/i915_buddy.c struct i915_buddy_block *prev; prev 199 drivers/gpu/drm/i915/selftests/i915_buddy.c prev = NULL; prev 229 drivers/gpu/drm/i915/selftests/i915_buddy.c if (prev) { prev 234 drivers/gpu/drm/i915/selftests/i915_buddy.c prev_offset = i915_buddy_block_offset(prev); prev 235 drivers/gpu/drm/i915/selftests/i915_buddy.c prev_block_size = i915_buddy_block_size(mm, prev); prev 255 drivers/gpu/drm/i915/selftests/i915_buddy.c prev = root; prev 268 drivers/gpu/drm/i915/selftests/i915_buddy.c if (prev) { prev 270 drivers/gpu/drm/i915/selftests/i915_buddy.c igt_dump_block(mm, prev); prev 939 drivers/gpu/drm/i915/selftests/i915_request.c struct i915_request *prev = NULL; prev 978 drivers/gpu/drm/i915/selftests/i915_request.c if (prev) { prev 980 drivers/gpu/drm/i915/selftests/i915_request.c &prev->fence); prev 1006 drivers/gpu/drm/i915/selftests/i915_request.c prev = request[id]; prev 32 drivers/gpu/drm/i915/selftests/intel_uncore.c s32 prev; prev 34 drivers/gpu/drm/i915/selftests/intel_uncore.c for (i = 0, prev = -1; i < num_ranges; i++, ranges++) { prev 36 drivers/gpu/drm/i915/selftests/intel_uncore.c if (is_watertight && (prev + 1) != (s32)ranges->start) { prev 38 drivers/gpu/drm/i915/selftests/intel_uncore.c __func__, i, ranges->start, ranges->end, prev); prev 43 drivers/gpu/drm/i915/selftests/intel_uncore.c if (prev >= (s32)ranges->start) { prev 45 drivers/gpu/drm/i915/selftests/intel_uncore.c __func__, i, ranges->start, ranges->end, prev); prev 56 drivers/gpu/drm/i915/selftests/intel_uncore.c prev = ranges->end; prev 73 drivers/gpu/drm/i915/selftests/intel_uncore.c s32 prev; prev 77 drivers/gpu/drm/i915/selftests/intel_uncore.c for (i = 0, prev = -1; i < reg_lists[j].size; i++, reg++) { prev 80 drivers/gpu/drm/i915/selftests/intel_uncore.c if (prev >= (s32)offset) { prev 82 drivers/gpu/drm/i915/selftests/intel_uncore.c __func__, i, offset, prev); prev 86 drivers/gpu/drm/i915/selftests/intel_uncore.c prev = offset; prev 268 drivers/gpu/drm/mga/mga_dma.c entry->prev = dev_priv->head; prev 273 drivers/gpu/drm/mga/mga_dma.c dev_priv->head->next->prev = entry; prev 326 drivers/gpu/drm/mga/mga_dma.c drm_mga_freelist_t *prev; prev 342 drivers/gpu/drm/mga/mga_dma.c prev = dev_priv->tail->prev; prev 344 drivers/gpu/drm/mga/mga_dma.c prev->next = NULL; prev 345 drivers/gpu/drm/mga/mga_dma.c next->prev = next->next = NULL; prev 346 drivers/gpu/drm/mga/mga_dma.c dev_priv->tail = prev; prev 359 drivers/gpu/drm/mga/mga_dma.c drm_mga_freelist_t *head, *entry, *prev; prev 371 drivers/gpu/drm/mga/mga_dma.c prev = dev_priv->tail; prev 372 drivers/gpu/drm/mga/mga_dma.c prev->next = entry; prev 373 drivers/gpu/drm/mga/mga_dma.c entry->prev = prev; prev 376 drivers/gpu/drm/mga/mga_dma.c prev = head->next; prev 378 drivers/gpu/drm/mga/mga_dma.c prev->prev = entry; prev 379 drivers/gpu/drm/mga/mga_dma.c entry->prev = head; prev 380 drivers/gpu/drm/mga/mga_dma.c entry->next = prev; prev 81 drivers/gpu/drm/mga/mga_drv.h struct drm_mga_freelist *prev; prev 111 drivers/gpu/drm/nouveau/include/nvif/list.h struct list_head *next, *prev; prev 130 drivers/gpu/drm/nouveau/include/nvif/list.h list->next = list->prev = list; prev 135 drivers/gpu/drm/nouveau/include/nvif/list.h struct list_head *prev, struct list_head *next) prev 137 drivers/gpu/drm/nouveau/include/nvif/list.h next->prev = entry; prev 139 drivers/gpu/drm/nouveau/include/nvif/list.h entry->prev = prev; prev 140 drivers/gpu/drm/nouveau/include/nvif/list.h prev->next = entry; prev 182 drivers/gpu/drm/nouveau/include/nvif/list.h __list_add(entry, head->prev, head); prev 186 drivers/gpu/drm/nouveau/include/nvif/list.h __list_del(struct list_head *prev, struct list_head *next) prev 188 drivers/gpu/drm/nouveau/include/nvif/list.h next->prev = prev; prev 189 drivers/gpu/drm/nouveau/include/nvif/list.h prev->next = next; prev 209 drivers/gpu/drm/nouveau/include/nvif/list.h __list_del(entry->prev, entry->next); prev 215 drivers/gpu/drm/nouveau/include/nvif/list.h __list_del(entry->prev, entry->next); prev 222 drivers/gpu/drm/nouveau/include/nvif/list.h __list_del(list->prev, list->next); prev 292 drivers/gpu/drm/nouveau/include/nvif/list.h list_entry((ptr)->prev, type, member) prev 334 drivers/gpu/drm/nouveau/include/nvif/list.h for (pos = __container_of((head)->prev, pos, member); \ prev 336 drivers/gpu/drm/nouveau/include/nvif/list.h pos = __container_of(pos->member.prev, pos, member)) prev 344 drivers/gpu/drm/nouveau/include/nvif/list.h for (pos = __container_of(pos->member.prev, pos, member); \ prev 346 drivers/gpu/drm/nouveau/include/nvif/list.h pos = __container_of(pos->member.prev, pos, member)) prev 53 drivers/gpu/drm/nouveau/nouveau_bo.h struct nouveau_bo *prev; prev 57 drivers/gpu/drm/nouveau/nouveau_bo.h prev = *pnvbo; prev 65 drivers/gpu/drm/nouveau/nouveau_bo.h if (prev) prev 66 drivers/gpu/drm/nouveau/nouveau_bo.h ttm_bo_put(&prev->bo); prev 353 drivers/gpu/drm/nouveau/nouveau_fence.c struct nouveau_channel *prev = NULL; prev 359 drivers/gpu/drm/nouveau/nouveau_fence.c prev = rcu_dereference(f->channel); prev 360 drivers/gpu/drm/nouveau/nouveau_fence.c if (prev && (prev == chan || fctx->sync(f, prev, chan) == 0)) prev 375 drivers/gpu/drm/nouveau/nouveau_fence.c struct nouveau_channel *prev = NULL; prev 384 drivers/gpu/drm/nouveau/nouveau_fence.c prev = rcu_dereference(f->channel); prev 385 drivers/gpu/drm/nouveau/nouveau_fence.c if (prev && (prev == chan || fctx->sync(f, prev, chan) == 0)) prev 54 drivers/gpu/drm/nouveau/nv04_fence.c struct nouveau_channel *prev, struct nouveau_channel *chan) prev 45 drivers/gpu/drm/nouveau/nv10_fence.c struct nouveau_channel *prev, struct nouveau_channel *chan) prev 35 drivers/gpu/drm/nouveau/nv17_fence.c struct nouveau_channel *prev, struct nouveau_channel *chan) prev 37 drivers/gpu/drm/nouveau/nv17_fence.c struct nouveau_cli *cli = (void *)prev->user.client; prev 51 drivers/gpu/drm/nouveau/nv17_fence.c ret = RING_SPACE(prev, 5); prev 53 drivers/gpu/drm/nouveau/nv17_fence.c BEGIN_NV04(prev, 0, NV11_SUBCHAN_DMA_SEMAPHORE, 4); prev 54 drivers/gpu/drm/nouveau/nv17_fence.c OUT_RING (prev, fctx->sema.handle); prev 55 drivers/gpu/drm/nouveau/nv17_fence.c OUT_RING (prev, 0); prev 56 drivers/gpu/drm/nouveau/nv17_fence.c OUT_RING (prev, value + 0); prev 57 drivers/gpu/drm/nouveau/nv17_fence.c OUT_RING (prev, value + 1); prev 58 drivers/gpu/drm/nouveau/nv17_fence.c FIRE_RING (prev); prev 79 drivers/gpu/drm/nouveau/nv84_fence.c struct nouveau_channel *prev, struct nouveau_channel *chan) prev 82 drivers/gpu/drm/nouveau/nv84_fence.c u64 addr = fctx->vma->addr + prev->chid * 16; prev 53 drivers/gpu/drm/nouveau/nvkm/core/mm.c struct nvkm_mm_node *prev = node(this, prev); prev 56 drivers/gpu/drm/nouveau/nvkm/core/mm.c if (prev && prev->type == NVKM_MM_TYPE_NONE) { prev 57 drivers/gpu/drm/nouveau/nvkm/core/mm.c prev->length += this->length; prev 59 drivers/gpu/drm/nouveau/nvkm/core/mm.c kfree(this); this = prev; prev 72 drivers/gpu/drm/nouveau/nvkm/core/mm.c list_for_each_entry(prev, &mm->free, fl_entry) { prev 73 drivers/gpu/drm/nouveau/nvkm/core/mm.c if (this->offset < prev->offset) prev 77 drivers/gpu/drm/nouveau/nvkm/core/mm.c list_add_tail(&this->fl_entry, &prev->fl_entry); prev 114 drivers/gpu/drm/nouveau/nvkm/core/mm.c struct nvkm_mm_node *prev, *this, *next; prev 129 drivers/gpu/drm/nouveau/nvkm/core/mm.c prev = node(this, prev); prev 130 drivers/gpu/drm/nouveau/nvkm/core/mm.c if (prev && prev->type != type) prev 189 drivers/gpu/drm/nouveau/nvkm/core/mm.c struct nvkm_mm_node *prev, *this, *next; prev 203 drivers/gpu/drm/nouveau/nvkm/core/mm.c prev = node(this, prev); prev 204 drivers/gpu/drm/nouveau/nvkm/core/mm.c if (prev && prev->type != type) prev 242 drivers/gpu/drm/nouveau/nvkm/core/mm.c struct nvkm_mm_node *node, *prev; prev 246 drivers/gpu/drm/nouveau/nvkm/core/mm.c prev = list_last_entry(&mm->nodes, typeof(*node), nl_entry); prev 247 drivers/gpu/drm/nouveau/nvkm/core/mm.c next = prev->offset + prev->length; prev 60 drivers/gpu/drm/nouveau/nvkm/engine/disp/baseg84.c .prev = 0x000004, prev 94 drivers/gpu/drm/nouveau/nvkm/engine/disp/basegf119.c .prev = -0x020000, prev 105 drivers/gpu/drm/nouveau/nvkm/engine/disp/basenv50.c .prev = 0x000004, prev 48 drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.c u32 prev = nvkm_rd32(device, list->data[i].addr + base + c); prev 53 drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.c if (prev != next) prev 60 drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.c mthd, prev, mods, name ? " // " : "", prev 100 drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.c nv50_disp_mthd_list(disp, debug, base, mthd->prev, prev 162 drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.h s32 prev; prev 94 drivers/gpu/drm/nouveau/nvkm/engine/disp/coreg84.c .prev = 0x000004, prev 40 drivers/gpu/drm/nouveau/nvkm/engine/disp/coreg94.c .prev = 0x000004, prev 158 drivers/gpu/drm/nouveau/nvkm/engine/disp/coregf119.c .prev = -0x020000, prev 109 drivers/gpu/drm/nouveau/nvkm/engine/disp/coregk104.c .prev = -0x020000, prev 125 drivers/gpu/drm/nouveau/nvkm/engine/disp/coregv100.c .prev = 0x008000, prev 155 drivers/gpu/drm/nouveau/nvkm/engine/disp/corenv50.c .prev = 0x000004, prev 58 drivers/gpu/drm/nouveau/nvkm/engine/disp/ovlyg84.c .prev = 0x000004, prev 82 drivers/gpu/drm/nouveau/nvkm/engine/disp/ovlygf119.c .prev = -0x020000, prev 84 drivers/gpu/drm/nouveau/nvkm/engine/disp/ovlygk104.c .prev = -0x020000, prev 61 drivers/gpu/drm/nouveau/nvkm/engine/disp/ovlygt200.c .prev = 0x000004, prev 94 drivers/gpu/drm/nouveau/nvkm/engine/disp/ovlynv50.c .prev = 0x000004, prev 123 drivers/gpu/drm/nouveau/nvkm/engine/disp/wndwgv100.c .prev = 0x000800, prev 48 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c } prev, next, *chan; prev 67 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c status->prev.tsg = !!(stat & 0x00001000); prev 68 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c status->prev.id = (stat & 0x00000fff); prev 76 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c status->chan = &status->prev; prev 81 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c status->chan = &status->prev; prev 85 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c status->chan = &status->prev; prev 92 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c status->prev.tsg ? "tsg" : "ch", status->prev.id, prev 93 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c status->chan == &status->prev ? "*" : " ", prev 1653 drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c u32 addr = ~0, prev = ~0, xfer = 0; prev 1667 drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c if (head != prev + 4 || xfer >= 32) { prev 1676 drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c prev = head; prev 1117 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv04.c struct nv04_gr_chan *prev = NULL; prev 1124 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv04.c prev = nv04_gr_channel(gr); prev 1125 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv04.c if (prev) prev 1126 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv04.c nv04_gr_unload_context(prev); prev 935 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv10.c struct nv10_gr_chan *prev = NULL; prev 942 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv10.c prev = nv10_gr_channel(gr); prev 943 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv10.c if (prev) prev 944 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv10.c nv10_gr_unload_context(prev); prev 116 drivers/gpu/drm/nouveau/nvkm/subdev/bios/base.c u8 *prev = bios->data; prev 118 drivers/gpu/drm/nouveau/nvkm/subdev/bios/base.c bios->data = prev; prev 121 drivers/gpu/drm/nouveau/nvkm/subdev/bios/base.c memcpy(bios->data, prev, bios->size); prev 123 drivers/gpu/drm/nouveau/nvkm/subdev/bios/base.c kfree(prev); prev 242 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c u32 prev = nvkm_rd32(device, addr); prev 243 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c u32 next = (prev & ~mask) | data; prev 883 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c nvkm_vmm_node_merge(struct nvkm_vmm *vmm, struct nvkm_vma *prev, prev 890 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c if (prev) { prev 891 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c prev->size += vma->size; prev 893 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c return prev; prev 897 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c BUG_ON(prev); prev 907 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c if (prev) { prev 910 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c prev->size += size; prev 915 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c prev->size += vma->size; prev 918 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c return prev; prev 928 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c struct nvkm_vma *prev = NULL; prev 931 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c prev = vma; prev 941 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c nvkm_vmm_node_merge(vmm, prev, vma, NULL, vma->size); prev 1154 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c struct nvkm_vma *prev = NULL; prev 1157 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c if (vma->addr == addr && vma->part && (prev = node(vma, prev))) { prev 1158 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c if (prev->memory || prev->mapped != map) prev 1159 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c prev = NULL; prev 1168 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c if (prev || next) prev 1169 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c return nvkm_vmm_node_merge(vmm, prev, vma, next, size); prev 1336 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c struct nvkm_vma *prev = NULL; prev 1343 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c if (vma->part && (prev = node(vma, prev)) && prev->mapped) prev 1344 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c prev = NULL; prev 1347 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c nvkm_vmm_node_merge(vmm, prev, vma, next, vma->size); prev 1524 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c struct nvkm_vma *prev, *next; prev 1526 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c if ((prev = node(vma, prev)) && !prev->used) { prev 1527 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c vma->addr = prev->addr; prev 1528 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c vma->size += prev->size; prev 1529 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c nvkm_vmm_free_delete(vmm, prev); prev 1706 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c struct nvkm_vma *prev = node(this, prev); prev 1711 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c if (vmm->func->page_block && prev && prev->page != p) prev 129 drivers/gpu/drm/nouveau/nvkm/subdev/therm/fan.c u32 cycles, cur, prev; prev 143 drivers/gpu/drm/nouveau/nvkm/subdev/therm/fan.c prev = nvkm_gpio_get(gpio, 0, therm->fan->tach.func, prev 151 drivers/gpu/drm/nouveau/nvkm/subdev/therm/fan.c if (prev != cur) { prev 155 drivers/gpu/drm/nouveau/nvkm/subdev/therm/fan.c prev = cur; prev 170 drivers/gpu/drm/nouveau/nvkm/subdev/volt/base.c int prev = nvkm_volt_get(volt); prev 171 drivers/gpu/drm/nouveau/nvkm/subdev/volt/base.c if (!condition || prev < 0 || prev 172 drivers/gpu/drm/nouveau/nvkm/subdev/volt/base.c (condition < 0 && ret < prev) || prev 173 drivers/gpu/drm/nouveau/nvkm/subdev/volt/base.c (condition > 0 && ret > prev)) { prev 798 drivers/gpu/drm/r128/r128_cce.c entry->prev = dev_priv->head; prev 810 drivers/gpu/drm/r128/r128_cce.c dev_priv->head->next->prev = entry; prev 71 drivers/gpu/drm/r128/r128_drv.h struct drm_r128_freelist *prev; prev 43 drivers/gpu/drm/radeon/mkregtable.c struct list_head *next, *prev; prev 50 drivers/gpu/drm/radeon/mkregtable.c list->prev = list; prev 61 drivers/gpu/drm/radeon/mkregtable.c struct list_head *prev, struct list_head *next) prev 63 drivers/gpu/drm/radeon/mkregtable.c next->prev = new; prev 65 drivers/gpu/drm/radeon/mkregtable.c new->prev = prev; prev 66 drivers/gpu/drm/radeon/mkregtable.c prev->next = new; prev 70 drivers/gpu/drm/radeon/mkregtable.c struct list_head *prev, struct list_head *next); prev 83 drivers/gpu/drm/radeon/mkregtable.c __list_add(new, head->prev, head); prev 146 drivers/gpu/drm/radeon/radeon_sa.c sa_manager->hole = sa_bo->olist.prev; prev 302 drivers/gpu/drm/radeon/radeon_sa.c sa_manager->hole = best_bo->olist.prev; prev 225 drivers/gpu/drm/savage/savage_bci.c dev_priv->head.prev = NULL; prev 229 drivers/gpu/drm/savage/savage_bci.c dev_priv->tail.prev = &dev_priv->head; prev 240 drivers/gpu/drm/savage/savage_bci.c entry->prev = &dev_priv->head; prev 241 drivers/gpu/drm/savage/savage_bci.c dev_priv->head.next->prev = entry; prev 251 drivers/gpu/drm/savage/savage_bci.c drm_savage_buf_priv_t *tail = dev_priv->tail.prev; prev 270 drivers/gpu/drm/savage/savage_bci.c drm_savage_buf_priv_t *prev = tail->prev; prev 271 drivers/gpu/drm/savage/savage_bci.c prev->next = next; prev 272 drivers/gpu/drm/savage/savage_bci.c next->prev = prev; prev 273 drivers/gpu/drm/savage/savage_bci.c tail->next = tail->prev = NULL; prev 284 drivers/gpu/drm/savage/savage_bci.c drm_savage_buf_priv_t *entry = buf->dev_private, *prev, *next; prev 288 drivers/gpu/drm/savage/savage_bci.c if (entry->next != NULL || entry->prev != NULL) { prev 293 drivers/gpu/drm/savage/savage_bci.c prev = &dev_priv->head; prev 294 drivers/gpu/drm/savage/savage_bci.c next = prev->next; prev 295 drivers/gpu/drm/savage/savage_bci.c prev->next = entry; prev 296 drivers/gpu/drm/savage/savage_bci.c next->prev = entry; prev 297 drivers/gpu/drm/savage/savage_bci.c entry->prev = prev; prev 1071 drivers/gpu/drm/savage/savage_bci.c buf_priv->next == NULL && buf_priv->prev == NULL) { prev 65 drivers/gpu/drm/savage/savage_drv.h struct drm_savage_buf_priv *prev; prev 318 drivers/gpu/drm/ttm/ttm_page_alloc.c __list_del(p->lru.prev, &pool->list); prev 594 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c list_splice_init(&ctx->preempted, restart_head[i].prev); prev 733 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c list_entry(fman->fence_list.prev, struct vmw_fence_obj, prev 60 drivers/gpu/host1x/intr.c struct host1x_waitlist *waiter, *next, *prev; prev 71 drivers/gpu/host1x/intr.c prev = list_entry(dest->prev, prev 73 drivers/gpu/host1x/intr.c if (prev->data == waiter->data) { prev 74 drivers/gpu/host1x/intr.c prev->count++; prev 137 drivers/greybus/operation.c int prev; prev 148 drivers/greybus/operation.c prev = operation->errno; prev 149 drivers/greybus/operation.c if (prev == -EBADR) prev 154 drivers/greybus/operation.c WARN_ON(prev != -EBADR); prev 172 drivers/greybus/operation.c prev = operation->errno; prev 173 drivers/greybus/operation.c if (prev == -EINPROGRESS) prev 177 drivers/greybus/operation.c return prev == -EINPROGRESS; prev 206 drivers/hid/hid-input.c __s32 prev; prev 231 drivers/hid/hid-input.c prev = physical_extents; prev 233 drivers/hid/hid-input.c if (physical_extents < prev) prev 249 drivers/hid/hid-input.c prev = logical_extents; prev 251 drivers/hid/hid-input.c if (logical_extents < prev) prev 265 drivers/hid/hid-input.c prev = logical_extents; prev 267 drivers/hid/hid-input.c if (logical_extents < prev) prev 272 drivers/hid/hid-input.c prev = physical_extents; prev 274 drivers/hid/hid-input.c if (physical_extents < prev) prev 1080 drivers/hwtracing/intel_th/msu.c struct msc_window *prev = list_last_entry(&msc->win_list, prev 1084 drivers/hwtracing/intel_th/msu.c win->pgoff = prev->pgoff + prev->nr_blocks; prev 122 drivers/infiniband/core/rw.c struct rdma_rw_reg_ctx *prev = NULL; prev 144 drivers/infiniband/core/rw.c if (prev) { prev 146 drivers/infiniband/core/rw.c prev->wr.wr.next = ®->inv_wr; prev 148 drivers/infiniband/core/rw.c prev->wr.wr.next = ®->reg_wr.wr; prev 171 drivers/infiniband/core/rw.c prev = reg; prev 175 drivers/infiniband/core/rw.c if (prev) prev 176 drivers/infiniband/core/rw.c prev->wr.wr.next = NULL; prev 4281 drivers/infiniband/hw/cxgb4/cm.c tmp->prev = NULL; prev 134 drivers/infiniband/hw/hfi1/aspm.c ktime_t now, prev; prev 141 drivers/infiniband/hw/hfi1/aspm.c prev = rcd->aspm_ts_last_intr; prev 146 drivers/infiniband/hw/hfi1/aspm.c close_interrupts = ktime_to_ns(ktime_sub(now, prev)) < ASPM_TRIGGER_NS; prev 58 drivers/infiniband/hw/hfi1/rc.c struct rvt_ack_entry *find_prev_entry(struct rvt_qp *qp, u32 psn, u8 *prev, prev 89 drivers/infiniband/hw/hfi1/rc.c if (prev) prev 90 drivers/infiniband/hw/hfi1/rc.c *prev = p; prev 2559 drivers/infiniband/hw/hfi1/rc.c u8 prev; prev 2607 drivers/infiniband/hw/hfi1/rc.c e = find_prev_entry(qp, psn, &prev, &mra, &old_req); prev 2653 drivers/infiniband/hw/hfi1/rc.c qp->s_acked_ack_queue = prev; prev 2654 drivers/infiniband/hw/hfi1/rc.c qp->s_tail_ack_queue = prev; prev 2668 drivers/infiniband/hw/hfi1/rc.c qp->s_acked_ack_queue = prev; prev 2669 drivers/infiniband/hw/hfi1/rc.c qp->s_tail_ack_queue = prev; prev 52 drivers/infiniband/hw/hfi1/rc.h struct rvt_ack_entry *find_prev_entry(struct rvt_qp *qp, u32 psn, u8 *prev, prev 1989 drivers/infiniband/hw/hfi1/tid_rdma.c u8 prev; prev 2008 drivers/infiniband/hw/hfi1/tid_rdma.c e = find_prev_entry(qp, psn, &prev, NULL, &old_req); prev 2086 drivers/infiniband/hw/hfi1/tid_rdma.c for (i = prev + 1; ; i++) { prev 2144 drivers/infiniband/hw/hfi1/tid_rdma.c for (i = prev + 1; ; i++) { prev 2177 drivers/infiniband/hw/hfi1/tid_rdma.c qp->s_acked_ack_queue = prev; prev 2178 drivers/infiniband/hw/hfi1/tid_rdma.c qp->s_tail_ack_queue = prev; prev 3203 drivers/infiniband/hw/hfi1/tid_rdma.c struct rvt_swqe *prev; prev 3209 drivers/infiniband/hw/hfi1/tid_rdma.c prev = rvt_get_swqe_ptr(qp, s_prev); prev 3218 drivers/infiniband/hw/hfi1/tid_rdma.c switch (prev->wr.opcode) { prev 3220 drivers/infiniband/hw/hfi1/tid_rdma.c req = wqe_to_tid_req(prev); prev 3228 drivers/infiniband/hw/hfi1/tid_rdma.c if (prev->wr.opcode != IB_WR_TID_RDMA_WRITE) prev 3232 drivers/infiniband/hw/hfi1/tid_rdma.c switch (prev->wr.opcode) { prev 3238 drivers/infiniband/hw/hfi1/tid_rdma.c req = wqe_to_tid_req(prev); prev 5451 drivers/infiniband/hw/hfi1/tid_rdma.c struct rvt_ack_entry *prev; prev 5459 drivers/infiniband/hw/hfi1/tid_rdma.c prev = &qp->s_ack_queue[s_prev]; prev 5463 drivers/infiniband/hw/hfi1/tid_rdma.c prev->opcode == TID_OP(WRITE_REQ)) { prev 5464 drivers/infiniband/hw/hfi1/tid_rdma.c req = ack_to_tid_req(prev); prev 1222 drivers/infiniband/hw/i40iw/i40iw_puda.c buf = (struct i40iw_puda_buf *)(pbufl.prev); prev 1209 drivers/infiniband/hw/mlx4/mcg.c pend_req = list_entry(group->func[slave].pending.prev, struct mcast_req, group_list); prev 65 drivers/infiniband/hw/mthca/mthca_mcg.c u16 *hash, int *prev, int *index) prev 89 drivers/infiniband/hw/mthca/mthca_mcg.c *prev = -1; prev 109 drivers/infiniband/hw/mthca/mthca_mcg.c *prev = *index; prev 126 drivers/infiniband/hw/mthca/mthca_mcg.c int index, prev; prev 138 drivers/infiniband/hw/mthca/mthca_mcg.c err = find_mgm(dev, gid->raw, mailbox, &hash, &prev, &index); prev 191 drivers/infiniband/hw/mthca/mthca_mcg.c err = mthca_READ_MGM(dev, prev, mailbox); prev 199 drivers/infiniband/hw/mthca/mthca_mcg.c err = mthca_WRITE_MGM(dev, prev, mailbox); prev 220 drivers/infiniband/hw/mthca/mthca_mcg.c int prev, index; prev 231 drivers/infiniband/hw/mthca/mthca_mcg.c err = find_mgm(dev, gid->raw, mailbox, &hash, &prev, &index); prev 266 drivers/infiniband/hw/mthca/mthca_mcg.c if (prev == -1) { prev 291 drivers/infiniband/hw/mthca/mthca_mcg.c err = mthca_READ_MGM(dev, prev, mailbox); prev 299 drivers/infiniband/hw/mthca/mthca_mcg.c err = mthca_WRITE_MGM(dev, prev, mailbox); prev 1501 drivers/infiniband/hw/qib/qib_rc.c u8 i, prev; prev 1551 drivers/infiniband/hw/qib/qib_rc.c for (i = qp->r_head_ack_queue; ; i = prev) { prev 1555 drivers/infiniband/hw/qib/qib_rc.c prev = i - 1; prev 1557 drivers/infiniband/hw/qib/qib_rc.c prev = QIB_MAX_RDMA_ATOMIC; prev 1558 drivers/infiniband/hw/qib/qib_rc.c if (prev == qp->r_head_ack_queue) { prev 1562 drivers/infiniband/hw/qib/qib_rc.c e = &qp->s_ack_queue[prev]; prev 1568 drivers/infiniband/hw/qib/qib_rc.c if (prev == qp->s_tail_ack_queue && prev 1621 drivers/infiniband/hw/qib/qib_rc.c qp->s_tail_ack_queue = prev; prev 1634 drivers/infiniband/hw/qib/qib_rc.c qp->s_tail_ack_queue = prev; prev 1085 drivers/infiniband/hw/qib/qib_user_sdma.c pkt = list_entry(free_list.prev, prev 1492 drivers/infiniband/ulp/ipoib/ipoib_cm.c p = list_entry(priv->cm.passive_ids.prev, typeof(*p), list); prev 111 drivers/infiniband/ulp/srpt/ib_srpt.c enum rdma_ch_state prev; prev 115 drivers/infiniband/ulp/srpt/ib_srpt.c prev = ch->state; prev 116 drivers/infiniband/ulp/srpt/ib_srpt.c if (new > prev) { prev 894 drivers/infiniband/ulp/srpt/ib_srpt.c struct scatterlist *prev = NULL; prev 928 drivers/infiniband/ulp/srpt/ib_srpt.c if (prev) { prev 929 drivers/infiniband/ulp/srpt/ib_srpt.c sg_unmark_end(&prev[prev_nents - 1]); prev 930 drivers/infiniband/ulp/srpt/ib_srpt.c sg_chain(prev, prev_nents + 1, ctx->sg); prev 935 drivers/infiniband/ulp/srpt/ib_srpt.c prev = ctx->sg; prev 36 drivers/input/joystick/joydump.c struct joydump *dump, *prev; /* one entry each */ prev 107 drivers/input/joystick/joydump.c prev = dump; prev 116 drivers/input/joystick/joydump.c for (i = 1; i < t; i++, dump++, prev++) { prev 118 drivers/input/joystick/joydump.c i, dump->time - prev->time); prev 106 drivers/input/mouse/hgpk.c static int approx_half(int curr, int prev) prev 110 drivers/input/mouse/hgpk.c if (curr < 5 || prev < 5) prev 113 drivers/input/mouse/hgpk.c belowhalf = (prev * 8) / 20; prev 114 drivers/input/mouse/hgpk.c abovehalf = (prev * 12) / 20; prev 671 drivers/input/touchscreen/ads7846.c list_entry(m->transfers.prev, struct spi_transfer, transfer_list); prev 690 drivers/input/touchscreen/ads7846.c list_entry(m->transfers.prev, struct spi_transfer, transfer_list); prev 220 drivers/interconnect/core.c struct icc_node *next, *prev = NULL; prev 231 drivers/interconnect/core.c if (!prev || next->provider != prev->provider) { prev 232 drivers/interconnect/core.c prev = next; prev 237 drivers/interconnect/core.c ret = next->provider->set(prev, next); prev 241 drivers/interconnect/core.c prev = next; prev 816 drivers/iommu/dma-iommu.c struct scatterlist *s, *prev = NULL; prev 857 drivers/iommu/dma-iommu.c prev->length += pad_len; prev 862 drivers/iommu/dma-iommu.c prev = s; prev 184 drivers/iommu/iova.c struct rb_node *curr, *prev; prev 204 drivers/iommu/iova.c prev = curr; prev 219 drivers/iommu/iova.c iova_insert_rbtree(&iovad->rbroot, new, prev); prev 733 drivers/iommu/iova.c struct iova *prev = NULL, *next = NULL; prev 737 drivers/iommu/iova.c prev = alloc_and_init_iova(iova->pfn_lo, pfn_lo - 1); prev 738 drivers/iommu/iova.c if (prev == NULL) prev 750 drivers/iommu/iova.c if (prev) { prev 751 drivers/iommu/iova.c iova_insert_rbtree(&iovad->rbroot, prev, NULL); prev 764 drivers/iommu/iova.c if (prev) prev 765 drivers/iommu/iova.c free_iova_mem(prev); prev 787 drivers/iommu/iova.c struct iova_magazine *prev; prev 878 drivers/iommu/iova.c cpu_rcache->prev = iova_magazine_alloc(GFP_KERNEL); prev 903 drivers/iommu/iova.c } else if (!iova_magazine_full(cpu_rcache->prev)) { prev 904 drivers/iommu/iova.c swap(cpu_rcache->prev, cpu_rcache->loaded); prev 966 drivers/iommu/iova.c } else if (!iova_magazine_empty(cpu_rcache->prev)) { prev 967 drivers/iommu/iova.c swap(cpu_rcache->prev, cpu_rcache->loaded); prev 1019 drivers/iommu/iova.c iova_magazine_free(cpu_rcache->prev); prev 1042 drivers/iommu/iova.c iova_magazine_free_pfns(cpu_rcache->prev, iovad); prev 1463 drivers/isdn/mISDN/dsp_cmx.c other = (list_entry(conf->mlist.prev, prev 1475 drivers/lightnvm/pblk-core.c list_cut_position(&list, &l_mg->emeta_list, l_mg->emeta_list.prev); prev 43 drivers/lightnvm/pblk-gc.c list_cut_position(&w_list, &gc->w_list, gc->w_list.prev); prev 540 drivers/lightnvm/pblk-recovery.c __list_add(&line->list, t->list.prev, &t->list); prev 295 drivers/md/bcache/bset.c kfree(t->prev); prev 297 drivers/md/bcache/bset.c free_pages((unsigned long) t->prev, prev 308 drivers/md/bcache/bset.c t->prev = NULL; prev 334 drivers/md/bcache/bset.c t->prev = bset_prev_bytes(b) < PAGE_SIZE prev 337 drivers/md/bcache/bset.c if (!t->prev) prev 552 drivers/md/bcache/bset.c return (void *) (((uint64_t *) tree_to_bkey(t, j)) - t->prev[j]); prev 561 drivers/md/bcache/bset.c return cacheline_to_bkey(t, cacheline, t->prev[cacheline]); prev 645 drivers/md/bcache/bset.c t->prev = t[-1].prev + j; prev 662 drivers/md/bcache/bset.c t->prev[0] = bkey_to_cacheline_offset(t, 0, t->data->start); prev 695 drivers/md/bcache/bset.c struct bkey *prev = NULL, *k = t->data->start; prev 718 drivers/md/bcache/bset.c prev = k, k = bkey_next(k); prev 720 drivers/md/bcache/bset.c t->prev[j] = bkey_u64s(prev); prev 810 drivers/md/bcache/bset.c t->prev[j] += shift; prev 812 drivers/md/bcache/bset.c if (t->prev[j] > 7) { prev 818 drivers/md/bcache/bset.c t->prev[j] = bkey_to_cacheline_offset(t, j, k); prev 831 drivers/md/bcache/bset.c t->prev[t->size] = prev 885 drivers/md/bcache/bset.c struct bkey *m, *prev = NULL; prev 911 drivers/md/bcache/bset.c prev = m, m = bkey_next(m); prev 915 drivers/md/bcache/bset.c if (prev && prev 916 drivers/md/bcache/bset.c bch_bkey_try_merge(b, prev, k)) prev 183 drivers/md/bcache/bset.h uint8_t *prev; prev 239 drivers/md/bcache/journal.c seq = list_entry(list->prev, struct journal_replay, prev 245 drivers/md/bcache/journal.c if (seq != list_entry(list->prev, struct journal_replay, prev 292 drivers/md/bcache/journal.c c->journal.seq = list_entry(list->prev, prev 363 drivers/md/bcache/journal.c list_entry(list->prev, struct journal_replay, list); prev 1876 drivers/md/bcache/super.c j = &list_entry(journal.prev, struct journal_replay, list)->j; prev 1848 drivers/md/dm-bufio.c b = list_entry(global_queue.prev, struct dm_buffer, global_list); prev 40 drivers/md/dm-cache-policy-smq.c unsigned prev:28; prev 139 drivers/md/dm-cache-policy-smq.c return to_entry(es, e->prev); prev 152 drivers/md/dm-cache-policy-smq.c e->prev = INDEXER_NULL; prev 155 drivers/md/dm-cache-policy-smq.c head->prev = l->head = to_index(es, e); prev 168 drivers/md/dm-cache-policy-smq.c e->prev = l->tail; prev 182 drivers/md/dm-cache-policy-smq.c struct entry *prev = l_prev(es, old); prev 184 drivers/md/dm-cache-policy-smq.c if (!prev) prev 188 drivers/md/dm-cache-policy-smq.c e->prev = old->prev; prev 190 drivers/md/dm-cache-policy-smq.c prev->next = old->prev = to_index(es, e); prev 199 drivers/md/dm-cache-policy-smq.c struct entry *prev = l_prev(es, e); prev 202 drivers/md/dm-cache-policy-smq.c if (prev) prev 203 drivers/md/dm-cache-policy-smq.c prev->next = e->next; prev 208 drivers/md/dm-cache-policy-smq.c next->prev = e->prev; prev 210 drivers/md/dm-cache-policy-smq.c l->tail = e->prev; prev 629 drivers/md/dm-cache-policy-smq.c struct entry **prev) prev 633 drivers/md/dm-cache-policy-smq.c *prev = NULL; prev 638 drivers/md/dm-cache-policy-smq.c *prev = e; prev 645 drivers/md/dm-cache-policy-smq.c struct entry *e, struct entry *prev) prev 647 drivers/md/dm-cache-policy-smq.c if (prev) prev 648 drivers/md/dm-cache-policy-smq.c prev->hash_next = e->hash_next; prev 658 drivers/md/dm-cache-policy-smq.c struct entry *e, *prev; prev 661 drivers/md/dm-cache-policy-smq.c e = __h_lookup(ht, h, oblock, &prev); prev 662 drivers/md/dm-cache-policy-smq.c if (e && prev) { prev 667 drivers/md/dm-cache-policy-smq.c __h_unlink(ht, h, e, prev); prev 677 drivers/md/dm-cache-policy-smq.c struct entry *prev; prev 683 drivers/md/dm-cache-policy-smq.c e = __h_lookup(ht, h, e->oblock, &prev); prev 685 drivers/md/dm-cache-policy-smq.c __h_unlink(ht, h, e, prev); prev 720 drivers/md/dm-cache-policy-smq.c e->prev = INDEXER_NULL; prev 529 drivers/md/dm-table.c struct dm_target *prev; prev 534 drivers/md/dm-table.c prev = &table->targets[table->num_targets - 1]; prev 535 drivers/md/dm-table.c return (ti->begin == (prev->begin + prev->len)); prev 689 drivers/md/dm-writecache.c wc->lru.prev = LIST_POISON2; prev 691 drivers/md/dm-writecache.c wc->freelist.prev = LIST_POISON2; prev 758 drivers/md/dm-writecache.c if (unlikely(e->lru.prev == &wc->lru)) prev 760 drivers/md/dm-writecache.c e = container_of(e->lru.prev, struct wc_entry, lru); prev 1448 drivers/md/dm-writecache.c list.next->prev = list.prev->next = &list; prev 1510 drivers/md/dm-writecache.c e = container_of(wbl->list.prev, struct wc_entry, lru); prev 1535 drivers/md/dm-writecache.c f = container_of(wbl->list.prev, struct wc_entry, lru); prev 1568 drivers/md/dm-writecache.c e = container_of(wbl->list.prev, struct wc_entry, lru); prev 1587 drivers/md/dm-writecache.c f = container_of(wbl->list.prev, struct wc_entry, lru); prev 1647 drivers/md/dm-writecache.c e = container_of(wc->lru.prev, struct wc_entry, lru); prev 328 drivers/md/md-multipath.c mp_bh = list_entry(head->prev, struct multipath_bh, retry_list); prev 329 drivers/md/md-multipath.c list_del(head->prev); prev 2564 drivers/md/raid1.c r1_bio = list_entry(head->prev, struct r1bio, retry_list); prev 2565 drivers/md/raid1.c list_del(head->prev); prev 632 drivers/md/raid10.c geo = &conf->prev; prev 862 drivers/md/raid10.c (i < conf->geo.raid_disks || i < conf->prev.raid_disks) prev 1524 drivers/md/raid10.c sector_t chunk_mask = (conf->geo.chunk_mask & conf->prev.chunk_mask); prev 1542 drivers/md/raid10.c || conf->prev.near_copies < prev 1543 drivers/md/raid10.c conf->prev.raid_disks))) prev 1593 drivers/md/raid10.c disks = conf->prev.raid_disks; prev 1594 drivers/md/raid10.c ncopies = conf->prev.near_copies; prev 2728 drivers/md/raid10.c list_move(conf->bio_end_io_list.prev, &tmp); prev 2757 drivers/md/raid10.c r10_bio = list_entry(head->prev, struct r10bio, retry_list); prev 2758 drivers/md/raid10.c list_del(head->prev); prev 3532 drivers/md/raid10.c conf->prev.raid_disks); prev 3681 drivers/md/raid10.c conf->prev = conf->geo; prev 3684 drivers/md/raid10.c if (setup_geo(&conf->prev, mddev, geo_old) != conf->copies) { prev 3689 drivers/md/raid10.c if (conf->prev.far_offset) prev 3690 drivers/md/raid10.c conf->prev.stride = 1 << conf->prev.chunk_shift; prev 3693 drivers/md/raid10.c conf->prev.stride = conf->dev_sectors; prev 3783 drivers/md/raid10.c disk_idx >= conf->prev.raid_disks) prev 3835 drivers/md/raid10.c if (conf->prev.far_copies != 1 && prev 3836 drivers/md/raid10.c conf->prev.far_offset == 0) prev 3843 drivers/md/raid10.c || i < conf->prev.raid_disks; prev 3907 drivers/md/raid10.c before_length = ((1 << conf->prev.chunk_shift) * prev 3908 drivers/md/raid10.c conf->prev.far_copies); prev 4139 drivers/md/raid10.c for (i = 0; i < conf->prev.raid_disks; i++) { prev 4151 drivers/md/raid10.c if (conf->geo.raid_disks == conf->prev.raid_disks) prev 4165 drivers/md/raid10.c if (conf->geo.raid_disks <= conf->prev.raid_disks) prev 4202 drivers/md/raid10.c before_length = ((1 << conf->prev.chunk_shift) * prev 4203 drivers/md/raid10.c conf->prev.far_copies); prev 4234 drivers/md/raid10.c sizeof(struct raid10_info)*conf->prev.raid_disks); prev 4305 drivers/md/raid10.c conf->prev.raid_disks) prev 4313 drivers/md/raid10.c } else if (rdev->raid_disk >= conf->prev.raid_disks prev 4350 drivers/md/raid10.c conf->geo = conf->prev; prev 4478 drivers/md/raid10.c &conf->prev); prev 4485 drivers/md/raid10.c & conf->prev.chunk_mask); prev 4497 drivers/md/raid10.c safe = first_dev_address(conf->reshape_safe, &conf->prev); prev 4507 drivers/md/raid10.c & conf->prev.chunk_mask); prev 4731 drivers/md/raid10.c conf->prev = conf->geo; prev 4785 drivers/md/raid10.c __raid10_find_phys(&conf->prev, r10b); prev 57 drivers/md/raid10.h } prev, geo; prev 497 drivers/md/raid5-ppl.c struct bio *prev = bio; prev 501 drivers/md/raid5-ppl.c bio->bi_opf = prev->bi_opf; prev 502 drivers/md/raid5-ppl.c bio->bi_write_hint = prev->bi_write_hint; prev 503 drivers/md/raid5-ppl.c bio_copy_dev(bio, prev); prev 504 drivers/md/raid5-ppl.c bio->bi_iter.bi_sector = bio_end_sector(prev); prev 507 drivers/md/raid5-ppl.c bio_chain(bio, prev); prev 508 drivers/md/raid5-ppl.c ppl_submit_iounit_bio(io, prev); prev 133 drivers/media/common/saa7146/saa7146_fops.c buf, q->queue.prev, q->queue.next); prev 310 drivers/media/common/videobuf2/videobuf2-core.c struct vb2_buffer *prev = q->bufs[vb->index - 1]; prev 311 drivers/media/common/videobuf2/videobuf2-core.c struct vb2_plane *p = &prev->planes[prev->num_planes - 1]; prev 1614 drivers/media/pci/cx23885/cx23885-core.c struct cx23885_buffer *prev; prev 1631 drivers/media/pci/cx23885/cx23885-core.c prev = list_entry(cx88q->active.prev, struct cx23885_buffer, prev 1634 drivers/media/pci/cx23885/cx23885-core.c prev->risc.jmp[1] = cpu_to_le32(buf->risc.dma); prev 188 drivers/media/pci/cx23885/cx23885-vbi.c struct cx23885_buffer *prev; prev 206 drivers/media/pci/cx23885/cx23885-vbi.c prev = list_entry(q->active.prev, struct cx23885_buffer, prev 211 drivers/media/pci/cx23885/cx23885-vbi.c prev->risc.jmp[1] = cpu_to_le32(buf->risc.dma); prev 457 drivers/media/pci/cx23885/cx23885-video.c struct cx23885_buffer *prev; prev 474 drivers/media/pci/cx23885/cx23885-video.c prev = list_entry(q->active.prev, struct cx23885_buffer, prev 477 drivers/media/pci/cx23885/cx23885-video.c prev->risc.jmp[1] = cpu_to_le32(buf->risc.dma); prev 242 drivers/media/pci/cx25821/cx25821-video.c struct cx25821_buffer *prev; prev 254 drivers/media/pci/cx25821/cx25821-video.c prev = list_entry(q->active.prev, struct cx25821_buffer, prev 257 drivers/media/pci/cx25821/cx25821-video.c prev->risc.jmp[1] = cpu_to_le32(buf->risc.dma); prev 240 drivers/media/pci/cx88/cx88-mpeg.c struct cx88_buffer *prev; prev 258 drivers/media/pci/cx88/cx88-mpeg.c prev = list_entry(cx88q->active.prev, struct cx88_buffer, list); prev 260 drivers/media/pci/cx88/cx88-mpeg.c prev->risc.jmp[1] = cpu_to_le32(buf->risc.dma); prev 171 drivers/media/pci/cx88/cx88-vbi.c struct cx88_buffer *prev; prev 186 drivers/media/pci/cx88/cx88-vbi.c prev = list_entry(q->active.prev, struct cx88_buffer, list); prev 188 drivers/media/pci/cx88/cx88-vbi.c prev->risc.jmp[1] = cpu_to_le32(buf->risc.dma); prev 506 drivers/media/pci/cx88/cx88-video.c struct cx88_buffer *prev; prev 521 drivers/media/pci/cx88/cx88-video.c prev = list_entry(q->active.prev, struct cx88_buffer, list); prev 523 drivers/media/pci/cx88/cx88-video.c prev->risc.jmp[1] = cpu_to_le32(buf->risc.dma); prev 133 drivers/media/pci/ivtv/ivtv-queue.c struct ivtv_buffer *buf = list_entry(steal->list.prev, struct ivtv_buffer, list); prev 141 drivers/media/pci/ivtv/ivtv-queue.c list_move_tail(steal->list.prev, &from->list); prev 152 drivers/media/pci/ivtv/ivtv-queue.c buf = list_entry(steal->list.prev, struct ivtv_buffer, list); prev 316 drivers/media/pci/saa7134/saa7134-core.c buf, q->queue.prev, q->queue.next); prev 323 drivers/media/pci/saa7134/saa7134-core.c q->queue.prev, q->queue.next); prev 409 drivers/media/pci/tw68/tw68-video.c struct tw68_buf *prev; prev 419 drivers/media/pci/tw68/tw68-video.c prev = list_entry(dev->active.prev, struct tw68_buf, list); prev 421 drivers/media/pci/tw68/tw68-video.c prev->jmp[1] = cpu_to_le32(buf->dma); prev 292 drivers/media/platform/fsl-viu.c struct viu_buf *buf, *prev; prev 303 drivers/media/platform/fsl-viu.c list_for_each_entry_safe(buf, prev, &vidq->active, vb.queue) { prev 312 drivers/media/platform/fsl-viu.c prev = NULL; prev 317 drivers/media/platform/fsl-viu.c if (prev == NULL) { prev 329 drivers/media/platform/fsl-viu.c } else if (prev->vb.width == buf->vb.width && prev 330 drivers/media/platform/fsl-viu.c prev->vb.height == buf->vb.height && prev 331 drivers/media/platform/fsl-viu.c prev->fmt == buf->fmt) { prev 339 drivers/media/platform/fsl-viu.c prev = buf; prev 498 drivers/media/platform/fsl-viu.c struct viu_buf *prev; prev 506 drivers/media/platform/fsl-viu.c vidq->queued.prev); prev 522 drivers/media/platform/fsl-viu.c prev = list_entry(vidq->active.prev, struct viu_buf, vb.queue); prev 523 drivers/media/platform/fsl-viu.c if (prev->vb.width == buf->vb.width && prev 524 drivers/media/platform/fsl-viu.c prev->vb.height == buf->vb.height && prev 525 drivers/media/platform/fsl-viu.c prev->fmt == buf->fmt) { prev 149 drivers/media/platform/omap3isp/isppreview.c preview_config_luma_enhancement(struct isp_prev_device *prev, prev 152 drivers/media/platform/omap3isp/isppreview.c struct isp_device *isp = to_isp_device(prev); prev 168 drivers/media/platform/omap3isp/isppreview.c preview_enable_luma_enhancement(struct isp_prev_device *prev, bool enable) prev 170 drivers/media/platform/omap3isp/isppreview.c struct isp_device *isp = to_isp_device(prev); prev 183 drivers/media/platform/omap3isp/isppreview.c static void preview_enable_invalaw(struct isp_prev_device *prev, bool enable) prev 185 drivers/media/platform/omap3isp/isppreview.c struct isp_device *isp = to_isp_device(prev); prev 198 drivers/media/platform/omap3isp/isppreview.c static void preview_config_hmed(struct isp_prev_device *prev, prev 201 drivers/media/platform/omap3isp/isppreview.c struct isp_device *isp = to_isp_device(prev); prev 213 drivers/media/platform/omap3isp/isppreview.c static void preview_enable_hmed(struct isp_prev_device *prev, bool enable) prev 215 drivers/media/platform/omap3isp/isppreview.c struct isp_device *isp = to_isp_device(prev); prev 233 drivers/media/platform/omap3isp/isppreview.c static void preview_config_cfa(struct isp_prev_device *prev, prev 242 drivers/media/platform/omap3isp/isppreview.c const unsigned int *order = cfa_coef_order[prev->params.cfa_order]; prev 244 drivers/media/platform/omap3isp/isppreview.c struct isp_device *isp = to_isp_device(prev); prev 269 drivers/media/platform/omap3isp/isppreview.c preview_config_chroma_suppression(struct isp_prev_device *prev, prev 272 drivers/media/platform/omap3isp/isppreview.c struct isp_device *isp = to_isp_device(prev); prev 285 drivers/media/platform/omap3isp/isppreview.c preview_enable_chroma_suppression(struct isp_prev_device *prev, bool enable) prev 287 drivers/media/platform/omap3isp/isppreview.c struct isp_device *isp = to_isp_device(prev); prev 303 drivers/media/platform/omap3isp/isppreview.c preview_config_whitebalance(struct isp_prev_device *prev, prev 306 drivers/media/platform/omap3isp/isppreview.c struct isp_device *isp = to_isp_device(prev); prev 342 drivers/media/platform/omap3isp/isppreview.c preview_config_blkadj(struct isp_prev_device *prev, prev 345 drivers/media/platform/omap3isp/isppreview.c struct isp_device *isp = to_isp_device(prev); prev 358 drivers/media/platform/omap3isp/isppreview.c preview_config_rgb_blending(struct isp_prev_device *prev, prev 361 drivers/media/platform/omap3isp/isppreview.c struct isp_device *isp = to_isp_device(prev); prev 396 drivers/media/platform/omap3isp/isppreview.c preview_config_csc(struct isp_prev_device *prev, prev 399 drivers/media/platform/omap3isp/isppreview.c struct isp_device *isp = to_isp_device(prev); prev 428 drivers/media/platform/omap3isp/isppreview.c preview_config_yc_range(struct isp_prev_device *prev, prev 431 drivers/media/platform/omap3isp/isppreview.c struct isp_device *isp = to_isp_device(prev); prev 446 drivers/media/platform/omap3isp/isppreview.c preview_config_dcor(struct isp_prev_device *prev, prev 449 drivers/media/platform/omap3isp/isppreview.c struct isp_device *isp = to_isp_device(prev); prev 468 drivers/media/platform/omap3isp/isppreview.c static void preview_enable_dcor(struct isp_prev_device *prev, bool enable) prev 470 drivers/media/platform/omap3isp/isppreview.c struct isp_device *isp = to_isp_device(prev); prev 484 drivers/media/platform/omap3isp/isppreview.c preview_enable_drkframe_capture(struct isp_prev_device *prev, bool enable) prev 486 drivers/media/platform/omap3isp/isppreview.c struct isp_device *isp = to_isp_device(prev); prev 499 drivers/media/platform/omap3isp/isppreview.c static void preview_enable_drkframe(struct isp_prev_device *prev, bool enable) prev 501 drivers/media/platform/omap3isp/isppreview.c struct isp_device *isp = to_isp_device(prev); prev 515 drivers/media/platform/omap3isp/isppreview.c preview_config_noisefilter(struct isp_prev_device *prev, prev 518 drivers/media/platform/omap3isp/isppreview.c struct isp_device *isp = to_isp_device(prev); prev 535 drivers/media/platform/omap3isp/isppreview.c preview_enable_noisefilter(struct isp_prev_device *prev, bool enable) prev 537 drivers/media/platform/omap3isp/isppreview.c struct isp_device *isp = to_isp_device(prev); prev 551 drivers/media/platform/omap3isp/isppreview.c preview_config_gammacorrn(struct isp_prev_device *prev, prev 554 drivers/media/platform/omap3isp/isppreview.c struct isp_device *isp = to_isp_device(prev); prev 584 drivers/media/platform/omap3isp/isppreview.c preview_enable_gammacorrn(struct isp_prev_device *prev, bool enable) prev 586 drivers/media/platform/omap3isp/isppreview.c struct isp_device *isp = to_isp_device(prev); prev 602 drivers/media/platform/omap3isp/isppreview.c preview_config_contrast(struct isp_prev_device *prev, prev 605 drivers/media/platform/omap3isp/isppreview.c struct isp_device *isp = to_isp_device(prev); prev 616 drivers/media/platform/omap3isp/isppreview.c preview_config_brightness(struct isp_prev_device *prev, prev 619 drivers/media/platform/omap3isp/isppreview.c struct isp_device *isp = to_isp_device(prev); prev 633 drivers/media/platform/omap3isp/isppreview.c preview_update_contrast(struct isp_prev_device *prev, u8 contrast) prev 638 drivers/media/platform/omap3isp/isppreview.c spin_lock_irqsave(&prev->params.lock, flags); prev 639 drivers/media/platform/omap3isp/isppreview.c params = (prev->params.active & OMAP3ISP_PREV_CONTRAST) prev 640 drivers/media/platform/omap3isp/isppreview.c ? &prev->params.params[0] : &prev->params.params[1]; prev 646 drivers/media/platform/omap3isp/isppreview.c spin_unlock_irqrestore(&prev->params.lock, flags); prev 655 drivers/media/platform/omap3isp/isppreview.c preview_update_brightness(struct isp_prev_device *prev, u8 brightness) prev 660 drivers/media/platform/omap3isp/isppreview.c spin_lock_irqsave(&prev->params.lock, flags); prev 661 drivers/media/platform/omap3isp/isppreview.c params = (prev->params.active & OMAP3ISP_PREV_BRIGHTNESS) prev 662 drivers/media/platform/omap3isp/isppreview.c ? &prev->params.params[0] : &prev->params.params[1]; prev 668 drivers/media/platform/omap3isp/isppreview.c spin_unlock_irqrestore(&prev->params.lock, flags); prev 672 drivers/media/platform/omap3isp/isppreview.c preview_params_lock(struct isp_prev_device *prev, u32 update, bool shadow) prev 674 drivers/media/platform/omap3isp/isppreview.c u32 active = prev->params.active; prev 678 drivers/media/platform/omap3isp/isppreview.c prev->params.params[0].busy |= ~active & update; prev 679 drivers/media/platform/omap3isp/isppreview.c prev->params.params[1].busy |= active & update; prev 682 drivers/media/platform/omap3isp/isppreview.c update = (prev->params.params[0].update & active) prev 683 drivers/media/platform/omap3isp/isppreview.c | (prev->params.params[1].update & ~active); prev 685 drivers/media/platform/omap3isp/isppreview.c prev->params.params[0].busy |= active & update; prev 686 drivers/media/platform/omap3isp/isppreview.c prev->params.params[1].busy |= ~active & update; prev 693 drivers/media/platform/omap3isp/isppreview.c preview_params_unlock(struct isp_prev_device *prev, u32 update, bool shadow) prev 695 drivers/media/platform/omap3isp/isppreview.c u32 active = prev->params.active; prev 701 drivers/media/platform/omap3isp/isppreview.c prev->params.params[0].update |= (~active & update); prev 702 drivers/media/platform/omap3isp/isppreview.c prev->params.params[1].update |= (active & update); prev 703 drivers/media/platform/omap3isp/isppreview.c prev->params.params[0].busy &= active; prev 704 drivers/media/platform/omap3isp/isppreview.c prev->params.params[1].busy &= ~active; prev 709 drivers/media/platform/omap3isp/isppreview.c prev->params.params[0].update &= ~(active & update); prev 710 drivers/media/platform/omap3isp/isppreview.c prev->params.params[1].update &= ~(~active & update); prev 711 drivers/media/platform/omap3isp/isppreview.c prev->params.params[0].busy &= ~active; prev 712 drivers/media/platform/omap3isp/isppreview.c prev->params.params[1].busy &= active; prev 716 drivers/media/platform/omap3isp/isppreview.c static void preview_params_switch(struct isp_prev_device *prev) prev 724 drivers/media/platform/omap3isp/isppreview.c to_switch = (prev->params.params[0].update & ~prev->params.active) prev 725 drivers/media/platform/omap3isp/isppreview.c | (prev->params.params[1].update & prev->params.active); prev 726 drivers/media/platform/omap3isp/isppreview.c to_switch &= ~(prev->params.params[0].busy | prev 727 drivers/media/platform/omap3isp/isppreview.c prev->params.params[1].busy); prev 731 drivers/media/platform/omap3isp/isppreview.c prev->params.active ^= to_switch; prev 736 drivers/media/platform/omap3isp/isppreview.c prev->params.params[0].update &= ~(~prev->params.active & to_switch); prev 737 drivers/media/platform/omap3isp/isppreview.c prev->params.params[1].update &= ~(prev->params.active & to_switch); prev 859 drivers/media/platform/omap3isp/isppreview.c static int preview_config(struct isp_prev_device *prev, prev 872 drivers/media/platform/omap3isp/isppreview.c spin_lock_irqsave(&prev->params.lock, flags); prev 873 drivers/media/platform/omap3isp/isppreview.c preview_params_lock(prev, cfg->update, true); prev 874 drivers/media/platform/omap3isp/isppreview.c active = prev->params.active; prev 875 drivers/media/platform/omap3isp/isppreview.c spin_unlock_irqrestore(&prev->params.lock, flags); prev 887 drivers/media/platform/omap3isp/isppreview.c params = &prev->params.params[!!(active & bit)]; prev 909 drivers/media/platform/omap3isp/isppreview.c spin_lock_irqsave(&prev->params.lock, flags); prev 910 drivers/media/platform/omap3isp/isppreview.c preview_params_unlock(prev, update, true); prev 911 drivers/media/platform/omap3isp/isppreview.c preview_params_switch(prev); prev 912 drivers/media/platform/omap3isp/isppreview.c spin_unlock_irqrestore(&prev->params.lock, flags); prev 925 drivers/media/platform/omap3isp/isppreview.c static void preview_setup_hw(struct isp_prev_device *prev, u32 update, prev 941 drivers/media/platform/omap3isp/isppreview.c params = &prev->params.params[!(active & bit)]; prev 945 drivers/media/platform/omap3isp/isppreview.c attr->config(prev, params); prev 947 drivers/media/platform/omap3isp/isppreview.c attr->enable(prev, true); prev 950 drivers/media/platform/omap3isp/isppreview.c attr->enable(prev, false); prev 960 drivers/media/platform/omap3isp/isppreview.c static void preview_config_ycpos(struct isp_prev_device *prev, u32 pixelcode) prev 962 drivers/media/platform/omap3isp/isppreview.c struct isp_device *isp = to_isp_device(prev); prev 985 drivers/media/platform/omap3isp/isppreview.c static void preview_config_averager(struct isp_prev_device *prev, u8 average) prev 987 drivers/media/platform/omap3isp/isppreview.c struct isp_device *isp = to_isp_device(prev); prev 1008 drivers/media/platform/omap3isp/isppreview.c static void preview_config_input_format(struct isp_prev_device *prev, prev 1011 drivers/media/platform/omap3isp/isppreview.c struct isp_device *isp = to_isp_device(prev); prev 1023 drivers/media/platform/omap3isp/isppreview.c prev->params.cfa_order = 0; prev 1026 drivers/media/platform/omap3isp/isppreview.c prev->params.cfa_order = 1; prev 1029 drivers/media/platform/omap3isp/isppreview.c prev->params.cfa_order = 2; prev 1032 drivers/media/platform/omap3isp/isppreview.c prev->params.cfa_order = 3; prev 1045 drivers/media/platform/omap3isp/isppreview.c params = (prev->params.active & OMAP3ISP_PREV_CFA) prev 1046 drivers/media/platform/omap3isp/isppreview.c ? &prev->params.params[0] : &prev->params.params[1]; prev 1048 drivers/media/platform/omap3isp/isppreview.c preview_config_cfa(prev, params); prev 1062 drivers/media/platform/omap3isp/isppreview.c static void preview_config_input_size(struct isp_prev_device *prev, u32 active) prev 1064 drivers/media/platform/omap3isp/isppreview.c const struct v4l2_mbus_framefmt *format = &prev->formats[PREV_PAD_SINK]; prev 1065 drivers/media/platform/omap3isp/isppreview.c struct isp_device *isp = to_isp_device(prev); prev 1066 drivers/media/platform/omap3isp/isppreview.c unsigned int sph = prev->crop.left; prev 1067 drivers/media/platform/omap3isp/isppreview.c unsigned int eph = prev->crop.left + prev->crop.width - 1; prev 1068 drivers/media/platform/omap3isp/isppreview.c unsigned int slv = prev->crop.top; prev 1069 drivers/media/platform/omap3isp/isppreview.c unsigned int elv = prev->crop.top + prev->crop.height - 1; prev 1080 drivers/media/platform/omap3isp/isppreview.c features = (prev->params.params[0].features & active) prev 1081 drivers/media/platform/omap3isp/isppreview.c | (prev->params.params[1].features & ~active); prev 1113 drivers/media/platform/omap3isp/isppreview.c preview_config_inlineoffset(struct isp_prev_device *prev, u32 offset) prev 1115 drivers/media/platform/omap3isp/isppreview.c struct isp_device *isp = to_isp_device(prev); prev 1127 drivers/media/platform/omap3isp/isppreview.c static void preview_set_inaddr(struct isp_prev_device *prev, u32 addr) prev 1129 drivers/media/platform/omap3isp/isppreview.c struct isp_device *isp = to_isp_device(prev); prev 1140 drivers/media/platform/omap3isp/isppreview.c static void preview_config_outlineoffset(struct isp_prev_device *prev, prev 1143 drivers/media/platform/omap3isp/isppreview.c struct isp_device *isp = to_isp_device(prev); prev 1155 drivers/media/platform/omap3isp/isppreview.c static void preview_set_outaddr(struct isp_prev_device *prev, u32 addr) prev 1157 drivers/media/platform/omap3isp/isppreview.c struct isp_device *isp = to_isp_device(prev); prev 1162 drivers/media/platform/omap3isp/isppreview.c static void preview_adjust_bandwidth(struct isp_prev_device *prev) prev 1164 drivers/media/platform/omap3isp/isppreview.c struct isp_pipeline *pipe = to_isp_pipeline(&prev->subdev.entity); prev 1165 drivers/media/platform/omap3isp/isppreview.c struct isp_device *isp = to_isp_device(prev); prev 1166 drivers/media/platform/omap3isp/isppreview.c const struct v4l2_mbus_framefmt *ifmt = &prev->formats[PREV_PAD_SINK]; prev 1176 drivers/media/platform/omap3isp/isppreview.c if (prev->input != PREVIEW_INPUT_MEMORY) { prev 1215 drivers/media/platform/omap3isp/isppreview.c int omap3isp_preview_busy(struct isp_prev_device *prev) prev 1217 drivers/media/platform/omap3isp/isppreview.c struct isp_device *isp = to_isp_device(prev); prev 1228 drivers/media/platform/omap3isp/isppreview.c struct isp_prev_device *prev = &isp->isp_prev; prev 1231 drivers/media/platform/omap3isp/isppreview.c prev->params.params[0].update = prev->params.active & update; prev 1232 drivers/media/platform/omap3isp/isppreview.c prev->params.params[1].update = ~prev->params.active & update; prev 1234 drivers/media/platform/omap3isp/isppreview.c preview_setup_hw(prev, update, prev->params.active); prev 1236 drivers/media/platform/omap3isp/isppreview.c prev->params.params[0].update = 0; prev 1237 drivers/media/platform/omap3isp/isppreview.c prev->params.params[1].update = 0; prev 1247 drivers/media/platform/omap3isp/isppreview.c static void preview_print_status(struct isp_prev_device *prev) prev 1249 drivers/media/platform/omap3isp/isppreview.c struct isp_device *isp = to_isp_device(prev); prev 1297 drivers/media/platform/omap3isp/isppreview.c static void preview_init_params(struct isp_prev_device *prev) prev 1302 drivers/media/platform/omap3isp/isppreview.c spin_lock_init(&prev->params.lock); prev 1304 drivers/media/platform/omap3isp/isppreview.c prev->params.active = ~0; prev 1305 drivers/media/platform/omap3isp/isppreview.c prev->params.params[0].busy = 0; prev 1306 drivers/media/platform/omap3isp/isppreview.c prev->params.params[0].update = OMAP3ISP_PREV_FEATURES_END - 1; prev 1307 drivers/media/platform/omap3isp/isppreview.c prev->params.params[1].busy = 0; prev 1308 drivers/media/platform/omap3isp/isppreview.c prev->params.params[1].update = 0; prev 1310 drivers/media/platform/omap3isp/isppreview.c params = &prev->params.params[0]; prev 1361 drivers/media/platform/omap3isp/isppreview.c static unsigned int preview_max_out_width(struct isp_prev_device *prev) prev 1363 drivers/media/platform/omap3isp/isppreview.c struct isp_device *isp = to_isp_device(prev); prev 1378 drivers/media/platform/omap3isp/isppreview.c static void preview_configure(struct isp_prev_device *prev) prev 1380 drivers/media/platform/omap3isp/isppreview.c struct isp_device *isp = to_isp_device(prev); prev 1387 drivers/media/platform/omap3isp/isppreview.c spin_lock_irqsave(&prev->params.lock, flags); prev 1389 drivers/media/platform/omap3isp/isppreview.c update = preview_params_lock(prev, 0, false); prev 1390 drivers/media/platform/omap3isp/isppreview.c active = prev->params.active; prev 1391 drivers/media/platform/omap3isp/isppreview.c spin_unlock_irqrestore(&prev->params.lock, flags); prev 1394 drivers/media/platform/omap3isp/isppreview.c format = &prev->formats[PREV_PAD_SINK]; prev 1397 drivers/media/platform/omap3isp/isppreview.c preview_adjust_bandwidth(prev); prev 1399 drivers/media/platform/omap3isp/isppreview.c preview_config_input_format(prev, info); prev 1400 drivers/media/platform/omap3isp/isppreview.c preview_config_input_size(prev, active); prev 1402 drivers/media/platform/omap3isp/isppreview.c if (prev->input == PREVIEW_INPUT_CCDC) prev 1403 drivers/media/platform/omap3isp/isppreview.c preview_config_inlineoffset(prev, 0); prev 1405 drivers/media/platform/omap3isp/isppreview.c preview_config_inlineoffset(prev, ALIGN(format->width, 0x20) * prev 1408 drivers/media/platform/omap3isp/isppreview.c preview_setup_hw(prev, update, active); prev 1411 drivers/media/platform/omap3isp/isppreview.c format = &prev->formats[PREV_PAD_SOURCE]; prev 1413 drivers/media/platform/omap3isp/isppreview.c if (prev->output & PREVIEW_OUTPUT_MEMORY) prev 1420 drivers/media/platform/omap3isp/isppreview.c if (prev->output & PREVIEW_OUTPUT_RESIZER) prev 1427 drivers/media/platform/omap3isp/isppreview.c if (prev->output & PREVIEW_OUTPUT_MEMORY) prev 1428 drivers/media/platform/omap3isp/isppreview.c preview_config_outlineoffset(prev, prev 1431 drivers/media/platform/omap3isp/isppreview.c preview_config_averager(prev, 0); prev 1432 drivers/media/platform/omap3isp/isppreview.c preview_config_ycpos(prev, format->code); prev 1434 drivers/media/platform/omap3isp/isppreview.c spin_lock_irqsave(&prev->params.lock, flags); prev 1435 drivers/media/platform/omap3isp/isppreview.c preview_params_unlock(prev, update, false); prev 1436 drivers/media/platform/omap3isp/isppreview.c spin_unlock_irqrestore(&prev->params.lock, flags); prev 1443 drivers/media/platform/omap3isp/isppreview.c static void preview_enable_oneshot(struct isp_prev_device *prev) prev 1445 drivers/media/platform/omap3isp/isppreview.c struct isp_device *isp = to_isp_device(prev); prev 1451 drivers/media/platform/omap3isp/isppreview.c if (prev->input == PREVIEW_INPUT_MEMORY) prev 1459 drivers/media/platform/omap3isp/isppreview.c void omap3isp_preview_isr_frame_sync(struct isp_prev_device *prev) prev 1467 drivers/media/platform/omap3isp/isppreview.c if (prev->state == ISP_PIPELINE_STREAM_CONTINUOUS && prev 1468 drivers/media/platform/omap3isp/isppreview.c prev->video_out.dmaqueue_flags & ISP_VIDEO_DMAQUEUE_QUEUED) { prev 1469 drivers/media/platform/omap3isp/isppreview.c preview_enable_oneshot(prev); prev 1470 drivers/media/platform/omap3isp/isppreview.c isp_video_dmaqueue_flags_clr(&prev->video_out); prev 1474 drivers/media/platform/omap3isp/isppreview.c static void preview_isr_buffer(struct isp_prev_device *prev) prev 1476 drivers/media/platform/omap3isp/isppreview.c struct isp_pipeline *pipe = to_isp_pipeline(&prev->subdev.entity); prev 1480 drivers/media/platform/omap3isp/isppreview.c if (prev->output & PREVIEW_OUTPUT_MEMORY) { prev 1481 drivers/media/platform/omap3isp/isppreview.c buffer = omap3isp_video_buffer_next(&prev->video_out); prev 1483 drivers/media/platform/omap3isp/isppreview.c preview_set_outaddr(prev, buffer->dma); prev 1489 drivers/media/platform/omap3isp/isppreview.c if (prev->input == PREVIEW_INPUT_MEMORY) { prev 1490 drivers/media/platform/omap3isp/isppreview.c buffer = omap3isp_video_buffer_next(&prev->video_in); prev 1492 drivers/media/platform/omap3isp/isppreview.c preview_set_inaddr(prev, buffer->dma); prev 1496 drivers/media/platform/omap3isp/isppreview.c switch (prev->state) { prev 1508 drivers/media/platform/omap3isp/isppreview.c preview_enable_oneshot(prev); prev 1522 drivers/media/platform/omap3isp/isppreview.c void omap3isp_preview_isr(struct isp_prev_device *prev) prev 1528 drivers/media/platform/omap3isp/isppreview.c if (omap3isp_module_sync_is_stopping(&prev->wait, &prev->stopping)) prev 1531 drivers/media/platform/omap3isp/isppreview.c spin_lock_irqsave(&prev->params.lock, flags); prev 1532 drivers/media/platform/omap3isp/isppreview.c preview_params_switch(prev); prev 1533 drivers/media/platform/omap3isp/isppreview.c update = preview_params_lock(prev, 0, false); prev 1534 drivers/media/platform/omap3isp/isppreview.c active = prev->params.active; prev 1535 drivers/media/platform/omap3isp/isppreview.c spin_unlock_irqrestore(&prev->params.lock, flags); prev 1537 drivers/media/platform/omap3isp/isppreview.c preview_setup_hw(prev, update, active); prev 1538 drivers/media/platform/omap3isp/isppreview.c preview_config_input_size(prev, active); prev 1540 drivers/media/platform/omap3isp/isppreview.c if (prev->input == PREVIEW_INPUT_MEMORY || prev 1541 drivers/media/platform/omap3isp/isppreview.c prev->output & PREVIEW_OUTPUT_MEMORY) prev 1542 drivers/media/platform/omap3isp/isppreview.c preview_isr_buffer(prev); prev 1543 drivers/media/platform/omap3isp/isppreview.c else if (prev->state == ISP_PIPELINE_STREAM_CONTINUOUS) prev 1544 drivers/media/platform/omap3isp/isppreview.c preview_enable_oneshot(prev); prev 1546 drivers/media/platform/omap3isp/isppreview.c spin_lock_irqsave(&prev->params.lock, flags); prev 1547 drivers/media/platform/omap3isp/isppreview.c preview_params_unlock(prev, update, false); prev 1548 drivers/media/platform/omap3isp/isppreview.c spin_unlock_irqrestore(&prev->params.lock, flags); prev 1558 drivers/media/platform/omap3isp/isppreview.c struct isp_prev_device *prev = &video->isp->isp_prev; prev 1561 drivers/media/platform/omap3isp/isppreview.c preview_set_inaddr(prev, buffer->dma); prev 1564 drivers/media/platform/omap3isp/isppreview.c preview_set_outaddr(prev, buffer->dma); prev 1583 drivers/media/platform/omap3isp/isppreview.c struct isp_prev_device *prev = prev 1588 drivers/media/platform/omap3isp/isppreview.c preview_update_brightness(prev, ctrl->val); prev 1591 drivers/media/platform/omap3isp/isppreview.c preview_update_contrast(prev, ctrl->val); prev 1611 drivers/media/platform/omap3isp/isppreview.c struct isp_prev_device *prev = v4l2_get_subdevdata(sd); prev 1615 drivers/media/platform/omap3isp/isppreview.c return preview_config(prev, arg); prev 1630 drivers/media/platform/omap3isp/isppreview.c struct isp_prev_device *prev = v4l2_get_subdevdata(sd); prev 1631 drivers/media/platform/omap3isp/isppreview.c struct isp_video *video_out = &prev->video_out; prev 1632 drivers/media/platform/omap3isp/isppreview.c struct isp_device *isp = to_isp_device(prev); prev 1633 drivers/media/platform/omap3isp/isppreview.c struct device *dev = to_device(prev); prev 1635 drivers/media/platform/omap3isp/isppreview.c if (prev->state == ISP_PIPELINE_STREAM_STOPPED) { prev 1640 drivers/media/platform/omap3isp/isppreview.c preview_configure(prev); prev 1641 drivers/media/platform/omap3isp/isppreview.c atomic_set(&prev->stopping, 0); prev 1642 drivers/media/platform/omap3isp/isppreview.c preview_print_status(prev); prev 1647 drivers/media/platform/omap3isp/isppreview.c if (prev->output & PREVIEW_OUTPUT_MEMORY) prev 1651 drivers/media/platform/omap3isp/isppreview.c !(prev->output & PREVIEW_OUTPUT_MEMORY)) prev 1652 drivers/media/platform/omap3isp/isppreview.c preview_enable_oneshot(prev); prev 1658 drivers/media/platform/omap3isp/isppreview.c if (prev->input == PREVIEW_INPUT_MEMORY) prev 1660 drivers/media/platform/omap3isp/isppreview.c if (prev->output & PREVIEW_OUTPUT_MEMORY) prev 1663 drivers/media/platform/omap3isp/isppreview.c preview_enable_oneshot(prev); prev 1667 drivers/media/platform/omap3isp/isppreview.c if (omap3isp_module_sync_idle(&sd->entity, &prev->wait, prev 1668 drivers/media/platform/omap3isp/isppreview.c &prev->stopping)) prev 1677 drivers/media/platform/omap3isp/isppreview.c prev->state = enable; prev 1682 drivers/media/platform/omap3isp/isppreview.c __preview_get_format(struct isp_prev_device *prev, struct v4l2_subdev_pad_config *cfg, prev 1686 drivers/media/platform/omap3isp/isppreview.c return v4l2_subdev_get_try_format(&prev->subdev, cfg, pad); prev 1688 drivers/media/platform/omap3isp/isppreview.c return &prev->formats[pad]; prev 1692 drivers/media/platform/omap3isp/isppreview.c __preview_get_crop(struct isp_prev_device *prev, struct v4l2_subdev_pad_config *cfg, prev 1696 drivers/media/platform/omap3isp/isppreview.c return v4l2_subdev_get_try_crop(&prev->subdev, cfg, PREV_PAD_SINK); prev 1698 drivers/media/platform/omap3isp/isppreview.c return &prev->crop; prev 1731 drivers/media/platform/omap3isp/isppreview.c static void preview_try_format(struct isp_prev_device *prev, prev 1751 drivers/media/platform/omap3isp/isppreview.c if (prev->input == PREVIEW_INPUT_MEMORY) { prev 1753 drivers/media/platform/omap3isp/isppreview.c preview_max_out_width(prev)); prev 1773 drivers/media/platform/omap3isp/isppreview.c *fmt = *__preview_get_format(prev, cfg, PREV_PAD_SINK, which); prev 1791 drivers/media/platform/omap3isp/isppreview.c crop = __preview_get_crop(prev, cfg, which); prev 1814 drivers/media/platform/omap3isp/isppreview.c static void preview_try_crop(struct isp_prev_device *prev, prev 1828 drivers/media/platform/omap3isp/isppreview.c if (prev->input == PREVIEW_INPUT_CCDC) { prev 1892 drivers/media/platform/omap3isp/isppreview.c struct isp_prev_device *prev = v4l2_get_subdevdata(sd); prev 1901 drivers/media/platform/omap3isp/isppreview.c preview_try_format(prev, cfg, fse->pad, &format, fse->which); prev 1911 drivers/media/platform/omap3isp/isppreview.c preview_try_format(prev, cfg, fse->pad, &format, fse->which); prev 1932 drivers/media/platform/omap3isp/isppreview.c struct isp_prev_device *prev = v4l2_get_subdevdata(sd); prev 1945 drivers/media/platform/omap3isp/isppreview.c format = __preview_get_format(prev, cfg, PREV_PAD_SINK, prev 1947 drivers/media/platform/omap3isp/isppreview.c preview_try_crop(prev, format, &sel->r); prev 1951 drivers/media/platform/omap3isp/isppreview.c sel->r = *__preview_get_crop(prev, cfg, sel->which); prev 1975 drivers/media/platform/omap3isp/isppreview.c struct isp_prev_device *prev = v4l2_get_subdevdata(sd); prev 1983 drivers/media/platform/omap3isp/isppreview.c if (prev->state != ISP_PIPELINE_STREAM_STOPPED) prev 1991 drivers/media/platform/omap3isp/isppreview.c sel->r = *__preview_get_crop(prev, cfg, sel->which); prev 1995 drivers/media/platform/omap3isp/isppreview.c format = __preview_get_format(prev, cfg, PREV_PAD_SINK, sel->which); prev 1996 drivers/media/platform/omap3isp/isppreview.c preview_try_crop(prev, format, &sel->r); prev 1997 drivers/media/platform/omap3isp/isppreview.c *__preview_get_crop(prev, cfg, sel->which) = sel->r; prev 2000 drivers/media/platform/omap3isp/isppreview.c format = __preview_get_format(prev, cfg, PREV_PAD_SOURCE, sel->which); prev 2001 drivers/media/platform/omap3isp/isppreview.c preview_try_format(prev, cfg, PREV_PAD_SOURCE, format, sel->which); prev 2016 drivers/media/platform/omap3isp/isppreview.c struct isp_prev_device *prev = v4l2_get_subdevdata(sd); prev 2019 drivers/media/platform/omap3isp/isppreview.c format = __preview_get_format(prev, cfg, fmt->pad, fmt->which); prev 2037 drivers/media/platform/omap3isp/isppreview.c struct isp_prev_device *prev = v4l2_get_subdevdata(sd); prev 2041 drivers/media/platform/omap3isp/isppreview.c format = __preview_get_format(prev, cfg, fmt->pad, fmt->which); prev 2045 drivers/media/platform/omap3isp/isppreview.c preview_try_format(prev, cfg, fmt->pad, &fmt->format, fmt->which); prev 2051 drivers/media/platform/omap3isp/isppreview.c crop = __preview_get_crop(prev, cfg, fmt->which); prev 2057 drivers/media/platform/omap3isp/isppreview.c preview_try_crop(prev, &fmt->format, crop); prev 2060 drivers/media/platform/omap3isp/isppreview.c format = __preview_get_format(prev, cfg, PREV_PAD_SOURCE, prev 2062 drivers/media/platform/omap3isp/isppreview.c preview_try_format(prev, cfg, PREV_PAD_SOURCE, format, prev 2143 drivers/media/platform/omap3isp/isppreview.c struct isp_prev_device *prev = v4l2_get_subdevdata(sd); prev 2154 drivers/media/platform/omap3isp/isppreview.c if (prev->input == PREVIEW_INPUT_CCDC) prev 2156 drivers/media/platform/omap3isp/isppreview.c prev->input = PREVIEW_INPUT_MEMORY; prev 2158 drivers/media/platform/omap3isp/isppreview.c if (prev->input == PREVIEW_INPUT_MEMORY) prev 2159 drivers/media/platform/omap3isp/isppreview.c prev->input = PREVIEW_INPUT_NONE; prev 2166 drivers/media/platform/omap3isp/isppreview.c if (prev->input == PREVIEW_INPUT_MEMORY) prev 2168 drivers/media/platform/omap3isp/isppreview.c prev->input = PREVIEW_INPUT_CCDC; prev 2170 drivers/media/platform/omap3isp/isppreview.c if (prev->input == PREVIEW_INPUT_CCDC) prev 2171 drivers/media/platform/omap3isp/isppreview.c prev->input = PREVIEW_INPUT_NONE; prev 2183 drivers/media/platform/omap3isp/isppreview.c if (prev->output & ~PREVIEW_OUTPUT_MEMORY) prev 2185 drivers/media/platform/omap3isp/isppreview.c prev->output |= PREVIEW_OUTPUT_MEMORY; prev 2187 drivers/media/platform/omap3isp/isppreview.c prev->output &= ~PREVIEW_OUTPUT_MEMORY; prev 2194 drivers/media/platform/omap3isp/isppreview.c if (prev->output & ~PREVIEW_OUTPUT_RESIZER) prev 2196 drivers/media/platform/omap3isp/isppreview.c prev->output |= PREVIEW_OUTPUT_RESIZER; prev 2198 drivers/media/platform/omap3isp/isppreview.c prev->output &= ~PREVIEW_OUTPUT_RESIZER; prev 2215 drivers/media/platform/omap3isp/isppreview.c void omap3isp_preview_unregister_entities(struct isp_prev_device *prev) prev 2217 drivers/media/platform/omap3isp/isppreview.c v4l2_device_unregister_subdev(&prev->subdev); prev 2218 drivers/media/platform/omap3isp/isppreview.c omap3isp_video_unregister(&prev->video_in); prev 2219 drivers/media/platform/omap3isp/isppreview.c omap3isp_video_unregister(&prev->video_out); prev 2222 drivers/media/platform/omap3isp/isppreview.c int omap3isp_preview_register_entities(struct isp_prev_device *prev, prev 2228 drivers/media/platform/omap3isp/isppreview.c prev->subdev.dev = vdev->mdev->dev; prev 2229 drivers/media/platform/omap3isp/isppreview.c ret = v4l2_device_register_subdev(vdev, &prev->subdev); prev 2233 drivers/media/platform/omap3isp/isppreview.c ret = omap3isp_video_register(&prev->video_in, vdev); prev 2237 drivers/media/platform/omap3isp/isppreview.c ret = omap3isp_video_register(&prev->video_out, vdev); prev 2244 drivers/media/platform/omap3isp/isppreview.c omap3isp_preview_unregister_entities(prev); prev 2257 drivers/media/platform/omap3isp/isppreview.c static int preview_init_entities(struct isp_prev_device *prev) prev 2259 drivers/media/platform/omap3isp/isppreview.c struct v4l2_subdev *sd = &prev->subdev; prev 2260 drivers/media/platform/omap3isp/isppreview.c struct media_pad *pads = prev->pads; prev 2264 drivers/media/platform/omap3isp/isppreview.c prev->input = PREVIEW_INPUT_NONE; prev 2270 drivers/media/platform/omap3isp/isppreview.c v4l2_set_subdevdata(sd, prev); prev 2273 drivers/media/platform/omap3isp/isppreview.c v4l2_ctrl_handler_init(&prev->ctrls, 2); prev 2274 drivers/media/platform/omap3isp/isppreview.c v4l2_ctrl_new_std(&prev->ctrls, &preview_ctrl_ops, V4L2_CID_BRIGHTNESS, prev 2277 drivers/media/platform/omap3isp/isppreview.c v4l2_ctrl_new_std(&prev->ctrls, &preview_ctrl_ops, V4L2_CID_CONTRAST, prev 2280 drivers/media/platform/omap3isp/isppreview.c v4l2_ctrl_handler_setup(&prev->ctrls); prev 2281 drivers/media/platform/omap3isp/isppreview.c sd->ctrl_handler = &prev->ctrls; prev 2298 drivers/media/platform/omap3isp/isppreview.c prev->video_in.type = V4L2_BUF_TYPE_VIDEO_OUTPUT; prev 2299 drivers/media/platform/omap3isp/isppreview.c prev->video_in.ops = &preview_video_ops; prev 2300 drivers/media/platform/omap3isp/isppreview.c prev->video_in.isp = to_isp_device(prev); prev 2301 drivers/media/platform/omap3isp/isppreview.c prev->video_in.capture_mem = PAGE_ALIGN(4096 * 4096) * 2 * 3; prev 2302 drivers/media/platform/omap3isp/isppreview.c prev->video_in.bpl_alignment = 64; prev 2303 drivers/media/platform/omap3isp/isppreview.c prev->video_out.type = V4L2_BUF_TYPE_VIDEO_CAPTURE; prev 2304 drivers/media/platform/omap3isp/isppreview.c prev->video_out.ops = &preview_video_ops; prev 2305 drivers/media/platform/omap3isp/isppreview.c prev->video_out.isp = to_isp_device(prev); prev 2306 drivers/media/platform/omap3isp/isppreview.c prev->video_out.capture_mem = PAGE_ALIGN(4096 * 4096) * 2 * 3; prev 2307 drivers/media/platform/omap3isp/isppreview.c prev->video_out.bpl_alignment = 32; prev 2309 drivers/media/platform/omap3isp/isppreview.c ret = omap3isp_video_init(&prev->video_in, "preview"); prev 2313 drivers/media/platform/omap3isp/isppreview.c ret = omap3isp_video_init(&prev->video_out, "preview"); prev 2320 drivers/media/platform/omap3isp/isppreview.c omap3isp_video_cleanup(&prev->video_in); prev 2322 drivers/media/platform/omap3isp/isppreview.c media_entity_cleanup(&prev->subdev.entity); prev 2333 drivers/media/platform/omap3isp/isppreview.c struct isp_prev_device *prev = &isp->isp_prev; prev 2335 drivers/media/platform/omap3isp/isppreview.c init_waitqueue_head(&prev->wait); prev 2337 drivers/media/platform/omap3isp/isppreview.c preview_init_params(prev); prev 2339 drivers/media/platform/omap3isp/isppreview.c return preview_init_entities(prev); prev 2344 drivers/media/platform/omap3isp/isppreview.c struct isp_prev_device *prev = &isp->isp_prev; prev 2346 drivers/media/platform/omap3isp/isppreview.c v4l2_ctrl_handler_free(&prev->ctrls); prev 2347 drivers/media/platform/omap3isp/isppreview.c omap3isp_video_cleanup(&prev->video_in); prev 2348 drivers/media/platform/omap3isp/isppreview.c omap3isp_video_cleanup(&prev->video_out); prev 2349 drivers/media/platform/omap3isp/isppreview.c media_entity_cleanup(&prev->subdev.entity); prev 154 drivers/media/platform/omap3isp/isppreview.h void omap3isp_preview_isr_frame_sync(struct isp_prev_device *prev); prev 155 drivers/media/platform/omap3isp/isppreview.h void omap3isp_preview_isr(struct isp_prev_device *prev); prev 1089 drivers/media/platform/pxa_camera.c last_buf = list_entry(pcdev->capture.prev, prev 830 drivers/media/platform/s5p-mfc/s5p_mfc_dec.c buf = list_entry(ctx->src_queue.prev, prev 2294 drivers/media/platform/s5p-mfc/s5p_mfc_enc.c buf = list_entry(ctx->src_queue.prev, prev 125 drivers/media/platform/sti/hva/hva-debugfs.c ktime_t prev = dbg->begin; prev 131 drivers/media/platform/sti/hva/hva-debugfs.c div = (u64)ktime_us_delta(dbg->begin, prev); prev 1577 drivers/media/platform/ti-vpe/cal.c struct device_node *prev) prev 1584 drivers/media/platform/ti-vpe/cal.c if (!prev) { prev 1601 drivers/media/platform/ti-vpe/cal.c ports = of_get_parent(prev); prev 1606 drivers/media/platform/ti-vpe/cal.c port = of_get_next_child(ports, prev); prev 1611 drivers/media/platform/ti-vpe/cal.c prev = port; prev 1621 drivers/media/platform/ti-vpe/cal.c struct device_node *prev) prev 1629 drivers/media/platform/ti-vpe/cal.c ep = of_get_next_child(parent, prev); prev 1632 drivers/media/platform/ti-vpe/cal.c prev = ep; prev 66 drivers/media/platform/vsp1/vsp1_drm.c struct vsp1_entity *prev, unsigned int prev_pad, prev 77 drivers/media/platform/vsp1/vsp1_drm.c prev->sink = next; prev 78 drivers/media/platform/vsp1/vsp1_drm.c prev->sink_pad = next_pad; prev 82 drivers/media/platform/vsp1/vsp1_drm.c prev->sink = uif; prev 83 drivers/media/platform/vsp1/vsp1_drm.c prev->sink_pad = UIF_PAD_SINK; prev 89 drivers/media/platform/vsp1/vsp1_drm.c ret = v4l2_subdev_call(&prev->subdev, pad, get_fmt, NULL, &format); prev 92 drivers/media/usb/hdpvr/hdpvr-video.c list_splice_init(&dev->rec_buff_list, dev->free_buff_list.prev); prev 1496 drivers/media/usb/uvc/uvc_driver.c struct uvc_entity *entity, struct uvc_entity *prev) prev 1510 drivers/media/usb/uvc/uvc_driver.c if (forward == prev) prev 1512 drivers/media/usb/uvc/uvc_driver.c if (forward->chain.next || forward->chain.prev) { prev 1598 drivers/media/usb/uvc/uvc_driver.c if (term->chain.next || term->chain.prev) { prev 1648 drivers/media/usb/uvc/uvc_driver.c struct uvc_entity *entity, *prev; prev 1653 drivers/media/usb/uvc/uvc_driver.c prev = NULL; prev 1657 drivers/media/usb/uvc/uvc_driver.c if (entity->chain.next || entity->chain.prev) { prev 1668 drivers/media/usb/uvc/uvc_driver.c if (uvc_scan_chain_forward(chain, entity, prev) < 0) prev 1672 drivers/media/usb/uvc/uvc_driver.c prev = entity; prev 1751 drivers/media/usb/uvc/uvc_driver.c struct uvc_entity *prev; prev 1782 drivers/media/usb/uvc/uvc_driver.c prev = oterm; prev 1802 drivers/media/usb/uvc/uvc_driver.c prev->baSourceID[0] = entity->id; prev 1803 drivers/media/usb/uvc/uvc_driver.c prev = entity; prev 1809 drivers/media/usb/uvc/uvc_driver.c prev->baSourceID[0] = iterm->id; prev 1843 drivers/media/usb/uvc/uvc_driver.c if (term->chain.next || term->chain.prev) prev 2310 drivers/media/v4l2-core/v4l2-ctrls.c if (list_empty(&hdl->ctrl_refs) || id > node2id(hdl->ctrl_refs.prev)) { prev 2324 drivers/media/v4l2-core/v4l2-ctrls.c list_add(&new_ref->node, ref->node.prev); prev 2966 drivers/media/v4l2-core/v4l2-ctrls.c if (id >= node2id(hdl->ctrl_refs.prev)) { prev 407 drivers/misc/habanalabs/memory.c struct hl_vm_va_block *prev, *next; prev 409 drivers/misc/habanalabs/memory.c prev = list_prev_entry(va_block, node); prev 410 drivers/misc/habanalabs/memory.c if (&prev->node != va_list && prev->end + 1 == va_block->start) { prev 411 drivers/misc/habanalabs/memory.c prev->end = va_block->end; prev 412 drivers/misc/habanalabs/memory.c prev->size = prev->end - prev->start; prev 415 drivers/misc/habanalabs/memory.c va_block = prev; prev 22 drivers/misc/mic/scif/scif_rma_list.c struct scif_window *prev = list_entry(head, struct scif_window, list); prev 28 drivers/misc/mic/scif/scif_rma_list.c curr = list_entry(head->prev, struct scif_window, list); prev 38 drivers/misc/mic/scif/scif_rma_list.c prev = curr; prev 40 drivers/misc/mic/scif/scif_rma_list.c list_add(&window->list, &prev->list); prev 51 drivers/misc/mic/scif/scif_rma_list.c struct scif_window *curr = NULL, *prev = NULL; prev 59 drivers/misc/mic/scif/scif_rma_list.c prev = curr; prev 61 drivers/misc/mic/scif/scif_rma_list.c if (!prev) prev 64 drivers/misc/mic/scif/scif_rma_list.c list_add(&window->list, &prev->list); prev 233 drivers/mmc/core/sdio_cis.c struct sdio_func_tuple *this, **prev; prev 257 drivers/mmc/core/sdio_cis.c prev = &func->tuples; prev 259 drivers/mmc/core/sdio_cis.c prev = &card->tuples; prev 261 drivers/mmc/core/sdio_cis.c if (*prev) prev 314 drivers/mmc/core/sdio_cis.c *prev = this; prev 315 drivers/mmc/core/sdio_cis.c prev = &this->next; prev 344 drivers/mmc/core/sdio_cis.c *prev = card->tuples; prev 1795 drivers/mtd/nand/onenand/onenand_base.c int prev = 0, prevlen = 0, prev_subpage = 0, first = 1; prev 1879 drivers/mtd/nand/onenand/onenand_base.c onenand_update_bufferram(mtd, prev, !ret && !prev_subpage); prev 1945 drivers/mtd/nand/onenand/onenand_base.c prev = to; prev 189 drivers/mtd/parsers/redboot.c struct fis_list *new_fl, **prev; prev 216 drivers/mtd/parsers/redboot.c prev = &fl; prev 217 drivers/mtd/parsers/redboot.c while(*prev && (*prev)->img->flash_base < new_fl->img->flash_base) prev 218 drivers/mtd/parsers/redboot.c prev = &(*prev)->next; prev 219 drivers/mtd/parsers/redboot.c new_fl->next = *prev; prev 220 drivers/mtd/parsers/redboot.c *prev = new_fl; prev 566 drivers/mtd/spi-nor/aspeed-smc.c u32 prev = readl(SEGMENT_ADDR_REG(controller, chip->cs - 1)); prev 568 drivers/mtd/spi-nor/aspeed-smc.c start = SEGMENT_ADDR_END(prev); prev 84 drivers/net/bonding/bond_alb.c entry->prev = TLB_NULL_INDEX; prev 209 drivers/net/bonding/bond_alb.c hash_table[hash_index].prev = TLB_NULL_INDEX; prev 212 drivers/net/bonding/bond_alb.c hash_table[next_index].prev = hash_index; prev 343 drivers/net/ethernet/brocade/bna/bnad.c struct bnad_rx_unmap *unmap, *prev; prev 363 drivers/net/ethernet/brocade/bna/bnad.c prev = &unmap_q->unmap[unmap_q->reuse_pi]; prev 364 drivers/net/ethernet/brocade/bna/bnad.c page = prev->page; prev 365 drivers/net/ethernet/brocade/bna/bnad.c page_offset = prev->page_offset + unmap_q->map_size; prev 615 drivers/net/ethernet/cavium/liquidio/octeon_network.h if (root->prev == root && root->next == root) prev 840 drivers/net/ethernet/cisco/enic/enic_main.c buf = wq->to_use->prev; prev 847 drivers/net/ethernet/cisco/enic/enic_main.c buf = buf->prev; prev 51 drivers/net/ethernet/cisco/enic/vnic_wq.c buf->next->prev = buf; prev 55 drivers/net/ethernet/cisco/enic/vnic_wq.c buf->next->prev = buf; prev 58 drivers/net/ethernet/cisco/enic/vnic_wq.c buf->next->prev = buf; prev 65 drivers/net/ethernet/cisco/enic/vnic_wq.h struct vnic_wq_buf *prev; prev 56 drivers/net/ethernet/freescale/dpaa/dpaa_eth_sysfs.c struct dpaa_fq *prev = NULL; prev 93 drivers/net/ethernet/freescale/dpaa/dpaa_eth_sysfs.c if (prev && (abs(fq->fqid - prev->fqid) != 1 || prev 97 drivers/net/ethernet/freescale/dpaa/dpaa_eth_sysfs.c "%s: %d\n", prevstr, prev->fqid); prev 104 drivers/net/ethernet/freescale/dpaa/dpaa_eth_sysfs.c if (prev && abs(fq->fqid - prev->fqid) == 1 && prev 112 drivers/net/ethernet/freescale/dpaa/dpaa_eth_sysfs.c prev = fq; prev 117 drivers/net/ethernet/freescale/dpaa/dpaa_eth_sysfs.c if (prev) { prev 120 drivers/net/ethernet/freescale/dpaa/dpaa_eth_sysfs.c prev->fqid); prev 1250 drivers/net/ethernet/i825xx/lib82596.c struct i596_cmd *prev = ptr; prev 1254 drivers/net/ethernet/i825xx/lib82596.c DMA_WBACK_INV(dev, prev, sizeof(struct i596_cmd)); prev 420 drivers/net/ethernet/intel/e100.c struct rx *next, *prev; prev 499 drivers/net/ethernet/intel/e100.c struct cb *next, *prev; prev 878 drivers/net/ethernet/intel/e100.c cb->prev->command &= cpu_to_le16(~cb_s); prev 1672 drivers/net/ethernet/intel/e100.c u32 prev = nic->adaptive_ifs; prev 1683 drivers/net/ethernet/intel/e100.c if (nic->adaptive_ifs != prev) prev 1895 drivers/net/ethernet/intel/e100.c cb->prev = (i == 0) ? nic->cbs + count - 1 : cb - 1; prev 1944 drivers/net/ethernet/intel/e100.c if (rx->prev->skb) { prev 1945 drivers/net/ethernet/intel/e100.c struct rfd *prev_rfd = (struct rfd *)rx->prev->skb->data; prev 1947 drivers/net/ethernet/intel/e100.c pci_dma_sync_single_for_device(nic->pdev, rx->prev->dma_addr, prev 2079 drivers/net/ethernet/intel/e100.c old_before_last_rx = nic->rx_to_use->prev->prev; prev 2088 drivers/net/ethernet/intel/e100.c new_before_last_rx = nic->rx_to_use->prev->prev; prev 2166 drivers/net/ethernet/intel/e100.c rx->prev = (i == 0) ? nic->rxs + count - 1 : rx - 1; prev 2179 drivers/net/ethernet/intel/e100.c rx = nic->rxs->prev->prev; prev 661 drivers/net/ethernet/intel/ice/ice_sched.c struct ice_sched_node *prev, *new_node; prev 726 drivers/net/ethernet/intel/ice/ice_sched.c prev = ice_sched_get_first_node(pi, tc_node, layer); prev 727 drivers/net/ethernet/intel/ice/ice_sched.c if (prev && prev != new_node) { prev 728 drivers/net/ethernet/intel/ice/ice_sched.c while (prev->sibling) prev 729 drivers/net/ethernet/intel/ice/ice_sched.c prev = prev->sibling; prev 730 drivers/net/ethernet/intel/ice/ice_sched.c prev->sibling = new_node; prev 697 drivers/net/ethernet/mellanox/mlx4/mcg.c int *prev, int *index) prev 723 drivers/net/ethernet/mellanox/mlx4/mcg.c *prev = -1; prev 742 drivers/net/ethernet/mellanox/mlx4/mcg.c *prev = *index; prev 1112 drivers/net/ethernet/mellanox/mlx4/mcg.c int index = -1, prev; prev 1126 drivers/net/ethernet/mellanox/mlx4/mcg.c mailbox, &prev, &index); prev 1180 drivers/net/ethernet/mellanox/mlx4/mcg.c err = mlx4_READ_ENTRY(dev, prev, mailbox); prev 1186 drivers/net/ethernet/mellanox/mlx4/mcg.c err = mlx4_WRITE_ENTRY(dev, prev, mailbox); prev 1221 drivers/net/ethernet/mellanox/mlx4/mcg.c int prev, index; prev 1235 drivers/net/ethernet/mellanox/mlx4/mcg.c mailbox, &prev, &index); prev 1282 drivers/net/ethernet/mellanox/mlx4/mcg.c if (prev == -1) { prev 1307 drivers/net/ethernet/mellanox/mlx4/mcg.c err = mlx4_READ_ENTRY(dev, prev, mailbox); prev 1313 drivers/net/ethernet/mellanox/mlx4/mcg.c err = mlx4_WRITE_ENTRY(dev, prev, mailbox); prev 1319 drivers/net/ethernet/mellanox/mlx4/mcg.c prev, index, dev->caps.num_mgms); prev 316 drivers/net/ethernet/mellanox/mlx5/core/en_main.c struct mlx5e_wqe_frag_info *prev = NULL; prev 331 drivers/net/ethernet/mellanox/mlx5/core/en_main.c if (prev) prev 332 drivers/net/ethernet/mellanox/mlx5/core/en_main.c prev->last_in_page = true; prev 338 drivers/net/ethernet/mellanox/mlx5/core/en_main.c prev = frag; prev 342 drivers/net/ethernet/mellanox/mlx5/core/en_main.c if (prev) prev 343 drivers/net/ethernet/mellanox/mlx5/core/en_main.c prev->last_in_page = true; prev 679 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c struct list_head *prev) prev 702 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c list_add(&fg->node.list, prev); prev 992 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c struct list_head *prev = &prio->node.children; prev 998 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c prev = &iter->node.list; prev 1000 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c list_add(&ft->node.list, prev); prev 1158 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c ft->node.children.prev); prev 1324 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c struct list_head *prev = &ft->node.children; prev 1345 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c prev = &fg->node.list; prev 1356 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c prev); prev 723 drivers/net/ethernet/mellanox/mlx5/core/health.c if (count == health->prev) prev 728 drivers/net/ethernet/mellanox/mlx5/core/health.c health->prev = count; prev 722 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c if (matcher->matcher_list.prev == &tbl->matcher_list) prev 91 drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c struct mlxsw_afa_set *prev; /* Pointer to the previous set, prev 218 drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c set->prev = NULL; prev 335 drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c block->cur_set->prev = block->first_set; prev 380 drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c prev_set = set->prev; prev 651 drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c set->prev = block->cur_set; prev 612 drivers/net/ethernet/mscc/ocelot_ace.c list_add(&rule->list, pos->prev); prev 44 drivers/net/ethernet/netronome/nfp/bpf/jit.c return meta->l.prev != &nfp_prog->insns; prev 2581 drivers/net/ethernet/netronome/nfp/bpf/jit.c struct nfp_insn_meta *prev = nfp_meta_prev(meta); prev 2585 drivers/net/ethernet/netronome/nfp/bpf/jit.c dst = prev->insn.dst_reg * 2; prev 2586 drivers/net/ethernet/netronome/nfp/bpf/jit.c imm_lo = prev->insn.imm; prev 857 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_cppcore.c cache = list_entry(cpp->area_cache_list.prev, prev 146 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c u32 cur, prev; prev 148 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c prev = adapter->ahw->idc.prev_state; prev 153 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c adapter->ahw->idc.name[prev]); prev 984 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c u32 cur, prev, next; prev 987 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c prev = adapter->ahw->idc.prev_state; prev 994 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c __func__, cur, prev, state); prev 999 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c (prev == QLC_83XX_IDC_DEV_UNKNOWN)) { prev 1004 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c __func__, cur, prev, next); prev 1010 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c if ((prev != QLC_83XX_IDC_DEV_INIT) && prev 1011 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c (prev != QLC_83XX_IDC_DEV_COLD) && prev 1012 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c (prev != QLC_83XX_IDC_DEV_NEED_RESET)) { prev 1015 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c __func__, cur, prev, next); prev 233 drivers/net/ethernet/sfc/falcon/mdio_10g.c struct ethtool_link_ksettings prev = { prev 239 drivers/net/ethernet/sfc/falcon/mdio_10g.c efx->phy_op->get_link_ksettings(efx, &prev); prev 244 drivers/net/ethernet/sfc/falcon/mdio_10g.c prev.link_modes.advertising); prev 246 drivers/net/ethernet/sfc/falcon/mdio_10g.c prev.link_modes.supported); prev 249 drivers/net/ethernet/sfc/falcon/mdio_10g.c cmd->base.speed == prev.base.speed && prev 250 drivers/net/ethernet/sfc/falcon/mdio_10g.c cmd->base.duplex == prev.base.duplex && prev 251 drivers/net/ethernet/sfc/falcon/mdio_10g.c cmd->base.port == prev.base.port && prev 252 drivers/net/ethernet/sfc/falcon/mdio_10g.c cmd->base.autoneg == prev.base.autoneg) prev 256 drivers/net/ethernet/sfc/falcon/mdio_10g.c if (prev.base.port != PORT_TP || cmd->base.port != PORT_TP) prev 182 drivers/net/ethernet/ti/cpmac.c struct cpmac_desc *prev; prev 452 drivers/net/ethernet/ti/cpmac.c desc->prev->hw_next = (u32)0; prev 453 drivers/net/ethernet/ti/cpmac.c priv->rx_head->prev->hw_next = priv->rx_head->mapping; prev 465 drivers/net/ethernet/ti/cpmac.c (priv->rx_head->prev->dataflags & (CPMAC_OWN|CPMAC_EOQ)) prev 471 drivers/net/ethernet/ti/cpmac.c priv->rx_head->prev->dataflags &= ~CPMAC_EOQ; prev 696 drivers/net/ethernet/ti/cpmac.c priv->rx_head->prev->hw_next = 0; prev 963 drivers/net/ethernet/ti/cpmac.c desc->next->prev = desc; prev 967 drivers/net/ethernet/ti/cpmac.c priv->rx_head->prev->hw_next = (u32)0; prev 984 drivers/net/ethernet/ti/davinci_cpdma.c struct cpdma_desc __iomem *prev = chan->tail; prev 1002 drivers/net/ethernet/ti/davinci_cpdma.c desc_write(prev, hw_next, desc_dma); prev 1007 drivers/net/ethernet/ti/davinci_cpdma.c mode = desc_read(prev, hw_mode); prev 1010 drivers/net/ethernet/ti/davinci_cpdma.c desc_write(prev, hw_mode, mode & ~CPDMA_DESC_EOQ); prev 491 drivers/net/ethernet/ti/netcp_core.c __list_add(&entry->list, next->list.prev, &next->list); prev 540 drivers/net/ethernet/ti/netcp_core.c __list_add(&entry->list, next->list.prev, &next->list); prev 330 drivers/net/ethernet/toshiba/ps3_gelic_net.c descr->prev = descr - 1; prev 334 drivers/net/ethernet/toshiba/ps3_gelic_net.c start_descr->prev = (descr - 1); prev 482 drivers/net/ethernet/toshiba/ps3_gelic_net.c chain->tail = card->rx_top->prev; /* point to the last */ prev 873 drivers/net/ethernet/toshiba/ps3_gelic_net.c descr->prev->next_descr_addr = cpu_to_be32(descr->bus_addr); prev 891 drivers/net/ethernet/toshiba/ps3_gelic_net.c descr->prev->next_descr_addr = 0; prev 1065 drivers/net/ethernet/toshiba/ps3_gelic_net.c descr->prev->next_descr_addr = cpu_to_be32(descr->bus_addr); prev 247 drivers/net/ethernet/toshiba/ps3_gelic_net.h struct gelic_descr *prev; prev 330 drivers/net/ethernet/toshiba/spider_net.c descr->prev = descr - 1; prev 336 drivers/net/ethernet/toshiba/spider_net.c chain->ring->prev = descr-1; prev 522 drivers/net/ethernet/toshiba/spider_net.c descr->prev->hwdescr->next_descr_addr = descr->bus_addr; prev 663 drivers/net/ethernet/toshiba/spider_net.c if (descr->next == chain->tail->prev) { prev 693 drivers/net/ethernet/toshiba/spider_net.c descr->prev->hwdescr->next_descr_addr = descr->bus_addr; prev 389 drivers/net/ethernet/toshiba/spider_net.h struct spider_net_descr *prev; prev 2521 drivers/net/ethernet/via/via-velocity.c int index, prev; prev 2593 drivers/net/ethernet/via/via-velocity.c prev = index - 1; prev 2594 drivers/net/ethernet/via/via-velocity.c if (prev < 0) prev 2595 drivers/net/ethernet/via/via-velocity.c prev = vptr->options.numtx - 1; prev 2603 drivers/net/ethernet/via/via-velocity.c td_ptr = &(vptr->tx.rings[qnum][prev]); prev 37 drivers/net/fddi/skfp/smttimer.c struct smt_timer **prev ; prev 47 drivers/net/fddi/skfp/smttimer.c for (prev = &smc->t.st_queue ; (tm = *prev) ; prev = &tm->tm_next ) { prev 49 drivers/net/fddi/skfp/smttimer.c *prev = tm->tm_next ; prev 61 drivers/net/fddi/skfp/smttimer.c struct smt_timer **prev ; prev 88 drivers/net/fddi/skfp/smttimer.c for (prev = &smc->t.st_queue ; (tm = *prev) ; prev = &tm->tm_next ) { prev 95 drivers/net/fddi/skfp/smttimer.c *prev = timer ; prev 1506 drivers/net/wan/cosa.c int i=0, id=0, prev=0, curr=0; prev 1524 drivers/net/wan/cosa.c for (i=0; i<COSA_MAX_ID_STRING-1; i++, prev=curr) { prev 1531 drivers/net/wan/cosa.c if (curr == 0x2e && prev == '\n') prev 336 drivers/net/wireless/ath/ath9k/channel.c struct ath_chanctx *prev, *cur; prev 344 drivers/net/wireless/ath/ath9k/channel.c prev = ath_chanctx_get_next(sc, cur); prev 346 drivers/net/wireless/ath/ath9k/channel.c if (!prev->switch_after_beacon) prev 353 drivers/net/wireless/ath/ath9k/channel.c prev_tsf = prev->last_beacon - (u32) prev->tsf_val + cur_tsf; prev 354 drivers/net/wireless/ath/ath9k/channel.c prev_tsf -= ath9k_hw_get_tsf_offset(&prev->tsf_ts, &ts); prev 366 drivers/net/wireless/ath/ath9k/channel.c prev->tsf_val += offset; prev 350 drivers/net/wireless/ath/ath9k/xmit.c int prev = fi->retries; prev 355 drivers/net/wireless/ath/ath9k/xmit.c if (prev > 0) prev 1995 drivers/net/wireless/ath/ath9k/xmit.c bf_last = list_entry(head->prev, struct ath_buf, list); prev 2616 drivers/net/wireless/ath/ath9k/xmit.c &txq->axq_q, lastbf->list.prev); prev 2718 drivers/net/wireless/ath/ath9k/xmit.c lastbf->list.prev); prev 166 drivers/net/wireless/ath/dfs_pri_detector.c return list_entry(l->prev, struct pulse_elem, head); prev 437 drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c struct brcmf_flowring_tdls_entry *prev; prev 446 drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c prev = NULL; prev 452 drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c prev = search; prev 469 drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c if (prev) prev 470 drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c prev->next = search->next; prev 1312 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c __skb_insert(p, p_tail->prev, p_tail, queue); prev 1676 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c pfirst->prev); prev 2240 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c pkt_prev = pkt_next->prev; prev 287 drivers/net/wireless/intel/iwlwifi/dvm/rx.c static void accum_stats(__le32 *prev, __le32 *cur, __le32 *delta, prev 294 drivers/net/wireless/intel/iwlwifi/dvm/rx.c i++, prev++, cur++, delta++, max_delta++, accum++) { prev 295 drivers/net/wireless/intel/iwlwifi/dvm/rx.c if (le32_to_cpu(*cur) > le32_to_cpu(*prev)) { prev 297 drivers/net/wireless/intel/iwlwifi/dvm/rx.c le32_to_cpu(*cur) - le32_to_cpu(*prev)); prev 875 drivers/net/wireless/intel/iwlwifi/mvm/tx.c tmp->prev = NULL; prev 803 drivers/net/wireless/intel/iwlwifi/mvm/utils.c bool prev; prev 807 drivers/net/wireless/intel/iwlwifi/mvm/utils.c prev = iwl_mvm_vif_low_latency(mvmvif); prev 812 drivers/net/wireless/intel/iwlwifi/mvm/utils.c if (low_latency == prev) prev 1069 drivers/net/wireless/intel/iwlwifi/mvm/utils.c bool low_latency, prev = mvmvif->low_latency & LOW_LATENCY_TRAFFIC; prev 1077 drivers/net/wireless/intel/iwlwifi/mvm/utils.c prev == low_latency) { prev 1082 drivers/net/wireless/intel/iwlwifi/mvm/utils.c if (prev != low_latency) { prev 1707 drivers/net/wireless/intel/iwlwifi/pcie/rx.c bool hw_rfkill, prev, report; prev 1710 drivers/net/wireless/intel/iwlwifi/pcie/rx.c prev = test_bit(STATUS_RFKILL_OPMODE, &trans->status); prev 1726 drivers/net/wireless/intel/iwlwifi/pcie/rx.c if (prev != report) prev 1072 drivers/net/wireless/intel/iwlwifi/pcie/trans.c bool prev = test_bit(STATUS_RFKILL_OPMODE, &trans->status); prev 1086 drivers/net/wireless/intel/iwlwifi/pcie/trans.c if (prev != report) prev 353 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c bss = list_entry(local->bss_list.prev, prev 378 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c bss = list_entry(local->bss_list.prev, prev 1069 drivers/net/wireless/intersil/hostap/hostap_ap.c struct add_sta_proc_data *entry, *prev; prev 1090 drivers/net/wireless/intersil/hostap/hostap_ap.c prev = entry; prev 1092 drivers/net/wireless/intersil/hostap/hostap_ap.c kfree(prev); prev 1956 drivers/net/wireless/intersil/hostap/hostap_ap.c struct wds_oper_data *entry, *prev; prev 1974 drivers/net/wireless/intersil/hostap/hostap_ap.c prev = entry; prev 1976 drivers/net/wireless/intersil/hostap/hostap_ap.c kfree(prev); prev 260 drivers/net/wireless/intersil/hostap/hostap_info.c struct hfa384x_hostscan_result *results, *prev; prev 290 drivers/net/wireless/intersil/hostap/hostap_info.c prev = local->last_scan_results; prev 294 drivers/net/wireless/intersil/hostap/hostap_info.c kfree(prev); prev 309 drivers/net/wireless/intersil/hostap/hostap_info.c struct hfa384x_hostscan_result *results, *prev; prev 356 drivers/net/wireless/intersil/hostap/hostap_info.c prev = local->last_scan_results; prev 360 drivers/net/wireless/intersil/hostap/hostap_info.c kfree(prev); prev 261 drivers/net/wireless/intersil/hostap/hostap_main.c struct hostap_tx_callback_info *cb, *prev = NULL; prev 266 drivers/net/wireless/intersil/hostap/hostap_main.c prev = cb; prev 270 drivers/net/wireless/intersil/hostap/hostap_main.c if (prev == NULL) prev 273 drivers/net/wireless/intersil/hostap/hostap_main.c prev->next = cb->next; prev 112 drivers/net/wireless/intersil/p54/txrx.c target_skb = entry->prev; prev 2137 drivers/net/wireless/intersil/prism54/isl_ioctl.c bss = list_entry(priv->bss_wpa_list.prev, prev 2162 drivers/net/wireless/intersil/prism54/isl_ioctl.c bss = list_entry(priv->bss_wpa_list.prev, prev 303 drivers/net/wireless/mediatek/mt76/dma.c skb->prev = skb->next = NULL; prev 236 drivers/net/wireless/mediatek/mt76/tx.c if (!skb->prev) { prev 782 drivers/net/wireless/mediatek/mt76/usb.c skb->prev = skb->next = NULL; prev 568 drivers/net/wireless/realtek/rtw88/pci.c struct sk_buff *prev = skb_dequeue(&ring->queue); prev 572 drivers/net/wireless/realtek/rtw88/pci.c if (!prev) prev 575 drivers/net/wireless/realtek/rtw88/pci.c tx_data = rtw_pci_get_tx_data(prev); prev 577 drivers/net/wireless/realtek/rtw88/pci.c pci_unmap_single(rtwpci->pdev, dma, prev->len, prev 579 drivers/net/wireless/realtek/rtw88/pci.c dev_kfree_skb_any(prev); prev 1599 drivers/net/wireless/st/cw1200/sta.c u32 bit, prev; prev 1608 drivers/net/wireless/st/cw1200/sta.c prev = priv->sta_asleep_mask & bit; prev 1612 drivers/net/wireless/st/cw1200/sta.c if (!prev) { prev 1621 drivers/net/wireless/st/cw1200/sta.c if (prev) { prev 321 drivers/net/wireless/st/cw1200/txrx.c entry = list_entry(cache->free.prev, prev 564 drivers/nvdimm/namespace_devs.c struct nd_label_id *label_id, struct resource *prev, prev 3650 drivers/nvme/host/core.c unsigned i, j, nsid, prev = 0; prev 3659 drivers/nvme/host/core.c ret = nvme_identify_ns_list(ctrl, prev, ns_list); prev 3670 drivers/nvme/host/core.c while (++prev < nsid) { prev 3671 drivers/nvme/host/core.c ns = nvme_find_get_ns(ctrl, prev); prev 3681 drivers/nvme/host/core.c nvme_remove_invalid_namespaces(ctrl, prev); prev 188 drivers/nvme/target/io-cmd-bdev.c struct bio *prev = bio; prev 195 drivers/nvme/target/io-cmd-bdev.c bio_chain(bio, prev); prev 196 drivers/nvme/target/io-cmd-bdev.c submit_bio(prev); prev 287 drivers/of/base.c struct device_node *__of_find_all_nodes(struct device_node *prev) prev 290 drivers/of/base.c if (!prev) { prev 292 drivers/of/base.c } else if (prev->child) { prev 293 drivers/of/base.c np = prev->child; prev 296 drivers/of/base.c np = prev; prev 312 drivers/of/base.c struct device_node *of_find_all_nodes(struct device_node *prev) prev 318 drivers/of/base.c np = __of_find_all_nodes(prev); prev 320 drivers/of/base.c of_node_put(prev); prev 729 drivers/of/base.c struct device_node *prev) prev 736 drivers/of/base.c next = prev ? prev->sibling : node->child; prev 740 drivers/of/base.c of_node_put(prev); prev 757 drivers/of/base.c struct device_node *prev) prev 763 drivers/of/base.c next = __of_get_next_child(node, prev); prev 778 drivers/of/base.c struct device_node *prev) prev 787 drivers/of/base.c next = prev ? prev->sibling : node->child; prev 794 drivers/of/base.c of_node_put(prev); prev 808 drivers/of/base.c struct device_node *of_get_next_cpu_node(struct device_node *prev) prev 814 drivers/of/base.c if (!prev) prev 818 drivers/of/base.c if (prev) prev 819 drivers/of/base.c next = prev->sibling; prev 831 drivers/of/base.c of_node_put(prev); prev 2228 drivers/of/base.c struct device_node *prev = NULL, *np = of_cpu_device_node_get(cpu); prev 2231 drivers/of/base.c prev = np; prev 2236 drivers/of/base.c of_property_read_u32(prev, "cache-level", &cache_level); prev 65 drivers/of/pdt.c static struct property * __init of_pdt_build_one_prop(phandle node, char *prev, prev 90 drivers/of/pdt.c err = of_pdt_prom_ops->nextprop(node, prev, p->name); prev 596 drivers/of/property.c struct device_node *prev) prev 609 drivers/of/property.c if (!prev) { prev 624 drivers/of/property.c port = of_get_parent(prev); prev 626 drivers/of/property.c __func__, prev)) prev 636 drivers/of/property.c endpoint = of_get_next_child(port, prev); prev 643 drivers/of/property.c prev = NULL; prev 935 drivers/of/property.c struct fwnode_handle *prev) prev 938 drivers/of/property.c to_of_node(prev))); prev 782 drivers/parport/share.c tmp->prev = NULL; prev 807 drivers/parport/share.c port->physport->devices->prev = tmp; prev 925 drivers/parport/share.c par_dev->prev = NULL; prev 951 drivers/parport/share.c port->physport->devices->prev = par_dev; prev 1021 drivers/parport/share.c dev->next->prev = dev->prev; prev 1022 drivers/parport/share.c if (dev->prev) prev 1023 drivers/parport/share.c dev->prev->next = dev->next; prev 545 drivers/pci/hotplug/acpiphp_glue.c struct pci_dev *dev, *prev; prev 554 drivers/pci/hotplug/acpiphp_glue.c list_for_each_entry_safe_reverse(dev, prev, &bus->devices, bus_list) prev 1739 drivers/pci/hotplug/ibmphp_res.c static struct bus_node *find_bus_wprev(u8 bus_number, struct bus_node **prev, u8 flag) prev 1745 drivers/pci/hotplug/ibmphp_res.c *prev = list_prev_entry(bus_cur, bus_list); prev 571 drivers/pci/pci-driver.c pci_power_t prev = pci_dev->current_state; prev 581 drivers/pci/pci-driver.c WARN_ONCE(pci_dev->current_state != prev, prev 598 drivers/pci/pci-driver.c pci_power_t prev = pci_dev->current_state; prev 608 drivers/pci/pci-driver.c WARN_ONCE(pci_dev->current_state != prev, prev 786 drivers/pci/pci-driver.c pci_power_t prev = pci_dev->current_state; prev 796 drivers/pci/pci-driver.c WARN_ONCE(pci_dev->current_state != prev, prev 834 drivers/pci/pci-driver.c pci_power_t prev = pci_dev->current_state; prev 844 drivers/pci/pci-driver.c WARN_ONCE(pci_dev->current_state != prev, prev 1285 drivers/pci/pci-driver.c pci_power_t prev = pci_dev->current_state; prev 1322 drivers/pci/pci-driver.c WARN_ONCE(pci_dev->current_state != prev, prev 206 drivers/perf/arm_smmuv3_pmu.c u64 delta, prev, now; prev 210 drivers/perf/arm_smmuv3_pmu.c prev = local64_read(&hwc->prev_count); prev 212 drivers/perf/arm_smmuv3_pmu.c } while (local64_cmpxchg(&hwc->prev_count, prev, now) != prev); prev 215 drivers/perf/arm_smmuv3_pmu.c delta = now - prev; prev 342 drivers/perf/qcom_l2_pmu.c u64 delta, prev, now; prev 346 drivers/perf/qcom_l2_pmu.c prev = local64_read(&hwc->prev_count); prev 348 drivers/perf/qcom_l2_pmu.c } while (local64_cmpxchg(&hwc->prev_count, prev, now) != prev); prev 354 drivers/perf/qcom_l2_pmu.c delta = now - prev; prev 244 drivers/perf/qcom_l3_pmu.c u64 prev, new; prev 247 drivers/perf/qcom_l3_pmu.c prev = local64_read(&event->hw.prev_count); prev 253 drivers/perf/qcom_l3_pmu.c } while (local64_cmpxchg(&event->hw.prev_count, prev, new) != prev); prev 255 drivers/perf/qcom_l3_pmu.c local64_add(new - prev, &event->count); prev 320 drivers/perf/qcom_l3_pmu.c u32 prev, new; prev 323 drivers/perf/qcom_l3_pmu.c prev = local64_read(&event->hw.prev_count); prev 325 drivers/perf/qcom_l3_pmu.c } while (local64_cmpxchg(&event->hw.prev_count, prev, new) != prev); prev 327 drivers/perf/qcom_l3_pmu.c local64_add(new - prev, &event->count); prev 317 drivers/perf/thunderx2_pmu.c s64 prev, delta, new = 0; prev 328 drivers/perf/thunderx2_pmu.c prev = local64_xchg(&hwc->prev_count, new); prev 331 drivers/perf/thunderx2_pmu.c delta = (u32)(((1UL << 32) - prev) + new); prev 1723 drivers/pinctrl/pinctrl-at91.c struct at91_gpio_chip *prev = NULL; prev 1775 drivers/pinctrl/pinctrl-at91.c prev = gpiochip_get_data(gpiochip_prev); prev 1779 drivers/pinctrl/pinctrl-at91.c if (prev->next) { prev 1780 drivers/pinctrl/pinctrl-at91.c prev = prev->next; prev 1782 drivers/pinctrl/pinctrl-at91.c prev->next = at91_gpio; prev 181 drivers/power/supply/ab8500_btemp.c static int prev; prev 188 drivers/power/supply/ab8500_btemp.c return prev; prev 190 drivers/power/supply/ab8500_btemp.c prev = vbtemp; prev 459 drivers/power/supply/ab8500_btemp.c static int prev; prev 488 drivers/power/supply/ab8500_btemp.c return prev; prev 499 drivers/power/supply/ab8500_btemp.c prev = temp; prev 833 drivers/power/supply/ab8500_fg.c static int prev; prev 840 drivers/power/supply/ab8500_fg.c return prev; prev 843 drivers/power/supply/ab8500_fg.c prev = vbat; prev 1081 drivers/power/supply/bq27xxx_battery.c u16 *prev = bq27xxx_dm_reg_ptr(buf, reg); prev 1083 drivers/power/supply/bq27xxx_battery.c if (prev == NULL) { prev 1096 drivers/power/supply/bq27xxx_battery.c if (be16_to_cpup(prev) == val) { prev 1113 drivers/power/supply/bq27xxx_battery.c "\n", str, be16_to_cpup(prev), val); prev 1119 drivers/power/supply/bq27xxx_battery.c *prev = cpu_to_be16(val); prev 525 drivers/rapidio/rio-scan.c u8 hopcount, struct rio_dev *prev, int prev_port) prev 549 drivers/rapidio/rio-scan.c if (rdev && prev && rio_is_switch(prev)) { prev 552 drivers/rapidio/rio-scan.c prev->rswitch->nextdev[prev_port] = rdev; prev 584 drivers/rapidio/rio-scan.c rdev->prev = prev; prev 585 drivers/rapidio/rio-scan.c if (prev && rio_is_switch(prev)) prev 586 drivers/rapidio/rio-scan.c prev->rswitch->nextdev[prev_port] = rdev; prev 729 drivers/rapidio/rio-scan.c u8 hopcount, struct rio_dev *prev, int prev_port) prev 737 drivers/rapidio/rio-scan.c rdev->prev = prev; prev 738 drivers/rapidio/rio-scan.c if (prev && rio_is_switch(prev)) prev 739 drivers/rapidio/rio-scan.c prev->rswitch->nextdev[prev_port] = rdev; prev 958 drivers/rapidio/rio-scan.c if (mport->nnode.next || mport->nnode.prev) prev 62 drivers/rapidio/rio-sysfs.c (rdev->prev) ? rio_name(rdev->prev) : "root"); prev 958 drivers/rapidio/rio.c struct rio_dev *prev = NULL; prev 961 drivers/rapidio/rio.c while (rdev->prev && (rdev->prev->pef & RIO_PEF_SWITCH)) { prev 962 drivers/rapidio/rio.c if (!rio_read_config_32(rdev->prev, RIO_DEV_ID_CAR, &result)) { prev 963 drivers/rapidio/rio.c prev = rdev->prev; prev 966 drivers/rapidio/rio.c rdev = rdev->prev; prev 969 drivers/rapidio/rio.c if (!prev) prev 972 drivers/rapidio/rio.c p_port = prev->rswitch->route_table[rdev->destid]; prev 976 drivers/rapidio/rio.c rio_name(prev), p_port); prev 977 drivers/rapidio/rio.c *nrdev = prev; prev 286 drivers/ras/cec.c u64 prev = 0; prev 292 drivers/ras/cec.c if (WARN(prev > this, "prev: 0x%016llx <-> this: 0x%016llx\n", prev, this)) prev 295 drivers/ras/cec.c prev = this; prev 734 drivers/s390/char/raw3270.c __list_add(&rp->list, l->prev, l); prev 379 drivers/s390/char/sclp.c req->list.prev == &sclp_req_queue) { prev 428 drivers/s390/char/sclp_cmd.c struct list_head *prev; prev 437 drivers/s390/char/sclp_cmd.c prev = &sclp_mem_list; prev 444 drivers/s390/char/sclp_cmd.c prev = &incr->list; prev 452 drivers/s390/char/sclp_cmd.c list_add(&new_incr->list, prev); prev 940 drivers/s390/char/tape_34xx.c tape_34xx_append_new_sbid(bid, l->prev); prev 946 drivers/s390/char/tape_34xx.c tape_34xx_append_new_sbid(bid, l->prev); prev 484 drivers/s390/char/tty3270.c if (tp->rcl_walk && tp->rcl_walk->prev != &tp->rcl_lines) prev 485 drivers/s390/char/tty3270.c tp->rcl_walk = tp->rcl_walk->prev; prev 487 drivers/s390/char/tty3270.c tp->rcl_walk = tp->rcl_lines.prev; prev 189 drivers/s390/cio/qdio_setup.c struct qdio_q *prev; prev 201 drivers/s390/cio/qdio_setup.c prev = (q->is_input_q) ? irq_ptr->input_qs[i - 1] prev 203 drivers/s390/cio/qdio_setup.c prev->slib->nsliba = (unsigned long)q->slib; prev 644 drivers/s390/net/lcs.c int prev, next; prev 647 drivers/s390/net/lcs.c prev = (index - 1) & (LCS_NUM_BUFFS - 1); prev 652 drivers/s390/net/lcs.c if (!(channel->ccws[prev].flags & CCW_FLAG_SUSPEND)) prev 689 drivers/s390/net/lcs.c int index, prev, next; prev 695 drivers/s390/net/lcs.c prev = (index - 1) & (LCS_NUM_BUFFS - 1); prev 701 drivers/s390/net/lcs.c if (channel->iob[prev].state == LCS_BUF_STATE_READY) { prev 708 drivers/s390/net/lcs.c __lcs_ready_buffer_bits(channel, prev); prev 4001 drivers/scsi/aacraid/aachba.c psg->sg[0].prev = 0; prev 4015 drivers/scsi/aacraid/aachba.c psg->sg[i].prev = 0; prev 477 drivers/scsi/aacraid/aacraid.h __le32 prev; /* reserved for F/W use */ prev 485 drivers/scsi/aacraid/aacraid.h u32 prev; /* reserved for F/W use */ prev 663 drivers/scsi/aha152x.c struct scsi_cmnd *ptr, *prev; prev 665 drivers/scsi/aha152x.c for (ptr = *SC, prev = NULL; prev 667 drivers/scsi/aha152x.c prev = ptr, ptr = SCNEXT(ptr)) prev 671 drivers/scsi/aha152x.c if (prev) prev 672 drivers/scsi/aha152x.c SCNEXT(prev) = SCNEXT(ptr); prev 685 drivers/scsi/aha152x.c struct scsi_cmnd *ptr, *prev; prev 687 drivers/scsi/aha152x.c for (ptr = *SC, prev = NULL; prev 689 drivers/scsi/aha152x.c prev = ptr, ptr = SCNEXT(ptr)) prev 693 drivers/scsi/aha152x.c if (prev) prev 694 drivers/scsi/aha152x.c SCNEXT(prev) = SCNEXT(ptr); prev 200 drivers/scsi/aic7xxx/aic79xx_core.c u_int prev, u_int next, u_int tid); prev 8383 drivers/scsi/aic7xxx/aic79xx_core.c u_int prev; prev 8388 drivers/scsi/aic7xxx/aic79xx_core.c prev = SCB_LIST_NULL; prev 8410 drivers/scsi/aic7xxx/aic79xx_core.c prev = scbid; prev 8421 drivers/scsi/aic7xxx/aic79xx_core.c ahd_rem_wscb(ahd, scbid, prev, next, tid); prev 8422 drivers/scsi/aic7xxx/aic79xx_core.c *list_tail = prev; prev 8423 drivers/scsi/aic7xxx/aic79xx_core.c if (SCBID_IS_NULL(prev)) prev 8430 drivers/scsi/aic7xxx/aic79xx_core.c prev = scbid; prev 8482 drivers/scsi/aic7xxx/aic79xx_core.c u_int prev, u_int next, u_int tid) prev 8487 drivers/scsi/aic7xxx/aic79xx_core.c if (!SCBID_IS_NULL(prev)) { prev 8488 drivers/scsi/aic7xxx/aic79xx_core.c ahd_set_scbptr(ahd, prev); prev 8502 drivers/scsi/aic7xxx/aic79xx_core.c ahd_outw(ahd, tail_offset, prev); prev 207 drivers/scsi/aic7xxx/aic7xxx_core.c u_int prev, u_int scbptr); prev 210 drivers/scsi/aic7xxx/aic7xxx_core.c u_int scbpos, u_int prev); prev 5839 drivers/scsi/aic7xxx/aic7xxx_core.c uint8_t prev; prev 5969 drivers/scsi/aic7xxx/aic7xxx_core.c prev = SCB_LIST_NULL; prev 6014 drivers/scsi/aic7xxx/aic7xxx_core.c next = ahc_rem_wscb(ahc, next, prev); prev 6017 drivers/scsi/aic7xxx/aic7xxx_core.c prev = next; prev 6023 drivers/scsi/aic7xxx/aic7xxx_core.c prev = next; prev 6142 drivers/scsi/aic7xxx/aic7xxx_core.c u_int prev; prev 6148 drivers/scsi/aic7xxx/aic7xxx_core.c prev = SCB_LIST_NULL; prev 6170 drivers/scsi/aic7xxx/aic7xxx_core.c if (next == prev) { prev 6173 drivers/scsi/aic7xxx/aic7xxx_core.c next, prev); prev 6181 drivers/scsi/aic7xxx/aic7xxx_core.c ahc_rem_scb_from_disc_list(ahc, prev, next); prev 6183 drivers/scsi/aic7xxx/aic7xxx_core.c prev = next; prev 6189 drivers/scsi/aic7xxx/aic7xxx_core.c prev = next; prev 6203 drivers/scsi/aic7xxx/aic7xxx_core.c ahc_rem_scb_from_disc_list(struct ahc_softc *ahc, u_int prev, u_int scbptr) prev 6214 drivers/scsi/aic7xxx/aic7xxx_core.c if (prev != SCB_LIST_NULL) { prev 6215 drivers/scsi/aic7xxx/aic7xxx_core.c ahc_outb(ahc, SCBPTR, prev); prev 6248 drivers/scsi/aic7xxx/aic7xxx_core.c ahc_rem_wscb(struct ahc_softc *ahc, u_int scbpos, u_int prev) prev 6266 drivers/scsi/aic7xxx/aic7xxx_core.c if (prev == SCB_LIST_NULL) { prev 6280 drivers/scsi/aic7xxx/aic7xxx_core.c ahc_outb(ahc, SCBPTR, prev); prev 1099 drivers/scsi/aic94xx/aic94xx_hwi.c struct asd_ascb *last = list_entry(first->list.prev, prev 1133 drivers/scsi/aic94xx/aic94xx_hwi.c struct asd_ascb *last = list_entry(ascb->list.prev, prev 1209 drivers/scsi/aic94xx/aic94xx_hwi.c __list_add(&list, ascb->list.prev, &ascb->list); prev 1214 drivers/scsi/aic94xx/aic94xx_hwi.c list_splice_init(&list, asd_ha->seq.pend_q.prev); prev 348 drivers/scsi/aic94xx/aic94xx_hwi.h __list_add(&list, ascb_list->list.prev, &ascb_list->list); prev 548 drivers/scsi/aic94xx/aic94xx_task.c __list_add(&alist, ascb->list.prev, &ascb->list); prev 588 drivers/scsi/aic94xx/aic94xx_task.c __list_add(&alist, ascb->list.prev, &ascb->list); prev 497 drivers/scsi/arm/acornscsi.c unsigned long prev; prev 507 drivers/scsi/arm/acornscsi.c prev = host->status[target][ptr].when; prev 526 drivers/scsi/arm/acornscsi.c time_diff = host->status[target][ptr].when - prev; prev 527 drivers/scsi/arm/acornscsi.c prev = host->status[target][ptr].when; prev 2832 drivers/scsi/arm/acornscsi.c unsigned int statptr, prev; prev 2840 drivers/scsi/arm/acornscsi.c prev = host->status[devidx][statptr].when; prev 2848 drivers/scsi/arm/acornscsi.c (host->status[devidx][statptr].when - prev) < 100 ? prev 2849 drivers/scsi/arm/acornscsi.c (host->status[devidx][statptr].when - prev) : 99); prev 2850 drivers/scsi/arm/acornscsi.c prev = host->status[devidx][statptr].when; prev 118 drivers/scsi/bfa/bfa_cs.h #define bfa_q_prev(_qe) (((struct list_head *) (_qe))->prev) prev 69 drivers/scsi/csiostor/csio_defs.h return ((list->next == list) && (list->prev == list)); prev 73 drivers/scsi/csiostor/csio_defs.h #define csio_list_prev(elem) (((struct list_head *)(elem))->prev) prev 69 drivers/scsi/dpt/dpti_i2o.h struct i2o_device *prev; prev 1582 drivers/scsi/dpt_i2o.c d->prev=NULL; prev 1584 drivers/scsi/dpt_i2o.c pHba->devices->prev=d; prev 767 drivers/scsi/initio.c struct scsi_ctrl_blk *tmp, *prev; prev 773 drivers/scsi/initio.c prev = tmp = host->first_pending; prev 780 drivers/scsi/initio.c prev->next = tmp->next; prev 782 drivers/scsi/initio.c host->last_pending = prev; prev 787 drivers/scsi/initio.c prev = tmp; prev 837 drivers/scsi/initio.c struct scsi_ctrl_blk *tmp, *prev; prev 843 drivers/scsi/initio.c prev = tmp = host->first_busy; prev 850 drivers/scsi/initio.c prev->next = tmp->next; prev 852 drivers/scsi/initio.c host->last_busy = prev; prev 861 drivers/scsi/initio.c prev = tmp; prev 869 drivers/scsi/initio.c struct scsi_ctrl_blk *tmp, *prev; prev 873 drivers/scsi/initio.c prev = tmp = host->first_busy; prev 879 drivers/scsi/initio.c prev = tmp; prev 923 drivers/scsi/initio.c struct scsi_ctrl_blk *tmp, *prev; prev 939 drivers/scsi/initio.c prev = tmp = host->first_pending; /* Check Pend queue */ prev 950 drivers/scsi/initio.c prev->next = tmp->next; prev 952 drivers/scsi/initio.c host->last_pending = prev; prev 961 drivers/scsi/initio.c prev = tmp; prev 965 drivers/scsi/initio.c prev = tmp = host->first_busy; /* Check Busy queue */ prev 980 drivers/scsi/initio.c prev->next = tmp->next; prev 982 drivers/scsi/initio.c host->last_busy = prev; prev 995 drivers/scsi/initio.c prev = tmp; prev 2358 drivers/scsi/initio.c struct scsi_ctrl_blk *tmp, *prev; prev 2373 drivers/scsi/initio.c prev = tmp = host->first_busy; /* Check Busy queue */ prev 2381 drivers/scsi/initio.c prev->next = tmp->next; prev 2383 drivers/scsi/initio.c host->last_busy = prev; prev 2390 drivers/scsi/initio.c prev = tmp; prev 2815 drivers/scsi/initio.c struct scsi_ctrl_blk *scb, *tmp, *prev = NULL /* silence gcc */; prev 2878 drivers/scsi/initio.c prev->next = tmp; prev 2879 drivers/scsi/initio.c prev = tmp; prev 2881 drivers/scsi/initio.c prev->next = NULL; prev 2884 drivers/scsi/initio.c host->last_avail = prev; prev 2570 drivers/scsi/libfc/fc_exch.c return list_entry(lport->ema_list.prev, prev 985 drivers/scsi/lpfc/lpfc_bsg.c iocbq = list_entry(head.prev, typeof(*iocbq), list); prev 1130 drivers/scsi/lpfc/lpfc_bsg.c list_move(evt->events_to_see.prev, &evt->events_to_get); prev 1323 drivers/scsi/lpfc/lpfc_bsg.c evt_dat = list_entry(evt->events_to_get.prev, prev 2757 drivers/scsi/lpfc/lpfc_bsg.c list_move(evt->events_to_see.prev, &evt->events_to_get); prev 2759 drivers/scsi/lpfc/lpfc_bsg.c *rxxri = (list_entry(evt->events_to_get.prev, prev 3337 drivers/scsi/lpfc/lpfc_bsg.c list_move(evt->events_to_see.prev, &evt->events_to_get); prev 3338 drivers/scsi/lpfc/lpfc_bsg.c evdat = list_entry(evt->events_to_get.prev, prev 2183 drivers/scsi/lpfc/lpfc_hbadisc.c if (fcf_pri->list.prev == &phba->fcf.fcf_pri_list) prev 2189 drivers/scsi/lpfc/lpfc_hbadisc.c fcf_pri->list.prev)->list); prev 10964 drivers/scsi/lpfc/lpfc_sli.c slp->next, slp->prev, pring->postbufq_cnt); prev 11008 drivers/scsi/lpfc/lpfc_sli.c slp->next, slp->prev, pring->postbufq_cnt); prev 17338 drivers/scsi/lpfc/lpfc_sli.c d_buf = list_entry(seq_dmabuf->dbuf.list.prev, typeof(*d_buf), list); prev 17355 drivers/scsi/lpfc/lpfc_sli.c d_buf = list_entry(d_buf->list.prev, typeof(*d_buf), list); prev 560 drivers/scsi/mesh.c struct scsi_cmnd *cmd, *prev, *next; prev 569 drivers/scsi/mesh.c prev = NULL; prev 575 drivers/scsi/mesh.c prev = cmd; prev 578 drivers/scsi/mesh.c if (prev == NULL) prev 581 drivers/scsi/mesh.c prev->host_scribble = (void *) next; prev 583 drivers/scsi/mesh.c ms->request_qtail = prev; prev 861 drivers/scsi/mesh.c int b, t, prev; prev 899 drivers/scsi/mesh.c prev = ms->conn_tgt; prev 964 drivers/scsi/mesh.c dlog(ms, "resel prev tgt=%d", prev); prev 6551 drivers/scsi/ncr53c8xx.c qp = lp->busy_ccbq.prev; prev 6554 drivers/scsi/ncr53c8xx.c qp = qp->prev; prev 565 drivers/scsi/scsi.c struct scsi_device *prev) prev 567 drivers/scsi/scsi.c struct list_head *list = (prev ? &prev->siblings : &shost->__devices); prev 582 drivers/scsi/scsi.c if (prev) prev 583 drivers/scsi/scsi.c scsi_device_put(prev); prev 5528 drivers/scsi/scsi_debug.c sdbg_host = list_entry(sdebug_host_list.prev, prev 639 drivers/scsi/ses.c struct enclosure_device *prev = NULL; prev 641 drivers/scsi/ses.c while ((edev = enclosure_find(&sdev->host->shost_gendev, prev)) != NULL) { prev 643 drivers/scsi/ses.c prev = edev; prev 787 drivers/scsi/ses.c struct enclosure_device *edev, *prev = NULL; prev 789 drivers/scsi/ses.c while ((edev = enclosure_find(&sdev->host->shost_gendev, prev)) != NULL) { prev 790 drivers/scsi/ses.c prev = edev; prev 480 drivers/scsi/wd33c93.c struct scsi_cmnd *cmd, *prev; prev 494 drivers/scsi/wd33c93.c prev = NULL; prev 499 drivers/scsi/wd33c93.c prev = cmd; prev 512 drivers/scsi/wd33c93.c if (prev) prev 513 drivers/scsi/wd33c93.c prev->host_scribble = cmd->host_scribble; prev 564 drivers/scsi/wd33c93.c for (prev = (struct scsi_cmnd *) hostdata->input_Q; prev; prev 565 drivers/scsi/wd33c93.c prev = (struct scsi_cmnd *) prev->host_scribble) { prev 566 drivers/scsi/wd33c93.c if ((prev->device->id != cmd->device->id) || prev 567 drivers/scsi/wd33c93.c (prev->device->lun != cmd->device->lun)) { prev 568 drivers/scsi/wd33c93.c for (prev = (struct scsi_cmnd *) hostdata->input_Q; prev; prev 569 drivers/scsi/wd33c93.c prev = (struct scsi_cmnd *) prev->host_scribble) prev 570 drivers/scsi/wd33c93.c prev->SCp.phase = 1; prev 1608 drivers/scsi/wd33c93.c struct scsi_cmnd *tmp, *prev; prev 1622 drivers/scsi/wd33c93.c prev = NULL; prev 1625 drivers/scsi/wd33c93.c if (prev) prev 1626 drivers/scsi/wd33c93.c prev->host_scribble = cmd->host_scribble; prev 1639 drivers/scsi/wd33c93.c prev = tmp; prev 421 drivers/sh/clk/core.c if (clk->node.next || clk->node.prev) prev 749 drivers/siox/siox-core.c sdevice = container_of(smaster->devices.prev, prev 858 drivers/siox/siox-core.c sdevice = container_of(smaster->devices.prev, struct siox_device, node); prev 207 drivers/soc/fsl/qbman/qman_test_stash.c static inline u32 do_lfsr(u32 prev) prev 209 drivers/soc/fsl/qbman/qman_test_stash.c return (prev >> 1) ^ (-(prev & 1u) & 0xd0000001u); prev 385 drivers/spi/spi-mxs.c flag = (&t->transfer_list == m->transfers.prev) ^ t->cs_change ? prev 1401 drivers/spi/spi-pl022.c previous = list_entry(transfer->transfer_list.prev, prev 1521 drivers/spi/spi-pl022.c list_entry(transfer->transfer_list.prev, prev 2832 drivers/spi/spi.c rxfer->replaced_after = xfer_first->transfer_list.prev; prev 366 drivers/staging/android/vsoc.c int prev = 0; prev 372 drivers/staging/android/vsoc.c prev = atomic_xchg(owner_ptr, VSOC_REGION_FREE); prev 373 drivers/staging/android/vsoc.c if (prev != perm->owned_value) prev 378 drivers/staging/android/vsoc.c perm->owned_value, prev); prev 1148 drivers/staging/comedi/drivers.c struct comedi_driver *prev; prev 1156 drivers/staging/comedi/drivers.c for (prev = comedi_drivers; prev->next; prev = prev->next) { prev 1157 drivers/staging/comedi/drivers.c if (prev->next == driver) { prev 1158 drivers/staging/comedi/drivers.c prev->next = driver->next; prev 530 drivers/staging/exfat/exfat.h struct buf_cache_t *prev; prev 43 drivers/staging/exfat/exfat_cache.c bp->prev = list; prev 44 drivers/staging/exfat/exfat_cache.c list->next->prev = bp; prev 50 drivers/staging/exfat/exfat_cache.c bp->prev = list->prev; prev 52 drivers/staging/exfat/exfat_cache.c list->prev->next = bp; prev 53 drivers/staging/exfat/exfat_cache.c list->prev = bp; prev 58 drivers/staging/exfat/exfat_cache.c bp->prev->next = bp->next; prev 59 drivers/staging/exfat/exfat_cache.c bp->next->prev = bp->prev; prev 65 drivers/staging/exfat/exfat_cache.c bp->prev->next = bp->next; prev 66 drivers/staging/exfat/exfat_cache.c bp->next->prev = bp->prev; prev 75 drivers/staging/exfat/exfat_cache.c bp = p_fs->FAT_cache_lru_list.prev; prev 139 drivers/staging/exfat/exfat_cache.c p_fs->FAT_cache_lru_list.prev = &p_fs->FAT_cache_lru_list; prev 146 drivers/staging/exfat/exfat_cache.c p_fs->FAT_cache_array[i].prev = NULL; prev 153 drivers/staging/exfat/exfat_cache.c p_fs->buf_cache_lru_list.prev = &p_fs->buf_cache_lru_list; prev 160 drivers/staging/exfat/exfat_cache.c p_fs->buf_cache_array[i].prev = NULL; prev 556 drivers/staging/exfat/exfat_cache.c bp = p_fs->buf_cache_lru_list.prev; prev 558 drivers/staging/exfat/exfat_cache.c bp = bp->prev; prev 307 drivers/staging/exfat/exfat_core.c u32 clu, prev; prev 330 drivers/staging/exfat/exfat_core.c prev = clu; prev 334 drivers/staging/exfat/exfat_core.c if (FAT_write(sb, prev, CLUSTER_32(0)) < 0) prev 1188 drivers/staging/fwserial/fwserial.c struct async_icount *prev) prev 1195 drivers/staging/fwserial/fwserial.c delta = ((mask & TIOCM_RNG && prev->rng != now.rng) || prev 1196 drivers/staging/fwserial/fwserial.c (mask & TIOCM_DSR && prev->dsr != now.dsr) || prev 1197 drivers/staging/fwserial/fwserial.c (mask & TIOCM_CAR && prev->dcd != now.dcd) || prev 1198 drivers/staging/fwserial/fwserial.c (mask & TIOCM_CTS && prev->cts != now.cts)); prev 1200 drivers/staging/fwserial/fwserial.c *prev = now; prev 1207 drivers/staging/fwserial/fwserial.c struct async_icount prev; prev 1209 drivers/staging/fwserial/fwserial.c prev = port->icount; prev 1212 drivers/staging/fwserial/fwserial.c check_msr_delta(port, mask, &prev)); prev 123 drivers/staging/gdm724x/gdm_mux.c r = list_entry(rx->rx_free_list.prev, struct mux_rx, free_list); prev 1753 drivers/staging/isdn/gigaset/bas-gigaset.c cs->cmdbuf->prev = NULL; prev 2014 drivers/staging/isdn/gigaset/bas-gigaset.c cb->prev = cs->lastcmdbuf; prev 519 drivers/staging/isdn/gigaset/gigaset.h struct cmdbuf_t *next, *prev; prev 135 drivers/staging/isdn/gigaset/ser-gigaset.c cb->prev = NULL; prev 248 drivers/staging/isdn/gigaset/ser-gigaset.c cb->prev = cs->lastcmdbuf; prev 439 drivers/staging/isdn/gigaset/usb-gigaset.c cs->cmdbuf->prev = NULL; prev 499 drivers/staging/isdn/gigaset/usb-gigaset.c cb->prev = cs->lastcmdbuf; prev 1703 drivers/staging/ks7010/ks_wlan_net.c pmk = list_entry(priv->pmklist.head.prev, struct pmk, prev 193 drivers/staging/media/imx/imx-media-vdic.c struct imx_media_buffer *prev; prev 201 drivers/staging/media/imx/imx-media-vdic.c prev = priv->prev_in_buf ? priv->prev_in_buf : curr; prev 203 drivers/staging/media/imx/imx-media-vdic.c prev_vb = &prev->vbuf.vb2_buf; prev 36 drivers/staging/rtl8192e/rtl819x_TSProc.c list_entry(pRxTs->RxPendingPktList.prev, prev 413 drivers/staging/rtl8192e/rtl819x_TSProc.c list_entry(pRxTS->RxPendingPktList.prev, prev 467 drivers/staging/rtl8192e/rtllib_rx.c pReorderEntry->List.next->prev = &pReorderEntry->List; prev 468 drivers/staging/rtl8192e/rtllib_rx.c pReorderEntry->List.prev = pList; prev 549 drivers/staging/rtl8192e/rtllib_rx.c list_entry(pTS->RxPendingPktList.prev, prev 701 drivers/staging/rtl8192e/rtllib_rx.c list_entry(pTS->RxPendingPktList.prev, prev 512 drivers/staging/rtl8192e/rtllib_softmac_wx.c short prev = ieee->raw_tx; prev 525 drivers/staging/rtl8192e/rtllib_softmac_wx.c if (prev == 0 && ieee->raw_tx) { prev 532 drivers/staging/rtl8192e/rtllib_softmac_wx.c if (prev && ieee->raw_tx == 1) prev 517 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c pReorderEntry->List.next->prev = &pReorderEntry->List; prev 518 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c pReorderEntry->List.prev = pList; prev 696 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c pReorderEntry = list_entry(pTS->rx_pending_pkt_list.prev, struct rx_reorder_entry, List); prev 453 drivers/staging/rtl8192u/ieee80211/ieee80211_softmac_wx.c short prev = ieee->raw_tx; prev 466 drivers/staging/rtl8192u/ieee80211/ieee80211_softmac_wx.c if (prev == 0 && ieee->raw_tx) { prev 473 drivers/staging/rtl8192u/ieee80211/ieee80211_softmac_wx.c if (prev && ieee->raw_tx == 1) prev 43 drivers/staging/rtl8192u/ieee80211/rtl819x_TSProc.c pReorderEntry = list_entry(pRxTs->rx_pending_pkt_list.prev, struct rx_reorder_entry, List); prev 426 drivers/staging/rtl8192u/ieee80211/rtl819x_TSProc.c pRxReorderEntry = list_entry(pRxTS->rx_pending_pkt_list.prev, struct rx_reorder_entry, List); prev 868 drivers/staging/speakup/main.c static int say_sentence_num(int num, int prev) prev 872 drivers/staging/speakup/main.c if (prev && --bn == -1) prev 1195 drivers/staging/unisys/visornic/visornic_main.c struct sk_buff *skb, *prev, *curr; prev 1293 drivers/staging/unisys/visornic/visornic_main.c for (cc = 1, prev = NULL; prev 1298 drivers/staging/unisys/visornic/visornic_main.c if (!prev) prev 1301 drivers/staging/unisys/visornic/visornic_main.c prev->next = curr; prev 1302 drivers/staging/unisys/visornic/visornic_main.c prev = curr; prev 18 drivers/staging/wusbcore/host/whci/asl.c struct whc_qset **next, struct whc_qset **prev) prev 27 drivers/staging/wusbcore/host/whci/asl.c p = qset->list_node.prev; prev 29 drivers/staging/wusbcore/host/whci/asl.c p = p->prev; prev 32 drivers/staging/wusbcore/host/whci/asl.c *prev = container_of(p, struct whc_qset, list_node); prev 44 drivers/staging/wusbcore/host/whci/asl.c struct whc_qset *next, *prev; prev 49 drivers/staging/wusbcore/host/whci/asl.c qset_get_next_prev(whc, qset, &next, &prev); prev 51 drivers/staging/wusbcore/host/whci/asl.c whc_qset_set_link_ptr(&prev->qh.link, qset->qset_dma); prev 57 drivers/staging/wusbcore/host/whci/asl.c struct whc_qset *prev, *next; prev 59 drivers/staging/wusbcore/host/whci/asl.c qset_get_next_prev(whc, qset, &next, &prev); prev 72 drivers/staging/wusbcore/host/whci/asl.c whc_qset_set_link_ptr(&prev->qh.link, next->qset_dma); prev 32 drivers/staging/wusbcore/host/whci/debug.c if (&qset->list_node == qset->whc->async_list.prev) { prev 1747 drivers/staging/wusbcore/wa-xfer.c wa->xfer_delayed_list.prev); prev 1779 drivers/staging/wusbcore/wa-xfer.c wa->xfer_errored_list.prev); prev 119 drivers/target/iscsi/cxgbit/cxgbit_cm.c struct np_info *p, **prev = &cdev->np_hash_tab[bucket]; prev 122 drivers/target/iscsi/cxgbit/cxgbit_cm.c for (p = *prev; p; prev = &p->next, p = p->next) { prev 125 drivers/target/iscsi/cxgbit/cxgbit_cm.c *prev = p->next; prev 787 drivers/target/iscsi/iscsi_target_erl1.c ooo_tail = list_entry(sess->sess_ooo_cmdsn_list.prev, prev 807 drivers/target/iscsi/iscsi_target_erl1.c ooo_tmp->ooo_list.prev); prev 1028 drivers/target/iscsi/iscsi_target_erl1.c ooo_tail = list_entry(sess->sess_ooo_cmdsn_list.prev, prev 532 drivers/target/target_core_transport.c se_nacl->acl_sess_list.prev, prev 2860 drivers/target/target_core_transport.c char *prev = *str; prev 2864 drivers/target/target_core_transport.c kfree(prev); prev 671 drivers/thunderbolt/property.c struct tb_property *prev) prev 673 drivers/thunderbolt/property.c if (prev) { prev 674 drivers/thunderbolt/property.c if (list_is_last(&prev->list, &dir->properties)) prev 676 drivers/thunderbolt/property.c return list_next_entry(prev, list); prev 766 drivers/thunderbolt/switch.c struct tb_port *prev) prev 770 drivers/thunderbolt/switch.c if (!prev) prev 773 drivers/thunderbolt/switch.c if (prev->sw == end->sw) { prev 774 drivers/thunderbolt/switch.c if (prev == end) prev 780 drivers/thunderbolt/switch.c if (prev->remote && prev 781 drivers/thunderbolt/switch.c prev->remote->sw->config.depth > prev->sw->config.depth) prev 782 drivers/thunderbolt/switch.c next = prev->remote; prev 784 drivers/thunderbolt/switch.c next = tb_port_at(tb_route(end->sw), prev->sw); prev 786 drivers/thunderbolt/switch.c if (tb_is_upstream_port(prev)) { prev 787 drivers/thunderbolt/switch.c next = prev->remote; prev 789 drivers/thunderbolt/switch.c next = tb_upstream_port(prev->sw); prev 795 drivers/thunderbolt/switch.c next->link_nr != prev->link_nr) { prev 603 drivers/thunderbolt/tb.h struct tb_port *prev); prev 2313 drivers/tty/synclink_gt.c struct cond_wait *w, *prev; prev 2316 drivers/tty/synclink_gt.c for (w = info->gpio_wait_q, prev = NULL ; w != NULL ; w = w->next) { prev 2320 drivers/tty/synclink_gt.c if (prev != NULL) prev 2321 drivers/tty/synclink_gt.c prev->next = w->next; prev 2325 drivers/tty/synclink_gt.c prev = w; prev 3007 drivers/tty/synclink_gt.c struct cond_wait *w, *prev; prev 3010 drivers/tty/synclink_gt.c for (w = *head, prev = NULL ; w != NULL ; prev = w, w = w->next) { prev 3012 drivers/tty/synclink_gt.c if (prev != NULL) prev 3013 drivers/tty/synclink_gt.c prev->next = w->next; prev 1408 drivers/tty/vt/vt_ioctl.c int prev; prev 1416 drivers/tty/vt/vt_ioctl.c prev = fg_console; prev 1439 drivers/tty/vt/vt_ioctl.c return prev; prev 271 drivers/usb/c67x00/c67x00-sched.c struct c67x00_ep_data *prev; prev 273 drivers/usb/c67x00/c67x00-sched.c list_for_each_entry(prev, &c67x00->list[type], node) { prev 274 drivers/usb/c67x00/c67x00-sched.c if (prev->hep->desc.bEndpointAddress > prev 276 drivers/usb/c67x00/c67x00-sched.c list_add(&ep_data->node, prev->node.prev); prev 407 drivers/usb/c67x00/c67x00-sched.c last_urb = list_entry(urbp->ep_data->queue.prev, prev 383 drivers/usb/chipidea/udc.c lastnode = list_entry(hwreq->tds.prev, prev 461 drivers/usb/chipidea/udc.c lastnode = list_entry(hwreq->tds.prev, prev 477 drivers/usb/chipidea/udc.c hwreqprev = list_entry(hwep->qh.queue.prev, prev 479 drivers/usb/chipidea/udc.c prevlastnode = list_entry(hwreqprev->tds.prev, prev 1796 drivers/usb/core/hcd.c urb = list_entry (ep->urb_list.prev, struct urb, prev 790 drivers/usb/core/urb.c victim = list_entry(anchor->urb_list.prev, struct urb, prev 823 drivers/usb/core/urb.c victim = list_entry(anchor->urb_list.prev, struct urb, prev 976 drivers/usb/core/urb.c victim = list_entry(anchor->urb_list.prev, struct urb, prev 259 drivers/usb/dwc2/debugfs.c ep->queue.next, ep->queue.prev); prev 433 drivers/usb/gadget/udc/aspeed-vhub/hub.c u16 prev; prev 436 drivers/usb/gadget/udc/aspeed-vhub/hub.c prev = p->status; prev 437 drivers/usb/gadget/udc/aspeed-vhub/hub.c p->status = (prev & ~clr_flags) | set_flags; prev 439 drivers/usb/gadget/udc/aspeed-vhub/hub.c port + 1, prev, p->status, set_c); prev 443 drivers/usb/gadget/udc/aspeed-vhub/hub.c u16 chg = p->status ^ prev; prev 738 drivers/usb/gadget/udc/fsl_udc_core.c lastreq = list_entry(ep->queue.prev, struct fsl_req, queue); prev 975 drivers/usb/gadget/udc/fsl_udc_core.c prev_req = list_entry(req->queue.prev, struct fsl_req, queue); prev 1222 drivers/usb/gadget/udc/goku_udc.c if (ep->dma && req->queue.prev == &ep->queue) { prev 269 drivers/usb/gadget/udc/mv_udc_core.c lastreq = list_entry(ep->queue.prev, struct mv_req, queue); prev 832 drivers/usb/gadget/udc/mv_udc_core.c prev_req = list_entry(req->queue.prev, struct mv_req, queue); prev 495 drivers/usb/host/ehci-q.c if (stopped && qtd->qtd_list.prev != &qh->qtd_list) { prev 496 drivers/usb/host/ehci-q.c last = list_entry (qtd->qtd_list.prev, prev 1085 drivers/usb/host/ehci-q.c qtd = list_entry (qh->qtd_list.prev, prev 1259 drivers/usb/host/ehci-q.c struct ehci_qh *prev; prev 1266 drivers/usb/host/ehci-q.c prev = ehci->async; prev 1267 drivers/usb/host/ehci-q.c while (prev->qh_next.qh != qh) prev 1268 drivers/usb/host/ehci-q.c prev = prev->qh_next.qh; prev 1270 drivers/usb/host/ehci-q.c prev->hw->hw_next = qh->hw->hw_next; prev 1271 drivers/usb/host/ehci-q.c prev->qh_next = qh->qh_next; prev 550 drivers/usb/host/ehci-sched.c union ehci_shadow *prev = &ehci->pshadow[i]; prev 552 drivers/usb/host/ehci-sched.c union ehci_shadow here = *prev; prev 560 drivers/usb/host/ehci-sched.c prev = periodic_next_shadow(ehci, prev, type); prev 562 drivers/usb/host/ehci-sched.c here = *prev; prev 571 drivers/usb/host/ehci-sched.c prev = &here.qh->qh_next; prev 573 drivers/usb/host/ehci-sched.c here = *prev; prev 581 drivers/usb/host/ehci-sched.c prev->qh = qh; prev 1726 drivers/usb/host/ehci-sched.c union ehci_shadow *prev = &ehci->pshadow[frame]; prev 1728 drivers/usb/host/ehci-sched.c union ehci_shadow here = *prev; prev 1736 drivers/usb/host/ehci-sched.c prev = periodic_next_shadow(ehci, prev, type); prev 1738 drivers/usb/host/ehci-sched.c here = *prev; prev 1743 drivers/usb/host/ehci-sched.c prev->itd = itd; prev 280 drivers/usb/host/ehci-timer.c ehci->cached_itd_list.prev, prev 283 drivers/usb/host/ehci-timer.c ehci->cached_sitd_list.prev, prev 1232 drivers/usb/host/fotg210-hcd.c fotg210->cached_itd_list.prev, prev 2443 drivers/usb/host/fotg210-hcd.c if (stopped && qtd->qtd_list.prev != &qh->qtd_list) { prev 2444 drivers/usb/host/fotg210-hcd.c last = list_entry(qtd->qtd_list.prev, prev 3007 drivers/usb/host/fotg210-hcd.c qtd = list_entry(qh->qtd_list.prev, prev 3077 drivers/usb/host/fotg210-hcd.c struct fotg210_qh *prev; prev 3088 drivers/usb/host/fotg210-hcd.c prev = fotg210->async; prev 3089 drivers/usb/host/fotg210-hcd.c while (prev->qh_next.qh != qh) prev 3090 drivers/usb/host/fotg210-hcd.c prev = prev->qh_next.qh; prev 3092 drivers/usb/host/fotg210-hcd.c prev->hw->hw_next = qh->hw->hw_next; prev 3093 drivers/usb/host/fotg210-hcd.c prev->qh_next = qh->qh_next; prev 3500 drivers/usb/host/fotg210-hcd.c union fotg210_shadow *prev = &fotg210->pshadow[i]; prev 3502 drivers/usb/host/fotg210-hcd.c union fotg210_shadow here = *prev; prev 3510 drivers/usb/host/fotg210-hcd.c prev = periodic_next_shadow(fotg210, prev, type); prev 3512 drivers/usb/host/fotg210-hcd.c here = *prev; prev 3521 drivers/usb/host/fotg210-hcd.c prev = &here.qh->qh_next; prev 3523 drivers/usb/host/fotg210-hcd.c here = *prev; prev 3531 drivers/usb/host/fotg210-hcd.c prev->qh = qh; prev 4331 drivers/usb/host/fotg210-hcd.c union fotg210_shadow *prev = &fotg210->pshadow[frame]; prev 4333 drivers/usb/host/fotg210-hcd.c union fotg210_shadow here = *prev; prev 4341 drivers/usb/host/fotg210-hcd.c prev = periodic_next_shadow(fotg210, prev, type); prev 4343 drivers/usb/host/fotg210-hcd.c here = *prev; prev 4348 drivers/usb/host/fotg210-hcd.c prev->itd = itd; prev 807 drivers/usb/host/imx21-hcd.c urb->start_frame = wrap_frame(list_entry(ep_priv->td_list.prev, prev 326 drivers/usb/host/isp116x-hcd.c struct isp116x_ep **prev = &isp116x->periodic[i]; prev 328 drivers/usb/host/isp116x-hcd.c while (*prev && ((temp = *prev) != ep)) prev 329 drivers/usb/host/isp116x-hcd.c prev = &temp->next; prev 330 drivers/usb/host/isp116x-hcd.c if (*prev) prev 331 drivers/usb/host/isp116x-hcd.c *prev = ep->next; prev 565 drivers/usb/host/isp116x-hcd.c if ((&isp116x->async)->next != (&isp116x->async)->prev) prev 797 drivers/usb/host/isp116x-hcd.c struct isp116x_ep **prev = &isp116x->periodic[i]; prev 798 drivers/usb/host/isp116x-hcd.c struct isp116x_ep *here = *prev; prev 803 drivers/usb/host/isp116x-hcd.c prev = &here->next; prev 804 drivers/usb/host/isp116x-hcd.c here = *prev; prev 808 drivers/usb/host/isp116x-hcd.c *prev = ep; prev 781 drivers/usb/host/isp1362-hcd.c if (isp1362_hcd->async.next != isp1362_hcd->async.prev) { prev 1952 drivers/usb/host/max3421-hcd.c struct max3421_hcd *max3421_hcd = NULL, **prev; prev 1956 drivers/usb/host/max3421-hcd.c for (prev = &max3421_hcd_list; *prev; prev = &(*prev)->next) { prev 1957 drivers/usb/host/max3421-hcd.c max3421_hcd = *prev; prev 1973 drivers/usb/host/max3421-hcd.c *prev = max3421_hcd->next; prev 113 drivers/usb/host/ohci-mem.c struct td **prev = &hc->td_hash [TD_HASH_FUNC (td->td_dma)]; prev 116 drivers/usb/host/ohci-mem.c while (*prev && *prev != td) prev 117 drivers/usb/host/ohci-mem.c prev = &(*prev)->td_hash; prev 118 drivers/usb/host/ohci-mem.c if (*prev) prev 119 drivers/usb/host/ohci-mem.c *prev = td->td_hash; prev 152 drivers/usb/host/ohci-q.c struct ed **prev = &ohci->periodic [i]; prev 154 drivers/usb/host/ohci-q.c struct ed *here = *prev; prev 163 drivers/usb/host/ohci-q.c prev = &here->ed_next; prev 165 drivers/usb/host/ohci-q.c here = *prev; prev 172 drivers/usb/host/ohci-q.c *prev = ed; prev 276 drivers/usb/host/ohci-q.c struct ed **prev = &ohci->periodic [i]; prev 279 drivers/usb/host/ohci-q.c while (*prev && (temp = *prev) != ed) { prev 281 drivers/usb/host/ohci-q.c prev = &temp->ed_next; prev 283 drivers/usb/host/ohci-q.c if (*prev) { prev 285 drivers/usb/host/ohci-q.c *prev = ed->ed_next; prev 984 drivers/usb/host/ohci-q.c __hc32 *prev; prev 1036 drivers/usb/host/ohci-q.c prev = &ed->hwHeadP; prev 1049 drivers/usb/host/ohci-q.c prev = &td->hwNextTD; prev 1054 drivers/usb/host/ohci-q.c savebits = *prev & ~cpu_to_hc32 (ohci, TD_MASK); prev 1055 drivers/usb/host/ohci-q.c *prev = td->hwNextTD | savebits; prev 1536 drivers/usb/host/oxu210hp-hcd.c if (stopped && qtd->qtd_list.prev != &qh->qtd_list) { prev 1537 drivers/usb/host/oxu210hp-hcd.c last = list_entry(qtd->qtd_list.prev, prev 2015 drivers/usb/host/oxu210hp-hcd.c list_splice(qtd_list, qh->qtd_list.prev); prev 2022 drivers/usb/host/oxu210hp-hcd.c qtd = list_entry(qh->qtd_list.prev, prev 2128 drivers/usb/host/oxu210hp-hcd.c struct ehci_qh *prev; prev 2153 drivers/usb/host/oxu210hp-hcd.c prev = oxu->async; prev 2154 drivers/usb/host/oxu210hp-hcd.c while (prev->qh_next.qh != qh) prev 2155 drivers/usb/host/oxu210hp-hcd.c prev = prev->qh_next.qh; prev 2157 drivers/usb/host/oxu210hp-hcd.c prev->hw_next = qh->hw_next; prev 2158 drivers/usb/host/oxu210hp-hcd.c prev->qh_next = qh->qh_next; prev 2367 drivers/usb/host/oxu210hp-hcd.c union ehci_shadow *prev = &oxu->pshadow[i]; prev 2369 drivers/usb/host/oxu210hp-hcd.c union ehci_shadow here = *prev; prev 2377 drivers/usb/host/oxu210hp-hcd.c prev = periodic_next_shadow(prev, type); prev 2379 drivers/usb/host/oxu210hp-hcd.c here = *prev; prev 2388 drivers/usb/host/oxu210hp-hcd.c prev = &here.qh->qh_next; prev 2390 drivers/usb/host/oxu210hp-hcd.c here = *prev; prev 2398 drivers/usb/host/oxu210hp-hcd.c prev->qh = qh; prev 454 drivers/usb/host/sl811-hcd.c struct sl811h_ep **prev = &sl811->periodic[i]; prev 456 drivers/usb/host/sl811-hcd.c while (*prev && ((temp = *prev) != ep)) prev 457 drivers/usb/host/sl811-hcd.c prev = &temp->next; prev 458 drivers/usb/host/sl811-hcd.c if (*prev) prev 459 drivers/usb/host/sl811-hcd.c *prev = ep->next; prev 927 drivers/usb/host/sl811-hcd.c struct sl811h_ep **prev = &sl811->periodic[i]; prev 928 drivers/usb/host/sl811-hcd.c struct sl811h_ep *here = *prev; prev 933 drivers/usb/host/sl811-hcd.c prev = &here->next; prev 934 drivers/usb/host/sl811-hcd.c here = *prev; prev 938 drivers/usb/host/sl811-hcd.c *prev = ep; prev 55 drivers/usb/host/uhci-q.c lqh = list_entry(uhci->skel_async_qh->node.prev, prev 67 drivers/usb/host/uhci-q.c lqh = list_entry(uhci->skel_async_qh->node.prev, prev 167 drivers/usb/host/uhci-q.c ltd = list_entry(ftd->fl_list.prev, struct uhci_td, fl_list); prev 207 drivers/usb/host/uhci-q.c ptd = list_entry(td->fl_list.prev, struct uhci_td, fl_list); prev 224 drivers/usb/host/uhci-q.c ltd = list_entry(ftd->fl_list.prev, struct uhci_td, fl_list); prev 229 drivers/usb/host/uhci-q.c list_del_init(ftd->fl_list.prev); prev 338 drivers/usb/host/uhci-q.c purbp = list_entry(urbp->node.prev, struct urb_priv, node); prev 340 drivers/usb/host/uhci-q.c ptd = list_entry(purbp->td_list.prev, struct uhci_td, prev 342 drivers/usb/host/uhci-q.c td = list_entry(urbp->td_list.prev, struct uhci_td, prev 400 drivers/usb/host/uhci-q.c td = list_entry(urbp->td_list.prev, struct uhci_td, prev 441 drivers/usb/host/uhci-q.c pqh = list_entry(qh->node.prev, struct uhci_qh, node); prev 525 drivers/usb/host/uhci-q.c pqh = list_entry(qh->node.prev, struct uhci_qh, node); prev 538 drivers/usb/host/uhci-q.c pqh = list_entry(qh->node.prev, struct uhci_qh, node); prev 1129 drivers/usb/host/uhci-q.c td = list_entry(urbp->td_list.prev, struct uhci_td, list); prev 1137 drivers/usb/host/uhci-q.c tmp = td->list.prev; prev 1152 drivers/usb/host/uhci-q.c tmp = urbp->td_list.prev; prev 1159 drivers/usb/host/uhci-q.c tmp = tmp->prev; prev 1213 drivers/usb/host/uhci-q.c if (td->list.next != urbp->td_list.prev) prev 1222 drivers/usb/host/uhci-q.c else if (&td->list != urbp->td_list.prev) prev 1297 drivers/usb/host/uhci-q.c lurb = list_entry(qh->queue.prev, prev 1534 drivers/usb/host/uhci-q.c urbp->node.prev == &qh->queue && prev 99 drivers/usb/host/xhci-mem.c static void xhci_link_segments(struct xhci_hcd *xhci, struct xhci_segment *prev, prev 104 drivers/usb/host/xhci-mem.c if (!prev || !next) prev 106 drivers/usb/host/xhci-mem.c prev->next = next; prev 108 drivers/usb/host/xhci-mem.c prev->trbs[TRBS_PER_SEGMENT-1].link.segment_ptr = prev 112 drivers/usb/host/xhci-mem.c val = le32_to_cpu(prev->trbs[TRBS_PER_SEGMENT-1].link.control); prev 121 drivers/usb/host/xhci-mem.c prev->trbs[TRBS_PER_SEGMENT-1].link.control = cpu_to_le32(val); prev 323 drivers/usb/host/xhci-mem.c struct xhci_segment *prev; prev 325 drivers/usb/host/xhci-mem.c prev = xhci_segment_alloc(xhci, cycle_state, max_packet, flags); prev 326 drivers/usb/host/xhci-mem.c if (!prev) prev 330 drivers/usb/host/xhci-mem.c *first = prev; prev 336 drivers/usb/host/xhci-mem.c prev = *first; prev 337 drivers/usb/host/xhci-mem.c while (prev) { prev 338 drivers/usb/host/xhci-mem.c next = prev->next; prev 339 drivers/usb/host/xhci-mem.c xhci_segment_free(xhci, prev); prev 340 drivers/usb/host/xhci-mem.c prev = next; prev 344 drivers/usb/host/xhci-mem.c xhci_link_segments(xhci, prev, next, type); prev 346 drivers/usb/host/xhci-mem.c prev = next; prev 349 drivers/usb/host/xhci-mem.c xhci_link_segments(xhci, prev, *first, type); prev 350 drivers/usb/host/xhci-mem.c *last = prev; prev 32 drivers/usb/image/microtek.h struct mts_desc *prev; prev 375 drivers/usb/musb/musb_host.c head = qh->ring.prev; prev 2415 drivers/usb/musb/musb_host.c || urb->urb_list.prev != &qh->hep->urb_list prev 130 drivers/usb/usbip/vhci_hcd.c u32 prev = prev_status & bit; prev 134 drivers/usb/usbip/vhci_hcd.c if (!prev && new) prev 136 drivers/usb/usbip/vhci_hcd.c else if (prev && !new) prev 141 drivers/usb/usbip/vhci_hcd.c if (prev || new) { prev 1430 drivers/vfio/pci/vfio_pci_config.c u8 pos, *prev, cap; prev 1446 drivers/vfio/pci/vfio_pci_config.c prev = &vdev->vconfig[PCI_CAPABILITY_LIST]; prev 1475 drivers/vfio/pci/vfio_pci_config.c *prev = next; prev 1496 drivers/vfio/pci/vfio_pci_config.c prev = &vdev->vconfig[pos + PCI_CAP_LIST_NEXT]; prev 1515 drivers/vfio/pci/vfio_pci_config.c __le32 *prev = NULL; prev 1551 drivers/vfio/pci/vfio_pci_config.c if (prev) { prev 1553 drivers/vfio/pci/vfio_pci_config.c *prev &= cpu_to_le32(~(0xffcU << 20)); prev 1554 drivers/vfio/pci/vfio_pci_config.c *prev |= cpu_to_le32(val << 20); prev 1598 drivers/vfio/pci/vfio_pci_config.c prev = (__le32 *)&vdev->vconfig[epos]; prev 1039 drivers/video/fbdev/core/fbmem.c if (info->modelist.prev && info->modelist.next && prev 1627 drivers/video/fbdev/core/fbmem.c if (!fb_info->modelist.prev || !fb_info->modelist.next) prev 1151 drivers/video/fbdev/core/modedb.c if (!head->prev || !head->next || list_empty(head)) prev 303 drivers/video/fbdev/omap/hwa742.c list_splice_init(head, hwa742.pending_req_list.prev); prev 470 drivers/video/fbdev/omap/hwa742.c last = list_entry(req_list.prev, struct hwa742_request, entry); prev 501 drivers/video/fbdev/omap/hwa742.c last = list_entry(req_list.prev, struct hwa742_request, entry); prev 20 drivers/video/fbdev/omap2/omapfb/dss/dss-of.c struct device_node *prev) prev 27 drivers/video/fbdev/omap2/omapfb/dss/dss-of.c if (!prev) { prev 44 drivers/video/fbdev/omap2/omapfb/dss/dss-of.c ports = of_get_parent(prev); prev 49 drivers/video/fbdev/omap2/omapfb/dss/dss-of.c port = of_get_next_child(ports, prev); prev 54 drivers/video/fbdev/omap2/omapfb/dss/dss-of.c prev = port; prev 66 drivers/video/fbdev/omap2/omapfb/dss/dss-of.c struct device_node *prev) prev 74 drivers/video/fbdev/omap2/omapfb/dss/dss-of.c ep = of_get_next_child(parent, prev); prev 77 drivers/video/fbdev/omap2/omapfb/dss/dss-of.c prev = ep; prev 427 drivers/virtio/virtio_ring.c unsigned int i, n, avail, descs_used, uninitialized_var(prev), err_idx; prev 490 drivers/virtio/virtio_ring.c prev = i; prev 503 drivers/virtio/virtio_ring.c prev = i; prev 508 drivers/virtio/virtio_ring.c desc[prev].flags &= cpu_to_virtio16(_vq->vdev, ~VRING_DESC_F_NEXT); prev 1105 drivers/virtio/virtio_ring.c u16 head, id, uninitialized_var(prev), curr, avail_used_flags; prev 1171 drivers/virtio/virtio_ring.c prev = curr; prev 1197 drivers/virtio/virtio_ring.c vq->packed.desc_state[id].last = prev; prev 1021 drivers/vme/bridges/vme_ca91cx42.c struct ca91cx42_dma_entry *entry, *prev; prev 1143 drivers/vme/bridges/vme_ca91cx42.c if (entry->list.prev != &list->entries) { prev 1144 drivers/vme/bridges/vme_ca91cx42.c prev = list_entry(entry->list.prev, struct ca91cx42_dma_entry, prev 1148 drivers/vme/bridges/vme_ca91cx42.c prev->descriptor.dcpp = desc_ptr & ~CA91CX42_DCPP_M; prev 1627 drivers/vme/bridges/vme_tsi148.c struct tsi148_dma_entry *entry, *prev; prev 1761 drivers/vme/bridges/vme_tsi148.c if (entry->list.prev != &list->entries) { prev 1764 drivers/vme/bridges/vme_tsi148.c prev = list_entry(entry->list.prev, struct tsi148_dma_entry, prev 1766 drivers/vme/bridges/vme_tsi148.c prev->descriptor.dnlau = cpu_to_be32(address_high); prev 1767 drivers/vme/bridges/vme_tsi148.c prev->descriptor.dnlal = cpu_to_be32(address_low); prev 457 fs/afs/file.c for (p = first->lru.prev; p != pages; p = p->prev) { prev 97 fs/autofs/expire.c static struct dentry *get_next_positive_subdir(struct dentry *prev, prev 105 fs/autofs/expire.c q = positive_after(root, prev); prev 108 fs/autofs/expire.c dput(prev); prev 115 fs/autofs/expire.c static struct dentry *get_next_positive_dentry(struct dentry *prev, prev 119 fs/autofs/expire.c struct dentry *p = prev, *ret = NULL, *d = NULL; prev 121 fs/autofs/expire.c if (prev == NULL) prev 140 fs/autofs/expire.c dput(prev); prev 679 fs/autofs/root.c d_child->prev == &parent->d_subdirs) prev 233 fs/btrfs/check-integrity.c struct btrfsic_stack_frame *prev; prev 963 fs/btrfs/check-integrity.c sf->prev = NULL; prev 1073 fs/btrfs/check-integrity.c next_stack->prev = sf; prev 1166 fs/btrfs/check-integrity.c next_stack->prev = sf; prev 1176 fs/btrfs/check-integrity.c if (NULL != sf->prev) { prev 1177 fs/btrfs/check-integrity.c struct btrfsic_stack_frame *const prev = sf->prev; prev 1183 fs/btrfs/check-integrity.c prev->error = sf->error; prev 1185 fs/btrfs/check-integrity.c sf = prev; prev 1190 fs/btrfs/check-integrity.c sf = prev; prev 343 fs/btrfs/delayed-inode.c struct btrfs_delayed_item **prev, prev 365 fs/btrfs/delayed-inode.c if (prev) { prev 367 fs/btrfs/delayed-inode.c *prev = NULL; prev 369 fs/btrfs/delayed-inode.c *prev = delayed_item; prev 371 fs/btrfs/delayed-inode.c *prev = rb_entry(node, struct btrfs_delayed_item, prev 374 fs/btrfs/delayed-inode.c *prev = NULL; prev 840 fs/btrfs/delayed-inode.c struct btrfs_delayed_item *curr, *prev; prev 855 fs/btrfs/delayed-inode.c prev = curr; prev 856 fs/btrfs/delayed-inode.c curr = __btrfs_next_delayed_item(prev); prev 857 fs/btrfs/delayed-inode.c if (curr && btrfs_is_continuous_delayed_item(prev, curr)) { prev 862 fs/btrfs/delayed-inode.c btrfs_release_delayed_item(prev); prev 942 fs/btrfs/delayed-inode.c struct btrfs_delayed_item *curr, *prev; prev 962 fs/btrfs/delayed-inode.c prev = curr; prev 963 fs/btrfs/delayed-inode.c curr = __btrfs_next_delayed_item(prev); prev 964 fs/btrfs/delayed-inode.c btrfs_release_delayed_item(prev); prev 388 fs/btrfs/extent_io.c struct rb_node *prev = NULL; prev 394 fs/btrfs/extent_io.c prev = *n; prev 395 fs/btrfs/extent_io.c entry = rb_entry(prev, struct tree_entry, rb_node); prev 409 fs/btrfs/extent_io.c *parent_ret = prev; prev 412 fs/btrfs/extent_io.c orig_prev = prev; prev 413 fs/btrfs/extent_io.c while (prev && offset > prev_entry->end) { prev 414 fs/btrfs/extent_io.c prev = rb_next(prev); prev 415 fs/btrfs/extent_io.c prev_entry = rb_entry(prev, struct tree_entry, rb_node); prev 417 fs/btrfs/extent_io.c *next_ret = prev; prev 418 fs/btrfs/extent_io.c prev = orig_prev; prev 422 fs/btrfs/extent_io.c prev_entry = rb_entry(prev, struct tree_entry, rb_node); prev 423 fs/btrfs/extent_io.c while (prev && offset < prev_entry->start) { prev 424 fs/btrfs/extent_io.c prev = rb_prev(prev); prev 425 fs/btrfs/extent_io.c prev_entry = rb_entry(prev, struct tree_entry, rb_node); prev 427 fs/btrfs/extent_io.c *prev_ret = prev; prev 1579 fs/btrfs/extent_io.c struct rb_node *node, *prev = NULL, *next; prev 1585 fs/btrfs/extent_io.c node = __etree_search(tree, start, &next, &prev, NULL, NULL); prev 1586 fs/btrfs/extent_io.c if (!node && !next && !prev) { prev 1599 fs/btrfs/extent_io.c state = rb_entry(prev, struct extent_state, rb_node); prev 1645 fs/btrfs/extent_io.c if (prev) { prev 1646 fs/btrfs/extent_io.c state = rb_entry(prev, struct extent_state, prev 150 fs/btrfs/extent_map.c struct rb_node *prev = NULL; prev 157 fs/btrfs/extent_map.c prev = n; prev 169 fs/btrfs/extent_map.c orig_prev = prev; prev 170 fs/btrfs/extent_map.c while (prev && offset >= extent_map_end(prev_entry)) { prev 171 fs/btrfs/extent_map.c prev = rb_next(prev); prev 172 fs/btrfs/extent_map.c prev_entry = rb_entry(prev, struct extent_map, rb_node); prev 174 fs/btrfs/extent_map.c *prev_ret = prev; prev 175 fs/btrfs/extent_map.c prev = orig_prev; prev 179 fs/btrfs/extent_map.c prev_entry = rb_entry(prev, struct extent_map, rb_node); prev 180 fs/btrfs/extent_map.c while (prev && offset < prev_entry->start) { prev 181 fs/btrfs/extent_map.c prev = rb_prev(prev); prev 182 fs/btrfs/extent_map.c prev_entry = rb_entry(prev, struct extent_map, rb_node); prev 184 fs/btrfs/extent_map.c *next_ret = prev; prev 190 fs/btrfs/extent_map.c static int mergable_maps(struct extent_map *prev, struct extent_map *next) prev 192 fs/btrfs/extent_map.c if (test_bit(EXTENT_FLAG_PINNED, &prev->flags)) prev 199 fs/btrfs/extent_map.c if (test_bit(EXTENT_FLAG_COMPRESSED, &prev->flags)) prev 202 fs/btrfs/extent_map.c if (test_bit(EXTENT_FLAG_LOGGING, &prev->flags) || prev 211 fs/btrfs/extent_map.c if (!list_empty(&prev->list) || !list_empty(&next->list)) prev 215 fs/btrfs/extent_map.c prev->block_start != EXTENT_MAP_DELALLOC); prev 217 fs/btrfs/extent_map.c if (extent_map_end(prev) == next->start && prev 218 fs/btrfs/extent_map.c prev->flags == next->flags && prev 219 fs/btrfs/extent_map.c prev->bdev == next->bdev && prev 221 fs/btrfs/extent_map.c prev->block_start == EXTENT_MAP_HOLE) || prev 223 fs/btrfs/extent_map.c prev->block_start == EXTENT_MAP_INLINE) || prev 225 fs/btrfs/extent_map.c next->block_start == extent_map_block_end(prev)))) { prev 419 fs/btrfs/extent_map.c struct rb_node *prev = NULL; prev 423 fs/btrfs/extent_map.c rb_node = __tree_search(&tree->map.rb_root, start, &prev, &next); prev 425 fs/btrfs/extent_map.c if (prev) prev 426 fs/btrfs/extent_map.c rb_node = prev; prev 522 fs/btrfs/extent_map.c struct rb_node *prev; prev 524 fs/btrfs/extent_map.c prev = rb_prev(&em->rb_node); prev 525 fs/btrfs/extent_map.c if (!prev) prev 527 fs/btrfs/extent_map.c return container_of(prev, struct extent_map, rb_node); prev 541 fs/btrfs/extent_map.c struct extent_map *prev; prev 551 fs/btrfs/extent_map.c prev = prev_extent_map(next); prev 553 fs/btrfs/extent_map.c prev = existing; prev 554 fs/btrfs/extent_map.c next = next_extent_map(prev); prev 557 fs/btrfs/extent_map.c start = prev ? extent_map_end(prev) : em->start; prev 2910 fs/btrfs/file.c struct falloc_range *prev = NULL; prev 2920 fs/btrfs/file.c prev = list_entry(head->prev, struct falloc_range, list); prev 2921 fs/btrfs/file.c if (prev->start + prev->len == start) { prev 2922 fs/btrfs/file.c prev->len += len; prev 639 fs/btrfs/free-space-cache.c struct btrfs_free_space *e, *prev = NULL; prev 646 fs/btrfs/free-space-cache.c if (!prev) prev 648 fs/btrfs/free-space-cache.c if (e->bitmap || prev->bitmap) prev 650 fs/btrfs/free-space-cache.c if (prev->offset + prev->bytes == e->offset) { prev 651 fs/btrfs/free-space-cache.c unlink_free_space(ctl, prev); prev 653 fs/btrfs/free-space-cache.c prev->bytes += e->bytes; prev 655 fs/btrfs/free-space-cache.c link_free_space(ctl, prev); prev 656 fs/btrfs/free-space-cache.c prev = NULL; prev 661 fs/btrfs/free-space-cache.c prev = e; prev 1512 fs/btrfs/free-space-cache.c struct btrfs_free_space *entry, *prev = NULL; prev 1522 fs/btrfs/free-space-cache.c prev = entry; prev 1559 fs/btrfs/free-space-cache.c prev = rb_entry(n, struct btrfs_free_space, prev 1561 fs/btrfs/free-space-cache.c if (!prev->bitmap && prev 1562 fs/btrfs/free-space-cache.c prev->offset + prev->bytes > offset) prev 1563 fs/btrfs/free-space-cache.c entry = prev; prev 1569 fs/btrfs/free-space-cache.c if (!prev) prev 1573 fs/btrfs/free-space-cache.c entry = prev; prev 1591 fs/btrfs/free-space-cache.c prev = rb_entry(n, struct btrfs_free_space, prev 1593 fs/btrfs/free-space-cache.c if (!prev->bitmap && prev 1594 fs/btrfs/free-space-cache.c prev->offset + prev->bytes > offset) prev 1595 fs/btrfs/free-space-cache.c return prev; prev 2713 fs/btrfs/inode.c struct sa_defrag_extent_backref *prev, prev 2737 fs/btrfs/inode.c if (prev && prev->root_id == backref->root_id && prev 2738 fs/btrfs/inode.c prev->inum == backref->inum && prev 2739 fs/btrfs/inode.c prev->file_pos + prev->num_bytes == backref->file_pos) prev 2931 fs/btrfs/inode.c struct sa_defrag_extent_backref *prev = NULL; prev 2953 fs/btrfs/inode.c ret = relink_extent_backref(path, prev, backref); prev 2956 fs/btrfs/inode.c kfree(prev); prev 2959 fs/btrfs/inode.c prev = backref; prev 2961 fs/btrfs/inode.c prev = NULL; prev 2964 fs/btrfs/inode.c kfree(prev); prev 4446 fs/btrfs/inode.c struct rb_node *prev; prev 4457 fs/btrfs/inode.c prev = NULL; prev 4459 fs/btrfs/inode.c prev = node; prev 4470 fs/btrfs/inode.c while (prev) { prev 4471 fs/btrfs/inode.c entry = rb_entry(prev, struct btrfs_inode, rb_node); prev 4473 fs/btrfs/inode.c node = prev; prev 4476 fs/btrfs/inode.c prev = rb_next(prev); prev 71 fs/btrfs/ordered-data.c struct rb_node *prev = NULL; prev 78 fs/btrfs/ordered-data.c prev = n; prev 91 fs/btrfs/ordered-data.c while (prev && file_offset >= entry_end(prev_entry)) { prev 92 fs/btrfs/ordered-data.c test = rb_next(prev); prev 100 fs/btrfs/ordered-data.c prev = test; prev 102 fs/btrfs/ordered-data.c if (prev) prev 103 fs/btrfs/ordered-data.c prev_entry = rb_entry(prev, struct btrfs_ordered_extent, prev 105 fs/btrfs/ordered-data.c while (prev && file_offset < entry_end(prev_entry)) { prev 106 fs/btrfs/ordered-data.c test = rb_prev(prev); prev 111 fs/btrfs/ordered-data.c prev = test; prev 113 fs/btrfs/ordered-data.c *prev_ret = prev; prev 145 fs/btrfs/ordered-data.c struct rb_node *prev = NULL; prev 155 fs/btrfs/ordered-data.c ret = __tree_search(root, file_offset, &prev); prev 157 fs/btrfs/ordered-data.c ret = prev; prev 811 fs/btrfs/ordered-data.c struct rb_node *prev = NULL; prev 868 fs/btrfs/ordered-data.c prev = tree_search(tree, offset); prev 873 fs/btrfs/ordered-data.c if (prev) { prev 874 fs/btrfs/ordered-data.c test = rb_entry(prev, struct btrfs_ordered_extent, prev 878 fs/btrfs/ordered-data.c node = prev; prev 494 fs/btrfs/raid56.c found = list_entry(table->stripe_cache.prev, prev 1554 fs/btrfs/relocation.c struct rb_node *prev; prev 1561 fs/btrfs/relocation.c prev = NULL; prev 1563 fs/btrfs/relocation.c prev = node; prev 1574 fs/btrfs/relocation.c while (prev) { prev 1575 fs/btrfs/relocation.c entry = rb_entry(prev, struct btrfs_inode, rb_node); prev 1577 fs/btrfs/relocation.c node = prev; prev 1580 fs/btrfs/relocation.c prev = rb_next(prev); prev 2059 fs/btrfs/transaction.c if (cur_trans->list.prev != &fs_info->trans_list) { prev 2060 fs/btrfs/transaction.c prev_trans = list_entry(cur_trans->list.prev, prev 2394 fs/btrfs/transaction.c unsigned long prev; prev 2397 fs/btrfs/transaction.c prev = xchg(&fs_info->pending_changes, 0); prev 2398 fs/btrfs/transaction.c if (!prev) prev 2402 fs/btrfs/transaction.c if (prev & bit) prev 2404 fs/btrfs/transaction.c prev &= ~bit; prev 2407 fs/btrfs/transaction.c if (prev & bit) prev 2409 fs/btrfs/transaction.c prev &= ~bit; prev 2412 fs/btrfs/transaction.c if (prev & bit) prev 2414 fs/btrfs/transaction.c prev &= ~bit; prev 2416 fs/btrfs/transaction.c if (prev) prev 2418 fs/btrfs/transaction.c "unknown pending changes left 0x%lx, ignoring", prev); prev 724 fs/buffer.c bh = BH_ENTRY(tmp.prev); prev 383 fs/ceph/addr.c page = list_entry(page_list->prev, struct page, lru); prev 1694 fs/ceph/caps.c struct ceph_cap_flush *prev; prev 1698 fs/ceph/caps.c if (wake && cf->g_list.prev != &mdsc->cap_flush_list) { prev 1699 fs/ceph/caps.c prev = list_prev_entry(cf, g_list); prev 1700 fs/ceph/caps.c prev->wake = true; prev 1705 fs/ceph/caps.c if (wake && cf->i_list.prev != &ci->i_cap_flush_list) { prev 1706 fs/ceph/caps.c prev = list_prev_entry(cf, i_list); prev 1707 fs/ceph/caps.c prev->wake = true; prev 3042 fs/ceph/caps.c struct dentry *dn, *prev = NULL; prev 3056 fs/ceph/caps.c if (dn == prev) { prev 3061 fs/ceph/caps.c if (prev) prev 3062 fs/ceph/caps.c dput(prev); prev 3063 fs/ceph/caps.c prev = dn; prev 3065 fs/ceph/caps.c if (prev) prev 3066 fs/ceph/caps.c dput(prev); prev 1496 fs/ceph/mds_client.c struct ceph_cap *cap, *prev = NULL; prev 1508 fs/ceph/mds_client.c if (cap == prev) prev 1510 fs/ceph/mds_client.c prev = cap; prev 100 fs/char_dev.c struct char_device_struct *cd, *curr, *prev = NULL; prev 134 fs/char_dev.c for (curr = chrdevs[i]; curr; prev = curr, curr = curr->next) { prev 155 fs/char_dev.c if (!prev) { prev 159 fs/char_dev.c cd->next = prev->next; prev 160 fs/char_dev.c prev->next = cd; prev 1110 fs/cifs/file.c (lock->blist.prev == &lock->blist) && prev 123 fs/crypto/hkdf.c const u8 *prev = NULL; prev 141 fs/crypto/hkdf.c if (prev) { prev 142 fs/crypto/hkdf.c err = crypto_shash_update(desc, prev, HKDF_HASHLEN); prev 168 fs/crypto/hkdf.c prev = &okm[i]; prev 1101 fs/dcache.c dentry = list_entry(list->prev, struct dentry, d_lru); prev 1307 fs/dlm/lock.c __list_add(new, lkb->lkb_statequeue.prev, &lkb->lkb_statequeue); prev 1434 fs/dlm/lowcomms.c e = list_entry(con->writequeue.prev, struct writequeue_entry, list); prev 307 fs/dlm/member.c newlist->prev = tmp->prev; prev 309 fs/dlm/member.c tmp->prev->next = newlist; prev 310 fs/dlm/member.c tmp->prev = newlist; prev 296 fs/erofs/data.c page = list_entry(pages->prev, struct page, lru); prev 1137 fs/eventpoll.c struct list_head *prev; prev 1156 fs/eventpoll.c prev = xchg(&head->prev, new); prev 1163 fs/eventpoll.c prev->next = new; prev 1164 fs/eventpoll.c new->prev = prev; prev 699 fs/exec.c struct vm_area_struct *prev = NULL; prev 759 fs/exec.c ret = mprotect_fixup(vma, &prev, vma->vm_start, vma->vm_end, prev 763 fs/exec.c BUG_ON(prev != vma); prev 205 fs/ext2/balloc.c struct ext2_reserve_window_node *rsv, *prev; prev 211 fs/ext2/balloc.c prev = NULL; prev 225 fs/ext2/balloc.c if (prev && prev->rsv_end >= rsv->rsv_start) { prev 238 fs/ext2/balloc.c prev = rsv; prev 784 fs/ext2/balloc.c struct ext2_reserve_window_node *rsv, *prev; prev 811 fs/ext2/balloc.c prev = rsv; prev 841 fs/ext2/balloc.c if ((prev != my_rsv) && (!rsv_is_empty(&my_rsv->rsv_window))) prev 855 fs/ext2/balloc.c if (prev != my_rsv) prev 407 fs/ext4/extents.c ext4_lblk_t prev = 0; prev 416 fs/ext4/extents.c if ((lblock <= prev) && prev) { prev 423 fs/ext4/extents.c prev = lblock + len - 1; prev 505 fs/ext4/extents.c ext4_lblk_t prev = 0; prev 513 fs/ext4/extents.c if (prev && (prev != lblk)) prev 514 fs/ext4/extents.c ext4_es_cache_extent(inode, prev, lblk - prev, ~0, prev 521 fs/ext4/extents.c prev = lblk + len; prev 373 fs/ext4/fsmap.c struct ext4_fsmap *prev = NULL; prev 377 fs/ext4/fsmap.c if (!prev) { prev 378 fs/ext4/fsmap.c prev = p; prev 382 fs/ext4/fsmap.c if (prev->fmr_owner == p->fmr_owner && prev 383 fs/ext4/fsmap.c prev->fmr_physical + prev->fmr_length == p->fmr_physical) { prev 384 fs/ext4/fsmap.c prev->fmr_length += p->fmr_length; prev 388 fs/ext4/fsmap.c prev = p; prev 1787 fs/ext4/namei.c struct ext4_dir_entry_2 *next, *to, *prev, *de = (struct ext4_dir_entry_2 *) base; prev 1790 fs/ext4/namei.c prev = to = de; prev 1798 fs/ext4/namei.c prev = to; prev 1803 fs/ext4/namei.c return prev; prev 2988 fs/ext4/namei.c struct list_head *prev; prev 3012 fs/ext4/namei.c prev = ei->i_orphan.prev; prev 3025 fs/ext4/namei.c if (prev == &sbi->s_orphan) { prev 3039 fs/ext4/namei.c &list_entry(prev, struct ext4_inode_info, i_orphan)->vfs_inode; prev 149 fs/ext4/page-io.c io, inode->i_ino, io->list.next, io->list.prev); prev 177 fs/ext4/page-io.c before = cur->prev; prev 366 fs/f2fs/checkpoint.c pgoff_t index = 0, prev = ULONG_MAX; prev 386 fs/f2fs/checkpoint.c if (prev == ULONG_MAX) prev 387 fs/f2fs/checkpoint.c prev = page->index - 1; prev 388 fs/f2fs/checkpoint.c if (nr_to_write != LONG_MAX && page->index != prev + 1) { prev 415 fs/f2fs/checkpoint.c prev = page->index; prev 503 fs/f2fs/extent_cache.c struct extent_info ei, dei, prev; prev 522 fs/f2fs/extent_cache.c prev = et->largest; prev 609 fs/f2fs/extent_cache.c prev.len < F2FS_MIN_EXTENT_LEN && prev 2749 fs/f2fs/node.c list_add(&nes->set_list, cur->set_list.prev); prev 169 fs/fat/cache.c struct list_head *p = MSDOS_I(inode)->cache_lru.prev; prev 1290 fs/fat/dir.c struct buffer_head *bh, *prev, *bhs[3]; /* 32*slots (672bytes) */ prev 1299 fs/fat/dir.c bh = prev = NULL; prev 1308 fs/fat/dir.c if (prev != bh) { prev 1310 fs/fat/dir.c bhs[nr_bhs] = prev = bh; prev 1319 fs/fat/dir.c prev = NULL; prev 299 fs/fat/namei_vfat.c unsigned char prev = buf[0]; prev 308 fs/fat/namei_vfat.c if (buf[0] == prev) prev 435 fs/fs-writeback.c inode_io_list_move_locked(inode, new_wb, pos->i_io_list.prev); prev 1247 fs/fs-writeback.c inode = wb_inode(delaying_queue->prev); prev 1270 fs/fs-writeback.c sb = wb_inode(tmp.prev)->i_sb; prev 1643 fs/fs-writeback.c struct inode *inode = wb_inode(wb->b_io.prev); prev 1773 fs/fs-writeback.c struct inode *inode = wb_inode(wb->b_io.prev); prev 1920 fs/fs-writeback.c inode = wb_inode(wb->b_more_io.prev); prev 2248 fs/gfs2/bmap.c jext = list_entry(jd->extent_list.prev, struct gfs2_journal_extent, list); prev 563 fs/gfs2/dir.c struct gfs2_dirent *dent, *prev; prev 573 fs/gfs2/dir.c prev = NULL; prev 585 fs/gfs2/dir.c prev = dent; prev 599 fs/gfs2/dir.c return prev ? prev : dent; prev 673 fs/gfs2/dir.c struct gfs2_dirent *prev, struct gfs2_dirent *cur) prev 688 fs/gfs2/dir.c if (!prev) { prev 696 fs/gfs2/dir.c prev_rec_len = be16_to_cpu(prev->de_rec_len); prev 699 fs/gfs2/dir.c if ((char *)prev + prev_rec_len != (char *)cur) prev 705 fs/gfs2/dir.c prev->de_rec_len = cpu_to_be16(prev_rec_len); prev 1009 fs/gfs2/dir.c struct gfs2_dirent *dent = NULL, *prev = NULL, *next = NULL, *new; prev 1108 fs/gfs2/dir.c dirent_del(dip, obh, prev, dent); prev 1114 fs/gfs2/dir.c if (!prev) prev 1115 fs/gfs2/dir.c prev = dent; prev 1117 fs/gfs2/dir.c prev = dent; prev 1878 fs/gfs2/dir.c struct gfs2_dirent *dent, *prev = NULL; prev 1895 fs/gfs2/dir.c prev = dent; prev 1896 fs/gfs2/dir.c dent = (struct gfs2_dirent *)((char *)dent + be16_to_cpu(prev->de_rec_len)); prev 1899 fs/gfs2/dir.c dirent_del(dip, bh, prev, dent); prev 360 fs/gfs2/glock.c if (gh->gh_list.prev == &gl->gl_holders && prev 385 fs/gfs2/glock.c if (gh->gh_list.prev == &gl->gl_holders) prev 283 fs/gfs2/log.c bd = list_entry(head->prev, struct gfs2_bufdata, prev 491 fs/gfs2/log.c tr = list_entry(sdp->sd_ail1_list.prev, struct gfs2_trans, prev 474 fs/gfs2/lops.c static struct bio *gfs2_chain_bio(struct bio *prev, unsigned int nr_iovecs) prev 479 fs/gfs2/lops.c bio_copy_dev(new, prev); prev 480 fs/gfs2/lops.c new->bi_iter.bi_sector = bio_end_sector(prev); prev 481 fs/gfs2/lops.c new->bi_opf = prev->bi_opf; prev 482 fs/gfs2/lops.c new->bi_write_hint = prev->bi_write_hint; prev 483 fs/gfs2/lops.c bio_chain(new, prev); prev 484 fs/gfs2/lops.c submit_bio(prev); prev 1443 fs/gfs2/quota.c qd = list_entry(head->prev, struct gfs2_quota_data, qd_list); prev 75 fs/gfs2/xattr.c struct gfs2_ea_header *prev, void *private); prev 80 fs/gfs2/xattr.c struct gfs2_ea_header *ea, *prev = NULL; prev 86 fs/gfs2/xattr.c for (ea = GFS2_EA_BH2FIRST(bh);; prev = ea, ea = GFS2_EA2NEXT(ea)) { prev 95 fs/gfs2/xattr.c error = ea_call(ip, bh, ea, prev, data); prev 165 fs/gfs2/xattr.c struct gfs2_ea_header *ea, struct gfs2_ea_header *prev, prev 180 fs/gfs2/xattr.c el->el_prev = prev; prev 225 fs/gfs2/xattr.c struct gfs2_ea_header *prev, void *private) prev 294 fs/gfs2/xattr.c if (prev && !leave) { prev 297 fs/gfs2/xattr.c len = GFS2_EA_REC_LEN(prev) + GFS2_EA_REC_LEN(ea); prev 298 fs/gfs2/xattr.c prev->ea_rec_len = cpu_to_be32(len); prev 301 fs/gfs2/xattr.c prev->ea_flags |= GFS2_EAFLAG_LAST; prev 319 fs/gfs2/xattr.c struct gfs2_ea_header *prev, int leave) prev 331 fs/gfs2/xattr.c error = ea_dealloc_unstuffed(ip, bh, ea, prev, (leave) ? &error : NULL); prev 344 fs/gfs2/xattr.c struct gfs2_ea_header *ea, struct gfs2_ea_header *prev, prev 830 fs/gfs2/xattr.c struct gfs2_ea_header *prev = el->el_prev; prev 835 fs/gfs2/xattr.c if (!prev || !GFS2_EA_IS_STUFFED(ea)) { prev 838 fs/gfs2/xattr.c } else if (GFS2_EA2NEXT(prev) != ea) { prev 839 fs/gfs2/xattr.c prev = GFS2_EA2NEXT(prev); prev 840 fs/gfs2/xattr.c gfs2_assert_withdraw(GFS2_SB(&ip->i_inode), GFS2_EA2NEXT(prev) == ea); prev 843 fs/gfs2/xattr.c len = GFS2_EA_REC_LEN(prev) + GFS2_EA_REC_LEN(ea); prev 844 fs/gfs2/xattr.c prev->ea_rec_len = cpu_to_be32(len); prev 847 fs/gfs2/xattr.c prev->ea_flags |= GFS2_EAFLAG_LAST; prev 910 fs/gfs2/xattr.c struct gfs2_ea_header *ea, struct gfs2_ea_header *prev, prev 925 fs/gfs2/xattr.c error = ea_remove_unstuffed(ip, bh, ea, prev, 1); prev 1077 fs/gfs2/xattr.c struct gfs2_ea_header *prev = el->el_prev; prev 1086 fs/gfs2/xattr.c if (prev) { prev 1089 fs/gfs2/xattr.c len = GFS2_EA_REC_LEN(prev) + GFS2_EA_REC_LEN(ea); prev 1090 fs/gfs2/xattr.c prev->ea_rec_len = cpu_to_be32(len); prev 1093 fs/gfs2/xattr.c prev->ea_flags |= GFS2_EAFLAG_LAST; prev 177 fs/hfs/bfind.c idx = bnode->prev; prev 143 fs/hfs/bnode.c be32_to_cpu(desc.next), be32_to_cpu(desc.prev), prev 178 fs/hfs/bnode.c if (node->prev) { prev 179 fs/hfs/bnode.c tmp = hfs_bnode_find(tree, node->prev); prev 193 fs/hfs/bnode.c tmp->prev = node->prev; prev 194 fs/hfs/bnode.c cnid = cpu_to_be32(tmp->prev); prev 195 fs/hfs/bnode.c hfs_bnode_write(tmp, &cnid, offsetof(struct hfs_bnode_desc, prev), 4); prev 198 fs/hfs/bnode.c tree->leaf_tail = node->prev; prev 201 fs/hfs/bnode.c if (!node->prev && !node->next) { prev 339 fs/hfs/bnode.c node->prev = be32_to_cpu(desc->prev); prev 248 fs/hfs/brec.c new_node->prev = node->this; prev 318 fs/hfs/brec.c node_desc.prev = cpu_to_be32(new_node->prev); prev 334 fs/hfs/brec.c next_node->prev = new_node->this; prev 336 fs/hfs/brec.c node_desc.prev = cpu_to_be32(next_node->prev); prev 488 fs/hfs/brec.c new_node->prev = 0; prev 492 fs/hfs/brec.c node_desc.prev = cpu_to_be32(new_node->prev); prev 188 fs/hfs/btree.c static struct hfs_bnode *hfs_bmap_new_bmap(struct hfs_bnode *prev, u32 idx) prev 190 fs/hfs/btree.c struct hfs_btree *tree = prev->tree; prev 202 fs/hfs/btree.c prev->next = idx; prev 204 fs/hfs/btree.c hfs_bnode_write(prev, &cnid, offsetof(struct hfs_bnode_desc, next), 4); prev 210 fs/hfs/btree.c desc.prev = 0; prev 49 fs/hfs/btree.h u32 prev; prev 128 fs/hfs/btree.h __be32 prev; /* (V) Number of the prev node at this level */ prev 246 fs/hfsplus/bfind.c idx = bnode->prev; prev 307 fs/hfsplus/bnode.c be32_to_cpu(desc.next), be32_to_cpu(desc.prev), prev 342 fs/hfsplus/bnode.c if (node->prev) { prev 343 fs/hfsplus/bnode.c tmp = hfs_bnode_find(tree, node->prev); prev 358 fs/hfsplus/bnode.c tmp->prev = node->prev; prev 359 fs/hfsplus/bnode.c cnid = cpu_to_be32(tmp->prev); prev 361 fs/hfsplus/bnode.c offsetof(struct hfs_bnode_desc, prev), 4); prev 364 fs/hfsplus/bnode.c tree->leaf_tail = node->prev; prev 367 fs/hfsplus/bnode.c if (!node->prev && !node->next) prev 507 fs/hfsplus/bnode.c node->prev = be32_to_cpu(desc->prev); prev 252 fs/hfsplus/brec.c new_node->prev = node->this; prev 322 fs/hfsplus/brec.c node_desc.prev = cpu_to_be32(new_node->prev); prev 338 fs/hfsplus/brec.c next_node->prev = new_node->this; prev 340 fs/hfsplus/brec.c node_desc.prev = cpu_to_be32(next_node->prev); prev 492 fs/hfsplus/brec.c new_node->prev = 0; prev 496 fs/hfsplus/brec.c node_desc.prev = cpu_to_be32(new_node->prev); prev 312 fs/hfsplus/btree.c static struct hfs_bnode *hfs_bmap_new_bmap(struct hfs_bnode *prev, u32 idx) prev 314 fs/hfsplus/btree.c struct hfs_btree *tree = prev->tree; prev 324 fs/hfsplus/btree.c prev->next = idx; prev 326 fs/hfsplus/btree.c hfs_bnode_write(prev, &cnid, offsetof(struct hfs_bnode_desc, next), 4); prev 332 fs/hfsplus/btree.c desc.prev = 0; prev 106 fs/hfsplus/hfsplus_fs.h u32 prev; prev 153 fs/hfsplus/hfsplus_raw.h __be32 prev; prev 1948 fs/io_uring.c struct io_kiocb *req, *prev; prev 1962 fs/io_uring.c prev = req; prev 1963 fs/io_uring.c list_for_each_entry_continue_reverse(prev, &ctx->timeout_list, list) prev 1964 fs/io_uring.c prev->sequence++; prev 2005 fs/io_uring.c entry = ctx->timeout_list.prev; prev 2577 fs/io_uring.c struct io_kiocb *prev = *link; prev 2587 fs/io_uring.c list_add_tail(&req->list, &prev->link_list); prev 388 fs/iomap/buffered-io.c loff_t pos = page_offset(list_entry(pages->prev, struct page, lru)); prev 12 fs/iomap/fiemap.c struct iomap prev; prev 55 fs/iomap/fiemap.c ret = iomap_to_fiemap(ctx->fi, &ctx->prev, 0); prev 56 fs/iomap/fiemap.c ctx->prev = *iomap; prev 75 fs/iomap/fiemap.c ctx.prev.type = IOMAP_HOLE; prev 102 fs/iomap/fiemap.c if (ctx.prev.type != IOMAP_HOLE) { prev 103 fs/iomap/fiemap.c ret = iomap_to_fiemap(fi, &ctx.prev, FIEMAP_EXTENT_LAST); prev 806 fs/jbd2/commit.c struct buffer_head *bh = list_entry(io_bufs.prev, prev 852 fs/jbd2/commit.c bh = list_entry(log_bufs.prev, struct buffer_head, b_assoc_buffers); prev 322 fs/jffs2/compr.c list_add(&comp->list, this->list.prev); prev 212 fs/jffs2/erase.c struct jffs2_raw_node_ref **prev; prev 214 fs/jffs2/erase.c prev = &ref->next_in_ino; prev 218 fs/jffs2/erase.c if (!(*prev)->next_in_ino) { prev 222 fs/jffs2/erase.c ic = (struct jffs2_inode_cache *)(*prev); prev 223 fs/jffs2/erase.c prev = &ic->nodes; prev 227 fs/jffs2/erase.c if (SECTOR_ADDR((*prev)->flash_offset) == jeb->offset) { prev 231 fs/jffs2/erase.c this = *prev; prev 232 fs/jffs2/erase.c *prev = this->next_in_ino; prev 241 fs/jffs2/erase.c prev = &((*prev)->next_in_ino); prev 28 fs/jffs2/nodelist.c struct jffs2_full_dirent **prev = list; prev 32 fs/jffs2/nodelist.c while ((*prev) && (*prev)->nhash <= new->nhash) { prev 33 fs/jffs2/nodelist.c if ((*prev)->nhash == new->nhash && !strcmp((*prev)->name, new->name)) { prev 35 fs/jffs2/nodelist.c if (new->version < (*prev)->version) { prev 37 fs/jffs2/nodelist.c (*prev)->name, (*prev)->ino); prev 42 fs/jffs2/nodelist.c (*prev)->name, (*prev)->ino); prev 43 fs/jffs2/nodelist.c new->next = (*prev)->next; prev 46 fs/jffs2/nodelist.c if ((*prev)->raw) prev 47 fs/jffs2/nodelist.c jffs2_mark_node_obsolete(c, ((*prev)->raw)); prev 48 fs/jffs2/nodelist.c jffs2_free_full_dirent(*prev); prev 49 fs/jffs2/nodelist.c *prev = new; prev 53 fs/jffs2/nodelist.c prev = &((*prev)->next); prev 55 fs/jffs2/nodelist.c new->next = *prev; prev 56 fs/jffs2/nodelist.c *prev = new; prev 386 fs/jffs2/nodelist.c struct jffs2_node_frag *prev = frag_prev(newfrag); prev 390 fs/jffs2/nodelist.c if (prev->node) prev 391 fs/jffs2/nodelist.c mark_ref_normal(prev->node->raw); prev 438 fs/jffs2/nodelist.c struct jffs2_inode_cache **prev; prev 446 fs/jffs2/nodelist.c prev = &c->inocache_list[new->ino % c->inocache_hashsize]; prev 448 fs/jffs2/nodelist.c while ((*prev) && (*prev)->ino < new->ino) { prev 449 fs/jffs2/nodelist.c prev = &(*prev)->next; prev 451 fs/jffs2/nodelist.c new->next = *prev; prev 452 fs/jffs2/nodelist.c *prev = new; prev 459 fs/jffs2/nodelist.c struct jffs2_inode_cache **prev; prev 467 fs/jffs2/nodelist.c prev = &c->inocache_list[old->ino % c->inocache_hashsize]; prev 469 fs/jffs2/nodelist.c while ((*prev) && (*prev)->ino < old->ino) { prev 470 fs/jffs2/nodelist.c prev = &(*prev)->next; prev 472 fs/jffs2/nodelist.c if ((*prev) == old) { prev 473 fs/jffs2/nodelist.c *prev = old->next; prev 529 fs/jffs2/nodelist.c struct jffs2_node_frag *prev = NULL; prev 541 fs/jffs2/nodelist.c if (!prev || frag->ofs > prev->ofs) prev 542 fs/jffs2/nodelist.c prev = frag; prev 554 fs/jffs2/nodelist.c if (prev) prev 556 fs/jffs2/nodelist.c prev->ofs, prev->ofs+prev->size); prev 560 fs/jffs2/nodelist.c return prev; prev 1172 fs/jfs/jfs_dtree.c sp->header.prev != 0 || skip > 1) { prev 1400 fs/jfs/jfs_dtree.c rp->header.prev = cpu_to_le64(addressPXD(&sp->header.self)); prev 1484 fs/jfs/jfs_dtree.c p->header.prev = cpu_to_le64(rbn); prev 1928 fs/jfs/jfs_dtree.c rp->header.prev = 0; prev 2410 fs/jfs/jfs_dtree.c ((p->header.flag & BT_ROOT) || p->header.prev == 0)) prev 2496 fs/jfs/jfs_dtree.c if (p->header.prev) { prev 2497 fs/jfs/jfs_dtree.c prevbn = le64_to_cpu(p->header.prev); prev 2537 fs/jfs/jfs_dtree.c rp->header.prev = cpu_to_le64(nxaddr); prev 2745 fs/jfs/jfs_dtree.c prevbn = le64_to_cpu(p->header.prev); prev 2772 fs/jfs/jfs_dtree.c p->header.prev = cpu_to_le64(prevbn); prev 183 fs/jfs/jfs_dtree.h __le64 prev; /* 8: previous sibling */ prev 355 fs/jfs/jfs_extent.c s64 prev; prev 367 fs/jfs/jfs_extent.c prev = ((offset & ~POFFSET) >> JFS_SBI(sb)->l2bsize) - nbperpage; prev 371 fs/jfs/jfs_extent.c if (prev < 0) prev 374 fs/jfs/jfs_extent.c rc = xtLookup(ip, prev, nbperpage, &xflag, &xaddr, &xlen, 0); prev 383 fs/jfs/jfs_extent.c XADoffset(xp, prev); prev 584 fs/jfs/jfs_logmgr.c tblk = list_entry(log->cqueue.prev, struct tblock, cqueue); prev 1534 fs/jfs/jfs_logmgr.c target = list_entry(log->cqueue.prev, struct tblock, cqueue); prev 1025 fs/jfs/jfs_xtree.c rp->header.prev = cpu_to_le64(addressPXD(&sp->header.self)); prev 1091 fs/jfs/jfs_xtree.c p->header.prev = cpu_to_le64(rbn); prev 1267 fs/jfs/jfs_xtree.c rp->header.prev = 0; prev 2727 fs/jfs/jfs_xtree.c if (p->header.prev) { prev 2728 fs/jfs/jfs_xtree.c prevbn = le64_to_cpu(p->header.prev); prev 2754 fs/jfs/jfs_xtree.c rp->header.prev = cpu_to_le64(nxaddr); prev 3006 fs/jfs/jfs_xtree.c prevbn = le64_to_cpu(p->header.prev); prev 3024 fs/jfs/jfs_xtree.c p->header.prev = cpu_to_le64(prevbn); prev 3535 fs/jfs/jfs_xtree.c struct tlock *prev; prev 3542 fs/jfs/jfs_xtree.c for (prev = lid_to_tlock(tblk->next); prev 3543 fs/jfs/jfs_xtree.c prev->next != lid; prev 3544 fs/jfs/jfs_xtree.c prev = lid_to_tlock(prev->next)) { prev 3545 fs/jfs/jfs_xtree.c assert(prev->next); prev 3547 fs/jfs/jfs_xtree.c prev->next = tlck->next; prev 74 fs/jfs/jfs_xtree.h __le64 prev; /* 8: */ prev 889 fs/namespace.c list_splice(&head, n->list.prev); prev 916 fs/namespace.c struct list_head *prev = p->mnt_mounts.prev; prev 917 fs/namespace.c while (prev != &p->mnt_mounts) { prev 918 fs/namespace.c p = list_entry(prev, struct mount, mnt_child); prev 919 fs/namespace.c prev = p->mnt_mounts.prev; prev 878 fs/nfs/blocklayout/blocklayout.c bl_pg_test_read(struct nfs_pageio_descriptor *pgio, struct nfs_page *prev, prev 883 fs/nfs/blocklayout/blocklayout.c return pnfs_generic_pg_test(pgio, prev, req); prev 941 fs/nfs/blocklayout/blocklayout.c bl_pg_test_write(struct nfs_pageio_descriptor *pgio, struct nfs_page *prev, prev 946 fs/nfs/blocklayout/blocklayout.c return pnfs_generic_pg_test(pgio, prev, req); prev 502 fs/nfs/delegation.c struct nfs_delegation *prev; prev 521 fs/nfs/delegation.c prev = NULL; prev 538 fs/nfs/delegation.c prev = delegation; prev 544 fs/nfs/delegation.c if (prev) { prev 547 fs/nfs/delegation.c tmp = nfs_delegation_grab_inode(prev); prev 551 fs/nfs/delegation.c place_holder_deleg = prev; prev 2374 fs/nfs/dir.c lh = rcu_dereference(nfsi->access_cache_entry_lru.prev); prev 847 fs/nfs/filelayout/filelayout.c filelayout_pg_test(struct nfs_pageio_descriptor *pgio, struct nfs_page *prev, prev 857 fs/nfs/filelayout/filelayout.c size = pnfs_generic_pg_test(pgio, prev, req); prev 862 fs/nfs/filelayout/filelayout.c if (prev) { prev 863 fs/nfs/filelayout/filelayout.c p_stripe = (u64)req_offset(prev) - segment_offset; prev 493 fs/nfs/nfs4client.c struct nfs_client **prev, struct nfs_net *nn) prev 511 fs/nfs/nfs4client.c nfs_put_client(*prev); prev 512 fs/nfs/nfs4client.c *prev = pos; prev 554 fs/nfs/nfs4client.c struct nfs_client *pos, *prev = NULL; prev 567 fs/nfs/nfs4client.c status = nfs4_match_client(pos, new, &prev, nn); prev 591 fs/nfs/nfs4client.c nfs_put_client(prev); prev 592 fs/nfs/nfs4client.c prev = pos; prev 603 fs/nfs/nfs4client.c prev = NULL; prev 623 fs/nfs/nfs4client.c nfs_put_client(prev); prev 720 fs/nfs/nfs4client.c struct nfs_client *pos, *prev = NULL; prev 729 fs/nfs/nfs4client.c status = nfs4_match_client(pos, new, &prev, nn); prev 753 fs/nfs/nfs4client.c nfs_put_client(prev); prev 2827 fs/nfs/nfs4proc.c struct nfs4_lock_state *lsp, *prev = NULL; prev 2841 fs/nfs/nfs4proc.c nfs4_put_lock_state(prev); prev 2842 fs/nfs/nfs4proc.c prev = lsp; prev 2856 fs/nfs/nfs4proc.c nfs4_put_lock_state(prev); prev 2863 fs/nfs/nfs4proc.c nfs4_put_lock_state(prev); prev 234 fs/nfs/pagelist.c nfs_page_group_init(struct nfs_page *req, struct nfs_page *prev) prev 237 fs/nfs/pagelist.c WARN_ON_ONCE(prev == req); prev 239 fs/nfs/pagelist.c if (!prev) { prev 245 fs/nfs/pagelist.c WARN_ON_ONCE(prev->wb_this_page != prev->wb_head); prev 246 fs/nfs/pagelist.c WARN_ON_ONCE(!test_bit(PG_HEADLOCK, &prev->wb_head->wb_flags)); prev 247 fs/nfs/pagelist.c req->wb_head = prev->wb_head; prev 248 fs/nfs/pagelist.c req->wb_this_page = prev->wb_this_page; prev 249 fs/nfs/pagelist.c prev->wb_this_page = req; prev 258 fs/nfs/pagelist.c if (test_bit(PG_INODE_REF, &prev->wb_head->wb_flags)) { prev 493 fs/nfs/pagelist.c struct nfs_page *prev, struct nfs_page *req) prev 916 fs/nfs/pagelist.c static bool nfs_can_coalesce_requests(struct nfs_page *prev, prev 923 fs/nfs/pagelist.c if (prev) { prev 924 fs/nfs/pagelist.c if (!nfs_match_open_context(nfs_req_openctx(req), nfs_req_openctx(prev))) prev 931 fs/nfs/pagelist.c prev->wb_lock_context)) prev 933 fs/nfs/pagelist.c if (req_offset(req) != req_offset(prev) + prev->wb_bytes) prev 935 fs/nfs/pagelist.c if (req->wb_page == prev->wb_page) { prev 936 fs/nfs/pagelist.c if (req->wb_pgbase != prev->wb_pgbase + prev->wb_bytes) prev 940 fs/nfs/pagelist.c prev->wb_pgbase + prev->wb_bytes != PAGE_SIZE) prev 944 fs/nfs/pagelist.c size = pgio->pg_ops->pg_test(pgio, prev, req); prev 964 fs/nfs/pagelist.c struct nfs_page *prev = NULL; prev 967 fs/nfs/pagelist.c prev = nfs_list_entry(mirror->pg_list.prev); prev 984 fs/nfs/pagelist.c if (!nfs_can_coalesce_requests(prev, req, desc)) prev 1299 fs/nfs/pagelist.c struct nfs_page *prev; prev 1305 fs/nfs/pagelist.c prev = nfs_list_entry(mirror->pg_list.prev); prev 1306 fs/nfs/pagelist.c if (index != prev->wb_index + 1) { prev 2566 fs/nfs/pnfs.c struct nfs_page *prev, struct nfs_page *req) prev 2571 fs/nfs/pnfs.c size = nfs_generic_pg_test(pgio, prev, req); prev 252 fs/nfs/pnfs.h struct nfs_page *prev, struct nfs_page *req); prev 202 fs/nilfs2/alloc.c struct nilfs_bh_assoc *prev, prev 208 fs/nilfs2/alloc.c if (prev->bh && blkoff == prev->blkoff) { prev 209 fs/nilfs2/alloc.c get_bh(prev->bh); prev 210 fs/nilfs2/alloc.c *bhp = prev->bh; prev 223 fs/nilfs2/alloc.c brelse(prev->bh); prev 225 fs/nilfs2/alloc.c prev->bh = *bhp; prev 226 fs/nilfs2/alloc.c prev->blkoff = blkoff; prev 240 fs/nilfs2/alloc.c struct nilfs_bh_assoc *prev, prev 244 fs/nilfs2/alloc.c if (prev->bh && blkoff == prev->blkoff) { prev 245 fs/nilfs2/alloc.c brelse(prev->bh); prev 246 fs/nilfs2/alloc.c prev->bh = NULL; prev 623 fs/nilfs2/cpfile.c __u64 curr, prev; prev 659 fs/nilfs2/cpfile.c prev = le64_to_cpu(list->ssl_prev); prev 660 fs/nilfs2/cpfile.c while (prev > cno) { prev 661 fs/nilfs2/cpfile.c prev_blkoff = nilfs_cpfile_get_blkoff(cpfile, prev); prev 662 fs/nilfs2/cpfile.c curr = prev; prev 676 fs/nilfs2/cpfile.c prev = le64_to_cpu(list->ssl_prev); prev 680 fs/nilfs2/cpfile.c if (prev != 0) { prev 681 fs/nilfs2/cpfile.c ret = nilfs_cpfile_get_checkpoint_block(cpfile, prev, 0, prev 699 fs/nilfs2/cpfile.c cp->cp_snapshot_list.ssl_prev = cpu_to_le64(prev); prev 705 fs/nilfs2/cpfile.c cpfile, prev, prev_bh, kaddr); prev 742 fs/nilfs2/cpfile.c __u64 next, prev; prev 768 fs/nilfs2/cpfile.c prev = le64_to_cpu(list->ssl_prev); prev 783 fs/nilfs2/cpfile.c if (prev != 0) { prev 784 fs/nilfs2/cpfile.c ret = nilfs_cpfile_get_checkpoint_block(cpfile, prev, 0, prev 796 fs/nilfs2/cpfile.c list->ssl_prev = cpu_to_le64(prev); prev 801 fs/nilfs2/cpfile.c cpfile, prev, prev_bh, kaddr); prev 78 fs/nilfs2/segbuf.c struct nilfs_segment_buffer *prev) prev 80 fs/nilfs2/segbuf.c segbuf->sb_segnum = prev->sb_segnum; prev 81 fs/nilfs2/segbuf.c segbuf->sb_fseg_start = prev->sb_fseg_start; prev 82 fs/nilfs2/segbuf.c segbuf->sb_fseg_end = prev->sb_fseg_end; prev 83 fs/nilfs2/segbuf.c segbuf->sb_pseg_start = prev->sb_pseg_start + prev->sb_sum.nblocks; prev 88 fs/nilfs2/segbuf.h #define NILFS_PREV_SEGBUF(segbuf) NILFS_LIST_SEGBUF((segbuf)->sb_list.prev) prev 89 fs/nilfs2/segbuf.h #define NILFS_LAST_SEGBUF(head) NILFS_LIST_SEGBUF((head)->prev) prev 111 fs/nilfs2/segbuf.h struct nilfs_segment_buffer *prev); prev 1153 fs/nilfs2/segment.c ii->i_dirty.prev, prev 1174 fs/nilfs2/segment.c list_entry(ii->i_dirty.prev, prev 1276 fs/nilfs2/segment.c struct nilfs_segment_buffer *segbuf, *prev; prev 1300 fs/nilfs2/segment.c prev = NILFS_LAST_SEGBUF(&sci->sc_write_logs); prev 1301 fs/nilfs2/segment.c nilfs_segbuf_map_cont(segbuf, prev); prev 1302 fs/nilfs2/segment.c segbuf->sb_sum.seg_seq = prev->sb_sum.seg_seq; prev 1303 fs/nilfs2/segment.c nextnum = prev->sb_nextnum; prev 1306 fs/nilfs2/segment.c nilfs_segbuf_map(segbuf, prev->sb_nextnum, 0, nilfs); prev 1336 fs/nilfs2/segment.c struct nilfs_segment_buffer *segbuf, *prev; prev 1342 fs/nilfs2/segment.c prev = NILFS_LAST_SEGBUF(&sci->sc_segbufs); prev 1349 fs/nilfs2/segment.c err = nilfs_sufile_mark_dirty(sufile, prev->sb_nextnum); prev 1361 fs/nilfs2/segment.c nilfs_segbuf_map(segbuf, prev->sb_nextnum, 0, nilfs); prev 1369 fs/nilfs2/segment.c segbuf->sb_sum.seg_seq = prev->sb_sum.seg_seq + 1; prev 1373 fs/nilfs2/segment.c prev = segbuf; prev 1392 fs/nilfs2/segment.c struct nilfs_segment_buffer *segbuf, *prev; prev 1414 fs/nilfs2/segment.c prev = segbuf; prev 1416 fs/nilfs2/segment.c if (prev->sb_nextnum != segbuf->sb_nextnum) { prev 1424 fs/nilfs2/segment.c prev = segbuf; prev 82 fs/notify/dnotify/dnotify.c struct dnotify_struct **prev; prev 96 fs/notify/dnotify/dnotify.c prev = &dn_mark->dn; prev 97 fs/notify/dnotify/dnotify.c while ((dn = *prev) != NULL) { prev 99 fs/notify/dnotify/dnotify.c prev = &dn->dn_next; prev 105 fs/notify/dnotify/dnotify.c prev = &dn->dn_next; prev 107 fs/notify/dnotify/dnotify.c *prev = dn->dn_next; prev 146 fs/notify/dnotify/dnotify.c struct dnotify_struct **prev; prev 162 fs/notify/dnotify/dnotify.c prev = &dn_mark->dn; prev 163 fs/notify/dnotify/dnotify.c while ((dn = *prev) != NULL) { prev 165 fs/notify/dnotify/dnotify.c *prev = dn->dn_next; prev 170 fs/notify/dnotify/dnotify.c prev = &dn->dn_next; prev 54 fs/notify/inotify/inotify_fsnotify.c last_event = list_entry(list->prev, struct fsnotify_event, list); prev 1526 fs/ocfs2/dir.c struct ocfs2_dir_block_trailer *trailer, *prev; prev 1538 fs/ocfs2/dir.c prev = ocfs2_trailer_from_bh(bh, dir->i_sb); prev 1539 fs/ocfs2/dir.c prev->db_free_next = trailer->db_free_next; prev 264 fs/ocfs2/extent_map.c emi = list_entry(em->em_list.prev, prev 1075 fs/ocfs2/quota_local.c chunk->qc_num = list_entry(chunk->qc_chunk.prev, prev 1111 fs/ocfs2/quota_local.c chunk = list_entry(oinfo->dqi_chunk.prev, prev 1359 fs/ocfs2/refcounttree.c static int ocfs2_refcount_rec_no_intersect(struct ocfs2_refcount_rec *prev, prev 1362 fs/ocfs2/refcounttree.c if (ocfs2_get_ref_rec_low_cpos(prev) + le32_to_cpu(prev->r_clusters) <= prev 469 fs/ocfs2/reservations.c struct rb_node *prev, *next; prev 561 fs/ocfs2/reservations.c prev = &prev_resv->r_node; prev 565 fs/ocfs2/reservations.c next = rb_next(prev); prev 609 fs/ocfs2/reservations.c prev = next; prev 610 fs/ocfs2/reservations.c prev_resv = rb_entry(prev, struct ocfs2_alloc_reservation, prev 165 fs/omfs/dir.c u64 block, prev; prev 177 fs/omfs/dir.c bh2 = omfs_scan_list(dir, block, name, namelen, &prev); prev 187 fs/omfs/dir.c if (prev != ~0) { prev 190 fs/omfs/dir.c bh = omfs_bread(dir->i_sb, prev); prev 201 fs/omfs/dir.c if (prev != ~0) { prev 202 fs/omfs/dir.c dirty = omfs_iget(dir->i_sb, prev); prev 653 fs/orangefs/devorangefs-req.c if (!orangefs_sb->list.prev) prev 625 fs/orangefs/super.c ORANGEFS_SB(sb)->list.prev = NULL; prev 29 fs/pnode.c return list_entry(p->mnt_slave_list.prev, struct mount, mnt_slave); prev 113 fs/pnode.c list_splice(&mnt->mnt_slave_list, master->mnt_slave_list.prev); prev 221 fs/readdir.c struct linux_dirent __user *dirent, *prev; prev 244 fs/readdir.c prev = (void __user *) dirent - prev_reclen; prev 245 fs/readdir.c if (!user_access_begin(prev, reclen + prev_reclen)) prev 249 fs/readdir.c unsafe_put_user(offset, &prev->d_off, efault_end); prev 312 fs/readdir.c struct linux_dirent64 __user *dirent, *prev; prev 329 fs/readdir.c prev = (void __user *)dirent - prev_reclen; prev 330 fs/readdir.c if (!user_access_begin(prev, reclen + prev_reclen)) prev 334 fs/readdir.c unsafe_put_user(offset, &prev->d_off, efault_end); prev 264 fs/reiserfs/item_ops.c __u32 prev = INT_MAX; prev 274 fs/reiserfs/item_ops.c if (sequence_finished(prev, &num, get_block_num(unp, j))) { prev 275 fs/reiserfs/item_ops.c print_sequence(prev, num); prev 276 fs/reiserfs/item_ops.c start_new_sequence(&prev, &num, get_block_num(unp, j)); prev 279 fs/reiserfs/item_ops.c print_sequence(prev, num); prev 358 fs/reiserfs/journal.c head[0].prev = NULL; prev 361 fs/reiserfs/journal.c head[i].prev = head + (i - 1); prev 386 fs/reiserfs/journal.c cn->next->prev = NULL; prev 408 fs/reiserfs/journal.c journal->j_cnode_free_list->prev = cn; prev 410 fs/reiserfs/journal.c cn->prev = NULL; /* not needed with the memset, but I might kill the memset, and forget to do this */ prev 851 fs/reiserfs/journal.c jh = JH_ENTRY(tmp.prev); prev 900 fs/reiserfs/journal.c entry = jl->j_list.prev; prev 908 fs/reiserfs/journal.c entry = other_jl->j_list.prev; prev 3361 fs/reiserfs/journal.c cn->prev = journal->j_last; prev 3428 fs/reiserfs/journal.c if (cn->prev) { prev 3429 fs/reiserfs/journal.c cn->prev->next = cn->next; prev 3432 fs/reiserfs/journal.c cn->next->prev = cn->prev; prev 3438 fs/reiserfs/journal.c journal->j_last = cn->prev; prev 3539 fs/reiserfs/journal.c entry = journal->j_journal_list.prev; prev 4126 fs/reiserfs/journal.c jl_cn->prev = last_cn; prev 1400 fs/reiserfs/lbalance.c int prev, next; prev 1408 fs/reiserfs/lbalance.c prev = (i != 0) ? deh_location(&deh[i - 1]) : 0; prev 1410 fs/reiserfs/lbalance.c if (prev && prev <= deh_location(&deh[i])) prev 239 fs/reiserfs/reiserfs.h struct reiserfs_journal_cnode *prev; /* prev in transaction list */ prev 113 fs/squashfs/decompressor_multi.c decomp_strm = list_entry(stream->strm_list.prev, prev 137 fs/squashfs/decompressor_multi.c decomp_strm = list_entry(stream->strm_list.prev, prev 664 fs/ubifs/commit.c i = list_entry(list.prev, struct idx_node, prev 708 fs/ubifs/commit.c i = list_entry(list.prev, struct idx_node, list); prev 1482 fs/ubifs/debug.c struct ubifs_znode *prev; prev 1501 fs/ubifs/debug.c prev = znode; prev 1510 fs/ubifs/debug.c last = prev->child_cnt - 1; prev 1511 fs/ubifs/debug.c if (prev->level == 0 && znode->level == 0 && !c->replaying && prev 1512 fs/ubifs/debug.c !keys_cmp(c, &prev->zbranch[last].key, prev 1514 fs/ubifs/debug.c err = dbg_check_key_order(c, &prev->zbranch[last], prev 1520 fs/ubifs/debug.c ubifs_dump_znode(c, prev); prev 106 fs/ubifs/master.c snod = list_entry(sleb->nodes.prev, struct ubifs_scan_node, prev 124 fs/ubifs/master.c snod = list_entry(sleb->nodes.prev, struct ubifs_scan_node, list); prev 500 fs/ubifs/recovery.c snod = list_entry(sleb->nodes.prev, prev 568 fs/ubifs/recovery.c snod = list_entry(sleb->nodes.prev, struct ubifs_scan_node, prev 596 fs/ubifs/recovery.c snod = list_entry(sleb->nodes.prev, struct ubifs_scan_node, prev 879 fs/userfaultfd.c struct vm_area_struct *vma, *prev; prev 900 fs/userfaultfd.c prev = NULL; prev 906 fs/userfaultfd.c prev = vma; prev 911 fs/userfaultfd.c prev = vma_merge(mm, prev, vma->vm_start, vma->vm_end, prev 916 fs/userfaultfd.c if (prev) prev 917 fs/userfaultfd.c vma = prev; prev 919 fs/userfaultfd.c prev = vma; prev 1306 fs/userfaultfd.c struct vm_area_struct *vma, *prev, *cur; prev 1356 fs/userfaultfd.c vma = find_vma_prev(mm, start, &prev); prev 1440 fs/userfaultfd.c prev = vma; prev 1464 fs/userfaultfd.c prev = vma_merge(mm, prev, start, vma_end, new_flags, prev 1468 fs/userfaultfd.c if (prev) { prev 1469 fs/userfaultfd.c vma = prev; prev 1492 fs/userfaultfd.c prev = vma; prev 1518 fs/userfaultfd.c struct vm_area_struct *vma, *prev, *cur; prev 1545 fs/userfaultfd.c vma = find_vma_prev(mm, start, &prev); prev 1591 fs/userfaultfd.c prev = vma; prev 1626 fs/userfaultfd.c prev = vma_merge(mm, prev, start, vma_end, new_flags, prev 1630 fs/userfaultfd.c if (prev) { prev 1631 fs/userfaultfd.c vma = prev; prev 1654 fs/userfaultfd.c prev = vma; prev 3068 fs/xfs/libxfs/xfs_bmap.c if (ap->eof && ap->prev.br_startoff != NULLFILEOFF && prev 3069 fs/xfs/libxfs/xfs_bmap.c !isnullstartblock(ap->prev.br_startblock) && prev 3070 fs/xfs/libxfs/xfs_bmap.c ISVALID(ap->prev.br_startblock + ap->prev.br_blockcount, prev 3071 fs/xfs/libxfs/xfs_bmap.c ap->prev.br_startblock)) { prev 3072 fs/xfs/libxfs/xfs_bmap.c ap->blkno = ap->prev.br_startblock + ap->prev.br_blockcount; prev 3077 fs/xfs/libxfs/xfs_bmap.c (ap->prev.br_startoff + ap->prev.br_blockcount); prev 3079 fs/xfs/libxfs/xfs_bmap.c ISVALID(ap->blkno + adjust, ap->prev.br_startblock)) prev 3097 fs/xfs/libxfs/xfs_bmap.c if (ap->prev.br_startoff != NULLFILEOFF && prev 3098 fs/xfs/libxfs/xfs_bmap.c !isnullstartblock(ap->prev.br_startblock) && prev 3099 fs/xfs/libxfs/xfs_bmap.c (prevbno = ap->prev.br_startblock + prev 3100 fs/xfs/libxfs/xfs_bmap.c ap->prev.br_blockcount) && prev 3101 fs/xfs/libxfs/xfs_bmap.c ISVALID(prevbno, ap->prev.br_startblock)) { prev 3106 fs/xfs/libxfs/xfs_bmap.c (ap->prev.br_startoff + prev 3107 fs/xfs/libxfs/xfs_bmap.c ap->prev.br_blockcount); prev 3118 fs/xfs/libxfs/xfs_bmap.c ap->prev.br_startblock)) prev 3417 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmap_extsize_align(mp, &ap->got, &ap->prev, prev 3944 fs/xfs/libxfs/xfs_bmap.c struct xfs_bmbt_irec prev; prev 3947 fs/xfs/libxfs/xfs_bmap.c if (!xfs_iext_peek_prev_extent(ifp, icur, &prev)) prev 3948 fs/xfs/libxfs/xfs_bmap.c prev.br_startoff = NULLFILEOFF; prev 3950 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmap_extsize_align(mp, got, &prev, extsz, 0, eof, prev 4031 fs/xfs/libxfs/xfs_bmap.c xfs_iext_peek_prev_extent(ifp, &bma->icur, &bma->prev); prev 4338 fs/xfs/libxfs/xfs_bmap.c if (!xfs_iext_peek_prev_extent(ifp, &bma.icur, &bma.prev)) prev 4339 fs/xfs/libxfs/xfs_bmap.c bma.prev.br_startoff = NULLFILEOFF; prev 4426 fs/xfs/libxfs/xfs_bmap.c bma.prev = bma.got; prev 4513 fs/xfs/libxfs/xfs_bmap.c if (!xfs_iext_peek_prev_extent(ifp, &bma.icur, &bma.prev)) prev 4514 fs/xfs/libxfs/xfs_bmap.c bma.prev.br_startoff = NULLFILEOFF; prev 5405 fs/xfs/libxfs/xfs_bmap.c struct xfs_bmbt_irec prev; prev 5413 fs/xfs/libxfs/xfs_bmap.c if (!xfs_iext_prev_extent(ifp, &icur, &prev)) prev 5415 fs/xfs/libxfs/xfs_bmap.c ASSERT(prev.br_state == XFS_EXT_NORM); prev 5416 fs/xfs/libxfs/xfs_bmap.c ASSERT(!isnullstartblock(prev.br_startblock)); prev 5418 fs/xfs/libxfs/xfs_bmap.c prev.br_startblock + prev.br_blockcount); prev 5419 fs/xfs/libxfs/xfs_bmap.c if (prev.br_startoff < start) { prev 5420 fs/xfs/libxfs/xfs_bmap.c mod = start - prev.br_startoff; prev 5421 fs/xfs/libxfs/xfs_bmap.c prev.br_blockcount -= mod; prev 5422 fs/xfs/libxfs/xfs_bmap.c prev.br_startblock += mod; prev 5423 fs/xfs/libxfs/xfs_bmap.c prev.br_startoff = start; prev 5425 fs/xfs/libxfs/xfs_bmap.c prev.br_state = XFS_EXT_UNWRITTEN; prev 5428 fs/xfs/libxfs/xfs_bmap.c &prev, &logflags); prev 5661 fs/xfs/libxfs/xfs_bmap.c struct xfs_bmbt_irec prev = *got; prev 5669 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_lookup_eq(cur, &prev, &i); prev 5685 fs/xfs/libxfs/xfs_bmap.c xfs_rmap_unmap_extent(tp, ip, whichfork, &prev); prev 5702 fs/xfs/libxfs/xfs_bmap.c struct xfs_bmbt_irec got, prev; prev 5740 fs/xfs/libxfs/xfs_bmap.c if (xfs_iext_peek_prev_extent(ifp, &icur, &prev)) { prev 5741 fs/xfs/libxfs/xfs_bmap.c if (new_startoff < prev.br_startoff + prev.br_blockcount) { prev 5746 fs/xfs/libxfs/xfs_bmap.c if (xfs_bmse_can_merge(&prev, &got, offset_shift_fsb)) { prev 5748 fs/xfs/libxfs/xfs_bmap.c offset_shift_fsb, &icur, &got, &prev, prev 24 fs/xfs/libxfs/xfs_bmap.h struct xfs_bmbt_irec prev; /* extent before the new one */ prev 124 fs/xfs/libxfs/xfs_iext_tree.c struct xfs_iext_leaf *prev; prev 268 fs/xfs/libxfs/xfs_iext_tree.c if (ifp->if_height > 1 && cur->leaf->prev) { prev 269 fs/xfs/libxfs/xfs_iext_tree.c cur->leaf = cur->leaf->prev; prev 405 fs/xfs/libxfs/xfs_iext_tree.c struct xfs_iext_leaf *prev = ifp->if_u1.if_root; prev 407 fs/xfs/libxfs/xfs_iext_tree.c node->keys[0] = xfs_iext_leaf_key(prev, 0); prev 408 fs/xfs/libxfs/xfs_iext_tree.c node->ptrs[0] = prev; prev 410 fs/xfs/libxfs/xfs_iext_tree.c struct xfs_iext_node *prev = ifp->if_u1.if_root; prev 414 fs/xfs/libxfs/xfs_iext_tree.c node->keys[0] = prev->keys[0]; prev 415 fs/xfs/libxfs/xfs_iext_tree.c node->ptrs[0] = prev; prev 572 fs/xfs/libxfs/xfs_iext_tree.c leaf->next->prev = new; prev 574 fs/xfs/libxfs/xfs_iext_tree.c new->prev = leaf; prev 687 fs/xfs/libxfs/xfs_iext_tree.c struct xfs_iext_node *prev = parent->ptrs[*pos - 1]; prev 688 fs/xfs/libxfs/xfs_iext_tree.c int nr_prev = xfs_iext_node_nr_entries(prev, 0), i; prev 692 fs/xfs/libxfs/xfs_iext_tree.c prev->keys[nr_prev + i] = node->keys[i]; prev 693 fs/xfs/libxfs/xfs_iext_tree.c prev->ptrs[nr_prev + i] = node->ptrs[i]; prev 803 fs/xfs/libxfs/xfs_iext_tree.c if (leaf->prev) { prev 804 fs/xfs/libxfs/xfs_iext_tree.c int nr_prev = xfs_iext_leaf_nr_entries(ifp, leaf->prev, 0), i; prev 808 fs/xfs/libxfs/xfs_iext_tree.c leaf->prev->recs[nr_prev + i] = leaf->recs[i]; prev 811 fs/xfs/libxfs/xfs_iext_tree.c cur->leaf = leaf->prev; prev 845 fs/xfs/libxfs/xfs_iext_tree.c if (leaf->prev) prev 846 fs/xfs/libxfs/xfs_iext_tree.c leaf->prev->next = leaf->next; prev 848 fs/xfs/libxfs/xfs_iext_tree.c leaf->next->prev = leaf->prev; prev 734 fs/xfs/xfs_aops.c struct bio *prev) prev 739 fs/xfs/xfs_aops.c bio_copy_dev(new, prev);/* also copies over blkcg information */ prev 740 fs/xfs/xfs_aops.c new->bi_iter.bi_sector = bio_end_sector(prev); prev 741 fs/xfs/xfs_aops.c new->bi_opf = prev->bi_opf; prev 742 fs/xfs/xfs_aops.c new->bi_write_hint = prev->bi_write_hint; prev 744 fs/xfs/xfs_aops.c bio_chain(prev, new); prev 745 fs/xfs/xfs_aops.c bio_get(prev); /* for xfs_destroy_ioend */ prev 746 fs/xfs/xfs_aops.c submit_bio(prev); prev 40 fs/xfs/xfs_bio_io.c struct bio *prev = bio; prev 43 fs/xfs/xfs_bio_io.c bio_copy_dev(bio, prev); prev 44 fs/xfs/xfs_bio_io.c bio->bi_iter.bi_sector = bio_end_sector(prev); prev 45 fs/xfs/xfs_bio_io.c bio->bi_opf = prev->bi_opf; prev 46 fs/xfs/xfs_bio_io.c bio_chain(prev, bio); prev 48 fs/xfs/xfs_bio_io.c submit_bio(prev); prev 86 fs/xfs/xfs_bmap_util.c error = xfs_bmap_extsize_align(mp, &ap->got, &ap->prev, prev 403 fs/xfs/xfs_iomap.c struct xfs_bmbt_irec prev; prev 423 fs/xfs/xfs_iomap.c !xfs_iext_peek_prev_extent(ifp, icur, &prev) || prev 424 fs/xfs/xfs_iomap.c prev.br_startoff + prev.br_blockcount < offset_fsb) prev 443 fs/xfs/xfs_iomap.c if (prev.br_blockcount <= (MAXEXTLEN >> 1)) prev 444 fs/xfs/xfs_iomap.c alloc_blocks = prev.br_blockcount << 1; prev 4201 fs/xfs/xfs_log_recover.c item = list_entry(trans->r_itemq.prev, xlog_recover_item_t, ri_list); prev 4271 fs/xfs/xfs_log_recover.c item = list_entry(trans->r_itemq.prev, xlog_recover_item_t, ri_list); prev 4276 fs/xfs/xfs_log_recover.c item = list_entry(trans->r_itemq.prev, prev 147 fs/xfs/xfs_mru_cache.c list_splice_init(lru_list, mru->reap_list.prev); prev 51 fs/xfs/xfs_trans_ail.c prev_lip = list_entry(lip->li_ail.prev, struct xfs_log_item, li_ail); prev 85 fs/xfs/xfs_trans_ail.c return list_entry(ailp->ail_head.prev, struct xfs_log_item, li_ail); prev 312 fs/xfs/xfs_trans_ail.c cur->item = list_entry(list->prev, struct xfs_log_item, li_ail); prev 18 include/asm-generic/cmpxchg-local.h unsigned long flags, prev; prev 28 include/asm-generic/cmpxchg-local.h case 1: prev = *(u8 *)ptr; prev 29 include/asm-generic/cmpxchg-local.h if (prev == old) prev 32 include/asm-generic/cmpxchg-local.h case 2: prev = *(u16 *)ptr; prev 33 include/asm-generic/cmpxchg-local.h if (prev == old) prev 36 include/asm-generic/cmpxchg-local.h case 4: prev = *(u32 *)ptr; prev 37 include/asm-generic/cmpxchg-local.h if (prev == old) prev 40 include/asm-generic/cmpxchg-local.h case 8: prev = *(u64 *)ptr; prev 41 include/asm-generic/cmpxchg-local.h if (prev == old) prev 48 include/asm-generic/cmpxchg-local.h return prev; prev 57 include/asm-generic/cmpxchg-local.h u64 prev; prev 61 include/asm-generic/cmpxchg-local.h prev = *(u64 *)ptr; prev 62 include/asm-generic/cmpxchg-local.h if (prev == old) prev 65 include/asm-generic/cmpxchg-local.h return prev; prev 35 include/asm-generic/mmu_context.h static inline void switch_mm(struct mm_struct *prev, prev 712 include/asm-generic/pgtable.h #define arch_start_context_switch(prev) do {} while (0) prev 21 include/asm-generic/switch_to.h #define switch_to(prev, next, last) \ prev 23 include/asm-generic/switch_to.h ((last) = __switch_to((prev), (next))); \ prev 133 include/crypto/drbg.h unsigned char *prev; /* FIPS 140-2 continuous test value */ prev 1201 include/linux/acpi.h struct fwnode_handle *prev) prev 71 include/linux/agp_backend.h struct agp_memory *prev; prev 93 include/linux/agpgart.h struct agp_client *prev; prev 101 include/linux/agpgart.h struct agp_controller *prev; prev 116 include/linux/agpgart.h struct agp_file_private *prev; prev 29 include/linux/dma-fence-chain.h struct dma_fence __rcu *prev; prev 69 include/linux/dma-fence-chain.h struct dma_fence *prev, prev 1512 include/linux/efi.h struct efivar_entry **prev); prev 98 include/linux/fwnode.h struct fwnode_handle *prev); prev 154 include/linux/interval_tree_generic.h struct rb_node *rb = node->ITRB.rb_right, *prev; \ prev 176 include/linux/interval_tree_generic.h prev = &node->ITRB; \ prev 179 include/linux/interval_tree_generic.h } while (prev == rb); \ prev 29 include/linux/list.h list->prev = list; prev 34 include/linux/list.h struct list_head *prev, prev 39 include/linux/list.h struct list_head *prev, prev 57 include/linux/list.h struct list_head *prev, prev 60 include/linux/list.h if (!__list_add_valid(new, prev, next)) prev 63 include/linux/list.h next->prev = new; prev 65 include/linux/list.h new->prev = prev; prev 66 include/linux/list.h WRITE_ONCE(prev->next, new); prev 93 include/linux/list.h __list_add(new, head->prev, head); prev 103 include/linux/list.h static inline void __list_del(struct list_head * prev, struct list_head * next) prev 105 include/linux/list.h next->prev = prev; prev 106 include/linux/list.h WRITE_ONCE(prev->next, next); prev 119 include/linux/list.h __list_del(entry->prev, entry->next); prev 120 include/linux/list.h entry->prev = NULL; prev 134 include/linux/list.h __list_del(entry->prev, entry->next); prev 141 include/linux/list.h entry->prev = LIST_POISON2; prev 155 include/linux/list.h new->next->prev = new; prev 156 include/linux/list.h new->prev = old->prev; prev 157 include/linux/list.h new->prev->next = new; prev 175 include/linux/list.h struct list_head *pos = entry2->prev; prev 230 include/linux/list.h first->prev->next = last->next; prev 231 include/linux/list.h last->next->prev = first->prev; prev 233 include/linux/list.h head->prev->next = first; prev 234 include/linux/list.h first->prev = head->prev; prev 237 include/linux/list.h head->prev = last; prev 248 include/linux/list.h return list->prev == head; prev 287 include/linux/list.h return (next == head) && (next == head->prev); prev 328 include/linux/list.h return !list_empty(head) && (head->next == head->prev); prev 336 include/linux/list.h list->next->prev = list; prev 337 include/linux/list.h list->prev = entry; prev 340 include/linux/list.h new_first->prev = head; prev 394 include/linux/list.h list->next->prev = list; prev 395 include/linux/list.h list->prev = entry->prev; prev 396 include/linux/list.h list->prev->next = list; prev 398 include/linux/list.h entry->prev = head; prev 402 include/linux/list.h struct list_head *prev, prev 406 include/linux/list.h struct list_head *last = list->prev; prev 408 include/linux/list.h first->prev = prev; prev 409 include/linux/list.h prev->next = first; prev 412 include/linux/list.h next->prev = last; prev 436 include/linux/list.h __list_splice(list, head->prev, head); prev 467 include/linux/list.h __list_splice(list, head->prev, head); prev 501 include/linux/list.h list_entry((ptr)->prev, type, member) prev 531 include/linux/list.h list_entry((pos)->member.prev, typeof(*(pos)), member) prev 547 include/linux/list.h for (pos = (head)->prev; pos != (head); pos = pos->prev) prev 566 include/linux/list.h for (pos = (head)->prev, n = pos->prev; \ prev 568 include/linux/list.h pos = n, n = pos->prev) prev 803 include/linux/list.h struct hlist_node *prev) prev 805 include/linux/list.h n->next = prev->next; prev 806 include/linux/list.h prev->next = n; prev 807 include/linux/list.h n->pprev = &prev->next; prev 105 include/linux/list_bl.h struct hlist_bl_node *prev) prev 107 include/linux/list_bl.h n->next = prev->next; prev 108 include/linux/list_bl.h n->pprev = &prev->next; prev 109 include/linux/list_bl.h prev->next = n; prev 939 include/linux/memcontrol.h struct mem_cgroup *prev, prev 946 include/linux/memcontrol.h struct mem_cgroup *prev) prev 435 include/linux/mlx5/driver.h u32 prev; prev 221 include/linux/mm.h #define lru_to_page(head) (list_entry((head)->prev, struct page, lru)) prev 2245 include/linux/mm.h struct vm_area_struct *prev, prev 2286 include/linux/mm.h struct vm_area_struct *prev, unsigned long addr, unsigned long end, prev 137 include/linux/nfs_page.h struct nfs_page *prev, prev 226 include/linux/of.h extern struct device_node *__of_find_all_nodes(struct device_node *prev); prev 227 include/linux/of.h extern struct device_node *of_find_all_nodes(struct device_node *prev); prev 289 include/linux/of.h struct device_node *prev); prev 291 include/linux/of.h const struct device_node *node, struct device_node *prev); prev 353 include/linux/of.h extern struct device_node *of_get_next_cpu_node(struct device_node *prev); prev 632 include/linux/of.h const struct device_node *node, struct device_node *prev) prev 638 include/linux/of.h const struct device_node *node, struct device_node *prev) prev 763 include/linux/of.h static inline struct device_node *of_get_next_cpu_node(struct device_node *prev) prev 18 include/linux/of_pdt.h int (*nextprop)(phandle node, char *prev, char *buf); prev 10 include/linux/osq_lock.h struct optimistic_spin_node *next, *prev; prev 148 include/linux/parport.h struct pardevice *prev; prev 894 include/linux/perf_event.h extern void __perf_event_task_sched_in(struct task_struct *prev, prev 896 include/linux/perf_event.h extern void __perf_event_task_sched_out(struct task_struct *prev, prev 1146 include/linux/perf_event.h static inline void perf_event_task_sched_in(struct task_struct *prev, prev 1150 include/linux/perf_event.h __perf_event_task_sched_in(prev, task); prev 1161 include/linux/perf_event.h static inline void perf_event_task_sched_out(struct task_struct *prev, prev 1167 include/linux/perf_event.h __perf_event_task_sched_out(prev, next); prev 1354 include/linux/perf_event.h perf_event_task_sched_in(struct task_struct *prev, prev 1357 include/linux/perf_event.h perf_event_task_sched_out(struct task_struct *prev, prev 294 include/linux/plist.h return list_entry(head->node_list.prev, prev 340 include/linux/property.h const struct fwnode_handle *fwnode, struct fwnode_handle *prev); prev 34 include/linux/rculist.h WRITE_ONCE(list->prev, list); prev 68 include/linux/rculist.h struct list_head *prev, struct list_head *next) prev 70 include/linux/rculist.h if (!__list_add_valid(new, prev, next)) prev 74 include/linux/rculist.h new->prev = prev; prev 75 include/linux/rculist.h rcu_assign_pointer(list_next_rcu(prev), new); prev 76 include/linux/rculist.h next->prev = new; prev 119 include/linux/rculist.h __list_add_rcu(new, head->prev, head); prev 149 include/linux/rculist.h entry->prev = LIST_POISON2; prev 192 include/linux/rculist.h new->prev = old->prev; prev 193 include/linux/rculist.h rcu_assign_pointer(list_next_rcu(new->prev), new); prev 194 include/linux/rculist.h new->next->prev = new; prev 195 include/linux/rculist.h old->prev = LIST_POISON2; prev 218 include/linux/rculist.h struct list_head *prev, prev 223 include/linux/rculist.h struct list_head *last = list->prev; prev 251 include/linux/rculist.h rcu_assign_pointer(list_next_rcu(prev), first); prev 252 include/linux/rculist.h first->prev = prev; prev 253 include/linux/rculist.h next->prev = last; prev 283 include/linux/rculist.h __list_splice_init_rcu(list, head->prev, head, sync); prev 620 include/linux/rculist.h struct hlist_node *prev) prev 622 include/linux/rculist.h n->next = prev->next; prev 623 include/linux/rculist.h n->pprev = &prev->next; prev 624 include/linux/rculist.h rcu_assign_pointer(hlist_next_rcu(prev), n); prev 202 include/linux/rio.h struct rio_dev *prev; prev 57 include/linux/sched/cputime.h extern void cputime_adjust(struct task_cputime *curr, struct prev_cputime *prev, prev 180 include/linux/sched/cputime.h static inline void prev_cputime_init(struct prev_cputime *prev) prev 183 include/linux/sched/cputime.h prev->utime = prev->stime = 0; prev 184 include/linux/sched/cputime.h raw_spin_lock_init(&prev->lock); prev 379 include/linux/sched/mm.h static inline void membarrier_arch_switch_mm(struct mm_struct *prev, prev 47 include/linux/sched/task.h extern asmlinkage void schedule_tail(struct task_struct *prev); prev 292 include/linux/skbuff.h struct sk_buff *prev; prev 690 include/linux/skbuff.h struct sk_buff *prev; prev 1536 include/linux/skbuff.h return skb->prev == (const struct sk_buff *) list; prev 1572 include/linux/skbuff.h return skb->prev; prev 1800 include/linux/skbuff.h struct sk_buff *skb = READ_ONCE(list_->prev); prev 1831 include/linux/skbuff.h list->prev = list->next = (struct sk_buff *)list; prev 1863 include/linux/skbuff.h struct sk_buff *prev, struct sk_buff *next, prev 1870 include/linux/skbuff.h WRITE_ONCE(newsk->prev, prev); prev 1871 include/linux/skbuff.h WRITE_ONCE(next->prev, newsk); prev 1872 include/linux/skbuff.h WRITE_ONCE(prev->next, newsk); prev 1877 include/linux/skbuff.h struct sk_buff *prev, prev 1881 include/linux/skbuff.h struct sk_buff *last = list->prev; prev 1883 include/linux/skbuff.h WRITE_ONCE(first->prev, prev); prev 1884 include/linux/skbuff.h WRITE_ONCE(prev->next, first); prev 1887 include/linux/skbuff.h WRITE_ONCE(next->prev, last); prev 1930 include/linux/skbuff.h __skb_queue_splice(list, head->prev, (struct sk_buff *) head); prev 1947 include/linux/skbuff.h __skb_queue_splice(list, head->prev, (struct sk_buff *) head); prev 1965 include/linux/skbuff.h struct sk_buff *prev, prev 1968 include/linux/skbuff.h __skb_insert(newsk, prev, prev->next, list); prev 1978 include/linux/skbuff.h __skb_insert(newsk, next->prev, next, list); prev 2022 include/linux/skbuff.h struct sk_buff *next, *prev; prev 2026 include/linux/skbuff.h prev = skb->prev; prev 2027 include/linux/skbuff.h skb->next = skb->prev = NULL; prev 2028 include/linux/skbuff.h WRITE_ONCE(next->prev, prev); prev 2029 include/linux/skbuff.h WRITE_ONCE(prev->next, next); prev 3436 include/linux/skbuff.h for (skb = (queue)->prev; \ prev 3438 include/linux/skbuff.h skb = skb->prev) prev 3441 include/linux/skbuff.h for (skb = (queue)->prev, tmp = skb->prev; \ prev 3443 include/linux/skbuff.h skb = tmp, tmp = skb->prev) prev 3446 include/linux/skbuff.h for (tmp = skb->prev; \ prev 3448 include/linux/skbuff.h skb = tmp, tmp = skb->prev) prev 160 include/linux/thunderbolt.h struct tb_property *prev); prev 182 include/linux/types.h struct list_head *next, *prev; prev 19 include/linux/user-return-notifier.h static inline void propagate_user_return_notify(struct task_struct *prev, prev 22 include/linux/user-return-notifier.h if (test_tsk_thread_flag(prev, TIF_USER_RETURN_NOTIFY)) { prev 23 include/linux/user-return-notifier.h clear_tsk_thread_flag(prev, TIF_USER_RETURN_NOTIFY); prev 39 include/linux/user-return-notifier.h static inline void propagate_user_return_notify(struct task_struct *prev, prev 50 include/linux/vtime.h extern void vtime_task_switch(struct task_struct *prev); prev 52 include/linux/vtime.h extern void vtime_common_task_switch(struct task_struct *prev); prev 53 include/linux/vtime.h static inline void vtime_task_switch(struct task_struct *prev) prev 56 include/linux/vtime.h vtime_common_task_switch(prev); prev 65 include/linux/vtime.h static inline void vtime_task_switch(struct task_struct *prev) { } prev 73 include/net/bond_alb.h u32 prev; /* The previous Hash table entry index, prev 64 include/net/bonding.h netdev_adjacent_get_private(bond_slave_list(bond)->prev) : \ prev 1026 include/net/sch_generic.h if (skb->prev) prev 1027 include/net/sch_generic.h skb->prev->next = *to_free; prev 415 include/net/sctp/sctp.h return (head->next != head) && (head->next == head->prev); prev 184 include/sound/gus.h struct snd_gf1_mem_block *prev; prev 49 include/sound/util_mem.h struct list_head *prev); prev 831 include/trace/events/rxrpc.h rxrpc_seq_t first, rxrpc_seq_t prev, u8 reason, u8 n_acks), prev 833 include/trace/events/rxrpc.h TP_ARGS(call, serial, ack_serial, first, prev, reason, n_acks), prev 840 include/trace/events/rxrpc.h __field(rxrpc_seq_t, prev ) prev 850 include/trace/events/rxrpc.h __entry->prev = prev; prev 861 include/trace/events/rxrpc.h __entry->prev, prev 141 include/trace/events/sched.h struct task_struct *prev, prev 144 include/trace/events/sched.h TP_ARGS(preempt, prev, next), prev 158 include/trace/events/sched.h __entry->prev_pid = prev->pid; prev 159 include/trace/events/sched.h __entry->prev_prio = prev->prio; prev 160 include/trace/events/sched.h __entry->prev_state = __trace_sched_switch_state(preempt, prev); prev 161 include/trace/events/sched.h memcpy(__entry->prev_comm, prev->comm, TASK_COMM_LEN); prev 117 include/uapi/drm/drm.h unsigned char prev; prev 151 include/uapi/drm/i810_drm.h unsigned char next, prev; /* indices to form a circular LRU */ prev 180 include/uapi/drm/via_drm.h unsigned char next, prev; /* indices to form a circular LRU */ prev 19 include/uapi/linux/netfilter/xt_limit.h unsigned long prev; /* moved to xt_limit_priv */ prev 20 include/uapi/linux/netfilter_bridge/ebt_limit.h unsigned long prev; prev 816 include/video/omapfb_dss.h struct device_node *prev); prev 820 include/video/omapfb_dss.h struct device_node *prev); prev 689 ipc/mqueue.c ptr = info->e_wait_q[sr].list.prev; prev 937 kernel/audit_tree.c while (barrier.prev != &tree_list) { prev 940 kernel/audit_tree.c tree = container_of(barrier.prev, struct audit_tree, list); prev 73 kernel/bpf/bpf_lru_list.c l->next_inactive_rotation = l->next_inactive_rotation->prev; prev 119 kernel/bpf/bpf_lru_list.c l->next_inactive_rotation = l->next_inactive_rotation->prev; prev 184 kernel/bpf/bpf_lru_list.c cur = cur->prev; prev 189 kernel/bpf/bpf_lru_list.c next = cur->prev; prev 643 kernel/bpf/core.c fp->aux->ksym_lnode.prev == LIST_POISON2; prev 662 kernel/bpf/cpumap.c if (!bq->flush_node.prev) prev 455 kernel/bpf/devmap.c if (!bq->flush_node.prev) prev 7543 kernel/bpf/verifier.c static bool reg_type_mismatch(enum bpf_reg_type src, enum bpf_reg_type prev) prev 7545 kernel/bpf/verifier.c return src != prev && (!reg_type_mismatch_ok(src) || prev 7546 kernel/bpf/verifier.c !reg_type_mismatch_ok(prev)); prev 186 kernel/bpf/xskmap.c if (!xs->flush_node.prev) prev 445 kernel/crash_core.c VMCOREINFO_OFFSET(list_head, prev); prev 748 kernel/debug/kdb/kdb_support.c struct debug_alloc_header *best, *bestprev, *prev, *h; prev 760 kernel/debug/kdb/kdb_support.c prev = best = bestprev = NULL; prev 764 kernel/debug/kdb/kdb_support.c bestprev = prev; prev 770 kernel/debug/kdb/kdb_support.c prev = h; prev 828 kernel/debug/kdb/kdb_support.c struct debug_alloc_header *prev; prev 830 kernel/debug/kdb/kdb_support.c prev = (struct debug_alloc_header *)(debug_alloc_pool + prev 833 kernel/debug/kdb/kdb_support.c if (!prev->next || prev->next > h_offset) prev 835 kernel/debug/kdb/kdb_support.c prev = (struct debug_alloc_header *) prev 836 kernel/debug/kdb/kdb_support.c (debug_alloc_pool + prev->next); prev 838 kernel/debug/kdb/kdb_support.c prev_offset = (char *)prev - debug_alloc_pool; prev 839 kernel/debug/kdb/kdb_support.c if (prev_offset + dah_overhead + prev->size == h_offset) { prev 840 kernel/debug/kdb/kdb_support.c prev->size += dah_overhead + h->size; prev 843 kernel/debug/kdb/kdb_support.c h = prev; prev 846 kernel/debug/kdb/kdb_support.c h->next = prev->next; prev 847 kernel/debug/kdb/kdb_support.c prev->next = h_offset; prev 873 kernel/events/core.c static inline void perf_cgroup_sched_in(struct task_struct *prev, prev 886 kernel/events/core.c cgrp2 = perf_cgroup_from_task(prev, NULL); prev 1021 kernel/events/core.c static inline void perf_cgroup_sched_in(struct task_struct *prev, prev 3275 kernel/events/core.c static void perf_pmu_sched_task(struct task_struct *prev, prev 3282 kernel/events/core.c if (prev == next) prev 3563 kernel/events/core.c void __perf_event_task_sched_in(struct task_struct *prev, prev 3577 kernel/events/core.c perf_cgroup_sched_in(prev, task); prev 3588 kernel/events/core.c perf_event_switch(task, prev, true); prev 3591 kernel/events/core.c perf_pmu_sched_task(prev, task, true); prev 9671 kernel/events/core.c s64 prev; prev 9675 kernel/events/core.c prev = local64_xchg(&event->hw.prev_count, now); prev 9676 kernel/events/core.c local64_add(now - prev, &event->count); prev 9748 kernel/events/core.c u64 prev; prev 9751 kernel/events/core.c prev = local64_xchg(&event->hw.prev_count, now); prev 9752 kernel/events/core.c delta = now - prev; prev 977 kernel/events/uprobes.c struct map_info *prev = NULL; prev 987 kernel/events/uprobes.c if (!prev && !more) { prev 992 kernel/events/uprobes.c prev = kmalloc(sizeof(struct map_info), prev 994 kernel/events/uprobes.c if (prev) prev 995 kernel/events/uprobes.c prev->next = NULL; prev 997 kernel/events/uprobes.c if (!prev) { prev 1005 kernel/events/uprobes.c info = prev; prev 1006 kernel/events/uprobes.c prev = prev->next; prev 1018 kernel/events/uprobes.c prev = curr; prev 1030 kernel/events/uprobes.c info->next = prev; prev 1031 kernel/events/uprobes.c prev = info; prev 1036 kernel/events/uprobes.c while (prev) prev 1037 kernel/events/uprobes.c prev = free_map_info(prev); prev 481 kernel/fork.c struct vm_area_struct *mpnt, *tmp, *prev, **pprev; prev 517 kernel/fork.c prev = NULL; prev 592 kernel/fork.c tmp->vm_prev = prev; prev 593 kernel/fork.c prev = tmp; prev 59 kernel/gcov/base.c struct gcov_info *prev = NULL; prev 68 kernel/gcov/base.c gcov_info_unlink(prev, info); prev 72 kernel/gcov/base.c prev = info; prev 206 kernel/gcov/clang.c void gcov_info_unlink(struct gcov_info *prev, struct gcov_info *info) prev 132 kernel/gcov/gcc_3_4.c void gcov_info_unlink(struct gcov_info *prev, struct gcov_info *info) prev 134 kernel/gcov/gcc_3_4.c if (prev) prev 135 kernel/gcov/gcc_3_4.c prev->next = info->next; prev 145 kernel/gcov/gcc_4_7.c void gcov_info_unlink(struct gcov_info *prev, struct gcov_info *info) prev 147 kernel/gcov/gcc_4_7.c if (prev) prev 148 kernel/gcov/gcc_4_7.c prev->next = info->next; prev 49 kernel/gcov/gcov.h void gcov_info_unlink(struct gcov_info *prev, struct gcov_info *info); prev 674 kernel/jump_label.c struct static_key_mod *jlm, **prev; prev 689 kernel/jump_label.c prev = &key->next; prev 693 kernel/jump_label.c prev = &jlm->next; prev 701 kernel/jump_label.c if (prev == &key->next) prev 704 kernel/jump_label.c *prev = jlm->next; prev 223 kernel/livepatch/transition.c struct klp_func *prev; prev 225 kernel/livepatch/transition.c prev = list_next_entry(func, stack_node); prev 226 kernel/livepatch/transition.c func_addr = (unsigned long)prev->new_func; prev 227 kernel/livepatch/transition.c func_size = prev->new_size; prev 2033 kernel/locking/lockdep.c struct held_lock *prev, prev 2057 kernel/locking/lockdep.c print_lock(prev); prev 2059 kernel/locking/lockdep.c print_lock_name(hlock_class(prev)); prev 2080 kernel/locking/lockdep.c hlock_class(prev), hlock_class(next)); prev 2227 kernel/locking/lockdep.c static int check_irq_usage(struct task_struct *curr, struct held_lock *prev, prev 2242 kernel/locking/lockdep.c this.class = hlock_class(prev); prev 2298 kernel/locking/lockdep.c prev, next, prev 2320 kernel/locking/lockdep.c struct held_lock *prev, struct held_lock *next) prev 2336 kernel/locking/lockdep.c struct lock_class *prev = hlock_class(prv); prev 2342 kernel/locking/lockdep.c __print_lock_name(prev); prev 2352 kernel/locking/lockdep.c print_deadlock_bug(struct task_struct *curr, struct held_lock *prev, prev 2367 kernel/locking/lockdep.c print_lock(prev); prev 2370 kernel/locking/lockdep.c print_deadlock_scenario(next, prev); prev 2388 kernel/locking/lockdep.c struct held_lock *prev; prev 2393 kernel/locking/lockdep.c prev = curr->held_locks + i; prev 2395 kernel/locking/lockdep.c if (prev->instance == next->nest_lock) prev 2396 kernel/locking/lockdep.c nest = prev; prev 2398 kernel/locking/lockdep.c if (hlock_class(prev) != hlock_class(next)) prev 2405 kernel/locking/lockdep.c if ((next->read == 2) && prev->read) prev 2415 kernel/locking/lockdep.c print_deadlock_bug(curr, prev, next); prev 2444 kernel/locking/lockdep.c check_prev_add(struct task_struct *curr, struct held_lock *prev, prev 2451 kernel/locking/lockdep.c if (!hlock_class(prev)->key || !hlock_class(next)->key) { prev 2458 kernel/locking/lockdep.c WARN_ONCE(!debug_locks_silent && !hlock_class(prev)->key, prev 2460 kernel/locking/lockdep.c hlock_class(prev), prev 2461 kernel/locking/lockdep.c hlock_class(prev)->name); prev 2479 kernel/locking/lockdep.c ret = check_noncircular(next, prev, trace); prev 2483 kernel/locking/lockdep.c if (!check_irq_usage(curr, prev, next)) prev 2494 kernel/locking/lockdep.c if (next->read == 2 || prev->read == 2) prev 2504 kernel/locking/lockdep.c list_for_each_entry(entry, &hlock_class(prev)->locks_after, entry) { prev 2516 kernel/locking/lockdep.c ret = check_redundant(prev, next); prev 2531 kernel/locking/lockdep.c ret = add_lock_to_list(hlock_class(next), hlock_class(prev), prev 2532 kernel/locking/lockdep.c &hlock_class(prev)->locks_after, prev 2538 kernel/locking/lockdep.c ret = add_lock_to_list(hlock_class(prev), hlock_class(next), prev 67 kernel/locking/mcs_spinlock.h struct mcs_spinlock *prev; prev 79 kernel/locking/mcs_spinlock.h prev = xchg(lock, node); prev 80 kernel/locking/mcs_spinlock.h if (likely(prev == NULL)) { prev 91 kernel/locking/mcs_spinlock.h WRITE_ONCE(prev->next, node); prev 76 kernel/locking/mutex-debug.c DEBUG_LOCKS_WARN_ON(!lock->wait_list.prev && !lock->wait_list.next); prev 14 kernel/locking/mutex.h __list_del((waiter)->list.prev, (waiter)->list.next) prev 44 kernel/locking/osq_lock.c struct optimistic_spin_node *prev) prev 55 kernel/locking/osq_lock.c old = prev ? prev->cpu : OSQ_UNLOCKED_VAL; prev 93 kernel/locking/osq_lock.c struct optimistic_spin_node *prev, *next; prev 111 kernel/locking/osq_lock.c prev = decode_cpu(old); prev 112 kernel/locking/osq_lock.c node->prev = prev; prev 126 kernel/locking/osq_lock.c WRITE_ONCE(prev->next, node); prev 143 kernel/locking/osq_lock.c if (need_resched() || vcpu_is_preempted(node_cpu(node->prev))) prev 160 kernel/locking/osq_lock.c if (prev->next == node && prev 161 kernel/locking/osq_lock.c cmpxchg(&prev->next, node, NULL) == node) prev 178 kernel/locking/osq_lock.c prev = READ_ONCE(node->prev); prev 188 kernel/locking/osq_lock.c next = osq_wait_next(lock, node, prev); prev 200 kernel/locking/osq_lock.c WRITE_ONCE(next->prev, prev); prev 201 kernel/locking/osq_lock.c WRITE_ONCE(prev->next, next); prev 273 kernel/locking/qspinlock.c struct mcs_spinlock *prev) { } prev 316 kernel/locking/qspinlock.c struct mcs_spinlock *prev, *next, *node; prev 467 kernel/locking/qspinlock.c prev = decode_tail(old); prev 470 kernel/locking/qspinlock.c WRITE_ONCE(prev->next, node); prev 472 kernel/locking/qspinlock.c pv_wait_node(node, prev); prev 267 kernel/locking/qspinlock_paravirt.h pv_wait_early(struct pv_node *prev, int loop) prev 272 kernel/locking/qspinlock_paravirt.h return READ_ONCE(prev->state) != vcpu_running; prev 293 kernel/locking/qspinlock_paravirt.h static void pv_wait_node(struct mcs_spinlock *node, struct mcs_spinlock *prev) prev 296 kernel/locking/qspinlock_paravirt.h struct pv_node *pp = (struct pv_node *)prev; prev 1398 kernel/module.c char *prev); prev 2542 kernel/module.c char *prev) prev 2555 kernel/module.c if (prev) { prev 2556 kernel/module.c size -= prev - modinfo; prev 2557 kernel/module.c modinfo = next_string(prev, &size); prev 958 kernel/power/snapshot.c region = list_entry(nosave_regions.prev, prev 688 kernel/rcu/tree_exp.h t = list_entry(rnp->exp_tasks->prev, prev 182 kernel/rcu/tree_stall.h t = list_entry(rnp->gp_tasks->prev, prev 208 kernel/rcu/tree_stall.h t = list_entry(rnp->gp_tasks->prev, prev 138 kernel/sched/autogroup.c struct autogroup *prev; prev 144 kernel/sched/autogroup.c prev = p->signal->autogroup; prev 145 kernel/sched/autogroup.c if (prev == ag) { prev 166 kernel/sched/autogroup.c autogroup_kref_put(prev); prev 3077 kernel/sched/core.c static inline void finish_task(struct task_struct *prev) prev 3090 kernel/sched/core.c smp_store_release(&prev->on_cpu, 0); prev 3148 kernel/sched/core.c prepare_task_switch(struct rq *rq, struct task_struct *prev, prev 3151 kernel/sched/core.c kcov_prepare_switch(prev); prev 3152 kernel/sched/core.c sched_info_switch(rq, prev, next); prev 3153 kernel/sched/core.c perf_event_task_sched_out(prev, next); prev 3154 kernel/sched/core.c rseq_preempt(prev); prev 3155 kernel/sched/core.c fire_sched_out_preempt_notifiers(prev, next); prev 3179 kernel/sched/core.c static struct rq *finish_task_switch(struct task_struct *prev) prev 3215 kernel/sched/core.c prev_state = prev->state; prev 3216 kernel/sched/core.c vtime_task_switch(prev); prev 3217 kernel/sched/core.c perf_event_task_sched_in(prev, current); prev 3218 kernel/sched/core.c finish_task(prev); prev 3241 kernel/sched/core.c if (prev->sched_class->task_dead) prev 3242 kernel/sched/core.c prev->sched_class->task_dead(prev); prev 3248 kernel/sched/core.c kprobe_flush_task(prev); prev 3251 kernel/sched/core.c put_task_stack(prev); prev 3253 kernel/sched/core.c put_task_struct_rcu_user(prev); prev 3301 kernel/sched/core.c asmlinkage __visible void schedule_tail(struct task_struct *prev) prev 3315 kernel/sched/core.c rq = finish_task_switch(prev); prev 3329 kernel/sched/core.c context_switch(struct rq *rq, struct task_struct *prev, prev 3332 kernel/sched/core.c prepare_task_switch(rq, prev, next); prev 3339 kernel/sched/core.c arch_start_context_switch(prev); prev 3349 kernel/sched/core.c enter_lazy_tlb(prev->active_mm, next); prev 3351 kernel/sched/core.c next->active_mm = prev->active_mm; prev 3352 kernel/sched/core.c if (prev->mm) // from user prev 3353 kernel/sched/core.c mmgrab(prev->active_mm); prev 3355 kernel/sched/core.c prev->active_mm = NULL; prev 3357 kernel/sched/core.c membarrier_switch_mm(rq, prev->active_mm, next->mm); prev 3366 kernel/sched/core.c switch_mm_irqs_off(prev->active_mm, next->mm, next); prev 3368 kernel/sched/core.c if (!prev->mm) { // from kernel prev 3370 kernel/sched/core.c rq->prev_mm = prev->active_mm; prev 3371 kernel/sched/core.c prev->active_mm = NULL; prev 3380 kernel/sched/core.c switch_to(prev, next, prev); prev 3383 kernel/sched/core.c return finish_task_switch(prev); prev 3842 kernel/sched/core.c static noinline void __schedule_bug(struct task_struct *prev) prev 3851 kernel/sched/core.c prev->comm, prev->pid, preempt_count()); prev 3853 kernel/sched/core.c debug_show_held_locks(prev); prev 3856 kernel/sched/core.c print_irqtrace_events(prev); prev 3873 kernel/sched/core.c static inline void schedule_debug(struct task_struct *prev, bool preempt) prev 3876 kernel/sched/core.c if (task_stack_end_corrupted(prev)) prev 3881 kernel/sched/core.c if (!preempt && prev->state && prev->non_block_count) { prev 3883 kernel/sched/core.c prev->comm, prev->pid, prev->non_block_count); prev 3890 kernel/sched/core.c __schedule_bug(prev); prev 3904 kernel/sched/core.c pick_next_task(struct rq *rq, struct task_struct *prev, struct rq_flags *rf) prev 3915 kernel/sched/core.c if (likely((prev->sched_class == &idle_sched_class || prev 3916 kernel/sched/core.c prev->sched_class == &fair_sched_class) && prev 3919 kernel/sched/core.c p = fair_sched_class.pick_next_task(rq, prev, rf); prev 3925 kernel/sched/core.c p = idle_sched_class.pick_next_task(rq, prev, rf); prev 3940 kernel/sched/core.c for_class_range(class, prev->sched_class, &idle_sched_class) { prev 3941 kernel/sched/core.c if (class->balance(rq, prev, rf)) prev 3946 kernel/sched/core.c put_prev_task(rq, prev); prev 3999 kernel/sched/core.c struct task_struct *prev, *next; prev 4007 kernel/sched/core.c prev = rq->curr; prev 4009 kernel/sched/core.c schedule_debug(prev, preempt); prev 4032 kernel/sched/core.c switch_count = &prev->nivcsw; prev 4033 kernel/sched/core.c if (!preempt && prev->state) { prev 4034 kernel/sched/core.c if (signal_pending_state(prev->state, prev)) { prev 4035 kernel/sched/core.c prev->state = TASK_RUNNING; prev 4037 kernel/sched/core.c deactivate_task(rq, prev, DEQUEUE_SLEEP | DEQUEUE_NOCLOCK); prev 4039 kernel/sched/core.c if (prev->in_iowait) { prev 4044 kernel/sched/core.c switch_count = &prev->nvcsw; prev 4047 kernel/sched/core.c next = pick_next_task(rq, prev, &rf); prev 4048 kernel/sched/core.c clear_tsk_need_resched(prev); prev 4051 kernel/sched/core.c if (likely(prev != next)) { prev 4074 kernel/sched/core.c trace_sched_switch(preempt, prev, next); prev 4077 kernel/sched/core.c rq = context_switch(rq, prev, next, &rf); prev 410 kernel/sched/cputime.c void vtime_common_task_switch(struct task_struct *prev) prev 412 kernel/sched/cputime.c if (is_idle_task(prev)) prev 413 kernel/sched/cputime.c vtime_account_idle(prev); prev 415 kernel/sched/cputime.c vtime_account_system(prev); prev 417 kernel/sched/cputime.c vtime_flush(prev); prev 418 kernel/sched/cputime.c arch_vtime_task_switch(prev); prev 444 kernel/sched/cputime.c void cputime_adjust(struct task_cputime *curr, struct prev_cputime *prev, prev 591 kernel/sched/cputime.c void cputime_adjust(struct task_cputime *curr, struct prev_cputime *prev, prev 598 kernel/sched/cputime.c raw_spin_lock_irqsave(&prev->lock, flags); prev 609 kernel/sched/cputime.c if (prev->stime + prev->utime >= rtime) prev 642 kernel/sched/cputime.c if (stime < prev->stime) prev 643 kernel/sched/cputime.c stime = prev->stime; prev 650 kernel/sched/cputime.c if (utime < prev->utime) { prev 651 kernel/sched/cputime.c utime = prev->utime; prev 655 kernel/sched/cputime.c prev->stime = stime; prev 656 kernel/sched/cputime.c prev->utime = utime; prev 658 kernel/sched/cputime.c *ut = prev->utime; prev 659 kernel/sched/cputime.c *st = prev->stime; prev 660 kernel/sched/cputime.c raw_spin_unlock_irqrestore(&prev->lock, flags); prev 807 kernel/sched/cputime.c void arch_vtime_task_switch(struct task_struct *prev) prev 809 kernel/sched/cputime.c struct vtime *vtime = &prev->vtime; prev 503 kernel/sched/deadline.c static inline bool need_pull_dl_task(struct rq *rq, struct task_struct *prev) prev 505 kernel/sched/deadline.c return dl_task(prev); prev 621 kernel/sched/deadline.c static inline bool need_pull_dl_task(struct rq *rq, struct task_struct *prev) prev 1777 kernel/sched/deadline.c pick_next_task_dl(struct rq *rq, struct task_struct *prev, struct rq_flags *rf) prev 1783 kernel/sched/deadline.c WARN_ON_ONCE(prev || rf); prev 372 kernel/sched/fair.c rq->tmp_alone_branch = cfs_rq->leaf_cfs_rq_list.prev; prev 3171 kernel/sched/fair.c struct cfs_rq *prev, struct cfs_rq *next) prev 3186 kernel/sched/fair.c if (!(se->avg.last_update_time && prev)) prev 3195 kernel/sched/fair.c p_last_update_time_copy = prev->load_last_update_time_copy; prev 3200 kernel/sched/fair.c p_last_update_time = prev->avg.last_update_time; prev 3207 kernel/sched/fair.c p_last_update_time = prev->avg.last_update_time; prev 4256 kernel/sched/fair.c static void put_prev_entity(struct cfs_rq *cfs_rq, struct sched_entity *prev) prev 4262 kernel/sched/fair.c if (prev->on_rq) prev 4268 kernel/sched/fair.c check_spread(cfs_rq, prev); prev 4270 kernel/sched/fair.c if (prev->on_rq) { prev 4271 kernel/sched/fair.c update_stats_wait_start(cfs_rq, prev); prev 4273 kernel/sched/fair.c __enqueue_entity(cfs_rq, prev); prev 4275 kernel/sched/fair.c update_load_avg(cfs_rq, prev, 0); prev 6014 kernel/sched/fair.c static int select_idle_sibling(struct task_struct *p, int prev, int target) prev 6025 kernel/sched/fair.c if (prev != target && cpus_share_cache(prev, target) && prev 6026 kernel/sched/fair.c (available_idle_cpu(prev) || sched_idle_cpu(prev))) prev 6027 kernel/sched/fair.c return prev; prev 6031 kernel/sched/fair.c if (recent_used_cpu != prev && prev 6040 kernel/sched/fair.c p->recent_used_cpu = prev; prev 6601 kernel/sched/fair.c balance_fair(struct rq *rq, struct task_struct *prev, struct rq_flags *rf) prev 6776 kernel/sched/fair.c pick_next_task_fair(struct rq *rq, struct task_struct *prev, struct rq_flags *rf) prev 6788 kernel/sched/fair.c if (!prev || prev->sched_class != &fair_sched_class) prev 6841 kernel/sched/fair.c if (prev != p) { prev 6842 kernel/sched/fair.c struct sched_entity *pse = &prev->se; prev 6865 kernel/sched/fair.c if (prev) prev 6866 kernel/sched/fair.c put_prev_task(rq, prev); prev 6922 kernel/sched/fair.c static void put_prev_task_fair(struct rq *rq, struct task_struct *prev) prev 6924 kernel/sched/fair.c struct sched_entity *se = &prev->se; prev 370 kernel/sched/idle.c balance_idle(struct rq *rq, struct task_struct *prev, struct rq_flags *rf) prev 384 kernel/sched/idle.c static void put_prev_task_idle(struct rq *rq, struct task_struct *prev) prev 395 kernel/sched/idle.c pick_next_task_idle(struct rq *rq, struct task_struct *prev, struct rq_flags *rf) prev 399 kernel/sched/idle.c if (prev) prev 400 kernel/sched/idle.c put_prev_task(rq, prev); prev 263 kernel/sched/rt.c static inline bool need_pull_rt_task(struct rq *rq, struct task_struct *prev) prev 266 kernel/sched/rt.c return rq->rt.highest_prio.curr > prev->prio; prev 418 kernel/sched/rt.c static inline bool need_pull_rt_task(struct rq *rq, struct task_struct *prev) prev 1571 kernel/sched/rt.c pick_next_task_rt(struct rq *rq, struct task_struct *prev, struct rq_flags *rf) prev 1575 kernel/sched/rt.c WARN_ON_ONCE(prev || rf); prev 2341 kernel/sched/rt.c if (rt_se->run_list.prev != rt_se->run_list.next) { prev 482 kernel/sched/sched.h struct cfs_rq *prev, struct cfs_rq *next); prev 485 kernel/sched/sched.h struct cfs_rq *prev, struct cfs_rq *next) { } prev 1734 kernel/sched/sched.h struct task_struct *prev, prev 1740 kernel/sched/sched.h int (*balance)(struct rq *rq, struct task_struct *prev, struct rq_flags *rf); prev 1780 kernel/sched/sched.h static inline void put_prev_task(struct rq *rq, struct task_struct *prev) prev 1782 kernel/sched/sched.h WARN_ON_ONCE(rq->curr != prev); prev 1783 kernel/sched/sched.h prev->sched_class->put_prev_task(rq, prev); prev 226 kernel/sched/stats.h __sched_info_switch(struct rq *rq, struct task_struct *prev, struct task_struct *next) prev 233 kernel/sched/stats.h if (prev != rq->idle) prev 234 kernel/sched/stats.h sched_info_depart(rq, prev); prev 241 kernel/sched/stats.h sched_info_switch(struct rq *rq, struct task_struct *prev, struct task_struct *next) prev 244 kernel/sched/stats.h __sched_info_switch(rq, prev, next); prev 20 kernel/sched/stop_task.c balance_stop(struct rq *rq, struct task_struct *prev, struct rq_flags *rf) prev 38 kernel/sched/stop_task.c pick_next_task_stop(struct rq *rq, struct task_struct *prev, struct rq_flags *rf) prev 40 kernel/sched/stop_task.c WARN_ON_ONCE(prev || rf); prev 66 kernel/sched/stop_task.c static void put_prev_task_stop(struct rq *rq, struct task_struct *prev) prev 131 kernel/seccomp.c struct seccomp_filter *prev; prev 271 kernel/seccomp.c for (; f; f = f->prev) { prev 322 kernel/seccomp.c for (; child; child = child->prev) prev 520 kernel/seccomp.c for (walker = current->seccomp.filter; walker; walker = walker->prev) prev 542 kernel/seccomp.c filter->prev = current->seccomp.filter; prev 579 kernel/seccomp.c orig = orig->prev; prev 1216 kernel/seccomp.c for (cur = current->seccomp.filter; cur; cur = cur->prev) { prev 1476 kernel/seccomp.c for (filter = orig; filter; filter = filter->prev) prev 1485 kernel/seccomp.c for (filter = orig; filter && count > 1; filter = filter->prev) prev 285 kernel/time/tick-sched.c int prev; prev 287 kernel/time/tick-sched.c prev = atomic_fetch_or(BIT(bit), dep); prev 288 kernel/time/tick-sched.c if (!prev) prev 312 kernel/time/tick-sched.c int prev; prev 317 kernel/time/tick-sched.c prev = atomic_fetch_or(BIT(bit), &ts->tick_dep_mask); prev 318 kernel/time/tick-sched.c if (!prev) { prev 390 kernel/trace/fgraph.c struct task_struct *prev, struct task_struct *next) prev 404 kernel/trace/fgraph.c prev->ftrace_timestamp = timestamp; prev 6424 kernel/trace/ftrace.c struct task_struct *prev, struct task_struct *next) prev 891 kernel/trace/ring_buffer.c struct list_head *list = page->list.prev; prev 923 kernel/trace/ring_buffer.c rb_set_list_to_head(cpu_buffer, head->list.prev); prev 950 kernel/trace/ring_buffer.c struct buffer_page *prev, prev 957 kernel/trace/ring_buffer.c list = &prev->list; prev 973 kernel/trace/ring_buffer.c struct buffer_page *prev, prev 976 kernel/trace/ring_buffer.c return rb_head_page_set(cpu_buffer, head, prev, prev 982 kernel/trace/ring_buffer.c struct buffer_page *prev, prev 985 kernel/trace/ring_buffer.c return rb_head_page_set(cpu_buffer, head, prev, prev 991 kernel/trace/ring_buffer.c struct buffer_page *prev, prev 994 kernel/trace/ring_buffer.c return rb_head_page_set(cpu_buffer, head, prev, prev 1019 kernel/trace/ring_buffer.c if (RB_WARN_ON(cpu_buffer, rb_list_head(list->prev->next) != list)) prev 1031 kernel/trace/ring_buffer.c if (rb_is_head_page(cpu_buffer, page, page->list.prev)) { prev 1047 kernel/trace/ring_buffer.c unsigned long *ptr = (unsigned long *)&old->list.prev->next; prev 1140 kernel/trace/ring_buffer.c if (RB_WARN_ON(cpu_buffer, rb_list_head(list->prev) != list->prev)) prev 1165 kernel/trace/ring_buffer.c if (RB_WARN_ON(cpu_buffer, head->next->prev != head)) prev 1167 kernel/trace/ring_buffer.c if (RB_WARN_ON(cpu_buffer, head->prev->next != head)) prev 1175 kernel/trace/ring_buffer.c bpage->list.next->prev != &bpage->list)) prev 1178 kernel/trace/ring_buffer.c bpage->list.prev->next != &bpage->list)) prev 1546 kernel/trace/ring_buffer.c next_page->prev = tail_page; prev 1637 kernel/trace/ring_buffer.c prev_page = head_page->prev; prev 1640 kernel/trace/ring_buffer.c last_page = pages->prev; prev 1646 kernel/trace/ring_buffer.c first_page->prev = prev_page; prev 1656 kernel/trace/ring_buffer.c head_page->prev = last_page; prev 3731 kernel/trace/ring_buffer.c cpu_buffer->reader_page->list.prev = reader->list.prev; prev 3738 kernel/trace/ring_buffer.c cpu_buffer->pages = reader->list.prev; prev 3779 kernel/trace/ring_buffer.c rb_list_head(reader->list.next)->prev = &cpu_buffer->reader_page->list; prev 2292 kernel/trace/trace.c void tracing_record_taskinfo_sched_switch(struct task_struct *prev, prev 2304 kernel/trace/trace.c done = !(flags & TRACE_RECORD_CMDLINE) || trace_save_cmdline(prev); prev 2306 kernel/trace/trace.c done &= !(flags & TRACE_RECORD_TGID) || trace_save_tgid(prev); prev 359 kernel/trace/trace.h tr = list_entry(ftrace_trace_arrays.prev, prev 542 kernel/trace/trace_events.c struct task_struct *prev, struct task_struct *next) prev 550 kernel/trace/trace_events.c trace_ignore_this_task(pid_list, prev) && prev 556 kernel/trace/trace_events.c struct task_struct *prev, struct task_struct *next) prev 1206 kernel/trace/trace_events.c node = node->prev; prev 25 kernel/trace/trace_sched_switch.c struct task_struct *prev, struct task_struct *next) prev 34 kernel/trace/trace_sched_switch.c tracing_record_taskinfo_sched_switch(prev, next, flags); prev 376 kernel/trace/trace_sched_wakeup.c struct task_struct *prev, prev 390 kernel/trace/trace_sched_wakeup.c entry->prev_pid = prev->pid; prev 391 kernel/trace/trace_sched_wakeup.c entry->prev_prio = prev->prio; prev 392 kernel/trace/trace_sched_wakeup.c entry->prev_state = task_state_index(prev); prev 432 kernel/trace/trace_sched_wakeup.c struct task_struct *prev, struct task_struct *next) prev 441 kernel/trace/trace_sched_wakeup.c tracing_record_cmdline(prev); prev 477 kernel/trace/trace_sched_wakeup.c tracing_sched_switch_trace(wakeup_trace, prev, next, flags, pc); prev 17 kernel/trace/trace_stat.h void *(*stat_next)(void *prev, int idx); prev 723 kernel/user_namespace.c struct uid_gid_extent *prev; prev 726 kernel/user_namespace.c prev = &new_map->extent[idx]; prev 728 kernel/user_namespace.c prev = &new_map->forward[idx]; prev 730 kernel/user_namespace.c prev_upper_first = prev->first; prev 731 kernel/user_namespace.c prev_lower_first = prev->lower_first; prev 732 kernel/user_namespace.c prev_upper_last = prev_upper_first + prev->count - 1; prev 733 kernel/user_namespace.c prev_lower_last = prev_lower_first + prev->count - 1; prev 2000 kernel/workqueue.c worker = list_entry(pool->idle_list.prev, struct worker, entry); prev 4175 kernel/workqueue.c wq->pwqs.prev != &wq->dfl_pwq->pwqs_node), prev 138 lib/dim/net_dim.c struct dim_stats *prev) prev 140 lib/dim/net_dim.c if (!prev->bpms) prev 143 lib/dim/net_dim.c if (IS_SIGNIFICANT_DIFF(curr->bpms, prev->bpms)) prev 144 lib/dim/net_dim.c return (curr->bpms > prev->bpms) ? DIM_STATS_BETTER : prev 147 lib/dim/net_dim.c if (!prev->ppms) prev 151 lib/dim/net_dim.c if (IS_SIGNIFICANT_DIFF(curr->ppms, prev->ppms)) prev 152 lib/dim/net_dim.c return (curr->ppms > prev->ppms) ? DIM_STATS_BETTER : prev 155 lib/dim/net_dim.c if (!prev->epms) prev 158 lib/dim/net_dim.c if (IS_SIGNIFICANT_DIFF(curr->epms, prev->epms)) prev 159 lib/dim/net_dim.c return (curr->epms < prev->epms) ? DIM_STATS_BETTER : prev 27 lib/dim/rdma_dim.c struct dim_stats *prev) prev 30 lib/dim/rdma_dim.c if (!prev->cpms) prev 33 lib/dim/rdma_dim.c if (IS_SIGNIFICANT_DIFF(curr->cpms, prev->cpms)) prev 34 lib/dim/rdma_dim.c return (curr->cpms > prev->cpms) ? DIM_STATS_BETTER : prev 37 lib/dim/rdma_dim.c if (IS_SIGNIFICANT_DIFF(curr->cpe_ratio, prev->cpe_ratio)) prev 38 lib/dim/rdma_dim.c return (curr->cpe_ratio > prev->cpe_ratio) ? DIM_STATS_BETTER : prev 337 lib/klist.c struct klist_node *prev; prev 343 lib/klist.c prev = to_klist_node(last->n_node.prev); prev 347 lib/klist.c prev = to_klist_node(i->i_klist->k_list.prev); prev 350 lib/klist.c while (prev != to_klist_node(&i->i_klist->k_list)) { prev 351 lib/klist.c if (likely(!knode_dead(prev))) { prev 352 lib/klist.c kref_get(&prev->n_ref); prev 353 lib/klist.c i->i_cur = prev; prev 356 lib/klist.c prev = to_klist_node(prev->n_node.prev); prev 20 lib/list_debug.c bool __list_add_valid(struct list_head *new, struct list_head *prev, prev 23 lib/list_debug.c if (CHECK_DATA_CORRUPTION(next->prev != prev, prev 25 lib/list_debug.c prev, next->prev, next) || prev 26 lib/list_debug.c CHECK_DATA_CORRUPTION(prev->next != next, prev 28 lib/list_debug.c next, prev->next, prev) || prev 29 lib/list_debug.c CHECK_DATA_CORRUPTION(new == prev || new == next, prev 31 lib/list_debug.c new, prev, next)) prev 40 lib/list_debug.c struct list_head *prev, *next; prev 42 lib/list_debug.c prev = entry->prev; prev 48 lib/list_debug.c CHECK_DATA_CORRUPTION(prev == LIST_POISON2, prev 51 lib/list_debug.c CHECK_DATA_CORRUPTION(prev->next != entry, prev 53 lib/list_debug.c entry, prev->next) || prev 54 lib/list_debug.c CHECK_DATA_CORRUPTION(next->prev != entry, prev 56 lib/list_debug.c entry, next->prev)) prev 65 lib/list_sort.c a->prev = tail; prev 72 lib/list_sort.c b->prev = tail; prev 94 lib/list_sort.c b->prev = tail; prev 101 lib/list_sort.c head->prev = tail; prev 195 lib/list_sort.c if (list == head->prev) /* Zero or one elements */ prev 199 lib/list_sort.c head->prev->next = NULL; prev 225 lib/list_sort.c tail = &(*tail)->prev; prev 228 lib/list_sort.c struct list_head *a = *tail, *b = a->prev; prev 232 lib/list_sort.c a->prev = b->prev; prev 237 lib/list_sort.c list->prev = pending; prev 246 lib/list_sort.c pending = pending->prev; prev 248 lib/list_sort.c struct list_head *next = pending->prev; prev 17 lib/lockref.c struct lockref new = old, prev = old; \ prev 22 lib/lockref.c if (likely(old.lock_count == prev.lock_count)) { \ prev 329 lib/lru_cache.c n = lc->lru.prev; prev 35 lib/plist.c WARN(n->prev != p || p->next != n, prev 39 lib/plist.c t, t->next, t->prev, prev 40 lib/plist.c p, p->next, p->prev, prev 41 lib/plist.c n, n->next, n->prev); prev 46 lib/plist.c struct list_head *prev = top, *next = top->next; prev 48 lib/plist.c plist_check_prev_next(top, prev, next); prev 50 lib/plist.c prev = next; prev 51 lib/plist.c next = prev->next; prev 52 lib/plist.c plist_check_prev_next(top, prev, next); prev 75 lib/plist.c struct plist_node *first, *iter, *prev = NULL; prev 93 lib/plist.c prev = iter; prev 98 lib/plist.c if (!prev || prev->prio != node->prio) prev 114 lib/rhashtable.c union nested_table __rcu **prev, prev 120 lib/rhashtable.c ntbl = rcu_dereference(*prev); prev 131 lib/rhashtable.c if (cmpxchg((union nested_table **)prev, NULL, ntbl) == NULL) prev 135 lib/rhashtable.c return rcu_dereference(*prev); prev 104 lib/test_list_sort.c if (cur->next->prev != cur) { prev 129 lib/test_list_sort.c if (head.prev != cur) { prev 2022 lib/xarray.c node->array, node->private_list.prev, node->private_list.next); prev 155 lib/zlib_deflate/deflate.c s->prev[(str) & s->w_mask] = match_head = s->head[s->ins_h], \ prev 233 lib/zlib_deflate/deflate.c s->prev = (Pos *) mem->prev_memory; prev 570 lib/zlib_deflate/deflate.c Pos *prev = s->prev; prev 686 lib/zlib_deflate/deflate.c } while ((cur_match = prev[cur_match & wmask]) > limit prev 780 lib/zlib_deflate/deflate.c p = &s->prev[n]; prev 104 lib/zlib_deflate/defutil.h Pos *prev; prev 290 mm/hugetlb.c list_add(&nrg->link, rg->link.prev); prev 302 mm/hugetlb.c list_for_each_entry_safe(rg, trg, rg->link.prev, link) { prev 415 mm/hugetlb.c list_add(&nrg->link, rg->link.prev); prev 426 mm/hugetlb.c list_for_each_entry(rg, rg->link.prev, link) { prev 293 mm/internal.h struct vm_area_struct *prev, struct rb_node *rb_parent); prev 29 mm/interval_tree.c struct vm_area_struct *prev, prev 36 mm/interval_tree.c VM_BUG_ON_VMA(vma_start_pgoff(node) != vma_start_pgoff(prev), node); prev 38 mm/interval_tree.c if (!prev->shared.rb.rb_right) { prev 39 mm/interval_tree.c parent = prev; prev 40 mm/interval_tree.c link = &prev->shared.rb.rb_right; prev 42 mm/interval_tree.c parent = rb_entry(prev->shared.rb.rb_right, prev 229 mm/ksm.c #define STABLE_NODE_DUP_HEAD ((struct list_head *)&migrate_nodes.prev) prev 66 mm/madvise.c struct vm_area_struct **prev, prev 130 mm/madvise.c *prev = vma; prev 135 mm/madvise.c *prev = vma_merge(mm, *prev, start, end, new_flags, vma->anon_vma, prev 138 mm/madvise.c if (*prev) { prev 139 mm/madvise.c vma = *prev; prev 143 mm/madvise.c *prev = vma; prev 254 mm/madvise.c struct vm_area_struct **prev, prev 260 mm/madvise.c *prev = vma; prev 289 mm/madvise.c *prev = NULL; /* tell sys_madvise we drop mmap_sem */ prev 492 mm/madvise.c struct vm_area_struct **prev, prev 498 mm/madvise.c *prev = vma; prev 541 mm/madvise.c struct vm_area_struct **prev, prev 547 mm/madvise.c *prev = vma; prev 761 mm/madvise.c struct vm_area_struct **prev, prev 765 mm/madvise.c *prev = vma; prev 770 mm/madvise.c *prev = NULL; /* mmap_sem has been dropped, prev is stale */ prev 821 mm/madvise.c struct vm_area_struct **prev, prev 828 mm/madvise.c *prev = NULL; /* tell sys_madvise we drop mmap_sem */ prev 934 mm/madvise.c madvise_vma(struct vm_area_struct *vma, struct vm_area_struct **prev, prev 939 mm/madvise.c return madvise_remove(vma, prev, start, end); prev 941 mm/madvise.c return madvise_willneed(vma, prev, start, end); prev 943 mm/madvise.c return madvise_cold(vma, prev, start, end); prev 945 mm/madvise.c return madvise_pageout(vma, prev, start, end); prev 948 mm/madvise.c return madvise_dontneed_free(vma, prev, start, end, behavior); prev 950 mm/madvise.c return madvise_behavior(vma, prev, start, end, behavior); prev 1056 mm/madvise.c struct vm_area_struct *vma, *prev; prev 1102 mm/madvise.c vma = find_vma_prev(current->mm, start, &prev); prev 1104 mm/madvise.c prev = vma; prev 1127 mm/madvise.c error = madvise_vma(vma, &prev, start, tmp, behavior); prev 1131 mm/madvise.c if (prev && start < prev->vm_end) prev 1132 mm/madvise.c start = prev->vm_end; prev 1136 mm/madvise.c if (prev) prev 1137 mm/madvise.c vma = prev->vm_next; prev 1039 mm/memcontrol.c struct mem_cgroup *prev, prev 1053 mm/memcontrol.c if (prev && !reclaim) prev 1054 mm/memcontrol.c pos = prev; prev 1057 mm/memcontrol.c if (prev) prev 1070 mm/memcontrol.c if (prev && reclaim->generation != iter->generation) prev 1101 mm/memcontrol.c if (!prev) prev 1135 mm/memcontrol.c else if (!prev) prev 1142 mm/memcontrol.c if (prev && prev != root) prev 1143 mm/memcontrol.c css_put(&prev->css); prev 1154 mm/memcontrol.c struct mem_cgroup *prev) prev 1158 mm/memcontrol.c if (prev && prev != root) prev 1159 mm/memcontrol.c css_put(&prev->css); prev 413 mm/mempolicy.c struct vm_area_struct *prev; prev 637 mm/mempolicy.c if (qp->prev && qp->prev->vm_end < vma->vm_start) prev 641 mm/mempolicy.c qp->prev = vma; prev 688 mm/mempolicy.c .prev = NULL, prev 735 mm/mempolicy.c struct vm_area_struct *prev; prev 746 mm/mempolicy.c prev = vma->vm_prev; prev 748 mm/mempolicy.c prev = vma; prev 750 mm/mempolicy.c for (; vma && vma->vm_start < end; prev = vma, vma = next) { prev 760 mm/mempolicy.c prev = vma_merge(mm, prev, vmstart, vmend, vma->vm_flags, prev 763 mm/mempolicy.c if (prev) { prev 764 mm/mempolicy.c vma = prev; prev 2306 mm/mempolicy.c struct rb_node *prev = rb_prev(n); prev 2307 mm/mempolicy.c if (!prev) prev 2309 mm/mempolicy.c w = rb_entry(prev, struct sp_node, nd); prev 2312 mm/mempolicy.c n = prev; prev 519 mm/mlock.c static int mlock_fixup(struct vm_area_struct *vma, struct vm_area_struct **prev, prev 536 mm/mlock.c *prev = vma_merge(mm, *prev, start, end, newflags, vma->anon_vma, prev 539 mm/mlock.c if (*prev) { prev 540 mm/mlock.c vma = *prev; prev 579 mm/mlock.c *prev = vma; prev 587 mm/mlock.c struct vm_area_struct * vma, * prev; prev 601 mm/mlock.c prev = vma->vm_prev; prev 603 mm/mlock.c prev = vma; prev 614 mm/mlock.c error = mlock_fixup(vma, &prev, nstart, tmp, newflags); prev 618 mm/mlock.c if (nstart < prev->vm_end) prev 619 mm/mlock.c nstart = prev->vm_end; prev 623 mm/mlock.c vma = prev->vm_next; prev 765 mm/mlock.c struct vm_area_struct * vma, * prev = NULL; prev 785 mm/mlock.c for (vma = current->mm->mmap; vma ; vma = prev->vm_next) { prev 792 mm/mlock.c mlock_fixup(vma, &prev, vma->vm_start, vma->vm_end, newflags); prev 77 mm/mmap.c struct vm_area_struct *vma, struct vm_area_struct *prev, prev 331 mm/mmap.c unsigned long prev = 0, pend = 0; prev 336 mm/mmap.c if (vma->vm_start < prev) { prev 338 mm/mmap.c vma->vm_start, prev); prev 361 mm/mmap.c prev = vma->vm_start; prev 633 mm/mmap.c struct vm_area_struct *prev, struct rb_node **rb_link, prev 636 mm/mmap.c __vma_link_list(mm, vma, prev, rb_parent); prev 641 mm/mmap.c struct vm_area_struct *prev, struct rb_node **rb_link, prev 651 mm/mmap.c __vma_link(mm, vma, prev, rb_link, rb_parent); prev 667 mm/mmap.c struct vm_area_struct *prev; prev 671 mm/mmap.c &prev, &rb_link, &rb_parent)) prev 673 mm/mmap.c __vma_link(mm, vma, prev, rb_link, rb_parent); prev 679 mm/mmap.c struct vm_area_struct *prev, prev 688 mm/mmap.c prev->vm_next = next; prev 690 mm/mmap.c prev = vma->vm_prev; prev 691 mm/mmap.c if (prev) prev 692 mm/mmap.c prev->vm_next = next; prev 697 mm/mmap.c next->vm_prev = prev; prev 705 mm/mmap.c struct vm_area_struct *prev) prev 707 mm/mmap.c __vma_unlink_common(mm, vma, prev, true, vma); prev 1136 mm/mmap.c struct vm_area_struct *prev, unsigned long addr, prev 1153 mm/mmap.c if (prev) prev 1154 mm/mmap.c next = prev->vm_next; prev 1162 mm/mmap.c VM_WARN_ON(prev && addr <= prev->vm_start); prev 1169 mm/mmap.c if (prev && prev->vm_end == addr && prev 1170 mm/mmap.c mpol_equal(vma_policy(prev), policy) && prev 1171 mm/mmap.c can_vma_merge_after(prev, vm_flags, prev 1183 mm/mmap.c is_mergeable_anon_vma(prev->anon_vma, prev 1186 mm/mmap.c err = __vma_adjust(prev, prev->vm_start, prev 1187 mm/mmap.c next->vm_end, prev->vm_pgoff, NULL, prev 1188 mm/mmap.c prev); prev 1190 mm/mmap.c err = __vma_adjust(prev, prev->vm_start, prev 1191 mm/mmap.c end, prev->vm_pgoff, NULL, prev); prev 1194 mm/mmap.c khugepaged_enter_vma_merge(prev, vm_flags); prev 1195 mm/mmap.c return prev; prev 1206 mm/mmap.c if (prev && addr < prev->vm_end) /* case 4 */ prev 1207 mm/mmap.c err = __vma_adjust(prev, prev->vm_start, prev 1208 mm/mmap.c addr, prev->vm_pgoff, NULL, next); prev 1716 mm/mmap.c struct vm_area_struct *vma, *prev; prev 1737 mm/mmap.c while (find_vma_links(mm, addr, addr + len, &prev, &rb_link, prev 1756 mm/mmap.c vma = vma_merge(mm, prev, addr, addr + len, vm_flags, prev 1819 mm/mmap.c vma_link(mm, vma, prev, rb_link, rb_parent); prev 1862 mm/mmap.c unmap_region(mm, vma, prev, vma->vm_start, vma->vm_end); prev 1947 mm/mmap.c struct rb_node *prev = &vma->vm_rb; prev 1948 mm/mmap.c if (!rb_parent(prev)) prev 1950 mm/mmap.c vma = rb_entry(rb_parent(prev), prev 1952 mm/mmap.c if (prev == vma->vm_rb.rb_left) { prev 2051 mm/mmap.c struct rb_node *prev = &vma->vm_rb; prev 2052 mm/mmap.c if (!rb_parent(prev)) prev 2054 mm/mmap.c vma = rb_entry(rb_parent(prev), prev 2056 mm/mmap.c if (prev == vma->vm_rb.rb_right) { prev 2105 mm/mmap.c struct vm_area_struct *vma, *prev; prev 2117 mm/mmap.c vma = find_vma_prev(mm, addr, &prev); prev 2120 mm/mmap.c (!prev || addr >= vm_end_gap(prev))) prev 2143 mm/mmap.c struct vm_area_struct *vma, *prev; prev 2158 mm/mmap.c vma = find_vma_prev(mm, addr, &prev); prev 2161 mm/mmap.c (!prev || addr >= vm_end_gap(prev))) prev 2434 mm/mmap.c struct vm_area_struct *prev; prev 2442 mm/mmap.c prev = vma->vm_prev; prev 2444 mm/mmap.c if (prev && !(prev->vm_flags & VM_GROWSDOWN) && prev 2445 mm/mmap.c (prev->vm_flags & (VM_WRITE|VM_READ|VM_EXEC))) { prev 2446 mm/mmap.c if (address - prev->vm_end < stack_guard_gap) prev 2529 mm/mmap.c struct vm_area_struct *vma, *prev; prev 2532 mm/mmap.c vma = find_vma_prev(mm, addr, &prev); prev 2536 mm/mmap.c if (!prev || !mmget_still_valid(mm) || expand_stack(prev, addr)) prev 2538 mm/mmap.c if (prev->vm_flags & VM_LOCKED) prev 2539 mm/mmap.c populate_vma_page_range(prev, addr, prev->vm_end, NULL); prev 2540 mm/mmap.c return prev; prev 2606 mm/mmap.c struct vm_area_struct *vma, struct vm_area_struct *prev, prev 2609 mm/mmap.c struct vm_area_struct *next = prev ? prev->vm_next : mm->mmap; prev 2616 mm/mmap.c free_pgtables(&tlb, vma, prev ? prev->vm_end : FIRST_USER_ADDRESS, prev 2627 mm/mmap.c struct vm_area_struct *prev, unsigned long end) prev 2632 mm/mmap.c insertion_point = (prev ? &prev->vm_next : &mm->mmap); prev 2642 mm/mmap.c vma->vm_prev = prev; prev 2645 mm/mmap.c mm->highest_vm_end = prev ? vm_end_gap(prev) : 0; prev 2738 mm/mmap.c struct vm_area_struct *vma, *prev, *last; prev 2759 mm/mmap.c prev = vma->vm_prev; prev 2787 mm/mmap.c prev = vma; prev 2797 mm/mmap.c vma = prev ? prev->vm_next : mm->mmap; prev 2830 mm/mmap.c detach_vmas_to_be_unmapped(mm, vma, prev, end); prev 2835 mm/mmap.c unmap_region(mm, vma, prev, start, end); prev 2995 mm/mmap.c struct vm_area_struct *vma, *prev; prev 3016 mm/mmap.c while (find_vma_links(mm, addr, addr + len, &prev, &rb_link, prev 3033 mm/mmap.c vma = vma_merge(mm, prev, addr, addr + len, flags, prev 3053 mm/mmap.c vma_link(mm, vma, prev, rb_link, rb_parent); prev 3173 mm/mmap.c struct vm_area_struct *prev; prev 3177 mm/mmap.c &prev, &rb_link, &rb_parent)) prev 3200 mm/mmap.c vma_link(mm, vma, prev, rb_link, rb_parent); prev 3215 mm/mmap.c struct vm_area_struct *new_vma, *prev; prev 3228 mm/mmap.c if (find_vma_links(mm, addr, addr + len, &prev, &rb_link, &rb_parent)) prev 3230 mm/mmap.c new_vma = vma_merge(mm, prev, addr, addr + len, vma->vm_flags, prev 3270 mm/mmap.c vma_link(mm, new_vma, prev, rb_link, rb_parent); prev 490 mm/mprotect.c struct vm_area_struct *vma, *prev; prev 530 mm/mprotect.c prev = vma->vm_prev; prev 549 mm/mprotect.c prev = vma; prev 587 mm/mprotect.c error = mprotect_fixup(vma, &prev, nstart, tmp, newflags); prev 592 mm/mprotect.c if (nstart < prev->vm_end) prev 593 mm/mprotect.c nstart = prev->vm_end; prev 597 mm/mprotect.c vma = prev->vm_next; prev 588 mm/nommu.c struct vm_area_struct *pvma, *prev; prev 640 mm/nommu.c prev = NULL; prev 642 mm/nommu.c prev = rb_entry(rb_prev, struct vm_area_struct, vm_rb); prev 644 mm/nommu.c __vma_link_list(mm, vma, prev, parent); prev 2192 mm/slab.c p = n->slabs_free.prev; prev 3801 mm/slab.c struct array_cache __percpu *cpu_cache, *prev; prev 3808 mm/slab.c prev = cachep->cpu_cache; prev 3814 mm/slab.c if (prev) prev 3822 mm/slab.c if (!prev) prev 3829 mm/slab.c struct array_cache *ac = per_cpu_ptr(prev, cpu); prev 3838 mm/slab.c free_percpu(prev); prev 240 mm/slob.c slob_t *prev, *cur, *aligned = NULL; prev 244 mm/slob.c for (prev = NULL, cur = sp->freelist; ; prev = cur, cur = slob_next(cur)) { prev 267 mm/slob.c prev = cur; prev 274 mm/slob.c if (prev) prev 275 mm/slob.c set_slob(prev, slob_units(prev), next); prev 279 mm/slob.c if (prev) prev 280 mm/slob.c set_slob(prev, slob_units(prev), cur + units); prev 385 mm/slob.c slob_t *prev, *next, *b = (slob_t *)block; prev 441 mm/slob.c prev = sp->freelist; prev 442 mm/slob.c next = slob_next(prev); prev 444 mm/slob.c prev = next; prev 445 mm/slob.c next = slob_next(prev); prev 448 mm/slob.c if (!slob_last(prev) && b + units == next) { prev 454 mm/slob.c if (prev + slob_units(prev) == b) { prev 455 mm/slob.c units = slob_units(b) + slob_units(prev); prev 456 mm/slob.c set_slob(prev, units, slob_next(b)); prev 458 mm/slob.c set_slob(prev, slob_units(prev), b); prev 1398 mm/swapfile.c struct swap_info_struct *p, *prev; prev 1404 mm/swapfile.c prev = NULL; prev 1415 mm/swapfile.c p = swap_info_get_cont(entries[i], prev); prev 1418 mm/swapfile.c prev = p; prev 2093 mm/swapfile.c unsigned int prev, bool frontswap) prev 2104 mm/swapfile.c for (i = prev + 1; i < si->max; i++) { prev 3677 mm/swapfile.c page = list_entry(page->lru.prev, struct page, lru); prev 3682 mm/swapfile.c page = list_entry(page->lru.prev, struct page, lru); prev 3702 mm/swapfile.c page = list_entry(page->lru.prev, struct page, lru); prev 3708 mm/swapfile.c page = list_entry(page->lru.prev, struct page, lru); prev 274 mm/util.c struct vm_area_struct *prev, struct rb_node *rb_parent) prev 278 mm/util.c vma->vm_prev = prev; prev 279 mm/util.c if (prev) { prev 280 mm/util.c next = prev->vm_next; prev 281 mm/util.c prev->vm_next = vma; prev 511 mm/vmalloc.c head = head->prev; prev 740 mm/vmalloc.c if (next->prev != head) { prev 741 mm/vmalloc.c sibling = list_entry(next->prev, struct vmap_area, list); prev 139 mm/vmscan.c if ((_page)->lru.prev != _base) { \ prev 140 mm/vmscan.c struct page *prev; \ prev 142 mm/vmscan.c prev = lru_to_page(&(_page->lru)); \ prev 143 mm/vmscan.c prefetch(&prev->_field); \ prev 153 mm/vmscan.c if ((_page)->lru.prev != _base) { \ prev 154 mm/vmscan.c struct page *prev; \ prev 156 mm/vmscan.c prev = lru_to_page(&(_page->lru)); \ prev 157 mm/vmscan.c prefetchw(&prev->_field); \ prev 1255 mm/zsmalloc.c static bool can_merge(struct size_class *prev, int pages_per_zspage, prev 1258 mm/zsmalloc.c if (prev->pages_per_zspage == pages_per_zspage && prev 1259 mm/zsmalloc.c prev->objs_per_zspage == objs_per_zspage) prev 860 net/9p/trans_fd.c m, m->mux_list.prev, m->mux_list.next); prev 376 net/atm/br2684.c brdev->brvccs.next != brdev->brvccs.prev) /* >1 VCC */ prev 709 net/atm/br2684.c brdev->number = BRPRIV(list_entry_brdev(br2684_devs.prev))->number + 1; prev 113 net/atm/mpoa_caches.c entry->prev = NULL; prev 115 net/atm/mpoa_caches.c client->in_cache->prev = entry; prev 200 net/atm/mpoa_caches.c if (entry->prev != NULL) prev 201 net/atm/mpoa_caches.c entry->prev->next = entry->next; prev 205 net/atm/mpoa_caches.c entry->next->prev = entry->prev; prev 433 net/atm/mpoa_caches.c if (entry->prev != NULL) prev 434 net/atm/mpoa_caches.c entry->prev->next = entry->next; prev 438 net/atm/mpoa_caches.c entry->next->prev = entry->prev; prev 473 net/atm/mpoa_caches.c entry->prev = NULL; prev 475 net/atm/mpoa_caches.c client->eg_cache->prev = entry; prev 19 net/atm/mpoa_caches.h struct in_cache_entry *prev; prev 56 net/atm/mpoa_caches.h struct eg_cache_entry *prev; prev 114 net/bridge/br_forward.c static int deliver_clone(const struct net_bridge_port *prev, prev 125 net/bridge/br_forward.c __br_forward(prev, skb, local_orig); prev 169 net/bridge/br_forward.c struct net_bridge_port *prev, struct net_bridge_port *p, prev 176 net/bridge/br_forward.c return prev; prev 178 net/bridge/br_forward.c if (!prev) prev 181 net/bridge/br_forward.c err = deliver_clone(prev, skb, local_orig); prev 194 net/bridge/br_forward.c struct net_bridge_port *prev = NULL; prev 223 net/bridge/br_forward.c prev = maybe_deliver(prev, p, skb, local_orig); prev 224 net/bridge/br_forward.c if (IS_ERR(prev)) prev 228 net/bridge/br_forward.c if (!prev) prev 232 net/bridge/br_forward.c deliver_clone(prev, skb, local_orig); prev 234 net/bridge/br_forward.c __br_forward(prev, skb, local_orig); prev 275 net/bridge/br_forward.c struct net_bridge_port *prev = NULL; prev 299 net/bridge/br_forward.c prev = maybe_deliver(prev, port, skb, local_orig); prev 300 net/bridge/br_forward.c if (IS_ERR(prev)) prev 309 net/bridge/br_forward.c if (!prev) prev 313 net/bridge/br_forward.c deliver_clone(prev, skb, local_orig); prev 315 net/bridge/br_forward.c __br_forward(prev, skb, local_orig); prev 42 net/bridge/netfilter/ebt_limit.c info->credit += (now - xchg(&info->prev, now)) * CREDITS_PER_JIFFY; prev 82 net/bridge/netfilter/ebt_limit.c info->prev = jiffies; prev 97 net/bridge/netfilter/ebt_limit.c compat_ulong_t prev; prev 109 net/bridge/netfilter/ebt_limit.c .usersize = offsetof(struct ebt_limit_info, prev), prev 129 net/can/j1939/address-claim.c struct j1939_ecu *ecu, *prev; prev 187 net/can/j1939/address-claim.c prev = j1939_ecu_get_by_addr_locked(priv, skcb->addr.sa); prev 188 net/can/j1939/address-claim.c if (prev) { prev 189 net/can/j1939/address-claim.c if (ecu->name > prev->name) { prev 191 net/can/j1939/address-claim.c j1939_ecu_put(prev); prev 195 net/can/j1939/address-claim.c j1939_ecu_unmap_locked(prev); prev 196 net/can/j1939/address-claim.c j1939_ecu_put(prev); prev 32 net/ceph/pagelist.c struct page *page = list_entry(pl->head.prev, struct page, lru); prev 140 net/ceph/pagelist.c c->page_lru = pl->head.prev; prev 158 net/ceph/pagelist.c while (pl->head.prev != c->page_lru) { prev 159 net/ceph/pagelist.c page = list_entry(pl->head.prev, struct page, lru); prev 166 net/ceph/pagelist.c page = list_entry(pl->head.prev, struct page, lru); prev 101 net/core/datagram.c if (READ_ONCE(sk->sk_receive_queue.prev) != skb) prev 152 net/core/datagram.c skb->prev->next = nskb; prev 153 net/core/datagram.c skb->next->prev = nskb; prev 154 net/core/datagram.c nskb->prev = skb->prev; prev 183 net/core/datagram.c *last = queue->prev; prev 282 net/core/datagram.c } while (READ_ONCE(sk->sk_receive_queue.prev) != *last); prev 409 net/core/dev.c list_add_rcu(&po->list, elem->list.prev); prev 3318 net/core/dev.c skb->prev = skb; prev 3331 net/core/dev.c tail = skb->prev; prev 205 net/core/dst.c unsigned long prev, new; prev 211 net/core/dst.c prev = cmpxchg(&dst->_metrics, old, new); prev 213 net/core/dst.c if (prev != old) { prev 215 net/core/dst.c p = (struct dst_metrics *)__DST_METRICS_PTR(prev); prev 216 net/core/dst.c if (prev & DST_METRICS_READ_ONLY) prev 218 net/core/dst.c } else if (prev & DST_METRICS_REFCOUNTED) { prev 231 net/core/dst.c unsigned long prev, new; prev 234 net/core/dst.c prev = cmpxchg(&dst->_metrics, old, new); prev 235 net/core/dst.c if (prev == old) prev 2479 net/core/filter.c int prev; prev 2482 net/core/filter.c prev = i; prev 2484 net/core/filter.c msg->sg.data[prev] = msg->sg.data[i]; prev 985 net/core/skbuff.c n->next = n->prev = NULL; prev 3942 net/core/skbuff.c segs->prev = tail; prev 34 net/ieee802154/6lowpan/reassembly.c struct sk_buff *prev, struct net_device *ldev); prev 792 net/ieee802154/socket.c struct sock *sk, *prev = NULL; prev 808 net/ieee802154/socket.c if (prev) { prev 813 net/ieee802154/socket.c dgram_rcv_skb(prev, clone); prev 816 net/ieee802154/socket.c prev = sk; prev 820 net/ieee802154/socket.c if (prev) { prev 821 net/ieee802154/socket.c dgram_rcv_skb(prev, skb); prev 268 net/ipv4/cipso_ipv4.c __list_del(entry->list.prev, entry->list.next); prev 270 net/ipv4/cipso_ipv4.c prev_entry->list.prev, prev 330 net/ipv4/cipso_ipv4.c old_entry = list_entry(cipso_v4_cache[bkt].list.prev, prev 2020 net/ipv4/igmp.c struct ip_sf_list *prev = NULL; prev 2025 net/ipv4/igmp.c prev = dpsf; prev 2028 net/ipv4/igmp.c if (prev) prev 2029 net/ipv4/igmp.c prev->sf_next = dpsf->sf_next; prev 298 net/ipv4/inet_fragment.c struct inet_frag_queue **prev) prev 305 net/ipv4/inet_fragment.c *prev = ERR_PTR(-ENOMEM); prev 310 net/ipv4/inet_fragment.c *prev = rhashtable_lookup_get_insert_key(&fqdir->rhashtable, &q->key, prev 312 net/ipv4/inet_fragment.c if (*prev) { prev 324 net/ipv4/inet_fragment.c struct inet_frag_queue *fq = NULL, *prev; prev 331 net/ipv4/inet_fragment.c prev = rhashtable_lookup(&fqdir->rhashtable, key, fqdir->f->rhash_params); prev 332 net/ipv4/inet_fragment.c if (!prev) prev 333 net/ipv4/inet_fragment.c fq = inet_frag_create(fqdir, key, &prev); prev 334 net/ipv4/inet_fragment.c if (!IS_ERR_OR_NULL(prev)) { prev 335 net/ipv4/inet_fragment.c fq = prev; prev 513 net/ipv4/inet_fragment.c fp->prev = NULL; prev 540 net/ipv4/inet_fragment.c head->prev = NULL; prev 1468 net/ipv4/route.c struct rtable *orig, *prev, **p; prev 1482 net/ipv4/route.c prev = cmpxchg(p, orig, rt); prev 1483 net/ipv4/route.c if (prev == orig) { prev 218 net/ipv4/tcp_cong.c const struct tcp_congestion_ops *prev; prev 228 net/ipv4/tcp_cong.c prev = xchg(&net->ipv4.tcp_congestion_control, ca); prev 229 net/ipv4/tcp_cong.c if (prev) prev 230 net/ipv4/tcp_cong.c module_put(prev->owner); prev 1282 net/ipv4/tcp_input.c static bool tcp_shifted_skb(struct sock *sk, struct sk_buff *prev, prev 1308 net/ipv4/tcp_input.c TCP_SKB_CB(prev)->end_seq += shifted; prev 1311 net/ipv4/tcp_input.c tcp_skb_pcount_add(prev, pcount); prev 1320 net/ipv4/tcp_input.c if (!TCP_SKB_CB(prev)->tcp_gso_size) prev 1321 net/ipv4/tcp_input.c TCP_SKB_CB(prev)->tcp_gso_size = mss; prev 1328 net/ipv4/tcp_input.c TCP_SKB_CB(prev)->sacked |= (TCP_SKB_CB(skb)->sacked & TCPCB_EVER_RETRANS); prev 1339 net/ipv4/tcp_input.c tp->retransmit_skb_hint = prev; prev 1341 net/ipv4/tcp_input.c tp->lost_skb_hint = prev; prev 1342 net/ipv4/tcp_input.c tp->lost_cnt_hint -= tcp_skb_pcount(prev); prev 1345 net/ipv4/tcp_input.c TCP_SKB_CB(prev)->tcp_flags |= TCP_SKB_CB(skb)->tcp_flags; prev 1346 net/ipv4/tcp_input.c TCP_SKB_CB(prev)->eor = TCP_SKB_CB(skb)->eor; prev 1348 net/ipv4/tcp_input.c TCP_SKB_CB(prev)->end_seq++; prev 1353 net/ipv4/tcp_input.c tcp_skb_collapse_tstamp(prev, skb); prev 1354 net/ipv4/tcp_input.c if (unlikely(TCP_SKB_CB(prev)->tx.delivered_mstamp)) prev 1355 net/ipv4/tcp_input.c TCP_SKB_CB(prev)->tx.delivered_mstamp = 0; prev 1402 net/ipv4/tcp_input.c struct sk_buff *prev; prev 1419 net/ipv4/tcp_input.c prev = skb_rb_prev(skb); prev 1420 net/ipv4/tcp_input.c if (!prev) prev 1423 net/ipv4/tcp_input.c if ((TCP_SKB_CB(prev)->sacked & TCPCB_TAGBITS) != TCPCB_SACKED_ACKED) prev 1426 net/ipv4/tcp_input.c if (!tcp_skb_can_collapse_to(prev)) prev 1440 net/ipv4/tcp_input.c if (mss != tcp_skb_seglen(prev)) prev 1481 net/ipv4/tcp_input.c if (mss != tcp_skb_seglen(prev)) prev 1498 net/ipv4/tcp_input.c if (!tcp_skb_shift(prev, skb, pcount, len)) prev 1500 net/ipv4/tcp_input.c if (!tcp_shifted_skb(sk, prev, skb, state, pcount, len, mss, dup_sack)) prev 1506 net/ipv4/tcp_input.c skb = skb_rb_next(prev); prev 1517 net/ipv4/tcp_input.c if (tcp_skb_shift(prev, skb, pcount, len)) prev 1518 net/ipv4/tcp_input.c tcp_shifted_skb(sk, prev, skb, state, pcount, prev 1522 net/ipv4/tcp_input.c return prev; prev 5077 net/ipv4/tcp_input.c struct rb_node *node, *prev; prev 5087 net/ipv4/tcp_input.c prev = rb_prev(node); prev 5091 net/ipv4/tcp_input.c if (!prev || goal <= 0) { prev 5098 net/ipv4/tcp_input.c node = prev; prev 5100 net/ipv4/tcp_input.c tp->ooo_last_skb = rb_to_skb(prev); prev 3432 net/ipv6/addrconf.c struct fib6_info *f6i, *prev; prev 3441 net/ipv6/addrconf.c prev = ifp->rt; prev 3445 net/ipv6/addrconf.c fib6_info_release(prev); prev 230 net/ipv6/calipso.c __list_del(entry->list.prev, entry->list.next); prev 232 net/ipv6/calipso.c prev_entry->list.prev, prev 294 net/ipv6/calipso.c old_entry = list_entry(calipso_cache[bkt].list.prev, prev 243 net/ipv6/ila/ila_xlat.c struct ila_map *tila = head, *prev = NULL; prev 254 net/ipv6/ila/ila_xlat.c prev = tila; prev 259 net/ipv6/ila/ila_xlat.c if (prev) { prev 262 net/ipv6/ila/ila_xlat.c rcu_assign_pointer(prev->next, ila); prev 286 net/ipv6/ila/ila_xlat.c struct ila_map *ila, *head, *prev; prev 296 net/ipv6/ila/ila_xlat.c prev = NULL; prev 300 net/ipv6/ila/ila_xlat.c prev = ila; prev 308 net/ipv6/ila/ila_xlat.c if (prev) { prev 310 net/ipv6/ila/ila_xlat.c rcu_assign_pointer(prev->next, ila->next); prev 1590 net/ipv6/ip6_fib.c struct fib6_node *fn, *prev = NULL; prev 1617 net/ipv6/ip6_fib.c prev = fn; prev 1632 net/ipv6/ip6_fib.c return prev; prev 2281 net/ipv6/mcast.c struct ip6_sf_list *prev = NULL; prev 2288 net/ipv6/mcast.c prev = dpsf; prev 2291 net/ipv6/mcast.c if (prev) prev 2292 net/ipv6/mcast.c prev->sf_next = dpsf->sf_next; prev 174 net/ipv6/netfilter/nf_conntrack_reasm.c struct sk_buff *prev; prev 255 net/ipv6/netfilter/nf_conntrack_reasm.c prev = fq->q.fragments_tail; prev 289 net/ipv6/netfilter/nf_conntrack_reasm.c err = nf_ct_frag6_reasm(fq, skb, prev, dev); prev 1411 net/ipv6/route.c struct rt6_info *prev, **p; prev 1414 net/ipv6/route.c prev = xchg(p, NULL); prev 1415 net/ipv6/route.c if (prev) { prev 1416 net/ipv6/route.c dst_dev_put(&prev->dst); prev 1417 net/ipv6/route.c dst_release(&prev->dst); prev 1429 net/ipv6/route.c struct rt6_info *pcpu_rt, *prev, **p; prev 1436 net/ipv6/route.c prev = cmpxchg(p, NULL, pcpu_rt); prev 1437 net/ipv6/route.c BUG_ON(prev); prev 4403 net/mac80211/rx.c struct ieee80211_sub_if_data *prev; prev 4475 net/mac80211/rx.c prev = NULL; prev 4491 net/mac80211/rx.c if (!prev) { prev 4492 net/mac80211/rx.c prev = sdata; prev 4496 net/mac80211/rx.c rx.sta = sta_info_get_bss(prev, hdr->addr2); prev 4497 net/mac80211/rx.c rx.sdata = prev; prev 4500 net/mac80211/rx.c prev = sdata; prev 4503 net/mac80211/rx.c if (prev) { prev 4504 net/mac80211/rx.c rx.sta = sta_info_get_bss(prev, hdr->addr2); prev 4505 net/mac80211/rx.c rx.sdata = prev; prev 3913 net/mac80211/tx.c skb->prev = NULL; prev 190 net/netfilter/ipset/ip_set_list_set.c struct set_elem *e, *next, *prev = NULL; prev 198 net/netfilter/ipset/ip_set_list_set.c prev = e; prev 209 net/netfilter/ipset/ip_set_list_set.c ret = prev && prev->id == d->refid; prev 237 net/netfilter/ipset/ip_set_list_set.c struct set_elem *e, *n, *prev, *next; prev 241 net/netfilter/ipset/ip_set_list_set.c n = prev = next = NULL; prev 253 net/netfilter/ipset/ip_set_list_set.c prev = e; prev 258 net/netfilter/ipset/ip_set_list_set.c (d->before < 0 && !prev)) prev 284 net/netfilter/ipset/ip_set_list_set.c if (prev->list.prev != &map->members) prev 285 net/netfilter/ipset/ip_set_list_set.c n = list_prev_entry(prev, list); prev 304 net/netfilter/ipset/ip_set_list_set.c else if (prev) prev 305 net/netfilter/ipset/ip_set_list_set.c list_add_rcu(&e->list, &prev->list); prev 319 net/netfilter/ipset/ip_set_list_set.c struct set_elem *e, *next, *prev = NULL; prev 326 net/netfilter/ipset/ip_set_list_set.c prev = e; prev 336 net/netfilter/ipset/ip_set_list_set.c if (!prev || prev->id != d->refid) prev 43 net/netfilter/ipvs/ip_vs_rr.c svc->sched_data = p->next->prev; prev 122 net/netfilter/nft_set_hash.c struct nft_rhash_elem *he, *prev; prev 137 net/netfilter/nft_set_hash.c prev = rhashtable_lookup_get_insert_key(&priv->ht, &arg, &he->node, prev 139 net/netfilter/nft_set_hash.c if (IS_ERR(prev)) prev 143 net/netfilter/nft_set_hash.c if (prev) { prev 145 net/netfilter/nft_set_hash.c he = prev; prev 169 net/netfilter/nft_set_hash.c struct nft_rhash_elem *prev; prev 171 net/netfilter/nft_set_hash.c prev = rhashtable_lookup_get_insert_key(&priv->ht, &arg, &he->node, prev 173 net/netfilter/nft_set_hash.c if (IS_ERR(prev)) prev 174 net/netfilter/nft_set_hash.c return PTR_ERR(prev); prev 175 net/netfilter/nft_set_hash.c if (prev) { prev 176 net/netfilter/nft_set_hash.c *ext = &prev->ext; prev 275 net/netfilter/x_tables.c u8 __user *prev, u8 rev) prev 281 net/netfilter/x_tables.c if (put_user(rev, prev)) prev 98 net/netfilter/xt_hashlimit.c unsigned long prev; /* last modification */ prev 532 net/netfilter/xt_hashlimit.c unsigned long delta = now - dh->rateinfo.prev; prev 544 net/netfilter/xt_hashlimit.c dh->rateinfo.prev = now; prev 553 net/netfilter/xt_hashlimit.c dh->rateinfo.prev = now; prev 576 net/netfilter/xt_hashlimit.c dh->rateinfo.prev = jiffies; prev 19 net/netfilter/xt_limit.c unsigned long prev; prev 72 net/netfilter/xt_limit.c priv->credit += (now - xchg(&priv->prev, now)) * CREDITS_PER_JIFFY; prev 119 net/netfilter/xt_limit.c priv->prev = jiffies; prev 142 net/netfilter/xt_limit.c compat_ulong_t prev; prev 157 net/netfilter/xt_limit.c .prev = cm->prev | (unsigned long)cm->master << 32, prev 171 net/netfilter/xt_limit.c .prev = m->prev, prev 175 net/netfilter/xt_limit.c .master = m->prev >> 32, prev 194 net/netfilter/xt_limit.c .usersize = offsetof(struct xt_rateinfo, prev), prev 165 net/netlabel/netlabel_addrlist.c iter->list.prev, prev 203 net/netlabel/netlabel_addrlist.c iter->list.prev, prev 560 net/netlabel/netlabel_kapi.c struct netlbl_lsm_catmap *prev = NULL; prev 567 net/netlabel/netlabel_kapi.c prev = iter; prev 587 net/netlabel/netlabel_kapi.c if (prev == NULL) { prev 591 net/netlabel/netlabel_kapi.c iter->next = prev->next; prev 592 net/netlabel/netlabel_kapi.c prev->next = iter; prev 666 net/netlabel/netlabel_kapi.c struct netlbl_lsm_catmap *prev = NULL; prev 692 net/netlabel/netlabel_kapi.c if (prev && idx == 0 && bit == 0) prev 693 net/netlabel/netlabel_kapi.c return prev->startbit + NETLBL_CATMAP_SIZE - 1; prev 700 net/netlabel/netlabel_kapi.c prev = iter; prev 1114 net/netlink/genetlink.c struct net *net, *prev = NULL; prev 1119 net/netlink/genetlink.c if (prev) { prev 1125 net/netlink/genetlink.c err = nlmsg_multicast(prev->genl_sock, tmp, prev 1133 net/netlink/genetlink.c prev = net; prev 1136 net/netlink/genetlink.c err = nlmsg_multicast(prev->genl_sock, skb, portid, group, flags); prev 765 net/packet/af_packet.c last_pkt = (struct tpacket3_hdr *)pkc1->prev; prev 834 net/packet/af_packet.c pkc1->prev = pkc1->nxt_offset; prev 996 net/packet/af_packet.c pkc->prev = curr; prev 1101 net/packet/af_packet.c unsigned int prev; prev 1103 net/packet/af_packet.c prev = rb->prb_bdqc.kactive_blk_num-1; prev 1105 net/packet/af_packet.c prev = rb->prb_bdqc.knum_blocks-1; prev 1106 net/packet/af_packet.c return prev; prev 38 net/packet/internal.h char *prev; prev 666 net/rds/af_rds.c &rs->rs_item != rs->rs_item.prev)); prev 82 net/rds/ib_recv.c struct list_head *from_last = from->prev; prev 490 net/rds/ib_send.c struct rds_ib_send_work *prev; prev 620 net/rds/ib_send.c prev = NULL; prev 686 net/rds/ib_send.c if (prev) prev 687 net/rds/ib_send.c prev->s_wr.next = &send->s_wr; prev 688 net/rds/ib_send.c prev = send; prev 704 net/rds/ib_send.c prev->s_op = ic->i_data_op; prev 705 net/rds/ib_send.c prev->s_wr.send_flags |= IB_SEND_SOLICITED; prev 706 net/rds/ib_send.c if (!(prev->s_wr.send_flags & IB_SEND_SIGNALED)) prev 707 net/rds/ib_send.c nr_sig += rds_ib_set_wr_signal_state(ic, prev, true); prev 733 net/rds/ib_send.c if (prev->s_op) { prev 734 net/rds/ib_send.c ic->i_data_op = prev->s_op; prev 735 net/rds/ib_send.c prev->s_op = NULL; prev 844 net/rds/ib_send.c struct rds_ib_send_work *prev; prev 890 net/rds/ib_send.c prev = NULL; prev 917 net/rds/ib_send.c if (prev) prev 918 net/rds/ib_send.c prev->s_rdma_wr.wr.next = &send->s_rdma_wr.wr; prev 939 net/rds/ib_send.c prev = send; prev 946 net/rds/ib_send.c prev->s_op = op; prev 311 net/rfkill/core.c bool prev, curr; prev 326 net/rfkill/core.c prev = rfkill->state & RFKILL_BLOCK_SW; prev 328 net/rfkill/core.c if (prev) prev 363 net/rfkill/core.c if (prev != curr) prev 528 net/rfkill/core.c bool ret, prev; prev 533 net/rfkill/core.c prev = !!(rfkill->state & RFKILL_BLOCK_HW); prev 544 net/rfkill/core.c if (rfkill->registered && prev != blocked) prev 568 net/rfkill/core.c bool prev, hwblock; prev 573 net/rfkill/core.c prev = !!(rfkill->state & RFKILL_BLOCK_SW); prev 582 net/rfkill/core.c if (prev != blocked && !hwblock) prev 476 net/rxrpc/call_object.c call->recvmsg_link.prev = NULL; prev 217 net/sched/sch_etf.c skb->prev = NULL; prev 241 net/sched/sch_etf.c skb->prev = NULL; prev 564 net/sched/sch_generic.c .prev = (struct sk_buff *)&noop_qdisc.gso_skb, prev 570 net/sched/sch_generic.c .prev = (struct sk_buff *)&noop_qdisc.skb_bad_txq, prev 191 net/sched/sch_hhf.c u32 prev = flow->hit_timestamp + q->hhf_evict_timeout; prev 193 net/sched/sch_hhf.c if (hhf_time_before(prev, now)) { prev 221 net/sched/sch_hhf.c u32 prev = flow->hit_timestamp + q->hhf_evict_timeout; prev 223 net/sched/sch_hhf.c if (hhf_time_before(prev, now)) prev 255 net/sched/sch_hhf.c u32 prev; prev 259 net/sched/sch_hhf.c prev = q->hhf_arrays_reset_timestamp + q->hhf_reset_timeout; prev 260 net/sched/sch_hhf.c if (hhf_time_before(prev, now)) { prev 448 net/sched/sch_netem.c skb->prev = NULL; prev 704 net/sched/sch_netem.c skb->prev = NULL; prev 97 net/sched/sch_sfq.c sfq_index prev; prev 213 net/sched/sch_sfq.c slot->dep.prev = p; prev 216 net/sched/sch_sfq.c sfq_dep_head(q, n)->prev = x; prev 222 net/sched/sch_sfq.c p = q->slots[x].dep.prev; \ prev 224 net/sched/sch_sfq.c sfq_dep_head(q, n)->prev = p; \ prev 261 net/sched/sch_sfq.c slot->skblist_prev = skb->prev; prev 262 net/sched/sch_sfq.c skb->prev->next = (struct sk_buff *)slot; prev 263 net/sched/sch_sfq.c skb->next = skb->prev = NULL; prev 273 net/sched/sch_sfq.c skb->next->prev = (struct sk_buff *)slot; prev 274 net/sched/sch_sfq.c skb->next = skb->prev = NULL; prev 287 net/sched/sch_sfq.c skb->prev = slot->skblist_prev; prev 747 net/sched/sch_sfq.c q->dep[i].prev = i + SFQ_MAX_FLOWS; prev 133 net/sched/sch_teql.c struct Qdisc *q, *prev; prev 137 net/sched/sch_teql.c prev = master->slaves; prev 138 net/sched/sch_teql.c if (prev) { prev 140 net/sched/sch_teql.c q = NEXT_SLAVE(prev); prev 142 net/sched/sch_teql.c NEXT_SLAVE(prev) = NEXT_SLAVE(q); prev 162 net/sched/sch_teql.c } while ((prev = q) != master->slaves); prev 332 net/sctp/outqueue.c list_add(new, pos->prev); prev 428 net/sctp/stream_interleave.c event_list = (struct sk_buff_head *)sctp_event2skb(event)->prev; prev 133 net/sctp/stream_sched_prio.c list_add(&soute->prio_list, prio_head->next->prio_list.prev); prev 142 net/sctp/stream_sched_prio.c list_add(&prio_head->prio_sched, prio->prio_sched.prev); prev 758 net/sctp/ulpqueue.c event_list = (struct sk_buff_head *) sctp_event2skb(event)->prev; prev 667 net/sunrpc/cache.c discard = list_entry(cache_defer_list.prev, prev 150 net/sunrpc/sched.c task->u.tk_wait.list.prev = NULL; prev 167 net/sunrpc/sched.c if (task->u.tk_wait.list.prev == NULL) { prev 616 net/sunrpc/svc_xprt.c xprt = list_entry(serv->sv_tempsocks.prev, prev 280 net/sunrpc/xprtmultipath.c struct rpc_xprt *pos, *prev = NULL; prev 284 net/sunrpc/xprtmultipath.c if (cur == prev) prev 288 net/sunrpc/xprtmultipath.c prev = pos; prev 528 net/sunrpc/xprtrdma/frwr_ops.c struct ib_send_wr *first, **prev, *last; prev 540 net/sunrpc/xprtrdma/frwr_ops.c prev = &first; prev 557 net/sunrpc/xprtrdma/frwr_ops.c *prev = last; prev 558 net/sunrpc/xprtrdma/frwr_ops.c prev = &last->next; prev 633 net/sunrpc/xprtrdma/frwr_ops.c struct ib_send_wr *first, *last, **prev; prev 643 net/sunrpc/xprtrdma/frwr_ops.c prev = &first; prev 660 net/sunrpc/xprtrdma/frwr_ops.c *prev = last; prev 661 net/sunrpc/xprtrdma/frwr_ops.c prev = &last->next; prev 390 net/tipc/group.c u16 prev = grp->bc_snd_nxt - 1; prev 399 net/tipc/group.c m->bc_acked = prev; prev 326 net/tipc/monitor.c struct tipc_peer *peer, *prev, *head; prev 332 net/tipc/monitor.c prev = peer_prev(peer); prev 338 net/tipc/monitor.c head = peer_head(prev); prev 341 net/tipc/monitor.c mon_update_neighbors(mon, prev); prev 360 net/tipc/monitor.c struct tipc_peer *cur, *prev, *p; prev 373 net/tipc/monitor.c prev = self; prev 375 net/tipc/monitor.c if ((addr > prev->addr) && (addr < cur->addr)) prev 377 net/tipc/monitor.c if (((addr < cur->addr) || (addr > prev->addr)) && prev 378 net/tipc/monitor.c (prev->addr > cur->addr)) prev 380 net/tipc/monitor.c prev = cur; prev 359 net/tls/tls_device_fallback.c nskb->prev = nskb; prev 10882 net/wireless/nl80211.c s32 prev = S32_MIN; prev 10886 net/wireless/nl80211.c if (thresholds[i] > 0 || thresholds[i] <= prev) prev 10889 net/wireless/nl80211.c prev = thresholds[i]; prev 694 net/xfrm/xfrm_policy.c struct xfrm_pol_inexact_bin *bin, *prev; prev 721 net/xfrm/xfrm_policy.c prev = rhashtable_lookup_get_insert_key(&xfrm_policy_inexact_table, prev 724 net/xfrm/xfrm_policy.c if (!prev) { prev 731 net/xfrm/xfrm_policy.c return IS_ERR(prev) ? NULL : prev; prev 37 samples/bpf/test_lru_dist.c struct list_head *next, *prev; prev 43 samples/bpf/test_lru_dist.c list->prev = list; prev 52 samples/bpf/test_lru_dist.c struct list_head *prev, prev 55 samples/bpf/test_lru_dist.c next->prev = new; prev 57 samples/bpf/test_lru_dist.c new->prev = prev; prev 58 samples/bpf/test_lru_dist.c prev->next = new; prev 66 samples/bpf/test_lru_dist.c static inline void __list_del(struct list_head *prev, struct list_head *next) prev 68 samples/bpf/test_lru_dist.c next->prev = prev; prev 69 samples/bpf/test_lru_dist.c prev->next = next; prev 74 samples/bpf/test_lru_dist.c __list_del(entry->prev, entry->next); prev 87 samples/bpf/test_lru_dist.c list_entry((ptr)->prev, type, member) prev 47 samples/bpf/xdp1_user.c __u64 values[nr_cpus], prev[UINT8_MAX] = { 0 }; prev 61 samples/bpf/xdp1_user.c if (sum > prev[key]) prev 63 samples/bpf/xdp1_user.c key, (sum - prev[key]) / interval); prev 64 samples/bpf/xdp1_user.c prev[key] = sum; prev 311 samples/bpf/xdp_monitor_user.c struct record_u64 *rec, *prev; prev 316 samples/bpf/xdp_monitor_user.c prev = &stats_prev->xdp_redirect[rec_i]; prev 317 samples/bpf/xdp_monitor_user.c t = calc_period_u64(rec, prev); prev 321 samples/bpf/xdp_monitor_user.c struct u64rec *p = &prev->cpu[i]; prev 329 samples/bpf/xdp_monitor_user.c pps = calc_pps_u64(&rec->total, &prev->total, t); prev 336 samples/bpf/xdp_monitor_user.c struct record_u64 *rec, *prev; prev 341 samples/bpf/xdp_monitor_user.c prev = &stats_prev->xdp_exception[rec_i]; prev 342 samples/bpf/xdp_monitor_user.c t = calc_period_u64(rec, prev); prev 346 samples/bpf/xdp_monitor_user.c struct u64rec *p = &prev->cpu[i]; prev 353 samples/bpf/xdp_monitor_user.c pps = calc_pps_u64(&rec->total, &prev->total, t); prev 363 samples/bpf/xdp_monitor_user.c struct record *rec, *prev; prev 368 samples/bpf/xdp_monitor_user.c prev = &stats_prev->xdp_cpumap_enqueue[to_cpu]; prev 369 samples/bpf/xdp_monitor_user.c t = calc_period(rec, prev); prev 372 samples/bpf/xdp_monitor_user.c struct datarec *p = &prev->cpu[i]; prev 385 samples/bpf/xdp_monitor_user.c pps = calc_pps(&rec->total, &prev->total, t); prev 387 samples/bpf/xdp_monitor_user.c drop = calc_drop(&rec->total, &prev->total, t); prev 388 samples/bpf/xdp_monitor_user.c info = calc_info(&rec->total, &prev->total, t); prev 402 samples/bpf/xdp_monitor_user.c struct record *rec, *prev; prev 407 samples/bpf/xdp_monitor_user.c prev = &stats_prev->xdp_cpumap_kthread; prev 408 samples/bpf/xdp_monitor_user.c t = calc_period(rec, prev); prev 411 samples/bpf/xdp_monitor_user.c struct datarec *p = &prev->cpu[i]; prev 422 samples/bpf/xdp_monitor_user.c pps = calc_pps(&rec->total, &prev->total, t); prev 423 samples/bpf/xdp_monitor_user.c drop = calc_drop(&rec->total, &prev->total, t); prev 424 samples/bpf/xdp_monitor_user.c info = calc_info(&rec->total, &prev->total, t); prev 434 samples/bpf/xdp_monitor_user.c struct record *rec, *prev; prev 440 samples/bpf/xdp_monitor_user.c prev = &stats_prev->xdp_devmap_xmit; prev 441 samples/bpf/xdp_monitor_user.c t = calc_period(rec, prev); prev 444 samples/bpf/xdp_monitor_user.c struct datarec *p = &prev->cpu[i]; prev 460 samples/bpf/xdp_monitor_user.c pps = calc_pps(&rec->total, &prev->total, t); prev 461 samples/bpf/xdp_monitor_user.c drop = calc_drop(&rec->total, &prev->total, t); prev 462 samples/bpf/xdp_monitor_user.c info = calc_info(&rec->total, &prev->total, t); prev 463 samples/bpf/xdp_monitor_user.c err = calc_err(&rec->total, &prev->total, t); prev 587 samples/bpf/xdp_monitor_user.c struct stats_record *rec, *prev; prev 590 samples/bpf/xdp_monitor_user.c prev = alloc_stats_record(); prev 612 samples/bpf/xdp_monitor_user.c swap(&prev, &rec); prev 614 samples/bpf/xdp_monitor_user.c stats_print(rec, prev, err_only); prev 620 samples/bpf/xdp_monitor_user.c free_stats_record(prev); prev 314 samples/bpf/xdp_redirect_cpu_user.c struct record *rec, *prev; prev 331 samples/bpf/xdp_redirect_cpu_user.c prev = &stats_prev->rx_cnt; prev 332 samples/bpf/xdp_redirect_cpu_user.c t = calc_period(rec, prev); prev 335 samples/bpf/xdp_redirect_cpu_user.c struct datarec *p = &prev->cpu[i]; prev 346 samples/bpf/xdp_redirect_cpu_user.c pps = calc_pps(&rec->total, &prev->total, t); prev 347 samples/bpf/xdp_redirect_cpu_user.c drop = calc_drop_pps(&rec->total, &prev->total, t); prev 348 samples/bpf/xdp_redirect_cpu_user.c err = calc_errs_pps(&rec->total, &prev->total, t); prev 359 samples/bpf/xdp_redirect_cpu_user.c prev = &stats_prev->enq[to_cpu]; prev 360 samples/bpf/xdp_redirect_cpu_user.c t = calc_period(rec, prev); prev 363 samples/bpf/xdp_redirect_cpu_user.c struct datarec *p = &prev->cpu[i]; prev 376 samples/bpf/xdp_redirect_cpu_user.c pps = calc_pps(&rec->total, &prev->total, t); prev 378 samples/bpf/xdp_redirect_cpu_user.c drop = calc_drop_pps(&rec->total, &prev->total, t); prev 379 samples/bpf/xdp_redirect_cpu_user.c err = calc_errs_pps(&rec->total, &prev->total, t); prev 396 samples/bpf/xdp_redirect_cpu_user.c prev = &stats_prev->kthread; prev 397 samples/bpf/xdp_redirect_cpu_user.c t = calc_period(rec, prev); prev 400 samples/bpf/xdp_redirect_cpu_user.c struct datarec *p = &prev->cpu[i]; prev 411 samples/bpf/xdp_redirect_cpu_user.c pps = calc_pps(&rec->total, &prev->total, t); prev 412 samples/bpf/xdp_redirect_cpu_user.c drop = calc_drop_pps(&rec->total, &prev->total, t); prev 413 samples/bpf/xdp_redirect_cpu_user.c err = calc_errs_pps(&rec->total, &prev->total, t); prev 425 samples/bpf/xdp_redirect_cpu_user.c prev = &stats_prev->redir_err; prev 426 samples/bpf/xdp_redirect_cpu_user.c t = calc_period(rec, prev); prev 429 samples/bpf/xdp_redirect_cpu_user.c struct datarec *p = &prev->cpu[i]; prev 436 samples/bpf/xdp_redirect_cpu_user.c pps = calc_pps(&rec->total, &prev->total, t); prev 437 samples/bpf/xdp_redirect_cpu_user.c drop = calc_drop_pps(&rec->total, &prev->total, t); prev 447 samples/bpf/xdp_redirect_cpu_user.c prev = &stats_prev->exception; prev 448 samples/bpf/xdp_redirect_cpu_user.c t = calc_period(rec, prev); prev 451 samples/bpf/xdp_redirect_cpu_user.c struct datarec *p = &prev->cpu[i]; prev 458 samples/bpf/xdp_redirect_cpu_user.c pps = calc_pps(&rec->total, &prev->total, t); prev 459 samples/bpf/xdp_redirect_cpu_user.c drop = calc_drop_pps(&rec->total, &prev->total, t); prev 580 samples/bpf/xdp_redirect_cpu_user.c struct stats_record *record, *prev; prev 583 samples/bpf/xdp_redirect_cpu_user.c prev = alloc_stats_record(); prev 591 samples/bpf/xdp_redirect_cpu_user.c swap(&prev, &record); prev 593 samples/bpf/xdp_redirect_cpu_user.c stats_print(record, prev, prog_name); prev 600 samples/bpf/xdp_redirect_cpu_user.c free_stats_record(prev); prev 66 samples/bpf/xdp_redirect_map_user.c __u64 values[nr_cpus], prev[nr_cpus]; prev 68 samples/bpf/xdp_redirect_map_user.c memset(prev, 0, sizeof(prev)); prev 78 samples/bpf/xdp_redirect_map_user.c sum += (values[i] - prev[i]); prev 82 samples/bpf/xdp_redirect_map_user.c memcpy(prev, values, sizeof(values)); prev 66 samples/bpf/xdp_redirect_user.c __u64 values[nr_cpus], prev[nr_cpus]; prev 68 samples/bpf/xdp_redirect_user.c memset(prev, 0, sizeof(prev)); prev 78 samples/bpf/xdp_redirect_user.c sum += (values[i] - prev[i]); prev 82 samples/bpf/xdp_redirect_user.c memcpy(prev, values, sizeof(values)); prev 520 samples/bpf/xdp_router_ipv4_user.c __u64 prev[nr_keys][nr_cpus]; prev 565 samples/bpf/xdp_router_ipv4_user.c memset(prev, 0, sizeof(prev)); prev 577 samples/bpf/xdp_router_ipv4_user.c sum += (values[i] - prev[key][i]); prev 581 samples/bpf/xdp_router_ipv4_user.c memcpy(prev[key], values, sizeof(values)); prev 348 samples/bpf/xdp_rxq_info_user.c struct record *rec, *prev; prev 367 samples/bpf/xdp_rxq_info_user.c prev = &stats_prev->stats; prev 368 samples/bpf/xdp_rxq_info_user.c t = calc_period(rec, prev); prev 371 samples/bpf/xdp_rxq_info_user.c struct datarec *p = &prev->cpu[i]; prev 381 samples/bpf/xdp_rxq_info_user.c pps = calc_pps (&rec->total, &prev->total, t); prev 382 samples/bpf/xdp_rxq_info_user.c err = calc_errs_pps(&rec->total, &prev->total, t); prev 401 samples/bpf/xdp_rxq_info_user.c prev = &stats_prev->rxq[rxq]; prev 402 samples/bpf/xdp_rxq_info_user.c t = calc_period(rec, prev); prev 405 samples/bpf/xdp_rxq_info_user.c struct datarec *p = &prev->cpu[i]; prev 419 samples/bpf/xdp_rxq_info_user.c pps = calc_pps (&rec->total, &prev->total, t); prev 420 samples/bpf/xdp_rxq_info_user.c err = calc_errs_pps(&rec->total, &prev->total, t); prev 439 samples/bpf/xdp_rxq_info_user.c struct stats_record *record, *prev; prev 442 samples/bpf/xdp_rxq_info_user.c prev = alloc_stats_record(); prev 446 samples/bpf/xdp_rxq_info_user.c swap(&prev, &record); prev 448 samples/bpf/xdp_rxq_info_user.c stats_print(record, prev, action, cfg_opt); prev 453 samples/bpf/xdp_rxq_info_user.c free_stats_record(prev); prev 56 samples/bpf/xdp_tx_iptunnel_user.c __u64 values[nr_cpus], prev[nr_protos][nr_cpus]; prev 60 samples/bpf/xdp_tx_iptunnel_user.c memset(prev, 0, sizeof(prev)); prev 71 samples/bpf/xdp_tx_iptunnel_user.c sum += (values[i] - prev[proto][i]); prev 76 samples/bpf/xdp_tx_iptunnel_user.c memcpy(prev[proto], values, sizeof(values)); prev 183 scripts/dtc/srcpos.c srcfile->prev = current_srcfile; prev 200 scripts/dtc/srcpos.c current_srcfile = srcfile->prev; prev 18 scripts/dtc/srcpos.h struct srcfile_state *prev; prev 25 scripts/kconfig/list.h struct list_head *next, *prev; prev 83 scripts/kconfig/list.h struct list_head *prev, prev 86 scripts/kconfig/list.h next->prev = _new; prev 88 scripts/kconfig/list.h _new->prev = prev; prev 89 scripts/kconfig/list.h prev->next = _new; prev 102 scripts/kconfig/list.h __list_add(_new, head->prev, head); prev 112 scripts/kconfig/list.h static inline void __list_del(struct list_head *prev, struct list_head *next) prev 114 scripts/kconfig/list.h next->prev = prev; prev 115 scripts/kconfig/list.h prev->next = next; prev 128 scripts/kconfig/list.h __list_del(entry->prev, entry->next); prev 130 scripts/kconfig/list.h entry->prev = (struct list_head*)LIST_POISON2; prev 461 scripts/kconfig/mconf.c list_del(trail.prev); prev 751 scripts/kconfig/mconf.c list_del(trail.prev); prev 734 scripts/kconfig/menu.c jump->index = list_entry(head->prev, struct jump_key, prev 348 scripts/kconfig/preprocess.c char *tmp, *name, *res, *endptr, *prev, *p; prev 368 scripts/kconfig/preprocess.c prev = p = tmp; prev 389 scripts/kconfig/preprocess.c new_argv[new_argc++] = prev; prev 390 scripts/kconfig/preprocess.c prev = p + 1; prev 399 scripts/kconfig/preprocess.c new_argv[new_argc++] = prev; prev 1008 scripts/kconfig/symbol.c struct dep_stack *prev, *next; prev 1019 scripts/kconfig/symbol.c stack->prev = check_top; prev 1026 scripts/kconfig/symbol.c check_top = check_top->prev; prev 1049 scripts/kconfig/symbol.c for (stack = check_top; stack != NULL; stack = stack->prev) prev 793 scripts/mod/modpost.c char *prev) prev 800 scripts/mod/modpost.c if (prev) { prev 801 scripts/mod/modpost.c size -= prev - modinfo; prev 802 scripts/mod/modpost.c modinfo = next_string(prev, &size); prev 53 security/apparmor/include/policy.h #define on_list_rcu(X) (!list_empty(X) && (X)->prev != LIST_POISON2) prev 226 security/apparmor/lib.c bool prev = false; prev 231 security/apparmor/lib.c if (!prev) { prev 232 security/apparmor/lib.c prev = true; prev 108 security/selinux/ibpkey.c sel_ib_pkey_hash[idx].list.prev, prev 168 security/selinux/netnode.c rcu_dereference_protected(sel_netnode_hash[idx].list.prev, prev 117 security/selinux/netport.c sel_netport_hash[idx].list.prev, prev 70 security/selinux/ss/avtab.c struct avtab_node *prev, struct avtab_node *cur, prev 92 security/selinux/ss/avtab.c if (prev) { prev 93 security/selinux/ss/avtab.c newnode->next = prev->next; prev 94 security/selinux/ss/avtab.c prev->next = newnode; prev 109 security/selinux/ss/avtab.c struct avtab_node *prev, *cur, *newnode; prev 116 security/selinux/ss/avtab.c for (prev = NULL, cur = h->htable[hvalue]; prev 118 security/selinux/ss/avtab.c prev = cur, cur = cur->next) { prev 139 security/selinux/ss/avtab.c newnode = avtab_insert_node(h, hvalue, prev, cur, key, datum); prev 154 security/selinux/ss/avtab.c struct avtab_node *prev, *cur; prev 160 security/selinux/ss/avtab.c for (prev = NULL, cur = h->htable[hvalue]; prev 162 security/selinux/ss/avtab.c prev = cur, cur = cur->next) { prev 178 security/selinux/ss/avtab.c return avtab_insert_node(h, hvalue, prev, cur, key, datum); prev 54 security/selinux/ss/ebitmap.c struct ebitmap_node *n, *new, *prev; prev 58 security/selinux/ss/ebitmap.c prev = NULL; prev 68 security/selinux/ss/ebitmap.c if (prev) prev 69 security/selinux/ss/ebitmap.c prev->next = new; prev 72 security/selinux/ss/ebitmap.c prev = new; prev 261 security/selinux/ss/ebitmap.c struct ebitmap_node *n, *prev, *new; prev 263 security/selinux/ss/ebitmap.c prev = NULL; prev 284 security/selinux/ss/ebitmap.c if (prev) prev 285 security/selinux/ss/ebitmap.c e->highbit = prev->startbit prev 290 security/selinux/ss/ebitmap.c if (prev) prev 291 security/selinux/ss/ebitmap.c prev->next = n->next; prev 298 security/selinux/ss/ebitmap.c prev = n; prev 316 security/selinux/ss/ebitmap.c if (prev) { prev 317 security/selinux/ss/ebitmap.c new->next = prev->next; prev 318 security/selinux/ss/ebitmap.c prev->next = new; prev 45 security/selinux/ss/hashtab.c struct hashtab_node *prev, *cur, *newnode; prev 53 security/selinux/ss/hashtab.c prev = NULL; prev 56 security/selinux/ss/hashtab.c prev = cur; prev 68 security/selinux/ss/hashtab.c if (prev) { prev 69 security/selinux/ss/hashtab.c newnode->next = prev->next; prev 70 security/selinux/ss/hashtab.c prev->next = newnode; prev 38 security/selinux/ss/mls.c int i, l, len, head, prev; prev 53 security/selinux/ss/mls.c prev = -2; prev 56 security/selinux/ss/mls.c if (i - prev > 1) { prev 58 security/selinux/ss/mls.c if (head != prev) { prev 59 security/selinux/ss/mls.c nm = sym_name(p, SYM_CATS, prev); prev 66 security/selinux/ss/mls.c prev = i; prev 68 security/selinux/ss/mls.c if (prev != head) { prev 69 security/selinux/ss/mls.c nm = sym_name(p, SYM_CATS, prev); prev 94 security/selinux/ss/mls.c int i, l, head, prev; prev 113 security/selinux/ss/mls.c prev = -2; prev 116 security/selinux/ss/mls.c if (i - prev > 1) { prev 118 security/selinux/ss/mls.c if (prev != head) { prev 119 security/selinux/ss/mls.c if (prev - head > 1) prev 123 security/selinux/ss/mls.c nm = sym_name(p, SYM_CATS, prev); prev 127 security/selinux/ss/mls.c if (prev < 0) prev 136 security/selinux/ss/mls.c prev = i; prev 139 security/selinux/ss/mls.c if (prev != head) { prev 140 security/selinux/ss/mls.c if (prev - head > 1) prev 144 security/selinux/ss/mls.c nm = sym_name(p, SYM_CATS, prev); prev 1914 security/smack/smackfs.c last = public->prev; prev 1917 security/smack/smackfs.c private->prev->next = public; prev 1918 security/smack/smackfs.c private->next->prev = public; prev 1920 security/smack/smackfs.c public->prev = private->prev; prev 1927 security/smack/smackfs.c private->prev = last; prev 1928 security/smack/smackfs.c first->prev = private; prev 474 security/tomoyo/gc.c list_add_rcu(element, element->prev); prev 558 sound/core/oss/pcm_oss.c plugin->prev = NULL; prev 560 sound/core/oss/pcm_oss.c runtime->oss.plugin_first->prev = plugin; prev 573 sound/core/oss/pcm_oss.c plugin->prev = runtime->oss.plugin_last; prev 776 sound/core/oss/pcm_oss.c unsigned int rate, prev; prev 800 sound/core/oss/pcm_oss.c prev = rate; prev 802 sound/core/oss/pcm_oss.c if (rate <= prev) prev 123 sound/core/oss/pcm_plugin.c while (plugin->prev) { prev 128 sound/core/oss/pcm_plugin.c plugin = plugin->prev; prev 214 sound/core/oss/pcm_plugin.c plugin_prev = plugin->prev; prev 269 sound/core/oss/pcm_plugin.c plugin_prev = plugin->prev; prev 59 sound/core/oss/pcm_plugin.h struct snd_pcm_plugin *prev; prev 75 sound/core/pcm.c list_add(&newpcm->list, pcm->list.prev); prev 640 sound/core/pcm.c struct snd_pcm_substream *substream, *prev; prev 664 sound/core/pcm.c prev = NULL; prev 665 sound/core/pcm.c for (idx = 0, prev = NULL; idx < substream_count; idx++) { prev 675 sound/core/pcm.c if (prev == NULL) prev 678 sound/core/pcm.c prev->next = substream; prev 685 sound/core/pcm.c if (prev == NULL) prev 688 sound/core/pcm.c prev->next = NULL; prev 697 sound/core/pcm.c prev = substream; prev 30 sound/core/pcm_lib.c #define trace_applptr(substream, prev, curr) prev 34 sound/core/pcm_native.c #define trace_hw_mask_param(substream, type, index, prev, curr) prev 35 sound/core/pcm_native.c #define trace_hw_interval_param(substream, type, index, prev, curr) prev 29 sound/core/pcm_param_trace.h TP_PROTO(struct snd_pcm_substream *substream, snd_pcm_hw_param_t type, int index, const struct snd_mask *prev, const struct snd_mask *curr), prev 30 sound/core/pcm_param_trace.h TP_ARGS(substream, type, index, prev, curr), prev 50 sound/core/pcm_param_trace.h memcpy(__entry->prev_bits, prev->bits, sizeof(__u32) * 8); prev 69 sound/core/pcm_param_trace.h TP_PROTO(struct snd_pcm_substream *substream, snd_pcm_hw_param_t type, int index, const struct snd_interval *prev, const struct snd_interval *curr), prev 70 sound/core/pcm_param_trace.h TP_ARGS(substream, type, index, prev, curr), prev 100 sound/core/pcm_param_trace.h __entry->prev_min = prev->min; prev 101 sound/core/pcm_param_trace.h __entry->prev_max = prev->max; prev 102 sound/core/pcm_param_trace.h __entry->prev_openmin = prev->openmin; prev 103 sound/core/pcm_param_trace.h __entry->prev_openmax = prev->openmax; prev 104 sound/core/pcm_param_trace.h __entry->prev_integer = prev->integer; prev 105 sound/core/pcm_param_trace.h __entry->prev_empty = prev->empty; prev 107 sound/core/pcm_trace.h TP_PROTO(struct snd_pcm_substream *substream, snd_pcm_uframes_t prev, snd_pcm_uframes_t curr), prev 108 sound/core/pcm_trace.h TP_ARGS(substream, prev, curr), prev 114 sound/core/pcm_trace.h __field( snd_pcm_uframes_t, prev ) prev 125 sound/core/pcm_trace.h __entry->prev = (prev); prev 136 sound/core/pcm_trace.h __entry->prev, prev 134 sound/core/seq/seq_prioq.c struct snd_seq_event_cell *cur, *prev; prev 164 sound/core/seq/seq_prioq.c prev = NULL; /* previous cell */ prev 179 sound/core/seq/seq_prioq.c prev = cur; prev 189 sound/core/seq/seq_prioq.c if (prev != NULL) prev 190 sound/core/seq/seq_prioq.c prev->next = cell; prev 279 sound/core/seq/seq_prioq.c struct snd_seq_event_cell *prev = NULL; prev 292 sound/core/seq/seq_prioq.c prev->next = cell->next; prev 314 sound/core/seq/seq_prioq.c prev = cell; prev 389 sound/core/seq/seq_prioq.c struct snd_seq_event_cell *prev = NULL; prev 405 sound/core/seq/seq_prioq.c prev->next = cell->next; prev 422 sound/core/seq/seq_prioq.c prev = cell; prev 1298 sound/core/timer.c int prev; prev 1302 sound/core/timer.c prev = tu->qtail == 0 ? tu->queue_size - 1 : tu->qtail - 1; prev 1303 sound/core/timer.c r = &tu->queue[prev]; prev 1376 sound/core/timer.c int prev, append = 0; prev 1406 sound/core/timer.c prev = tu->qtail == 0 ? tu->queue_size - 1 : tu->qtail - 1; prev 1407 sound/core/timer.c r = &tu->tqueue[prev]; prev 39 sound/isa/gus/gus_mem.c nblock->prev = pblock->prev; prev 41 sound/isa/gus/gus_mem.c pblock->prev = nblock; prev 45 sound/isa/gus/gus_mem.c nblock->prev->next = nblock; prev 53 sound/isa/gus/gus_mem.c nblock->prev = NULL; prev 56 sound/isa/gus/gus_mem.c nblock->prev = alloc->last; prev 73 sound/isa/gus/gus_mem.c block->next->prev = NULL; prev 75 sound/isa/gus/gus_mem.c block->prev->next = block->next; prev 77 sound/isa/gus/gus_mem.c block->next->prev = block->prev; prev 80 sound/isa/gus/gus_mem.c alloc->last = block->prev; prev 81 sound/isa/gus/gus_mem.c if (block->prev) prev 82 sound/isa/gus/gus_mem.c block->prev->next = NULL; prev 84 sound/isa/gus/gus_mem.c block->next->prev = block->prev; prev 85 sound/isa/gus/gus_mem.c if (block->prev) prev 86 sound/isa/gus/gus_mem.c block->prev->next = block->next; prev 139 sound/isa/gus/gus_mem.c block->prev = block->next = NULL; prev 261 sound/pci/au88x0/au88x0_core.c int temp, lifeboat = 0, prev; prev 269 sound/pci/au88x0/au88x0_core.c prev = VORTEX_MIXER_CHNBASE + (ch << 2); prev 270 sound/pci/au88x0/au88x0_core.c temp = hwread(vortex->mmio, prev); prev 272 sound/pci/au88x0/au88x0_core.c prev = VORTEX_MIXER_RTBASE + ((temp & 0xf) << 2); prev 273 sound/pci/au88x0/au88x0_core.c temp = hwread(vortex->mmio, prev); prev 282 sound/pci/au88x0/au88x0_core.c hwwrite(vortex->mmio, prev, (temp & 0xf) | 0x10); prev 657 sound/pci/au88x0/au88x0_core.c int temp, lifeboat = 0, prev; prev 666 sound/pci/au88x0/au88x0_core.c prev = VORTEX_SRC_CHNBASE + (ch << 2); /*ebp */ prev 667 sound/pci/au88x0/au88x0_core.c temp = hwread(vortex->mmio, prev); prev 670 sound/pci/au88x0/au88x0_core.c prev = VORTEX_SRC_RTBASE + ((temp & 0xf) << 2); /*esp12 */ prev 672 sound/pci/au88x0/au88x0_core.c temp = hwread(vortex->mmio, prev); prev 682 sound/pci/au88x0/au88x0_core.c hwwrite(vortex->mmio, prev, (temp & 0xf) | 0x10); prev 1632 sound/pci/au88x0/au88x0_core.c int temp, prev, lifeboat = 0; prev 1660 sound/pci/au88x0/au88x0_core.c prev = temp; prev 1672 sound/pci/au88x0/au88x0_core.c hwwrite(vortex->mmio, VORTEX_ADB_RTBASE + (prev << 2), route[0]); prev 1679 sound/pci/au88x0/au88x0_core.c int temp, lifeboat = 0, prev; prev 1696 sound/pci/au88x0/au88x0_core.c prev = temp; prev 1699 sound/pci/au88x0/au88x0_core.c VORTEX_ADB_RTBASE + (prev << 2)) & ADB_MASK; prev 1712 sound/pci/au88x0/au88x0_core.c hwwrite(vortex->mmio, VORTEX_ADB_RTBASE + (prev << 2), temp); prev 325 sound/pci/azt3328.c u8 prev = inb(reg), new; prev 327 sound/pci/azt3328.c new = (do_set) ? (prev|mask) : (prev & ~mask); prev 331 sound/pci/azt3328.c if (new != prev) prev 42 sound/pci/ctxfi/ctimap.c pre = pos->prev; prev 44 sound/pci/ctxfi/ctimap.c pre = head->prev; prev 46 sound/pci/ctxfi/ctimap.c __list_add(&entry->list, pos->prev, pos); prev 48 sound/pci/ctxfi/ctimap.c pre = head->prev; prev 75 sound/pci/ctxfi/ctimap.c pre = (entry->list.prev == head) ? head->prev : entry->list.prev; prev 100 sound/pci/ctxfi/ctvmem.c __list_add(&block->list, pos->prev, pos); prev 106 sound/pci/ctxfi/ctvmem.c pre = pos->prev; prev 117 sound/pci/ctxfi/ctvmem.c pre = pos->prev; prev 172 sound/pci/emu10k1/memory.c if ((p = blk->mapped_link.prev) != &emu->mapped_link_head) { prev 221 sound/pci/emu10k1/memory.c blk = (struct snd_emu10k1_memblk *)__snd_util_memblk_new(emu->memhdr, psize << PAGE_SHIFT, p->prev); prev 457 sound/pci/emu10k1/memory.c if ((p = blk->mem.list.prev) != &hdr->block) { prev 1385 sound/pci/es1968.c if (buf->list.prev != &chip->buf_list) { prev 1386 sound/pci/es1968.c chunk = list_entry(buf->list.prev, struct esm_memory, list); prev 944 sound/pci/hda/hda_intel.c int prev = power_save; prev 947 sound/pci/hda/hda_intel.c if (ret || prev == power_save) prev 144 sound/pci/trident/trident_memory.c blk = __snd_util_memblk_new(hdr, psize * ALIGN_PAGE_SIZE, p->prev); prev 447 sound/soc/intel/atom/sst/sst.c stream->resume_prev = stream->prev; prev 562 sound/soc/intel/atom/sst/sst.c stream->prev = stream->resume_prev; prev 185 sound/soc/intel/atom/sst/sst.h unsigned int prev; prev 320 sound/soc/intel/atom/sst/sst_drv_interface.c str_info->prev = str_info->status; prev 553 sound/soc/intel/atom/sst/sst_drv_interface.c str_info->prev = str_info->status; prev 571 sound/soc/intel/atom/sst/sst_drv_interface.c str_info->prev = STREAM_UN_INIT; prev 624 sound/soc/intel/atom/sst/sst_drv_interface.c stream->prev = stream->status; prev 215 sound/soc/intel/atom/sst/sst_pvt.c stream->prev = STREAM_UN_INIT; prev 73 sound/soc/intel/atom/sst/sst_stream.c sst_drv_ctx->streams[str_id].prev = STREAM_UN_INIT; prev 255 sound/soc/intel/atom/sst/sst_stream.c if (str_info->prev == STREAM_UN_INIT) prev 263 sound/soc/intel/atom/sst/sst_stream.c str_info->prev = str_info->status; prev 306 sound/soc/intel/atom/sst/sst_stream.c str_info->prev = STREAM_PAUSED; prev 317 sound/soc/intel/atom/sst/sst_stream.c str_info->prev = STREAM_PAUSED; prev 326 sound/soc/intel/atom/sst/sst_stream.c if (str_info->prev == STREAM_RUNNING) prev 330 sound/soc/intel/atom/sst/sst_stream.c str_info->prev = STREAM_PAUSED; prev 364 sound/soc/intel/atom/sst/sst_stream.c str_info->prev = STREAM_UN_INIT; prev 442 sound/soc/intel/atom/sst/sst_stream.c str_info->prev = str_info->status; prev 480 sound/synth/emux/soundfont.c struct snd_sf_zone *prev, *next, *p; prev 483 sound/synth/emux/soundfont.c prev = NULL; prev 489 sound/synth/emux/soundfont.c if (prev) prev 490 sound/synth/emux/soundfont.c prev->next = next; prev 496 sound/synth/emux/soundfont.c prev = p; prev 88 sound/synth/util_mem.c return __snd_util_memblk_new(hdr, units, p->prev); prev 98 sound/synth/util_mem.c struct list_head *prev) prev 107 sound/synth/util_mem.c if (prev == &hdr->block) prev 110 sound/synth/util_mem.c struct snd_util_memblk *p = get_memblk(prev); prev 114 sound/synth/util_mem.c list_add(&blk->list, prev); prev 136 sound/usb/stream.c struct audioformat *prev = fp; prev 138 sound/usb/stream.c list_for_each_entry_continue_reverse(prev, &subs->fmt_list, list) { prev 139 sound/usb/stream.c if (prev->chmap && prev 140 sound/usb/stream.c !memcmp(prev->chmap, fp->chmap, sizeof(*fp->chmap))) prev 3 tools/firewire/list.h struct list *next, *prev; prev 10 tools/firewire/list.h list->prev = list; prev 22 tools/firewire/list.h new_link->prev = link->prev; prev 24 tools/firewire/list.h new_link->prev->next = new_link; prev 25 tools/firewire/list.h new_link->next->prev = new_link; prev 43 tools/firewire/list.h link->prev->next = link->next; prev 44 tools/firewire/list.h link->next->prev = link->prev; prev 54 tools/firewire/list.h list_entry((list)->prev, type, member) prev 485 tools/firewire/nosy-dump.c struct subaction *sa, *prev; prev 494 tools/firewire/nosy-dump.c prev = list_tail(&t->request_list, prev 497 tools/firewire/nosy-dump.c if (!ACK_BUSY(prev->ack)) { prev 506 tools/firewire/nosy-dump.c if (prev->packet.common.tcode != sa->packet.common.tcode || prev 507 tools/firewire/nosy-dump.c prev->packet.common.tlabel != sa->packet.common.tlabel) { prev 550 tools/firewire/nosy-dump.c struct subaction *sa, *prev; prev 563 tools/firewire/nosy-dump.c prev = list_tail(&t->response_list, struct subaction, link); prev 565 tools/firewire/nosy-dump.c if (!ACK_BUSY(prev->ack)) { prev 572 tools/firewire/nosy-dump.c if (prev->packet.common.tcode != sa->packet.common.tcode || prev 573 tools/firewire/nosy-dump.c prev->packet.common.tlabel != sa->packet.common.tlabel) { prev 578 tools/firewire/nosy-dump.c prev = list_tail(&t->request_list, struct subaction, link); prev 579 tools/firewire/nosy-dump.c if (prev->ack != ACK_PENDING) { prev 586 tools/firewire/nosy-dump.c if (packet_info[prev->packet.common.tcode].response_tcode != prev 28 tools/include/linux/list.h list->prev = list; prev 39 tools/include/linux/list.h struct list_head *prev, prev 42 tools/include/linux/list.h next->prev = new; prev 44 tools/include/linux/list.h new->prev = prev; prev 45 tools/include/linux/list.h prev->next = new; prev 49 tools/include/linux/list.h struct list_head *prev, prev 77 tools/include/linux/list.h __list_add(new, head->prev, head); prev 87 tools/include/linux/list.h static inline void __list_del(struct list_head * prev, struct list_head * next) prev 89 tools/include/linux/list.h next->prev = prev; prev 90 tools/include/linux/list.h WRITE_ONCE(prev->next, next); prev 102 tools/include/linux/list.h __list_del(entry->prev, entry->next); prev 107 tools/include/linux/list.h __list_del(entry->prev, entry->next); prev 109 tools/include/linux/list.h entry->prev = LIST_POISON2; prev 127 tools/include/linux/list.h new->next->prev = new; prev 128 tools/include/linux/list.h new->prev = old->prev; prev 129 tools/include/linux/list.h new->prev->next = new; prev 208 tools/include/linux/list.h return (next == head) && (next == head->prev); prev 231 tools/include/linux/list.h return !list_empty(head) && (head->next == head->prev); prev 239 tools/include/linux/list.h list->next->prev = list; prev 240 tools/include/linux/list.h list->prev = entry; prev 243 tools/include/linux/list.h new_first->prev = head; prev 275 tools/include/linux/list.h struct list_head *prev, prev 279 tools/include/linux/list.h struct list_head *last = list->prev; prev 281 tools/include/linux/list.h first->prev = prev; prev 282 tools/include/linux/list.h prev->next = first; prev 285 tools/include/linux/list.h next->prev = last; prev 309 tools/include/linux/list.h __list_splice(list, head->prev, head); prev 340 tools/include/linux/list.h __list_splice(list, head->prev, head); prev 374 tools/include/linux/list.h list_entry((ptr)->prev, type, member) prev 401 tools/include/linux/list.h list_entry((pos)->member.prev, typeof(*(pos)), member) prev 417 tools/include/linux/list.h for (pos = (head)->prev; pos != (head); pos = pos->prev) prev 436 tools/include/linux/list.h for (pos = (head)->prev, n = pos->prev; \ prev 438 tools/include/linux/list.h pos = n, n = pos->prev) prev 660 tools/include/linux/list.h struct hlist_node *prev) prev 662 tools/include/linux/list.h n->next = prev->next; prev 663 tools/include/linux/list.h prev->next = n; prev 664 tools/include/linux/list.h n->pprev = &prev->next; prev 760 tools/include/linux/list.h begin->prev->next = end->next; prev 761 tools/include/linux/list.h end->next->prev = begin->prev; prev 70 tools/include/linux/types.h struct list_head *next, *prev; prev 117 tools/include/uapi/drm/drm.h unsigned char prev; prev 4227 tools/lib/bpf/libbpf.c bpf_object__next(struct bpf_object *prev) prev 4231 tools/lib/bpf/libbpf.c if (!prev) prev 4236 tools/lib/bpf/libbpf.c next = list_next_entry(prev, list); prev 4308 tools/lib/bpf/libbpf.c bpf_program__next(struct bpf_program *prev, const struct bpf_object *obj) prev 4310 tools/lib/bpf/libbpf.c struct bpf_program *prog = prev; prev 4750 tools/lib/bpf/libbpf.c bpf_map__next(const struct bpf_map *prev, const struct bpf_object *obj) prev 4752 tools/lib/bpf/libbpf.c if (prev == NULL) prev 4755 tools/lib/bpf/libbpf.c return __bpf_map__iter(prev, obj, 1); prev 113 tools/lib/bpf/libbpf.h LIBBPF_API struct bpf_object *bpf_object__next(struct bpf_object *prev); prev 51 tools/lib/traceevent/event-parse.h struct tep_record *prev; prev 132 tools/perf/builtin-annotate.c struct addr_map_symbol *prev = NULL; prev 147 tools/perf/builtin-annotate.c if (prev) prev 148 tools/perf/builtin-annotate.c process_basic_block(prev, &bi[i].from, &bi[i].flags); prev 149 tools/perf/builtin-annotate.c prev = &bi[i].to; prev 188 tools/perf/builtin-kvm.c void *prev; prev 196 tools/perf/builtin-kvm.c prev = event->vcpu; prev 200 tools/perf/builtin-kvm.c free(prev); prev 838 tools/perf/builtin-sched.c struct task_desc *prev, __maybe_unused *next; prev 863 tools/perf/builtin-sched.c prev = register_pid(sched, prev_pid, prev_comm); prev 868 tools/perf/builtin-sched.c add_sched_event_run(sched, prev, timestamp, delta); prev 869 tools/perf/builtin-sched.c add_sched_event_sleep(sched, prev, timestamp, prev_state); prev 1064 tools/perf/builtin-sched.c atom = list_entry(atoms->work_list.prev, struct work_atom, list); prev 1079 tools/perf/builtin-sched.c atom = list_entry(atoms->work_list.prev, struct work_atom, list); prev 1236 tools/perf/builtin-sched.c atom = list_entry(atoms->work_list.prev, struct work_atom, list); prev 1304 tools/perf/builtin-sched.c atom = list_entry(atoms->work_list.prev, struct work_atom, list); prev 724 tools/perf/builtin-timechart.c struct io_sample *prev; prev 735 tools/perf/builtin-timechart.c prev = c->io_samples; prev 737 tools/perf/builtin-timechart.c if (prev && prev->start_time && !prev->end_time) { prev 743 tools/perf/builtin-timechart.c c->io_samples = prev->next; prev 744 tools/perf/builtin-timechart.c free(prev); prev 768 tools/perf/builtin-timechart.c struct io_sample *sample, *prev; prev 792 tools/perf/builtin-timechart.c prev = sample->next; prev 798 tools/perf/builtin-timechart.c if (prev && sample->start_time < prev->end_time) { prev 799 tools/perf/builtin-timechart.c if (prev->err) /* try to make errors more visible */ prev 800 tools/perf/builtin-timechart.c sample->start_time = prev->end_time; prev 802 tools/perf/builtin-timechart.c prev->end_time = sample->start_time; prev 819 tools/perf/builtin-timechart.c if (prev && prev 820 tools/perf/builtin-timechart.c prev->type == sample->type && prev 821 tools/perf/builtin-timechart.c prev->err == sample->err && prev 822 tools/perf/builtin-timechart.c prev->fd == sample->fd && prev 823 tools/perf/builtin-timechart.c prev->end_time + tchart->merge_dist >= sample->start_time) { prev 825 tools/perf/builtin-timechart.c sample->bytes += prev->bytes; prev 826 tools/perf/builtin-timechart.c sample->merges += prev->merges + 1; prev 828 tools/perf/builtin-timechart.c sample->start_time = prev->start_time; prev 829 tools/perf/builtin-timechart.c sample->next = prev->next; prev 830 tools/perf/builtin-timechart.c free(prev); prev 966 tools/perf/builtin-timechart.c struct per_pid *new_list, *p, *cursor, *prev; prev 981 tools/perf/builtin-timechart.c prev = NULL; prev 987 tools/perf/builtin-timechart.c if (prev) { prev 988 tools/perf/builtin-timechart.c p->next = prev->next; prev 989 tools/perf/builtin-timechart.c prev->next = p; prev 1000 tools/perf/builtin-timechart.c prev = cursor; prev 1003 tools/perf/builtin-timechart.c prev->next = p; prev 91 tools/perf/lib/cpumap.c int n, cpu, prev; prev 95 tools/perf/lib/cpumap.c prev = -1; prev 100 tools/perf/lib/cpumap.c if (prev >= 0) { prev 101 tools/perf/lib/cpumap.c int new_max = nr_cpus + cpu - prev - 1; prev 114 tools/perf/lib/cpumap.c while (++prev < cpu) prev 115 tools/perf/lib/cpumap.c tmp_cpus[nr_cpus++] = prev; prev 127 tools/perf/lib/cpumap.c prev = cpu; prev 129 tools/perf/lib/cpumap.c prev = -1; prev 85 tools/perf/lib/evlist.c perf_evlist__next(struct perf_evlist *evlist, struct perf_evsel *prev) prev 89 tools/perf/lib/evlist.c if (!prev) { prev 94 tools/perf/lib/evlist.c next = list_next_entry(prev, node); prev 70 tools/perf/lib/include/internal/evlist.h return list_entry(evlist->entries.prev, struct perf_evsel, node); prev 24 tools/perf/lib/include/internal/mmap.h u64 prev; prev 97 tools/perf/ui/browser.c pos = pos->prev; prev 119 tools/perf/ui/browser.c pos = ui_browser__list_head_filter_prev_entries(browser, head->prev); prev 132 tools/perf/ui/browser.c pos = ui_browser__list_head_filter_prev_entries(browser, pos->prev); prev 276 tools/perf/ui/browsers/annotate.c pos = list_entry(pos->node.prev, struct annotation_line, node); prev 357 tools/perf/ui/browsers/hists.c chain = list_entry(node->val.prev, struct callchain_list, list); prev 3328 tools/perf/ui/browsers/hists.c if (pos->core.node.prev == &evlist->core.entries) prev 537 tools/perf/util/annotate.c char *s = strchr(ops->raw, ','), *target, *comment, prev; prev 560 tools/perf/util/annotate.c prev = *s; prev 564 tools/perf/util/annotate.c *s = prev; prev 598 tools/perf/util/annotate.c char *target, *comment, *s, prev; prev 604 tools/perf/util/annotate.c prev = *s; prev 608 tools/perf/util/annotate.c *s = prev; prev 1583 tools/perf/util/annotate.c dl = list_entry(list->prev, struct disasm_line, al.node); prev 2934 tools/perf/util/annotate.c int prev; prev 2938 tools/perf/util/annotate.c prev = obj__set_jumps_percent_color(obj, al->jump_sources, prev 2941 tools/perf/util/annotate.c obj__set_color(obj, prev); prev 77 tools/perf/util/auxtrace.c mm->prev = 0; prev 644 tools/perf/util/auxtrace.c auxtrace_index = list_entry(head->prev, struct auxtrace_index, list); prev 650 tools/perf/util/auxtrace.c auxtrace_index = list_entry(head->prev, struct auxtrace_index, prev 1237 tools/perf/util/auxtrace.c u64 head, old = mm->prev, offset, ref; prev 1326 tools/perf/util/auxtrace.c mm->prev = head; prev 276 tools/perf/util/auxtrace.h u64 prev; prev 804 tools/perf/util/callchain.c old_tail = parent->val.prev; prev 807 tools/perf/util/callchain.c new->val.prev = old_tail; prev 808 tools/perf/util/callchain.c to_split->list.prev = &new->val; prev 194 tools/perf/util/evlist.c evsel = list_entry(list->prev, struct evsel, core.node); prev 314 tools/perf/util/evsel.h return list_entry(evsel->core.node.prev, struct evsel, core.node); prev 2582 tools/perf/util/hist.c struct addr_map_symbol *prev = NULL; prev 2596 tools/perf/util/hist.c nonany_branch_mode ? NULL : prev, prev 2598 tools/perf/util/hist.c prev = &bi[i].to; prev 254 tools/perf/util/intel-bts.c if (b->list.prev == &queue->head) prev 256 tools/perf/util/intel-bts.c a = list_entry(b->list.prev, struct auxtrace_buffer, list); prev 37 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c struct intel_pt_blk *prev; prev 285 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c stack->blk = blk->prev; prev 318 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c blk->prev = stack->blk; prev 83 tools/perf/util/mem2node.c struct phys_entry *prev = &entries[j - 1]; prev 85 tools/perf/util/mem2node.c if ((prev->end == start) && prev 86 tools/perf/util/mem2node.c (prev->node == n->node)) { prev 87 tools/perf/util/mem2node.c prev->end += bsize; prev 103 tools/perf/util/mmap.c map->core.prev = map->core.start; prev 110 tools/perf/util/mmap.c return perf_mmap__read_head(map) == map->core.prev && !map->auxtrace_mmap.base; prev 129 tools/perf/util/mmap.c u64 old = map->core.prev; prev 372 tools/perf/util/mmap.c map->core.prev = 0; prev 447 tools/perf/util/mmap.c u64 old = md->core.prev; prev 462 tools/perf/util/mmap.c md->core.prev = head; prev 524 tools/perf/util/mmap.c md->core.prev = head; prev 544 tools/perf/util/mmap.c map->core.prev = perf_mmap__read_head(map); prev 53 tools/perf/util/ordered-events.c p = last->list.prev; prev 258 tools/perf/util/ordered-events.c oe->last = list_entry(head->prev, struct ordered_event, list); prev 2075 tools/perf/util/parse-events.c if (last->core.node.prev == &evlist->core.entries) prev 2077 tools/perf/util/parse-events.c last = list_entry(last->core.node.prev, struct evsel, core.node); prev 117 tools/perf/util/srccode.c h = list_entry(srcfile_list.prev, struct srcfile, nd); prev 222 tools/perf/util/symbol.c struct symbol *curr, *prev; prev 230 tools/perf/util/symbol.c prev = curr; prev 233 tools/perf/util/symbol.c if (prev->end == prev->start && prev->end != curr->start) prev 234 tools/perf/util/symbol.c arch__symbols__fixup_end(prev, curr); prev 33 tools/testing/selftests/bpf/progs/btf_dump_test_case_ordering.c struct list_head *prev; prev 644 tools/testing/selftests/kselftest_harness.h struct __test_metadata *prev, *next; prev 672 tools/testing/selftests/kselftest_harness.h t->prev = t; prev 677 tools/testing/selftests/kselftest_harness.h t->prev = __test_list->prev; prev 678 tools/testing/selftests/kselftest_harness.h t->prev->next = t; prev 679 tools/testing/selftests/kselftest_harness.h __test_list->prev = t; prev 682 tools/testing/selftests/kselftest_harness.h t->next->prev = t; prev 683 tools/testing/selftests/kselftest_harness.h t->prev = t; prev 48 tools/testing/selftests/kvm/include/sparsebit.h sparsebit_idx_t sparsebit_next_set(struct sparsebit *sbit, sparsebit_idx_t prev); prev 49 tools/testing/selftests/kvm/include/sparsebit.h sparsebit_idx_t sparsebit_next_clear(struct sparsebit *sbit, sparsebit_idx_t prev); prev 419 tools/testing/selftests/kvm/lib/kvm_util.c vcpu->next->prev = vcpu->prev; prev 420 tools/testing/selftests/kvm/lib/kvm_util.c if (vcpu->prev) prev 421 tools/testing/selftests/kvm/lib/kvm_util.c vcpu->prev->next = vcpu->next; prev 688 tools/testing/selftests/kvm/lib/kvm_util.c vm->userspace_mem_region_head->prev = region; prev 835 tools/testing/selftests/kvm/lib/kvm_util.c vm->vcpu_head->prev = vcpu; prev 27 tools/testing/selftests/kvm/lib/kvm_util_internal.h struct userspace_mem_region *next, *prev; prev 38 tools/testing/selftests/kvm/lib/kvm_util_internal.h struct vcpu *next, *prev; prev 335 tools/testing/selftests/kvm/lib/sparsebit.c struct node *nodep, *parentp, *prev; prev 381 tools/testing/selftests/kvm/lib/sparsebit.c prev = node_prev(s, nodep); prev 382 tools/testing/selftests/kvm/lib/sparsebit.c while (prev && prev->idx + MASK_BITS + prev->num_after - 1 >= nodep->idx) { prev 383 tools/testing/selftests/kvm/lib/sparsebit.c unsigned int n1 = (prev->idx + MASK_BITS + prev->num_after - 1) prev 385 tools/testing/selftests/kvm/lib/sparsebit.c assert(prev->num_after > 0); prev 389 tools/testing/selftests/kvm/lib/sparsebit.c prev->num_after--; prev 605 tools/testing/selftests/kvm/lib/sparsebit.c struct node *prev, *next, *tmp; prev 670 tools/testing/selftests/kvm/lib/sparsebit.c prev = node_prev(s, nodep); prev 671 tools/testing/selftests/kvm/lib/sparsebit.c if (prev) { prev 675 tools/testing/selftests/kvm/lib/sparsebit.c if (prev->mask == 0 && prev->num_after == 0) { prev 676 tools/testing/selftests/kvm/lib/sparsebit.c node_rm(s, prev); prev 687 tools/testing/selftests/kvm/lib/sparsebit.c prev->idx + MASK_BITS == nodep->idx) { prev 688 tools/testing/selftests/kvm/lib/sparsebit.c prev->num_after += MASK_BITS + nodep->num_after; prev 701 tools/testing/selftests/kvm/lib/sparsebit.c prev_highest_bit = prev->idx + MASK_BITS - 1 + prev->num_after; prev 716 tools/testing/selftests/kvm/lib/sparsebit.c prev->num_after += num_contiguous; prev 733 tools/testing/selftests/kvm/lib/sparsebit.c prev->num_after += nodep->num_after; prev 1156 tools/testing/selftests/kvm/lib/sparsebit.c sparsebit_idx_t prev) prev 1158 tools/testing/selftests/kvm/lib/sparsebit.c sparsebit_idx_t lowest_possible = prev + 1; prev 1249 tools/testing/selftests/kvm/lib/sparsebit.c sparsebit_idx_t prev) prev 1251 tools/testing/selftests/kvm/lib/sparsebit.c sparsebit_idx_t lowest_possible = prev + 1; prev 1688 tools/testing/selftests/kvm/lib/sparsebit.c struct node *nodep, *prev = NULL; prev 1694 tools/testing/selftests/kvm/lib/sparsebit.c prev = nodep, nodep = node_next(s, nodep)) { prev 1803 tools/testing/selftests/kvm/lib/sparsebit.c if (prev) { prev 1808 tools/testing/selftests/kvm/lib/sparsebit.c if (prev->idx >= nodep->idx) { prev 1813 tools/testing/selftests/kvm/lib/sparsebit.c prev, prev->idx, nodep, nodep->idx); prev 1822 tools/testing/selftests/kvm/lib/sparsebit.c if ((prev->idx + MASK_BITS + prev->num_after - 1) prev 1831 tools/testing/selftests/kvm/lib/sparsebit.c prev, prev->idx, prev->num_after, prev 1844 tools/testing/selftests/kvm/lib/sparsebit.c prev->idx + MASK_BITS + prev->num_after == nodep->idx) { prev 1853 tools/testing/selftests/kvm/lib/sparsebit.c prev, prev->idx, prev->num_after, prev 109 tools/testing/selftests/rcutorture/formal/srcu-cbmc/include/linux/types.h struct list_head *next, *prev; prev 20 tools/usb/usbip/libsrc/list.h struct list_head *next, *prev; prev 31 tools/usb/usbip/libsrc/list.h list->prev = list; prev 41 tools/usb/usbip/libsrc/list.h struct list_head *prev, prev 44 tools/usb/usbip/libsrc/list.h next->prev = new; prev 46 tools/usb/usbip/libsrc/list.h new->prev = prev; prev 47 tools/usb/usbip/libsrc/list.h prev->next = new; prev 70 tools/usb/usbip/libsrc/list.h static inline void __list_del(struct list_head * prev, struct list_head * next) prev 72 tools/usb/usbip/libsrc/list.h next->prev = prev; prev 73 tools/usb/usbip/libsrc/list.h prev->next = next; prev 88 tools/usb/usbip/libsrc/list.h __list_del(entry->prev, entry->next); prev 93 tools/usb/usbip/libsrc/list.h __list_del(entry->prev, entry->next); prev 95 tools/usb/usbip/libsrc/list.h entry->prev = LIST_POISON2; prev 21 tools/virtio/ringtest/ring.c unsigned short prev) prev 23 tools/virtio/ringtest/ring.c return (unsigned short)(next - event - 1) < (unsigned short)(next - prev);