new 202 arch/alpha/include/asm/atomic.h #define atomic64_cmpxchg(v, old, new) (cmpxchg(&((v)->counter), old, new)) new 203 arch/alpha/include/asm/atomic.h #define atomic64_xchg(v, new) (xchg(&((v)->counter), new)) new 205 arch/alpha/include/asm/atomic.h #define atomic_cmpxchg(v, old, new) (cmpxchg(&((v)->counter), old, new)) new 206 arch/alpha/include/asm/atomic.h #define atomic_xchg(v, new) (xchg(&((v)->counter), new)) new 219 arch/alpha/include/asm/atomic.h int c, new, old; new 232 arch/alpha/include/asm/atomic.h : [old] "=&r"(old), [new] "=&r"(new), [c] "=&r"(c) new 251 arch/alpha/include/asm/atomic.h s64 c, new, old; new 264 arch/alpha/include/asm/atomic.h : [old] "=&r"(old), [new] "=&r"(new), [c] "=&r"(c) new 129 arch/alpha/include/asm/xchg.h ____cmpxchg(_u8, volatile char *m, unsigned char old, unsigned char new) new 148 arch/alpha/include/asm/xchg.h : "=&r" (prev), "=&r" (new), "=&r" (tmp), "=&r" (cmp), "=&r" (addr64) new 149 arch/alpha/include/asm/xchg.h : "r" ((long)m), "Ir" (old), "1" (new) : "memory"); new 155 arch/alpha/include/asm/xchg.h ____cmpxchg(_u16, volatile short *m, unsigned short old, unsigned short new) new 174 arch/alpha/include/asm/xchg.h : "=&r" (prev), "=&r" (new), "=&r" (tmp), "=&r" (cmp), "=&r" (addr64) new 175 arch/alpha/include/asm/xchg.h : "r" ((long)m), "Ir" (old), "1" (new) : "memory"); new 181 arch/alpha/include/asm/xchg.h ____cmpxchg(_u32, volatile int *m, int old, int new) new 197 arch/alpha/include/asm/xchg.h : "r"((long) old), "r"(new), "m"(*m) : "memory"); new 203 arch/alpha/include/asm/xchg.h ____cmpxchg(_u64, volatile long *m, unsigned long old, unsigned long new) new 219 arch/alpha/include/asm/xchg.h : "r"((long) old), "r"(new), "m"(*m) : "memory"); new 229 arch/alpha/include/asm/xchg.h ____cmpxchg(, volatile void *ptr, unsigned long old, unsigned long new, new 234 arch/alpha/include/asm/xchg.h return ____cmpxchg(_u8, ptr, old, new); new 236 arch/alpha/include/asm/xchg.h return ____cmpxchg(_u16, ptr, old, new); new 238 arch/alpha/include/asm/xchg.h return ____cmpxchg(_u32, ptr, old, new); new 240 arch/alpha/include/asm/xchg.h return ____cmpxchg(_u64, ptr, old, new); new 280 arch/alpha/kernel/err_common.c cdl_register_subpacket_annotation(struct el_subpacket_annotation *new) new 284 arch/alpha/kernel/err_common.c if (a == NULL) subpacket_annotation_list = new; new 287 arch/alpha/kernel/err_common.c if ((a->class == new->class && a->type == new->type) || new 288 arch/alpha/kernel/err_common.c a == new) { new 294 arch/alpha/kernel/err_common.c a->next = new; new 296 arch/alpha/kernel/err_common.c new->next = NULL; new 302 arch/alpha/kernel/err_common.c cdl_register_subpacket_handler(struct el_subpacket_handler *new) new 306 arch/alpha/kernel/err_common.c if (h == NULL) subpacket_handler_list = new; new 309 arch/alpha/kernel/err_common.c if (h->class == new->class || h == new) { new 315 arch/alpha/kernel/err_common.c h->next = new; new 317 arch/alpha/kernel/err_common.c new->next = NULL; new 445 arch/arc/include/asm/atomic.h atomic64_cmpxchg(atomic64_t *ptr, s64 expected, s64 new) new 459 arch/arc/include/asm/atomic.h : "r"(ptr), "ir"(expected), "r"(new) new 467 arch/arc/include/asm/atomic.h static inline s64 atomic64_xchg(atomic64_t *ptr, s64 new) new 479 arch/arc/include/asm/atomic.h : "r"(ptr), "r"(new) new 17 arch/arc/include/asm/cmpxchg.h __cmpxchg(volatile void *ptr, unsigned long expected, unsigned long new) new 36 arch/arc/include/asm/cmpxchg.h "r"(new) /* can't be "ir". scond can't take LIMM for "b" */ new 47 arch/arc/include/asm/cmpxchg.h __cmpxchg(volatile void *ptr, unsigned long expected, unsigned long new) new 59 arch/arc/include/asm/cmpxchg.h *p = new; new 67 arch/arc/include/asm/cmpxchg.h __cmpxchg(volatile void *ptr, unsigned long expected, unsigned long new) new 81 arch/arc/include/asm/cmpxchg.h : "+r"(new) new 87 arch/arc/include/asm/cmpxchg.h return new; new 222 arch/arc/include/asm/cmpxchg.h #define atomic_xchg(v, new) (xchg(&((v)->counter), new)) new 44 arch/arc/kernel/process.c SYSCALL_DEFINE3(arc_usr_cmpxchg, int *, uaddr, int, expected, int, new) new 75 arch/arc/kernel/process.c ret = __put_user(new, uaddr); new 268 arch/arc/kernel/smp.c unsigned long old, new; new 280 arch/arc/kernel/smp.c new = old = READ_ONCE(*ipi_data_ptr); new 281 arch/arc/kernel/smp.c new |= 1U << msg; new 282 arch/arc/kernel/smp.c } while (cmpxchg(ipi_data_ptr, old, new) != old); new 108 arch/arm/include/asm/atomic.h static inline int atomic_cmpxchg_relaxed(atomic_t *ptr, int old, int new) new 122 arch/arm/include/asm/atomic.h : "r" (&ptr->counter), "Ir" (old), "r" (new) new 202 arch/arm/include/asm/atomic.h static inline int atomic_cmpxchg(atomic_t *v, int old, int new) new 210 arch/arm/include/asm/atomic.h v->counter = new; new 245 arch/arm/include/asm/atomic.h #define atomic_xchg(v, new) (xchg(&((v)->counter), new)) new 406 arch/arm/include/asm/atomic.h static inline s64 atomic64_cmpxchg_relaxed(atomic64_t *ptr, s64 old, s64 new) new 421 arch/arm/include/asm/atomic.h : "r" (&ptr->counter), "r" (old), "r" (new) new 429 arch/arm/include/asm/atomic.h static inline s64 atomic64_xchg_relaxed(atomic64_t *ptr, s64 new) new 442 arch/arm/include/asm/atomic.h : "r" (&ptr->counter), "r" (new) new 157 arch/arm/include/asm/cmpxchg.h unsigned long new, int size) new 173 arch/arm/include/asm/cmpxchg.h : "r" (ptr), "Ir" (old), "r" (new) new 185 arch/arm/include/asm/cmpxchg.h : "r" (ptr), "Ir" (old), "r" (new) new 198 arch/arm/include/asm/cmpxchg.h : "r" (ptr), "Ir" (old), "r" (new) new 219 arch/arm/include/asm/cmpxchg.h unsigned long new, int size) new 227 arch/arm/include/asm/cmpxchg.h ret = __cmpxchg_local_generic(ptr, old, new, size); new 231 arch/arm/include/asm/cmpxchg.h ret = __cmpxchg(ptr, old, new, size); new 246 arch/arm/include/asm/cmpxchg.h unsigned long long new) new 263 arch/arm/include/asm/cmpxchg.h : "r" (ptr), "r" (old), "r" (new) new 80 arch/arm/kernel/ftrace.c unsigned long new, bool validate) new 86 arch/arm/kernel/ftrace.c new = __opcode_to_mem_thumb32(new); new 89 arch/arm/kernel/ftrace.c new = __opcode_to_mem_arm(new); new 100 arch/arm/kernel/ftrace.c if (probe_kernel_write((void *)pc, &new, MCOUNT_INSN_SIZE)) new 111 arch/arm/kernel/ftrace.c unsigned long new; new 115 arch/arm/kernel/ftrace.c new = ftrace_call_replace(pc, (unsigned long)func); new 117 arch/arm/kernel/ftrace.c ret = ftrace_modify_code(pc, 0, new, false); new 122 arch/arm/kernel/ftrace.c new = ftrace_call_replace(pc, (unsigned long)func); new 124 arch/arm/kernel/ftrace.c ret = ftrace_modify_code(pc, 0, new, false); new 133 arch/arm/kernel/ftrace.c unsigned long new, old; new 138 arch/arm/kernel/ftrace.c new = ftrace_call_replace(ip, adjust_address(rec, addr)); new 140 arch/arm/kernel/ftrace.c return ftrace_modify_code(rec->ip, old, new, true); new 148 arch/arm/kernel/ftrace.c unsigned long new, old; new 153 arch/arm/kernel/ftrace.c new = ftrace_call_replace(ip, adjust_address(rec, addr)); new 155 arch/arm/kernel/ftrace.c return ftrace_modify_code(rec->ip, old, new, true); new 165 arch/arm/kernel/ftrace.c unsigned long new; new 169 arch/arm/kernel/ftrace.c new = ftrace_nop_replace(rec); new 170 arch/arm/kernel/ftrace.c ret = ftrace_modify_code(ip, old, new, true); new 213 arch/arm/kernel/ftrace.c unsigned long new = enable ? branch : nop; new 215 arch/arm/kernel/ftrace.c return ftrace_modify_code(pc, old, new, true); new 781 arch/arm/kernel/smp.c if ((val == CPUFREQ_PRECHANGE && freq->old < freq->new) || new 782 arch/arm/kernel/smp.c (val == CPUFREQ_POSTCHANGE && freq->old > freq->new)) { new 785 arch/arm/kernel/smp.c freq->new); new 788 arch/arm/kernel/smp.c per_cpu(l_p_j_ref_freq, first), freq->new); new 895 arch/arm/mach-pxa/viper.c if (freq->old < freq->new) { new 898 arch/arm/mach-pxa/viper.c viper_set_core_cpu_voltage(freq->new, 0); new 902 arch/arm/mach-pxa/viper.c if (freq->old > freq->new) { new 905 arch/arm/mach-pxa/viper.c viper_set_core_cpu_voltage(freq->new, 0); new 56 arch/arm/mach-s3c24xx/mach-osiris-dvs.c bool new_dvs = is_dvs(&freqs->new); new 64 arch/arm/mach-s3c24xx/mach-osiris-dvs.c freqs->new.armclk, freqs->new.hclk); new 61 arch/arm/plat-samsung/include/plat/cpu-freq.h struct s3c_freq new; new 127 arch/arm/plat-samsung/pm-gpio.c u32 gpcon, old, new, mask; new 141 arch/arm/plat-samsung/pm-gpio.c new = (gps_gpcon & mask) >> nr; new 145 arch/arm/plat-samsung/pm-gpio.c if (old == new) new 150 arch/arm/plat-samsung/pm-gpio.c if (is_sfn(old) && is_sfn(new)) new 155 arch/arm/plat-samsung/pm-gpio.c if (is_in(old) && is_out(new)) new 160 arch/arm/plat-samsung/pm-gpio.c if (is_sfn(old) && is_out(new)) new 204 arch/arm/plat-samsung/pm-gpio.c u32 old, new, mask; new 210 arch/arm/plat-samsung/pm-gpio.c new = (gps_gpcon & mask) >> nr; new 214 arch/arm/plat-samsung/pm-gpio.c if (old == new) new 219 arch/arm/plat-samsung/pm-gpio.c if (is_sfn(old) && is_sfn(new)) new 224 arch/arm/plat-samsung/pm-gpio.c if (is_in(old) && is_out(new)) new 229 arch/arm/plat-samsung/pm-gpio.c if (is_sfn(old) && is_out(new)) new 32 arch/arm/xen/p2m.c static int xen_add_phys_to_mach_entry(struct xen_p2m_entry *new) new 43 arch/arm/xen/p2m.c if (new->pfn == entry->pfn) new 46 arch/arm/xen/p2m.c if (new->pfn < entry->pfn) new 51 arch/arm/xen/p2m.c rb_link_node(&new->rbnode_phys, parent, link); new 52 arch/arm/xen/p2m.c rb_insert_color(&new->rbnode_phys, &phys_to_mach); new 58 arch/arm/xen/p2m.c __func__, &new->pfn, &new->mfn, &entry->pfn, &entry->mfn); new 147 arch/arm64/include/asm/atomic.h #define arch_atomic_xchg_relaxed(v, new) \ new 148 arch/arm64/include/asm/atomic.h arch_xchg_relaxed(&((v)->counter), (new)) new 149 arch/arm64/include/asm/atomic.h #define arch_atomic_xchg_acquire(v, new) \ new 150 arch/arm64/include/asm/atomic.h arch_xchg_acquire(&((v)->counter), (new)) new 151 arch/arm64/include/asm/atomic.h #define arch_atomic_xchg_release(v, new) \ new 152 arch/arm64/include/asm/atomic.h arch_xchg_release(&((v)->counter), (new)) new 153 arch/arm64/include/asm/atomic.h #define arch_atomic_xchg(v, new) \ new 154 arch/arm64/include/asm/atomic.h arch_xchg(&((v)->counter), (new)) new 156 arch/arm64/include/asm/atomic.h #define arch_atomic_cmpxchg_relaxed(v, old, new) \ new 157 arch/arm64/include/asm/atomic.h arch_cmpxchg_relaxed(&((v)->counter), (old), (new)) new 158 arch/arm64/include/asm/atomic.h #define arch_atomic_cmpxchg_acquire(v, old, new) \ new 159 arch/arm64/include/asm/atomic.h arch_cmpxchg_acquire(&((v)->counter), (old), (new)) new 160 arch/arm64/include/asm/atomic.h #define arch_atomic_cmpxchg_release(v, old, new) \ new 161 arch/arm64/include/asm/atomic.h arch_cmpxchg_release(&((v)->counter), (old), (new)) new 162 arch/arm64/include/asm/atomic.h #define arch_atomic_cmpxchg(v, old, new) \ new 163 arch/arm64/include/asm/atomic.h arch_cmpxchg(&((v)->counter), (old), (new)) new 263 arch/arm64/include/asm/atomic_ll_sc.h u##sz new) \ new 288 arch/arm64/include/asm/atomic_ll_sc.h : [old] __stringify(constraint) "r" (old), [new] "r" (new) \ new 344 arch/arm64/include/asm/atomic_lse.h u##sz new) \ new 348 arch/arm64/include/asm/atomic_lse.h register u##sz x2 asm ("x2") = new; \ new 358 arch/arm64/include/asm/atomic_lse.h : [old] "r" (x1), [new] "r" (x2) \ new 109 arch/arm64/include/asm/cmpxchg.h u##sz new) \ new 112 arch/arm64/include/asm/cmpxchg.h ptr, old, new); \ new 153 arch/arm64/include/asm/cmpxchg.h unsigned long new, \ new 158 arch/arm64/include/asm/cmpxchg.h return __cmpxchg_case##sfx##_8(ptr, old, new); \ new 160 arch/arm64/include/asm/cmpxchg.h return __cmpxchg_case##sfx##_16(ptr, old, new); \ new 162 arch/arm64/include/asm/cmpxchg.h return __cmpxchg_case##sfx##_32(ptr, old, new); \ new 164 arch/arm64/include/asm/cmpxchg.h return __cmpxchg_case##sfx##_64(ptr, old, new); \ new 471 arch/arm64/kernel/cpufeature.c static s64 arm64_ftr_safe_value(const struct arm64_ftr_bits *ftrp, s64 new, new 481 arch/arm64/kernel/cpufeature.c ret = new < cur ? new : cur; new 484 arch/arm64/kernel/cpufeature.c if (!cur || !new) new 488 arch/arm64/kernel/cpufeature.c ret = new > cur ? new : cur; new 512 arch/arm64/kernel/cpufeature.c static void __init init_cpu_ftr_reg(u32 sys_reg, u64 new) new 526 arch/arm64/kernel/cpufeature.c s64 ftr_new = arm64_ftr_value(ftrp, new); new 630 arch/arm64/kernel/cpufeature.c static void update_cpu_ftr_reg(struct arm64_ftr_reg *reg, u64 new) new 636 arch/arm64/kernel/cpufeature.c s64 ftr_new = arm64_ftr_value(ftrp, new); new 24 arch/arm64/kernel/ftrace.c static int ftrace_modify_code(unsigned long pc, u32 old, u32 new, new 44 arch/arm64/kernel/ftrace.c if (aarch64_insn_patch_text_nosync((void *)pc, new)) new 56 arch/arm64/kernel/ftrace.c u32 new; new 59 arch/arm64/kernel/ftrace.c new = aarch64_insn_gen_branch_imm(pc, (unsigned long)func, new 62 arch/arm64/kernel/ftrace.c return ftrace_modify_code(pc, 0, new, false); new 71 arch/arm64/kernel/ftrace.c u32 old, new; new 142 arch/arm64/kernel/ftrace.c new = aarch64_insn_gen_branch_imm(pc, addr, AARCH64_INSN_BRANCH_LINK); new 144 arch/arm64/kernel/ftrace.c return ftrace_modify_code(pc, old, new, true); new 155 arch/arm64/kernel/ftrace.c u32 old = 0, new; new 200 arch/arm64/kernel/ftrace.c new = aarch64_insn_gen_nop(); new 202 arch/arm64/kernel/ftrace.c return ftrace_modify_code(pc, old, new, validate); new 65 arch/arm64/kvm/inject_fault.c unsigned long old, new; new 68 arch/arm64/kvm/inject_fault.c new = 0; new 70 arch/arm64/kvm/inject_fault.c new |= (old & PSR_N_BIT); new 71 arch/arm64/kvm/inject_fault.c new |= (old & PSR_Z_BIT); new 72 arch/arm64/kvm/inject_fault.c new |= (old & PSR_C_BIT); new 73 arch/arm64/kvm/inject_fault.c new |= (old & PSR_V_BIT); new 77 arch/arm64/kvm/inject_fault.c new |= (old & PSR_DIT_BIT); new 85 arch/arm64/kvm/inject_fault.c new |= (old & PSR_PAN_BIT); new 87 arch/arm64/kvm/inject_fault.c new |= PSR_PAN_BIT; new 98 arch/arm64/kvm/inject_fault.c new |= PSR_SSBS_BIT; new 103 arch/arm64/kvm/inject_fault.c new |= PSR_D_BIT; new 104 arch/arm64/kvm/inject_fault.c new |= PSR_A_BIT; new 105 arch/arm64/kvm/inject_fault.c new |= PSR_I_BIT; new 106 arch/arm64/kvm/inject_fault.c new |= PSR_F_BIT; new 108 arch/arm64/kvm/inject_fault.c new |= PSR_MODE_EL1h; new 110 arch/arm64/kvm/inject_fault.c return new; new 117 arch/arm64/mm/mmu.c static bool pgattr_change_is_safe(u64 old, u64 new) new 126 arch/arm64/mm/mmu.c if (old == 0 || new == 0) new 130 arch/arm64/mm/mmu.c if ((old | new) & PTE_CONT) new 134 arch/arm64/mm/mmu.c if (old & ~new & PTE_NG) new 137 arch/arm64/mm/mmu.c return ((old ^ new) & ~mask) == 0; new 11 arch/csky/include/asm/cmpxchg.h #define __xchg(new, ptr, size) \ new 14 arch/csky/include/asm/cmpxchg.h __typeof__(new) __new = (new); \ new 38 arch/csky/include/asm/cmpxchg.h #define __cmpxchg(ptr, old, new, size) \ new 41 arch/csky/include/asm/cmpxchg.h __typeof__(new) __new = (new); \ new 42 arch/csky/include/asm/cmpxchg.h __typeof__(new) __tmp; \ new 72 arch/h8300/include/asm/atomic.h static inline int atomic_cmpxchg(atomic_t *v, int old, int new) new 80 arch/h8300/include/asm/atomic.h v->counter = new; new 64 arch/h8300/include/asm/cmpxchg.h #define atomic_xchg(v, new) (xchg(&((v)->counter), new)) new 19 arch/hexagon/include/asm/atomic.h static inline void atomic_set(atomic_t *v, int new) new 26 arch/hexagon/include/asm/atomic.h : "r" (&v->counter), "r" (new) new 46 arch/hexagon/include/asm/atomic.h #define atomic_xchg(v, new) (xchg(&((v)->counter), (new))) new 66 arch/hexagon/include/asm/atomic.h static inline int atomic_cmpxchg(atomic_t *v, int old, int new) new 78 arch/hexagon/include/asm/atomic.h : "r" (&v->counter), "r" (old), "r" (new) new 54 arch/hexagon/include/asm/cmpxchg.h #define cmpxchg(ptr, old, new) \ new 58 arch/hexagon/include/asm/cmpxchg.h __typeof__(*(ptr)) __new = (new); \ new 22 arch/ia64/include/asm/acenv.h unsigned int old, new, val; new 25 arch/ia64/include/asm/acenv.h new = (((old & ~0x3) + 2) + ((old >> 1) & 0x1)); new 26 arch/ia64/include/asm/acenv.h val = ia64_cmpxchg4_acq(lock, new, old); new 28 arch/ia64/include/asm/acenv.h return (new < 3) ? -1 : 0; new 34 arch/ia64/include/asm/acenv.h unsigned int old, new, val; new 37 arch/ia64/include/asm/acenv.h new = old & ~0x3; new 38 arch/ia64/include/asm/acenv.h val = ia64_cmpxchg4_acq(lock, new, old); new 35 arch/ia64/include/asm/atomic.h __s32 old, new; \ new 41 arch/ia64/include/asm/atomic.h new = old c_op i; \ new 42 arch/ia64/include/asm/atomic.h } while (ia64_cmpxchg(acq, v, old, new, sizeof(atomic_t)) != old); \ new 43 arch/ia64/include/asm/atomic.h return new; \ new 50 arch/ia64/include/asm/atomic.h __s32 old, new; \ new 56 arch/ia64/include/asm/atomic.h new = old c_op i; \ new 57 arch/ia64/include/asm/atomic.h } while (ia64_cmpxchg(acq, v, old, new, sizeof(atomic_t)) != old); \ new 130 arch/ia64/include/asm/atomic.h s64 old, new; \ new 136 arch/ia64/include/asm/atomic.h new = old c_op i; \ new 137 arch/ia64/include/asm/atomic.h } while (ia64_cmpxchg(acq, v, old, new, sizeof(atomic64_t)) != old); \ new 138 arch/ia64/include/asm/atomic.h return new; \ new 145 arch/ia64/include/asm/atomic.h s64 old, new; \ new 151 arch/ia64/include/asm/atomic.h new = old c_op i; \ new 152 arch/ia64/include/asm/atomic.h } while (ia64_cmpxchg(acq, v, old, new, sizeof(atomic64_t)) != old); \ new 211 arch/ia64/include/asm/atomic.h #define atomic_cmpxchg(v, old, new) (cmpxchg(&((v)->counter), old, new)) new 212 arch/ia64/include/asm/atomic.h #define atomic_xchg(v, new) (xchg(&((v)->counter), new)) new 214 arch/ia64/include/asm/atomic.h #define atomic64_cmpxchg(v, old, new) \ new 215 arch/ia64/include/asm/atomic.h (cmpxchg(&((v)->counter), old, new)) new 216 arch/ia64/include/asm/atomic.h #define atomic64_xchg(v, new) (xchg(&((v)->counter), new)) new 42 arch/ia64/include/asm/bitops.h __u32 bit, old, new; new 51 arch/ia64/include/asm/bitops.h new = old | bit; new 52 arch/ia64/include/asm/bitops.h } while (cmpxchg_acq(m, old, new) != old); new 83 arch/ia64/include/asm/bitops.h __u32 mask, old, new; new 92 arch/ia64/include/asm/bitops.h new = old & mask; new 93 arch/ia64/include/asm/bitops.h } while (cmpxchg_acq(m, old, new) != old); new 107 arch/ia64/include/asm/bitops.h __u32 mask, old, new; new 116 arch/ia64/include/asm/bitops.h new = old & mask; new 117 arch/ia64/include/asm/bitops.h } while (cmpxchg_rel(m, old, new) != old); new 132 arch/ia64/include/asm/bitops.h __u32 const new = *m & ~(1 << (nr & 31)); new 134 arch/ia64/include/asm/bitops.h ia64_st4_rel_nta(m, new); new 164 arch/ia64/include/asm/bitops.h __u32 bit, old, new; new 173 arch/ia64/include/asm/bitops.h new = old ^ bit; new 174 arch/ia64/include/asm/bitops.h } while (cmpxchg_acq(m, old, new) != old); new 203 arch/ia64/include/asm/bitops.h __u32 bit, old, new; new 212 arch/ia64/include/asm/bitops.h new = old | bit; new 213 arch/ia64/include/asm/bitops.h } while (cmpxchg_acq(m, old, new) != old); new 257 arch/ia64/include/asm/bitops.h __u32 mask, old, new; new 266 arch/ia64/include/asm/bitops.h new = old & mask; new 267 arch/ia64/include/asm/bitops.h } while (cmpxchg_acq(m, old, new) != old); new 302 arch/ia64/include/asm/bitops.h __u32 bit, old, new; new 311 arch/ia64/include/asm/bitops.h new = old ^ bit; new 312 arch/ia64/include/asm/bitops.h } while (cmpxchg_acq(m, old, new) != old); new 441 arch/ia64/include/asm/pgtable.h unsigned long new, old; new 445 arch/ia64/include/asm/pgtable.h new = pte_val(pte_wrprotect(__pte (old))); new 446 arch/ia64/include/asm/pgtable.h } while (cmpxchg((unsigned long *) ptep, old, new) != old); new 269 arch/ia64/include/asm/spinlock.h } old, new; new 270 arch/ia64/include/asm/spinlock.h old.lock = new.lock = *x; new 271 arch/ia64/include/asm/spinlock.h old.lock.write_lock = new.lock.write_lock = 0; new 272 arch/ia64/include/asm/spinlock.h ++new.lock.read_counter; new 273 arch/ia64/include/asm/spinlock.h return (u32)ia64_cmpxchg4_acq((__u32 *)(x), new.word, old.word) == old.word; new 71 arch/ia64/include/uapi/asm/cmpxchg.h #define ia64_cmpxchg(sem, ptr, old, new, size) \ new 93 arch/ia64/include/uapi/asm/cmpxchg.h _r_ = ia64_cmpxchg1_##sem((__u8 *) ptr, new, _o_); \ new 97 arch/ia64/include/uapi/asm/cmpxchg.h _r_ = ia64_cmpxchg2_##sem((__u16 *) ptr, new, _o_); \ new 101 arch/ia64/include/uapi/asm/cmpxchg.h _r_ = ia64_cmpxchg4_##sem((__u32 *) ptr, new, _o_); \ new 105 arch/ia64/include/uapi/asm/cmpxchg.h _r_ = ia64_cmpxchg8_##sem((__u64 *) ptr, new, _o_); \ new 308 arch/ia64/include/uapi/asm/gcc_intrin.h #define ia64_cmpxchg1_acq(ptr, new, old) \ new 313 arch/ia64/include/uapi/asm/gcc_intrin.h "=r"(ia64_intri_res) : "r"(ptr), "r"(new) : "memory"); \ new 317 arch/ia64/include/uapi/asm/gcc_intrin.h #define ia64_cmpxchg1_rel(ptr, new, old) \ new 322 arch/ia64/include/uapi/asm/gcc_intrin.h "=r"(ia64_intri_res) : "r"(ptr), "r"(new) : "memory"); \ new 326 arch/ia64/include/uapi/asm/gcc_intrin.h #define ia64_cmpxchg2_acq(ptr, new, old) \ new 331 arch/ia64/include/uapi/asm/gcc_intrin.h "=r"(ia64_intri_res) : "r"(ptr), "r"(new) : "memory"); \ new 335 arch/ia64/include/uapi/asm/gcc_intrin.h #define ia64_cmpxchg2_rel(ptr, new, old) \ new 341 arch/ia64/include/uapi/asm/gcc_intrin.h "=r"(ia64_intri_res) : "r"(ptr), "r"(new) : "memory"); \ new 345 arch/ia64/include/uapi/asm/gcc_intrin.h #define ia64_cmpxchg4_acq(ptr, new, old) \ new 350 arch/ia64/include/uapi/asm/gcc_intrin.h "=r"(ia64_intri_res) : "r"(ptr), "r"(new) : "memory"); \ new 354 arch/ia64/include/uapi/asm/gcc_intrin.h #define ia64_cmpxchg4_rel(ptr, new, old) \ new 359 arch/ia64/include/uapi/asm/gcc_intrin.h "=r"(ia64_intri_res) : "r"(ptr), "r"(new) : "memory"); \ new 363 arch/ia64/include/uapi/asm/gcc_intrin.h #define ia64_cmpxchg8_acq(ptr, new, old) \ new 368 arch/ia64/include/uapi/asm/gcc_intrin.h "=r"(ia64_intri_res) : "r"(ptr), "r"(new) : "memory"); \ new 372 arch/ia64/include/uapi/asm/gcc_intrin.h #define ia64_cmpxchg8_rel(ptr, new, old) \ new 378 arch/ia64/include/uapi/asm/gcc_intrin.h "=r"(ia64_intri_res) : "r"(ptr), "r"(new) : "memory"); \ new 163 arch/ia64/kernel/ftrace.c char *new; new 168 arch/ia64/kernel/ftrace.c new = ftrace_nop_replace(); new 169 arch/ia64/kernel/ftrace.c return ftrace_modify_code(rec->ip, NULL, new, 0); new 175 arch/ia64/kernel/ftrace.c unsigned char *old, *new; new 178 arch/ia64/kernel/ftrace.c new = ftrace_call_replace(ip, addr); new 179 arch/ia64/kernel/ftrace.c return ftrace_modify_code(ip, old, new, 1); new 137 arch/m68k/bvme6000/config.c unsigned long *new = (unsigned long *)vectors; new 144 arch/m68k/bvme6000/config.c *(new+4) = *(old+4); /* Illegal instruction */ new 145 arch/m68k/bvme6000/config.c *(new+9) = *(old+9); /* Trace */ new 146 arch/m68k/bvme6000/config.c *(new+47) = *(old+47); /* Trap #15 */ new 147 arch/m68k/bvme6000/config.c *(new+0x1f) = *(old+0x1f); /* ABORT switch */ new 67 arch/m68k/fpsp040/fpsp.h | handler exit code will reload the new frame and discard the old. new 166 arch/m68k/include/asm/atomic.h #define atomic_xchg(v, new) (xchg(&((v)->counter), new)) new 170 arch/m68k/include/asm/atomic.h static inline int atomic_cmpxchg(atomic_t *v, int old, int new) new 178 arch/m68k/include/asm/atomic.h atomic_set(v, new); new 183 arch/m68k/include/asm/atomic.h static inline int atomic_xchg(atomic_t *v, int new) new 190 arch/m68k/include/asm/atomic.h atomic_set(v, new); new 96 arch/m68k/include/asm/cmpxchg.h unsigned long new, int size) new 102 arch/m68k/include/asm/cmpxchg.h : "d" (new), "0" (old), "m" (*(char *)p)); new 107 arch/m68k/include/asm/cmpxchg.h : "d" (new), "0" (old), "m" (*(short *)p)); new 112 arch/m68k/include/asm/cmpxchg.h : "d" (new), "0" (old), "m" (*(int *)p)); new 115 arch/m68k/include/asm/cmpxchg.h old = __invalid_cmpxchg_size(p, old, new, size); new 75 arch/m68k/mm/memory.c ptable_desc *new; new 83 arch/m68k/mm/memory.c new = PD_PTABLE(page); new 84 arch/m68k/mm/memory.c PD_MARKBITS(new) = 0xfe; new 85 arch/m68k/mm/memory.c list_add_tail(new, dp); new 321 arch/m68k/mvme16x/config.c unsigned long *new = (unsigned long *)vectors; new 336 arch/m68k/mvme16x/config.c *(new+4) = *(old+4); /* Illegal instruction */ new 337 arch/m68k/mvme16x/config.c *(new+9) = *(old+9); /* Trace */ new 338 arch/m68k/mvme16x/config.c *(new+47) = *(old+47); /* Trap #15 */ new 341 arch/m68k/mvme16x/config.c *(new+0x5e) = *(old+0x5e); /* ABORT switch */ new 343 arch/m68k/mvme16x/config.c *(new+0x6e) = *(old+0x6e); /* ABORT switch */ new 248 arch/m68k/sun3/mmu_emu.c unsigned long new = 1; new 253 arch/m68k/sun3/mmu_emu.c new = next_to_die; new 254 arch/m68k/sun3/mmu_emu.c clear_context(new); new 259 arch/m68k/sun3/mmu_emu.c while(new < CONTEXTS_NUM) { new 260 arch/m68k/sun3/mmu_emu.c if(ctx_alloc[new]) new 261 arch/m68k/sun3/mmu_emu.c new++; new 266 arch/m68k/sun3/mmu_emu.c if(new == CONTEXTS_NUM) new 270 arch/m68k/sun3/mmu_emu.c ctx_alloc[new] = mm; new 273 arch/m68k/sun3/mmu_emu.c return new; new 102 arch/mips/boot/elf2ecoff.c static void combine(struct sect *base, struct sect *new, int pad) new 105 arch/mips/boot/elf2ecoff.c *base = *new; new 106 arch/mips/boot/elf2ecoff.c else if (new->len) { new 107 arch/mips/boot/elf2ecoff.c if (base->vaddr + base->len != new->vaddr) { new 109 arch/mips/boot/elf2ecoff.c base->len = new->vaddr - base->vaddr; new 116 arch/mips/boot/elf2ecoff.c base->len += new->len; new 231 arch/mips/include/asm/atomic.h #define atomic_xchg(v, new) (xchg(&((v)->counter), (new))) new 431 arch/mips/include/asm/atomic.h #define atomic64_xchg(v, new) (xchg(&((v)->counter), (new))) new 116 arch/mips/include/asm/cmpxchg.h #define __cmpxchg_asm(ld, st, m, old, new) \ new 137 arch/mips/include/asm/cmpxchg.h : GCC_OFF_SMALL_ASM() (*m), "Jr" (old), "Jr" (new) \ new 146 arch/mips/include/asm/cmpxchg.h *m = new; \ new 154 arch/mips/include/asm/cmpxchg.h unsigned long new, unsigned int size); new 158 arch/mips/include/asm/cmpxchg.h unsigned long new, unsigned int size) new 163 arch/mips/include/asm/cmpxchg.h return __cmpxchg_small(ptr, old, new, size); new 167 arch/mips/include/asm/cmpxchg.h (u32)old, new); new 175 arch/mips/include/asm/cmpxchg.h (u64)old, new); new 182 arch/mips/include/asm/cmpxchg.h #define cmpxchg_local(ptr, old, new) \ new 186 arch/mips/include/asm/cmpxchg.h (unsigned long)(__typeof__(*(ptr)))(new), \ new 189 arch/mips/include/asm/cmpxchg.h #define cmpxchg(ptr, old, new) \ new 194 arch/mips/include/asm/cmpxchg.h __res = cmpxchg_local((ptr), (old), (new)); \ new 221 arch/mips/include/asm/cmpxchg.h unsigned long long new) new 280 arch/mips/include/asm/cmpxchg.h "r" (new) new 2776 arch/mips/include/asm/mipsregs.h unsigned int res, new; \ new 2779 arch/mips/include/asm/mipsregs.h new = res | set; \ new 2780 arch/mips/include/asm/mipsregs.h write_##name(new); \ new 2788 arch/mips/include/asm/mipsregs.h unsigned int res, new; \ new 2791 arch/mips/include/asm/mipsregs.h new = res & ~clear; \ new 2792 arch/mips/include/asm/mipsregs.h write_##name(new); \ new 2800 arch/mips/include/asm/mipsregs.h unsigned int res, new; \ new 2803 arch/mips/include/asm/mipsregs.h new = res & ~change; \ new 2804 arch/mips/include/asm/mipsregs.h new |= (val & change); \ new 2805 arch/mips/include/asm/mipsregs.h write_##name(new); \ new 252 arch/mips/include/asm/txx9/tx4927.h static inline void tx4927_ccfg_change(__u64 change, __u64 new) new 256 arch/mips/include/asm/txx9/tx4927.h new, new 268 arch/mips/include/asm/txx9/tx4938.h #define tx4938_ccfg_change(change, new) tx4927_ccfg_change(change, new) new 488 arch/mips/include/asm/txx9/tx4939.h #define tx4939_ccfg_change(change, new) tx4938_ccfg_change(change, new) new 51 arch/mips/kernel/cmpxchg.c unsigned long new, unsigned int size) new 63 arch/mips/kernel/cmpxchg.c new &= mask; new 99 arch/mips/kernel/cmpxchg.c new32 = (load32 & ~mask) | (new << shift); new 181 arch/mips/kernel/ftrace.c unsigned int new; new 188 arch/mips/kernel/ftrace.c new = core_kernel_text(ip) ? INSN_NOP : INSN_B_1F; new 190 arch/mips/kernel/ftrace.c return ftrace_modify_code(ip, new); new 199 arch/mips/kernel/ftrace.c return ftrace_modify_code_2(ip, new, INSN_NOP); new 205 arch/mips/kernel/ftrace.c unsigned int new; new 208 arch/mips/kernel/ftrace.c new = core_kernel_text(ip) ? insn_jal_ftrace_caller : insn_la_mcount[0]; new 211 arch/mips/kernel/ftrace.c return ftrace_modify_code(ip, new); new 213 arch/mips/kernel/ftrace.c return ftrace_modify_code_2r(ip, new, core_kernel_text(ip) ? new 222 arch/mips/kernel/ftrace.c unsigned int new; new 224 arch/mips/kernel/ftrace.c new = INSN_JAL((unsigned long)func); new 226 arch/mips/kernel/ftrace.c return ftrace_modify_code(FTRACE_CALL_IP, new); new 96 arch/mips/kernel/syscall.c static inline int mips_atomic_set(unsigned long addr, unsigned long new) new 132 arch/mips/kernel/syscall.c [new] "r" (new), new 162 arch/mips/kernel/syscall.c [new] "r" (new), new 173 arch/mips/kernel/syscall.c err |= __put_user(new, (unsigned int *) addr); new 52 arch/mips/kernel/vpe-cmp.c unsigned long new; new 55 arch/mips/kernel/vpe-cmp.c ret = kstrtoul(buf, 0, &new); new 60 arch/mips/kernel/vpe-cmp.c if (new != 1) new 63 arch/mips/kernel/vpe-cmp.c vpe->ntcs = new; new 291 arch/mips/kernel/vpe-mt.c unsigned long new; new 294 arch/mips/kernel/vpe-mt.c ret = kstrtoul(buf, 0, &new); new 298 arch/mips/kernel/vpe-mt.c if (new == 0 || new > (hw_tcs - aprp_cpu_index())) new 301 arch/mips/kernel/vpe-mt.c vpe->ntcs = new; new 234 arch/mips/kvm/mips.c const struct kvm_memory_slot *new, new 254 arch/mips/kvm/mips.c new->flags & KVM_MEM_LOG_DIRTY_PAGES)) { new 257 arch/mips/kvm/mips.c needs_flush = kvm_mips_mkclean_gpa_pt(kvm, new->base_gfn, new 258 arch/mips/kvm/mips.c new->base_gfn + new->npages - 1); new 261 arch/mips/kvm/mips.c kvm_mips_callbacks->flush_shadow_memslot(kvm, new); new 314 arch/mips/kvm/mmu.c pte_t old, new; \ new 321 arch/mips/kvm/mmu.c new = op(old); \ new 322 arch/mips/kvm/mmu.c if (pte_val(new) == pte_val(old)) \ new 324 arch/mips/kvm/mmu.c set_pte(pte + i, new); \ new 118 arch/mips/txx9/generic/pci.c } *new = NULL; new 122 arch/mips/txx9/generic/pci.c new = kzalloc(sizeof(*new), GFP_KERNEL); new 123 arch/mips/txx9/generic/pci.c if (!new) new 125 arch/mips/txx9/generic/pci.c new->r_mem[0].name = "PCI mem"; new 126 arch/mips/txx9/generic/pci.c new->r_mem[1].name = "PCI mmio"; new 127 arch/mips/txx9/generic/pci.c new->r_io.name = "PCI io"; new 128 arch/mips/txx9/generic/pci.c new->c.mem_resource = new->r_mem; new 129 arch/mips/txx9/generic/pci.c new->c.io_resource = &new->r_io; new 130 arch/mips/txx9/generic/pci.c pcic = &new->c; new 220 arch/mips/txx9/generic/pci.c kfree(new); new 26 arch/openrisc/include/asm/cmpxchg.h unsigned long old, unsigned long new) new 38 arch/openrisc/include/asm/cmpxchg.h : "r"(ptr), "r"(old), "r"(new) new 59 arch/openrisc/include/asm/cmpxchg.h static inline u32 cmpxchg_small(volatile void *ptr, u32 old, u32 new, new 81 arch/openrisc/include/asm/cmpxchg.h new32 = (load32 & ~bitmask) | (new << bitoff); new 122 arch/openrisc/include/asm/cmpxchg.h unsigned long new, int size) new 127 arch/openrisc/include/asm/cmpxchg.h return cmpxchg_small(ptr, old, new, size); new 129 arch/openrisc/include/asm/cmpxchg.h return cmpxchg_u32(ptr, old, new); new 236 arch/openrisc/kernel/process.c struct task_struct *new) new 248 arch/openrisc/kernel/process.c new_ti = new->stack; new 78 arch/parisc/include/asm/atomic.h #define atomic_xchg(v, new) (xchg(&((v)->counter), new)) new 224 arch/parisc/include/asm/atomic.h #define atomic64_xchg(v, new) (xchg(&((v)->counter), new)) new 713 arch/parisc/kernel/drivers.c struct device *new = parse_tree_node(dev, d->index, d->modpath); new 714 arch/parisc/kernel/drivers.c if (new) new 715 arch/parisc/kernel/drivers.c d->dev = new; new 59 arch/parisc/lib/bitops.c u64 __cmpxchg_u64(volatile u64 *ptr, u64 old, u64 new) new 66 arch/parisc/lib/bitops.c *ptr = new; new 71 arch/parisc/lib/bitops.c unsigned long __cmpxchg_u32(volatile unsigned int *ptr, unsigned int old, unsigned int new) new 78 arch/parisc/lib/bitops.c *ptr = new; new 100 arch/powerpc/boot/simple_alloc.c void *new; new 116 arch/powerpc/boot/simple_alloc.c new = simple_malloc(size); new 117 arch/powerpc/boot/simple_alloc.c memcpy(new, ptr, p->size); new 119 arch/powerpc/boot/simple_alloc.c return new; new 201 arch/powerpc/include/asm/atomic.h #define atomic_xchg(v, new) (xchg(&((v)->counter), new)) new 202 arch/powerpc/include/asm/atomic.h #define atomic_xchg_relaxed(v, new) xchg_relaxed(&((v)->counter), (new)) new 493 arch/powerpc/include/asm/atomic.h #define atomic64_xchg(v, new) (xchg(&((v)->counter), new)) new 494 arch/powerpc/include/asm/atomic.h #define atomic64_xchg_relaxed(v, new) xchg_relaxed(&((v)->counter), (new)) new 102 arch/powerpc/include/asm/book3s/64/hugetlb.h static inline void hugepd_populate(hugepd_t *hpdp, pte_t *new, unsigned int pshift) new 104 arch/powerpc/include/asm/book3s/64/hugetlb.h *hpdp = __hugepd(__pa(new) | HUGEPD_VAL_BITS | (shift_to_mmu_psize(pshift) << 2)); new 44 arch/powerpc/include/asm/cmpxchg.h u32 __cmpxchg_##type##sfx(volatile void *p, u32 old, u32 new) \ new 52 arch/powerpc/include/asm/cmpxchg.h new <<= bitoff; \ new 70 arch/powerpc/include/asm/cmpxchg.h : "r" (p), "r" (old), "r" (new), "r" (prev_mask) \ new 223 arch/powerpc/include/asm/cmpxchg.h __cmpxchg_u32(volatile unsigned int *p, unsigned long old, unsigned long new) new 239 arch/powerpc/include/asm/cmpxchg.h : "r" (p), "r" (old), "r" (new) new 247 arch/powerpc/include/asm/cmpxchg.h unsigned long new) new 261 arch/powerpc/include/asm/cmpxchg.h : "r" (p), "r" (old), "r" (new) new 268 arch/powerpc/include/asm/cmpxchg.h __cmpxchg_u32_relaxed(u32 *p, unsigned long old, unsigned long new) new 281 arch/powerpc/include/asm/cmpxchg.h : "r" (p), "r" (old), "r" (new) new 296 arch/powerpc/include/asm/cmpxchg.h __cmpxchg_u32_acquire(u32 *p, unsigned long old, unsigned long new) new 311 arch/powerpc/include/asm/cmpxchg.h : "r" (p), "r" (old), "r" (new) new 319 arch/powerpc/include/asm/cmpxchg.h __cmpxchg_u64(volatile unsigned long *p, unsigned long old, unsigned long new) new 334 arch/powerpc/include/asm/cmpxchg.h : "r" (p), "r" (old), "r" (new) new 342 arch/powerpc/include/asm/cmpxchg.h unsigned long new) new 355 arch/powerpc/include/asm/cmpxchg.h : "r" (p), "r" (old), "r" (new) new 362 arch/powerpc/include/asm/cmpxchg.h __cmpxchg_u64_relaxed(u64 *p, unsigned long old, unsigned long new) new 374 arch/powerpc/include/asm/cmpxchg.h : "r" (p), "r" (old), "r" (new) new 381 arch/powerpc/include/asm/cmpxchg.h __cmpxchg_u64_acquire(u64 *p, unsigned long old, unsigned long new) new 395 arch/powerpc/include/asm/cmpxchg.h : "r" (p), "r" (old), "r" (new) new 403 arch/powerpc/include/asm/cmpxchg.h __cmpxchg(volatile void *ptr, unsigned long old, unsigned long new, new 408 arch/powerpc/include/asm/cmpxchg.h return __cmpxchg_u8(ptr, old, new); new 410 arch/powerpc/include/asm/cmpxchg.h return __cmpxchg_u16(ptr, old, new); new 412 arch/powerpc/include/asm/cmpxchg.h return __cmpxchg_u32(ptr, old, new); new 415 arch/powerpc/include/asm/cmpxchg.h return __cmpxchg_u64(ptr, old, new); new 423 arch/powerpc/include/asm/cmpxchg.h __cmpxchg_local(void *ptr, unsigned long old, unsigned long new, new 428 arch/powerpc/include/asm/cmpxchg.h return __cmpxchg_u8_local(ptr, old, new); new 430 arch/powerpc/include/asm/cmpxchg.h return __cmpxchg_u16_local(ptr, old, new); new 432 arch/powerpc/include/asm/cmpxchg.h return __cmpxchg_u32_local(ptr, old, new); new 435 arch/powerpc/include/asm/cmpxchg.h return __cmpxchg_u64_local(ptr, old, new); new 443 arch/powerpc/include/asm/cmpxchg.h __cmpxchg_relaxed(void *ptr, unsigned long old, unsigned long new, new 448 arch/powerpc/include/asm/cmpxchg.h return __cmpxchg_u8_relaxed(ptr, old, new); new 450 arch/powerpc/include/asm/cmpxchg.h return __cmpxchg_u16_relaxed(ptr, old, new); new 452 arch/powerpc/include/asm/cmpxchg.h return __cmpxchg_u32_relaxed(ptr, old, new); new 455 arch/powerpc/include/asm/cmpxchg.h return __cmpxchg_u64_relaxed(ptr, old, new); new 463 arch/powerpc/include/asm/cmpxchg.h __cmpxchg_acquire(void *ptr, unsigned long old, unsigned long new, new 468 arch/powerpc/include/asm/cmpxchg.h return __cmpxchg_u8_acquire(ptr, old, new); new 470 arch/powerpc/include/asm/cmpxchg.h return __cmpxchg_u16_acquire(ptr, old, new); new 472 arch/powerpc/include/asm/cmpxchg.h return __cmpxchg_u32_acquire(ptr, old, new); new 475 arch/powerpc/include/asm/cmpxchg.h return __cmpxchg_u64_acquire(ptr, old, new); new 26 arch/powerpc/include/asm/cputime.h #define cmpxchg_cputime(ptr, old, new) cmpxchg(ptr, old, new) new 215 arch/powerpc/include/asm/kvm_ppc.h const struct kvm_memory_slot *new, new 287 arch/powerpc/include/asm/kvm_ppc.h const struct kvm_memory_slot *new, new 33 arch/powerpc/include/asm/nohash/32/hugetlb-8xx.h static inline void hugepd_populate(hugepd_t *hpdp, pte_t *new, unsigned int pshift) new 35 arch/powerpc/include/asm/nohash/32/hugetlb-8xx.h *hpdp = __hugepd(__pa(new) | _PMD_USER | _PMD_PRESENT | new 240 arch/powerpc/include/asm/nohash/32/pgtable.h unsigned long new = (old & ~clr) | set; new 243 arch/powerpc/include/asm/nohash/32/pgtable.h p->pte = p->pte1 = p->pte2 = p->pte3 = new; new 245 arch/powerpc/include/asm/nohash/32/pgtable.h *p = __pte(new); new 31 arch/powerpc/include/asm/nohash/hugetlb-book3e.h static inline void hugepd_populate(hugepd_t *hpdp, pte_t *new, unsigned int pshift) new 34 arch/powerpc/include/asm/nohash/hugetlb-book3e.h *hpdp = __hugepd(((unsigned long)new & ~PD_HUGE) | pshift); new 81 arch/powerpc/include/asm/pgtable-be-types.h static inline bool pte_xchg(pte_t *ptep, pte_t old, pte_t new) new 88 arch/powerpc/include/asm/pgtable-be-types.h (__force unsigned long)pte_raw(new)); new 93 arch/powerpc/include/asm/pgtable-be-types.h static inline bool pmd_xchg(pmd_t *pmdp, pmd_t old, pmd_t new) new 99 arch/powerpc/include/asm/pgtable-be-types.h (__force unsigned long)pmd_raw(new)); new 61 arch/powerpc/include/asm/pgtable-types.h static inline bool pte_xchg(pte_t *ptep, pte_t old, pte_t new) new 66 arch/powerpc/include/asm/pgtable-types.h return pte_val(old) == __cmpxchg_u64(p, pte_val(old), pte_val(new)); new 86 arch/powerpc/kernel/eeh_driver.c enum pci_ers_result new) new 88 arch/powerpc/kernel/eeh_driver.c if (eeh_result_priority(new) > eeh_result_priority(old)) new 89 arch/powerpc/kernel/eeh_driver.c return new; new 947 arch/powerpc/kernel/process.c static inline void tm_recheckpoint_new_task(struct task_struct *new) new 960 arch/powerpc/kernel/process.c if (!tm_enabled(new)) new 963 arch/powerpc/kernel/process.c if (!MSR_TM_ACTIVE(new->thread.regs->msr)){ new 964 arch/powerpc/kernel/process.c tm_restore_sprs(&new->thread); new 969 arch/powerpc/kernel/process.c new->pid, new->thread.regs->msr); new 971 arch/powerpc/kernel/process.c tm_recheckpoint(&new->thread); new 978 arch/powerpc/kernel/process.c new->thread.regs->msr &= ~(MSR_FP | MSR_VEC | MSR_VSX); new 982 arch/powerpc/kernel/process.c new->pid, mfmsr()); new 986 arch/powerpc/kernel/process.c struct task_struct *new) new 989 arch/powerpc/kernel/process.c if (tm_enabled(prev) || tm_enabled(new)) new 999 arch/powerpc/kernel/process.c tm_recheckpoint_new_task(new); new 1047 arch/powerpc/kernel/process.c #define tm_recheckpoint_new_task(new) new 1048 arch/powerpc/kernel/process.c #define __switch_to_tm(prev, new) new 1123 arch/powerpc/kernel/process.c struct task_struct *new) new 1131 arch/powerpc/kernel/process.c new_thread = &new->thread; new 1147 arch/powerpc/kernel/process.c switch_booke_debug_regs(&new->thread.debug); new 1154 arch/powerpc/kernel/process.c if (unlikely(!hw_brk_match(this_cpu_ptr(¤t_brk), &new->thread.hw_brk))) new 1155 arch/powerpc/kernel/process.c __set_breakpoint(&new->thread.hw_brk); new 1168 arch/powerpc/kernel/process.c __switch_to_tm(prev, new); new 57 arch/powerpc/kernel/trace/ftrace.c ftrace_modify_code(unsigned long ip, unsigned int old, unsigned int new) new 81 arch/powerpc/kernel/trace/ftrace.c if (patch_instruction((unsigned int *)ip, new)) new 431 arch/powerpc/kernel/trace/ftrace.c unsigned int old, new; new 441 arch/powerpc/kernel/trace/ftrace.c new = PPC_INST_NOP; new 442 arch/powerpc/kernel/trace/ftrace.c return ftrace_modify_code(ip, old, new); new 664 arch/powerpc/kernel/trace/ftrace.c unsigned int old, new; new 674 arch/powerpc/kernel/trace/ftrace.c new = ftrace_call_replace(ip, addr, 1); new 675 arch/powerpc/kernel/trace/ftrace.c return ftrace_modify_code(ip, old, new); new 797 arch/powerpc/kernel/trace/ftrace.c unsigned int old, new; new 807 arch/powerpc/kernel/trace/ftrace.c new = ftrace_call_replace(ip, addr, 1); new 808 arch/powerpc/kernel/trace/ftrace.c return ftrace_modify_code(ip, old, new); new 837 arch/powerpc/kernel/trace/ftrace.c unsigned int old, new; new 841 arch/powerpc/kernel/trace/ftrace.c new = ftrace_call_replace(ip, (unsigned long)func, 1); new 842 arch/powerpc/kernel/trace/ftrace.c ret = ftrace_modify_code(ip, old, new); new 849 arch/powerpc/kernel/trace/ftrace.c new = ftrace_call_replace(ip, (unsigned long)func, 1); new 850 arch/powerpc/kernel/trace/ftrace.c ret = ftrace_modify_code(ip, old, new); new 922 arch/powerpc/kernel/trace/ftrace.c unsigned int old, new; new 925 arch/powerpc/kernel/trace/ftrace.c new = ftrace_call_replace(ip, addr, 0); new 927 arch/powerpc/kernel/trace/ftrace.c return ftrace_modify_code(ip, old, new); new 935 arch/powerpc/kernel/trace/ftrace.c unsigned int old, new; new 938 arch/powerpc/kernel/trace/ftrace.c new = ftrace_call_replace(ip, stub, 0); new 940 arch/powerpc/kernel/trace/ftrace.c return ftrace_modify_code(ip, old, new); new 864 arch/powerpc/kvm/book3s.c const struct kvm_memory_slot *new, new 867 arch/powerpc/kvm/book3s.c kvm->arch.kvm_ops->commit_memory_region(kvm, mem, old, new, change); new 1245 arch/powerpc/kvm/book3s_64_mmu_hv.c struct kvm_hpt_info *new = &resize->hpt; new 1247 arch/powerpc/kvm/book3s_64_mmu_hv.c unsigned long new_hash_mask = (1ULL << (new->order - 7)) - 1; new 1357 arch/powerpc/kvm/book3s_64_mmu_hv.c new_hptep = (__be64 *)(new->virt + (new_idx << 4)); new 1366 arch/powerpc/kvm/book3s_64_mmu_hv.c BUG_ON(new->order >= old->order); new 1385 arch/powerpc/kvm/book3s_64_mmu_hv.c new->rev[new_idx].guest_rpte = guest_rpte; new 4487 arch/powerpc/kvm/book3s_hv.c const struct kvm_memory_slot *new, new 4515 arch/powerpc/kvm/book3s_hv.c ((new->flags ^ old->flags) & KVM_MEM_LOG_DIRTY_PAGES)) new 86 arch/powerpc/kvm/book3s_hv_rm_xics.c union kvmppc_rm_state old, new; new 89 arch/powerpc/kvm/book3s_hv_rm_xics.c old = new = READ_ONCE(rm_core[core].rm_state); new 95 arch/powerpc/kvm/book3s_hv_rm_xics.c new.rm_action = action; new 98 arch/powerpc/kvm/book3s_hv_rm_xics.c old.raw, new.raw) == old.raw; new 183 arch/powerpc/kvm/book3s_hv_rm_xics.c union kvmppc_icp_state new) new 189 arch/powerpc/kvm/book3s_hv_rm_xics.c new.out_ee = (new.xisr && (new.pending_pri < new.cppr)); new 192 arch/powerpc/kvm/book3s_hv_rm_xics.c success = cmpxchg64(&icp->state.raw, old.raw, new.raw) == old.raw; new 211 arch/powerpc/kvm/book3s_hv_rm_xics.c if (new.out_ee) new 215 arch/powerpc/kvm/book3s_hv_rm_xics.c this_vcpu->arch.icp->rm_dbgstate = new; new 1916 arch/powerpc/kvm/book3s_pr.c const struct kvm_memory_slot *new, new 273 arch/powerpc/kvm/book3s_xics.c union kvmppc_icp_state new, new 279 arch/powerpc/kvm/book3s_xics.c new.out_ee = (new.xisr && (new.pending_pri < new.cppr)); new 282 arch/powerpc/kvm/book3s_xics.c success = cmpxchg64(&icp->state.raw, old.raw, new.raw) == old.raw; new 291 arch/powerpc/kvm/book3s_xics.c new.cppr, new.mfrr, new.pending_pri, new.xisr, new 292 arch/powerpc/kvm/book3s_xics.c new.need_resend, new.out_ee); new 308 arch/powerpc/kvm/book3s_xics.c if (new.out_ee) { new 1825 arch/powerpc/kvm/booke.c const struct kvm_memory_slot *new, new 707 arch/powerpc/kvm/powerpc.c const struct kvm_memory_slot *new, new 710 arch/powerpc/kvm/powerpc.c kvmppc_core_commit_memory_region(kvm, mem, old, new, change); new 49 arch/powerpc/mm/hugetlbpage.c pte_t *new; new 56 arch/powerpc/mm/hugetlbpage.c new = NULL; new 60 arch/powerpc/mm/hugetlbpage.c new = pte_alloc_one(mm); new 64 arch/powerpc/mm/hugetlbpage.c new = NULL; new 67 arch/powerpc/mm/hugetlbpage.c if (!cachep && !new) { new 73 arch/powerpc/mm/hugetlbpage.c new = kmem_cache_alloc(cachep, pgtable_gfp_flags(mm, GFP_KERNEL)); new 76 arch/powerpc/mm/hugetlbpage.c BUG_ON((unsigned long)new & HUGEPD_SHIFT_MASK); new 78 arch/powerpc/mm/hugetlbpage.c if (!new) new 98 arch/powerpc/mm/hugetlbpage.c hugepd_populate(hpdp, new, pshift); new 105 arch/powerpc/mm/hugetlbpage.c kmem_cache_free(cachep, new); new 107 arch/powerpc/mm/hugetlbpage.c pte_free(mm, new); new 109 arch/powerpc/mm/hugetlbpage.c kmemleak_ignore(new); new 107 arch/powerpc/mm/init-common.c struct kmem_cache *new; new 120 arch/powerpc/mm/init-common.c new = kmem_cache_create(name, table_size, align, 0, ctor(shift)); new 121 arch/powerpc/mm/init-common.c if (!new) new 125 arch/powerpc/mm/init-common.c pgtable_cache[shift] = new; new 42 arch/powerpc/mm/kasan/kasan_init_32.c pte_t *new; new 49 arch/powerpc/mm/kasan/kasan_init_32.c new = pte_alloc_one_kernel(&init_mm); new 51 arch/powerpc/mm/kasan/kasan_init_32.c new = memblock_alloc(PTE_FRAG_SIZE, PTE_FRAG_SIZE); new 53 arch/powerpc/mm/kasan/kasan_init_32.c if (!new) new 55 arch/powerpc/mm/kasan/kasan_init_32.c kasan_populate_pte(new, prot); new 62 arch/powerpc/mm/kasan/kasan_init_32.c pmd_populate_kernel(&init_mm, pmd, new); new 63 arch/powerpc/mm/kasan/kasan_init_32.c new = NULL; new 67 arch/powerpc/mm/kasan/kasan_init_32.c if (new && slab_is_available()) new 68 arch/powerpc/mm/kasan/kasan_init_32.c pte_free_kernel(&init_mm, new); new 68 arch/powerpc/oprofile/cell/vma_map.c struct vma_to_fileoffset_map *new = kzalloc(sizeof(*new), GFP_KERNEL); new 70 arch/powerpc/oprofile/cell/vma_map.c if (!new) { new 77 arch/powerpc/oprofile/cell/vma_map.c new->next = map; new 78 arch/powerpc/oprofile/cell/vma_map.c new->vma = vma; new 79 arch/powerpc/oprofile/cell/vma_map.c new->size = size; new 80 arch/powerpc/oprofile/cell/vma_map.c new->offset = offset; new 81 arch/powerpc/oprofile/cell/vma_map.c new->guard_ptr = guard_ptr; new 82 arch/powerpc/oprofile/cell/vma_map.c new->guard_val = guard_val; new 84 arch/powerpc/oprofile/cell/vma_map.c return new; new 1115 arch/powerpc/oprofile/op_model_cell.c if ((val == CPUFREQ_PRECHANGE && frq->old < frq->new) || new 1116 arch/powerpc/oprofile/op_model_cell.c (val == CPUFREQ_POSTCHANGE && frq->old > frq->new)) new 1117 arch/powerpc/oprofile/op_model_cell.c set_spu_profiling_frequency(frq->new, spu_cycle_reset); new 562 arch/powerpc/perf/hv-24x7.c struct rb_node **new = &(root->rb_node), *parent = NULL; new 566 arch/powerpc/perf/hv-24x7.c while (*new) { new 570 arch/powerpc/perf/hv-24x7.c it = rb_entry(*new, struct event_uniq, node); new 574 arch/powerpc/perf/hv-24x7.c parent = *new; new 576 arch/powerpc/perf/hv-24x7.c new = &((*new)->rb_left); new 578 arch/powerpc/perf/hv-24x7.c new = &((*new)->rb_right); new 599 arch/powerpc/perf/hv-24x7.c rb_link_node(&data->node, parent, new); new 853 arch/powerpc/platforms/cell/spufs/sched.c struct spu_context *new = NULL; new 856 arch/powerpc/platforms/cell/spufs/sched.c new = grab_runnable_context(max_prio, spu->node); new 857 arch/powerpc/platforms/cell/spufs/sched.c if (new || force) { new 858 arch/powerpc/platforms/cell/spufs/sched.c spu_unschedule(spu, ctx, new == NULL); new 859 arch/powerpc/platforms/cell/spufs/sched.c if (new) { new 860 arch/powerpc/platforms/cell/spufs/sched.c if (new->flags & SPU_CREATE_NOSCHED) new 861 arch/powerpc/platforms/cell/spufs/sched.c wake_up(&new->stop_wq); new 864 arch/powerpc/platforms/cell/spufs/sched.c spu_schedule(spu, new); new 873 arch/powerpc/platforms/cell/spufs/sched.c return new != NULL; new 909 arch/powerpc/platforms/cell/spufs/sched.c struct spu_context *new = NULL; new 929 arch/powerpc/platforms/cell/spufs/sched.c new = grab_runnable_context(ctx->prio + 1, spu->node); new 930 arch/powerpc/platforms/cell/spufs/sched.c if (new) { new 942 arch/powerpc/platforms/cell/spufs/sched.c if (new) new 943 arch/powerpc/platforms/cell/spufs/sched.c spu_schedule(spu, new); new 351 arch/powerpc/platforms/cell/spufs/spufs.h extern int spu_restore(struct spu_state *new, struct spu *spu); new 352 arch/powerpc/platforms/cell/spufs/spufs.h extern int spu_switch(struct spu_state *prev, struct spu_state *new, new 2118 arch/powerpc/platforms/cell/spufs/switch.c int spu_restore(struct spu_state *new, struct spu *spu) new 2125 arch/powerpc/platforms/cell/spufs/switch.c rc = __do_spu_restore(new, spu); new 262 arch/powerpc/platforms/powernv/idle.c u64 new, tmp; new 268 arch/powerpc/platforms/powernv/idle.c new = (s | thread) & ~PNV_CORE_IDLE_LOCK_BIT; new 269 arch/powerpc/platforms/powernv/idle.c tmp = cmpxchg(state, s, new); new 303 arch/powerpc/platforms/powernv/opal-core.c struct opalcore *new; new 397 arch/powerpc/platforms/powernv/opal-core.c new = get_new_element(); new 398 arch/powerpc/platforms/powernv/opal-core.c if (!new) new 400 arch/powerpc/platforms/powernv/opal-core.c new->paddr = oc_conf->ptload_addr[i]; new 401 arch/powerpc/platforms/powernv/opal-core.c new->size = oc_conf->ptload_size[i]; new 402 arch/powerpc/platforms/powernv/opal-core.c new->offset = opalcore_off; new 403 arch/powerpc/platforms/powernv/opal-core.c list_add_tail(&new->list, &opalcore_list); new 171 arch/powerpc/platforms/pseries/reconfig.c struct property *new = kzalloc(sizeof(*new), GFP_KERNEL); new 173 arch/powerpc/platforms/pseries/reconfig.c if (!new) new 176 arch/powerpc/platforms/pseries/reconfig.c if (!(new->name = kstrdup(name, GFP_KERNEL))) new 178 arch/powerpc/platforms/pseries/reconfig.c if (!(new->value = kmalloc(length + 1, GFP_KERNEL))) new 181 arch/powerpc/platforms/pseries/reconfig.c memcpy(new->value, value, length); new 182 arch/powerpc/platforms/pseries/reconfig.c *(((char *)new->value) + length) = 0; new 183 arch/powerpc/platforms/pseries/reconfig.c new->length = length; new 184 arch/powerpc/platforms/pseries/reconfig.c new->next = last; new 185 arch/powerpc/platforms/pseries/reconfig.c return new; new 188 arch/powerpc/platforms/pseries/reconfig.c kfree(new->name); new 189 arch/powerpc/platforms/pseries/reconfig.c kfree(new->value); new 190 arch/powerpc/platforms/pseries/reconfig.c kfree(new); new 42 arch/powerpc/platforms/pseries/vphn.c u16 new = be16_to_cpup(field++); new 50 arch/powerpc/platforms/pseries/vphn.c cpu_to_be32(last << 16 | new); new 52 arch/powerpc/platforms/pseries/vphn.c } else if (new == VPHN_FIELD_UNUSED) new 55 arch/powerpc/platforms/pseries/vphn.c else if (new & VPHN_FIELD_MSB) { new 58 arch/powerpc/platforms/pseries/vphn.c cpu_to_be32(new & VPHN_FIELD_MASK); new 64 arch/powerpc/platforms/pseries/vphn.c last = new; new 14 arch/riscv/include/asm/cmpxchg.h #define __xchg_relaxed(ptr, new, size) \ new 17 arch/riscv/include/asm/cmpxchg.h __typeof__(new) __new = (new); \ new 47 arch/riscv/include/asm/cmpxchg.h #define __xchg_acquire(ptr, new, size) \ new 50 arch/riscv/include/asm/cmpxchg.h __typeof__(new) __new = (new); \ new 82 arch/riscv/include/asm/cmpxchg.h #define __xchg_release(ptr, new, size) \ new 85 arch/riscv/include/asm/cmpxchg.h __typeof__(new) __new = (new); \ new 117 arch/riscv/include/asm/cmpxchg.h #define __xchg(ptr, new, size) \ new 120 arch/riscv/include/asm/cmpxchg.h __typeof__(new) __new = (new); \ new 166 arch/riscv/include/asm/cmpxchg.h #define __cmpxchg_relaxed(ptr, old, new, size) \ new 170 arch/riscv/include/asm/cmpxchg.h __typeof__(*(ptr)) __new = (new); \ new 210 arch/riscv/include/asm/cmpxchg.h #define __cmpxchg_acquire(ptr, old, new, size) \ new 214 arch/riscv/include/asm/cmpxchg.h __typeof__(*(ptr)) __new = (new); \ new 256 arch/riscv/include/asm/cmpxchg.h #define __cmpxchg_release(ptr, old, new, size) \ new 260 arch/riscv/include/asm/cmpxchg.h __typeof__(*(ptr)) __new = (new); \ new 302 arch/riscv/include/asm/cmpxchg.h #define __cmpxchg(ptr, old, new, size) \ new 306 arch/riscv/include/asm/cmpxchg.h __typeof__(*(ptr)) __new = (new); \ new 408 arch/riscv/include/asm/uaccess.h #define __cmpxchg_user(ptr, old, new, err, size, lrb, scb) \ new 412 arch/riscv/include/asm/uaccess.h __typeof__(*(ptr)) __new = (new); \ new 441 arch/riscv/include/asm/uaccess.h [new] "rJ" (__new), \ new 467 arch/riscv/include/asm/uaccess.h [new] "rJ" (__new), \ new 78 arch/s390/include/asm/atomic.h #define atomic_xchg(v, new) (xchg(&((v)->counter), new)) new 80 arch/s390/include/asm/atomic.h static inline int atomic_cmpxchg(atomic_t *v, int old, int new) new 82 arch/s390/include/asm/atomic.h return __atomic_cmpxchg(&v->counter, old, new); new 125 arch/s390/include/asm/atomic.h #define atomic64_xchg(v, new) (xchg(&((v)->counter), new)) new 127 arch/s390/include/asm/atomic.h static inline s64 atomic64_cmpxchg(atomic64_t *v, s64 old, s64 new) new 129 arch/s390/include/asm/atomic.h return __atomic64_cmpxchg((long *)&v->counter, old, new); new 67 arch/s390/include/asm/atomic_ops.h int old, new; \ new 74 arch/s390/include/asm/atomic_ops.h : [old] "=d" (old), [new] "=&d" (new), [ptr] "+Q" (*ptr)\ new 93 arch/s390/include/asm/atomic_ops.h long old, new; \ new 100 arch/s390/include/asm/atomic_ops.h : [old] "=d" (old), [new] "=&d" (new), [ptr] "+Q" (*ptr)\ new 123 arch/s390/include/asm/atomic_ops.h static inline int __atomic_cmpxchg(int *ptr, int old, int new) new 125 arch/s390/include/asm/atomic_ops.h return __sync_val_compare_and_swap(ptr, old, new); new 128 arch/s390/include/asm/atomic_ops.h static inline int __atomic_cmpxchg_bool(int *ptr, int old, int new) new 130 arch/s390/include/asm/atomic_ops.h return __sync_bool_compare_and_swap(ptr, old, new); new 133 arch/s390/include/asm/atomic_ops.h static inline long __atomic64_cmpxchg(long *ptr, long old, long new) new 135 arch/s390/include/asm/atomic_ops.h return __sync_val_compare_and_swap(ptr, old, new); new 138 arch/s390/include/asm/atomic_ops.h static inline long __atomic64_cmpxchg_bool(long *ptr, long old, long new) new 140 arch/s390/include/asm/atomic_ops.h return __sync_bool_compare_and_swap(ptr, old, new); new 48 arch/s390/include/asm/cmpxchg.h : [new] "d" (__new1), "d" (__new2), \ new 19 arch/s390/include/asm/cputime.h #define cmpxchg_cputime(ptr, old, new) cmpxchg64(ptr, old, new) new 548 arch/s390/include/asm/pgtable.h static inline void csp(unsigned int *ptr, unsigned int old, unsigned int new) new 551 arch/s390/include/asm/pgtable.h register unsigned long reg3 asm("3") = new; new 561 arch/s390/include/asm/pgtable.h static inline void cspg(unsigned long *ptr, unsigned long old, unsigned long new) new 564 arch/s390/include/asm/pgtable.h register unsigned long reg3 asm("3") = new; new 580 arch/s390/include/asm/pgtable.h static inline void crdte(unsigned long old, unsigned long new, new 585 arch/s390/include/asm/pgtable.h register unsigned long reg3 asm("3") = new; new 22 arch/s390/include/asm/preempt.h int old, new; new 26 arch/s390/include/asm/preempt.h new = (old & PREEMPT_NEED_RESCHED) | new 29 arch/s390/include/asm/preempt.h old, new) != old); new 109 arch/s390/kernel/ftrace.c struct ftrace_insn orig, new, old; new 116 arch/s390/kernel/ftrace.c ftrace_generate_nop_insn(&new); new 126 arch/s390/kernel/ftrace.c ftrace_generate_kprobe_nop_insn(&new); new 130 arch/s390/kernel/ftrace.c ftrace_generate_nop_insn(&new); new 135 arch/s390/kernel/ftrace.c s390_kernel_write((void *) rec->ip, &new, sizeof(new)); new 141 arch/s390/kernel/ftrace.c struct ftrace_insn orig, new, old; new 154 arch/s390/kernel/ftrace.c ftrace_generate_kprobe_call_insn(&new); new 158 arch/s390/kernel/ftrace.c ftrace_generate_call_insn(&new, rec->ip); new 163 arch/s390/kernel/ftrace.c s390_kernel_write((void *) rec->ip, &new, sizeof(new)); new 201 arch/s390/kernel/irq.c unsigned long old, new; new 205 arch/s390/kernel/irq.c new = S390_lowcore.async_stack; new 206 arch/s390/kernel/irq.c if (((new - old) >> (PAGE_SHIFT + THREAD_SIZE_ORDER)) != 0) { new 207 arch/s390/kernel/irq.c CALL_ON_STACK(__do_softirq, new, 0); new 33 arch/s390/kernel/jump_label.c struct insn *new) new 37 arch/s390/kernel/jump_label.c unsigned char *ipn = (unsigned char *)new; new 56 arch/s390/kernel/jump_label.c struct insn old, new; new 60 arch/s390/kernel/jump_label.c jump_label_make_branch(entry, &new); new 63 arch/s390/kernel/jump_label.c jump_label_make_nop(entry, &new); new 67 arch/s390/kernel/jump_label.c jump_label_bug(entry, &orignop, &new); new 70 arch/s390/kernel/jump_label.c jump_label_bug(entry, &old, &new); new 72 arch/s390/kernel/jump_label.c s390_kernel_write(code, &new, sizeof(new)); new 316 arch/s390/kernel/perf_cpum_cf.c u64 prev, new; new 321 arch/s390/kernel/perf_cpum_cf.c err = ecctr(event->hw.config, &new); new 330 arch/s390/kernel/perf_cpum_cf.c new = 0; new 332 arch/s390/kernel/perf_cpum_cf.c } while (local64_cmpxchg(&event->hw.prev_count, prev, new) != prev); new 339 arch/s390/kernel/perf_cpum_cf.c u64 prev, new, delta; new 344 arch/s390/kernel/perf_cpum_cf.c err = ecctr(event->hw.config, &new); new 347 arch/s390/kernel/perf_cpum_cf.c } while (local64_cmpxchg(&event->hw.prev_count, prev, new) != prev); new 349 arch/s390/kernel/perf_cpum_cf.c delta = (prev <= new) ? new - prev new 350 arch/s390/kernel/perf_cpum_cf.c : (-1ULL - prev) + new + 1; /* overflow */ new 548 arch/s390/kernel/perf_cpum_cf_events.c struct attribute **new; new 559 arch/s390/kernel/perf_cpum_cf_events.c new = kmalloc_array(j, sizeof(struct attribute *), GFP_KERNEL); new 560 arch/s390/kernel/perf_cpum_cf_events.c if (!new) new 564 arch/s390/kernel/perf_cpum_cf_events.c new[j++] = a[i]; new 566 arch/s390/kernel/perf_cpum_cf_events.c new[j++] = b[i]; new 568 arch/s390/kernel/perf_cpum_cf_events.c new[j++] = c[i]; new 569 arch/s390/kernel/perf_cpum_cf_events.c new[j] = NULL; new 571 arch/s390/kernel/perf_cpum_cf_events.c return new; new 196 arch/s390/kernel/perf_cpum_sf.c unsigned long *new, *tail, *tail_prev = NULL; new 227 arch/s390/kernel/perf_cpum_sf.c new = (unsigned long *) get_zeroed_page(gfp_flags); new 228 arch/s390/kernel/perf_cpum_sf.c if (!new) { new 234 arch/s390/kernel/perf_cpum_sf.c *tail = (unsigned long)(void *) new + 1; new 236 arch/s390/kernel/perf_cpum_sf.c tail = new; new 253 arch/s390/kernel/perf_cpum_sf.c free_page((unsigned long) new); new 260 arch/s390/kernel/perf_cpum_sf.c tail_prev = new = NULL; /* Allocated at least one SBD */ new 1696 arch/s390/kernel/perf_cpum_sf.c unsigned long *new, *tail; new 1745 arch/s390/kernel/perf_cpum_sf.c new = (unsigned long *) get_zeroed_page(GFP_KERNEL); new 1746 arch/s390/kernel/perf_cpum_sf.c if (!new) new 1748 arch/s390/kernel/perf_cpum_sf.c aux->sdbt_index[sfb->num_sdbt++] = (unsigned long)new; new 1750 arch/s390/kernel/perf_cpum_sf.c *tail = (unsigned long)(void *) new + 1; new 1751 arch/s390/kernel/perf_cpum_sf.c tail = new; new 49 arch/s390/kernel/ptrace.c struct per_regs old, new; new 87 arch/s390/kernel/ptrace.c new.control = thread->per_user.control; new 88 arch/s390/kernel/ptrace.c new.start = thread->per_user.start; new 89 arch/s390/kernel/ptrace.c new.end = thread->per_user.end; new 95 arch/s390/kernel/ptrace.c new.control |= PER_EVENT_BRANCH; new 97 arch/s390/kernel/ptrace.c new.control |= PER_EVENT_IFETCH; new 98 arch/s390/kernel/ptrace.c new.control |= PER_CONTROL_SUSPENSION; new 99 arch/s390/kernel/ptrace.c new.control |= PER_EVENT_TRANSACTION_END; new 101 arch/s390/kernel/ptrace.c new.control |= PER_EVENT_IFETCH; new 102 arch/s390/kernel/ptrace.c new.start = 0; new 103 arch/s390/kernel/ptrace.c new.end = -1UL; new 107 arch/s390/kernel/ptrace.c if (!(new.control & PER_EVENT_MASK)) { new 113 arch/s390/kernel/ptrace.c if (memcmp(&new, &old, sizeof(struct per_regs)) != 0) new 114 arch/s390/kernel/ptrace.c __ctl_load(new, 9, 11); new 345 arch/s390/kernel/setup.c unsigned long old, new; new 348 arch/s390/kernel/setup.c new = stack_alloc(); new 349 arch/s390/kernel/setup.c if (!new) new 351 arch/s390/kernel/setup.c S390_lowcore.async_stack = new + STACK_INIT_OFFSET; new 94 arch/s390/kernel/vtime.c static inline u64 update_tsk_timer(unsigned long *tsk_vtime, u64 new) new 98 arch/s390/kernel/vtime.c delta = new - *tsk_vtime; new 99 arch/s390/kernel/vtime.c *tsk_vtime = new; new 279 arch/s390/kvm/gaccess.c union ipte_control old, new, *ic; new 295 arch/s390/kvm/gaccess.c new = old; new 296 arch/s390/kvm/gaccess.c new.k = 1; new 297 arch/s390/kvm/gaccess.c } while (cmpxchg(&ic->val, old.val, new.val) != old.val); new 305 arch/s390/kvm/gaccess.c union ipte_control old, new, *ic; new 315 arch/s390/kvm/gaccess.c new = old; new 316 arch/s390/kvm/gaccess.c new.k = 0; new 317 arch/s390/kvm/gaccess.c } while (cmpxchg(&ic->val, old.val, new.val) != old.val); new 326 arch/s390/kvm/gaccess.c union ipte_control old, new, *ic; new 338 arch/s390/kvm/gaccess.c new = old; new 339 arch/s390/kvm/gaccess.c new.k = 1; new 340 arch/s390/kvm/gaccess.c new.kh++; new 341 arch/s390/kvm/gaccess.c } while (cmpxchg(&ic->val, old.val, new.val) != old.val); new 347 arch/s390/kvm/gaccess.c union ipte_control old, new, *ic; new 353 arch/s390/kvm/gaccess.c new = old; new 354 arch/s390/kvm/gaccess.c new.kh--; new 355 arch/s390/kvm/gaccess.c if (!new.kh) new 356 arch/s390/kvm/gaccess.c new.k = 0; new 357 arch/s390/kvm/gaccess.c } while (cmpxchg(&ic->val, old.val, new.val) != old.val); new 359 arch/s390/kvm/gaccess.c if (!new.kh) new 875 arch/s390/kvm/kvm-s390.c struct gmap *new = gmap_create(current->mm, new_limit); new 877 arch/s390/kvm/kvm-s390.c if (!new) { new 881 arch/s390/kvm/kvm-s390.c new->private = kvm; new 882 arch/s390/kvm/kvm-s390.c kvm->arch.gmap = new; new 4530 arch/s390/kvm/kvm-s390.c const struct kvm_memory_slot *new, new 84 arch/s390/lib/spinlock.c static inline int arch_cmpxchg_niai8(int *lock, int old, int new) new 92 arch/s390/lib/spinlock.c : "0" (old), "d" (new), "Q" (*lock) new 120 arch/s390/lib/spinlock.c int lockval, ix, node_id, tail_id, old, new, owner, count; new 141 arch/s390/lib/spinlock.c new = (old ? (old + _Q_LOCK_STEAL_ADD) : 0) | lockval; new 142 arch/s390/lib/spinlock.c if (__atomic_cmpxchg_bool(&lp->lock, old, new)) new 149 arch/s390/lib/spinlock.c new = node_id | (old & _Q_LOCK_MASK); new 150 arch/s390/lib/spinlock.c if (__atomic_cmpxchg_bool(&lp->lock, old, new)) new 186 arch/s390/lib/spinlock.c new = ((tail_id != node_id) ? tail_id : 0) | lockval; new 187 arch/s390/lib/spinlock.c if (__atomic_cmpxchg_bool(&lp->lock, old, new)) new 213 arch/s390/lib/spinlock.c int lockval, old, new, owner, count; new 228 arch/s390/lib/spinlock.c new = (old & _Q_TAIL_MASK) | lockval; new 229 arch/s390/lib/spinlock.c if (arch_cmpxchg_niai8(&lp->lock, old, new)) { new 309 arch/s390/mm/gmap.c unsigned long *new; new 315 arch/s390/mm/gmap.c new = (unsigned long *) page_to_phys(page); new 316 arch/s390/mm/gmap.c crst_table_init(new, init); new 320 arch/s390/mm/gmap.c *table = (unsigned long) new | _REGION_ENTRY_LENGTH | new 526 arch/s390/mm/gmap.c static void gmap_pmdp_xchg(struct gmap *gmap, pmd_t *old, pmd_t new, new 964 arch/s390/mm/gmap.c pmd_t new = *pmdp; new 971 arch/s390/mm/gmap.c pmd_val(new) |= _SEGMENT_ENTRY_INVALID; new 972 arch/s390/mm/gmap.c gmap_pmdp_xchg(gmap, pmdp, new, gaddr); new 976 arch/s390/mm/gmap.c pmd_val(new) &= ~_SEGMENT_ENTRY_INVALID; new 977 arch/s390/mm/gmap.c pmd_val(new) |= _SEGMENT_ENTRY_PROTECT; new 978 arch/s390/mm/gmap.c gmap_pmdp_xchg(gmap, pmdp, new, gaddr); new 1645 arch/s390/mm/gmap.c struct gmap *sg, *new; new 1660 arch/s390/mm/gmap.c new = gmap_alloc(limit); new 1661 arch/s390/mm/gmap.c if (!new) new 1663 arch/s390/mm/gmap.c new->mm = parent->mm; new 1664 arch/s390/mm/gmap.c new->parent = gmap_get(parent); new 1665 arch/s390/mm/gmap.c new->orig_asce = asce; new 1666 arch/s390/mm/gmap.c new->edat_level = edat_level; new 1667 arch/s390/mm/gmap.c new->initialized = false; new 1673 arch/s390/mm/gmap.c gmap_free(new); new 1689 arch/s390/mm/gmap.c refcount_set(&new->ref_count, 2); new 1690 arch/s390/mm/gmap.c list_add(&new->list, &parent->children); new 1693 arch/s390/mm/gmap.c new->initialized = true; new 1695 arch/s390/mm/gmap.c return new; new 1705 arch/s390/mm/gmap.c new->initialized = true; new 1707 arch/s390/mm/gmap.c list_del(&new->list); new 1708 arch/s390/mm/gmap.c gmap_free(new); new 1709 arch/s390/mm/gmap.c new = ERR_PTR(rc); new 1712 arch/s390/mm/gmap.c return new; new 2286 arch/s390/mm/gmap.c static void gmap_pmdp_xchg(struct gmap *gmap, pmd_t *pmdp, pmd_t new, new 2291 arch/s390/mm/gmap.c pmd_val(new) &= ~_SEGMENT_ENTRY_GMAP_IN; new 2299 arch/s390/mm/gmap.c *pmdp = new; new 57 arch/s390/mm/pageattr.c static void pgt_set(unsigned long *old, unsigned long new, unsigned long addr, new 76 arch/s390/mm/pageattr.c crdte(*old, new, table, dtt, addr, S390_lowcore.kernel_asce); new 78 arch/s390/mm/pageattr.c cspg(old, *old, new); new 80 arch/s390/mm/pageattr.c csp((unsigned int *)old + 1, *old, new); new 87 arch/s390/mm/pageattr.c pte_t *ptep, new; new 91 arch/s390/mm/pageattr.c new = *ptep; new 92 arch/s390/mm/pageattr.c if (pte_none(new)) new 95 arch/s390/mm/pageattr.c new = pte_wrprotect(new); new 97 arch/s390/mm/pageattr.c new = pte_mkwrite(pte_mkdirty(new)); new 99 arch/s390/mm/pageattr.c pte_val(new) |= _PAGE_NOEXEC; new 101 arch/s390/mm/pageattr.c pte_val(new) &= ~_PAGE_NOEXEC; new 102 arch/s390/mm/pageattr.c pgt_set((unsigned long *)ptep, pte_val(new), addr, CRDTE_DTT_PAGE); new 114 arch/s390/mm/pageattr.c pmd_t new; new 132 arch/s390/mm/pageattr.c pmd_val(new) = __pa(pt_dir) | _SEGMENT_ENTRY; new 133 arch/s390/mm/pageattr.c pgt_set((unsigned long *)pmdp, pmd_val(new), addr, CRDTE_DTT_SEGMENT); new 142 arch/s390/mm/pageattr.c pmd_t new = *pmdp; new 145 arch/s390/mm/pageattr.c new = pmd_wrprotect(new); new 147 arch/s390/mm/pageattr.c new = pmd_mkwrite(pmd_mkdirty(new)); new 149 arch/s390/mm/pageattr.c pmd_val(new) |= _SEGMENT_ENTRY_NOEXEC; new 151 arch/s390/mm/pageattr.c pmd_val(new) &= ~_SEGMENT_ENTRY_NOEXEC; new 152 arch/s390/mm/pageattr.c pgt_set((unsigned long *)pmdp, pmd_val(new), addr, CRDTE_DTT_SEGMENT); new 191 arch/s390/mm/pageattr.c pud_t new; new 209 arch/s390/mm/pageattr.c pud_val(new) = __pa(pm_dir) | _REGION3_ENTRY; new 210 arch/s390/mm/pageattr.c pgt_set((unsigned long *)pudp, pud_val(new), addr, CRDTE_DTT_REGION3); new 219 arch/s390/mm/pageattr.c pud_t new = *pudp; new 222 arch/s390/mm/pageattr.c new = pud_wrprotect(new); new 224 arch/s390/mm/pageattr.c new = pud_mkwrite(pud_mkdirty(new)); new 226 arch/s390/mm/pageattr.c pud_val(new) |= _REGION_ENTRY_NOEXEC; new 228 arch/s390/mm/pageattr.c pud_val(new) &= ~_REGION_ENTRY_NOEXEC; new 229 arch/s390/mm/pageattr.c pgt_set((unsigned long *)pudp, pud_val(new), addr, CRDTE_DTT_REGION3); new 157 arch/s390/mm/pgalloc.c unsigned int old, new; new 161 arch/s390/mm/pgalloc.c new = old ^ bits; new 162 arch/s390/mm/pgalloc.c } while (atomic_cmpxchg(v, old, new) != old); new 163 arch/s390/mm/pgalloc.c return new; new 110 arch/s390/mm/pgtable.c unsigned long new = 0; new 121 arch/s390/mm/pgtable.c : "=&d" (old), "=&d" (new), "=Q" (ptep[PTRS_PER_PTE]) new 124 arch/s390/mm/pgtable.c return __pgste(new); new 251 arch/s390/mm/pgtable.c pgste_t pgste, pte_t old, pte_t new) new 255 arch/s390/mm/pgtable.c pgste_set_key(ptep, pgste, new, mm); new 256 arch/s390/mm/pgtable.c if (pte_val(new) & _PAGE_INVALID) { new 262 arch/s390/mm/pgtable.c pgste = pgste_set_pte(ptep, pgste, new); new 265 arch/s390/mm/pgtable.c *ptep = new; new 271 arch/s390/mm/pgtable.c pte_t *ptep, pte_t new) new 281 arch/s390/mm/pgtable.c old = ptep_xchg_commit(mm, addr, ptep, pgste, old, new); new 288 arch/s390/mm/pgtable.c pte_t *ptep, pte_t new) new 298 arch/s390/mm/pgtable.c old = ptep_xchg_commit(mm, addr, ptep, pgste, old, new); new 434 arch/s390/mm/pgtable.c pmd_t *pmdp, pmd_t new) new 440 arch/s390/mm/pgtable.c *pmdp = new; new 447 arch/s390/mm/pgtable.c pmd_t *pmdp, pmd_t new) new 453 arch/s390/mm/pgtable.c *pmdp = new; new 504 arch/s390/mm/pgtable.c pud_t *pudp, pud_t new) new 510 arch/s390/mm/pgtable.c *pudp = new; new 759 arch/s390/mm/pgtable.c pgste_t old, new; new 790 arch/s390/mm/pgtable.c new = old = pgste_get_lock(ptep); new 791 arch/s390/mm/pgtable.c pgste_val(new) &= ~(PGSTE_GR_BIT | PGSTE_GC_BIT | new 794 arch/s390/mm/pgtable.c pgste_val(new) |= (keyul & (_PAGE_CHANGED | _PAGE_REFERENCED)) << 48; new 795 arch/s390/mm/pgtable.c pgste_val(new) |= (keyul & (_PAGE_ACC_BITS | _PAGE_FP_BIT)) << 56; new 806 arch/s390/mm/pgtable.c pgste_val(new) |= bits << 52; new 809 arch/s390/mm/pgtable.c if ((pgste_val(new) ^ pgste_val(old)) & new 811 arch/s390/mm/pgtable.c pgste_val(new) |= PGSTE_UC_BIT; new 813 arch/s390/mm/pgtable.c pgste_set_unlock(ptep, new); new 861 arch/s390/mm/pgtable.c pgste_t old, new; new 889 arch/s390/mm/pgtable.c new = old = pgste_get_lock(ptep); new 891 arch/s390/mm/pgtable.c pgste_val(new) &= ~PGSTE_GR_BIT; new 897 arch/s390/mm/pgtable.c pgste_val(new) |= ((unsigned long) cc << 53) & PGSTE_HR_BIT; new 902 arch/s390/mm/pgtable.c if ((pgste_val(new) ^ pgste_val(old)) & PGSTE_GR_BIT) new 903 arch/s390/mm/pgtable.c pgste_val(new) |= PGSTE_UC_BIT; new 905 arch/s390/mm/pgtable.c pgste_set_unlock(ptep, new); new 1075 arch/s390/mm/pgtable.c pgste_t new; new 1081 arch/s390/mm/pgtable.c new = pgste_get_lock(ptep); new 1083 arch/s390/mm/pgtable.c pgste_val(new) &= ~bits; new 1084 arch/s390/mm/pgtable.c pgste_val(new) |= value & bits; new 1086 arch/s390/mm/pgtable.c pgste_set_unlock(ptep, new); new 27 arch/s390/pci/pci_clp.c static void update_uid_checking(bool new) new 29 arch/s390/pci/pci_clp.c if (zpci_unique_uid != new) new 30 arch/s390/pci/pci_clp.c zpci_dbg(1, "uid checking:%d\n", new); new 32 arch/s390/pci/pci_clp.c zpci_unique_uid = new; new 35 arch/sh/drivers/heartbeat.c unsigned int new; new 37 arch/sh/drivers/heartbeat.c new = (1 << hd->bit_pos[bit]); new 39 arch/sh/drivers/heartbeat.c new = ~new; new 41 arch/sh/drivers/heartbeat.c new &= hd->mask; new 45 arch/sh/drivers/heartbeat.c new |= ioread32(hd->base) & ~hd->mask; new 46 arch/sh/drivers/heartbeat.c iowrite32(new, hd->base); new 49 arch/sh/drivers/heartbeat.c new |= ioread16(hd->base) & ~hd->mask; new 50 arch/sh/drivers/heartbeat.c iowrite16(new, hd->base); new 53 arch/sh/drivers/heartbeat.c new |= ioread8(hd->base) & ~hd->mask; new 54 arch/sh/drivers/heartbeat.c iowrite8(new, hd->base); new 35 arch/sh/include/asm/atomic.h #define atomic_xchg(v, new) (xchg(&((v)->counter), new)) new 5 arch/sh/include/asm/bitops-cas.h static inline unsigned __bo_cas(volatile unsigned *p, unsigned old, unsigned new) new 8 arch/sh/include/asm/bitops-cas.h : "+r"(new) new 11 arch/sh/include/asm/bitops-cas.h return new; new 6 arch/sh/include/asm/cmpxchg-cas.h __cmpxchg_u32(volatile u32 *m, unsigned long old, unsigned long new) new 9 arch/sh/include/asm/cmpxchg-cas.h : "+r"(new) new 12 arch/sh/include/asm/cmpxchg-cas.h return new; new 72 arch/sh/include/asm/cmpxchg-grb.h unsigned long new) new 88 arch/sh/include/asm/cmpxchg-grb.h "+r" (old), "+r" (new) /* old or new can be r15 */ new 41 arch/sh/include/asm/cmpxchg-irq.h unsigned long new) new 49 arch/sh/include/asm/cmpxchg-irq.h *m = new; new 27 arch/sh/include/asm/cmpxchg-llsc.h __cmpxchg_u32(volatile u32 *m, unsigned long old, unsigned long new) new 44 arch/sh/include/asm/cmpxchg-llsc.h : "r" (m), "r" (old), "r" (new) new 56 arch/sh/include/asm/cmpxchg.h unsigned long new, int size) new 60 arch/sh/include/asm/cmpxchg.h return __cmpxchg_u32(ptr, old, new); new 13 arch/sh/include/asm/spinlock-cas.h static inline unsigned __sl_cas(volatile unsigned *p, unsigned old, unsigned new) new 16 arch/sh/include/asm/spinlock-cas.h : "+r"(new) new 19 arch/sh/include/asm/spinlock-cas.h return new; new 105 arch/sh/kernel/ftrace.c int new = old & ~MOD_CODE_WRITE_FLAG; new 107 arch/sh/kernel/ftrace.c if (old == new) new 110 arch/sh/kernel/ftrace.c old = atomic_cmpxchg(&nmi_running, old, new); new 225 arch/sh/kernel/ftrace.c unsigned char old[MCOUNT_INSN_SIZE], *new; new 228 arch/sh/kernel/ftrace.c new = ftrace_call_replace(ip, (unsigned long)func); new 230 arch/sh/kernel/ftrace.c return ftrace_modify_code(ip, old, new); new 236 arch/sh/kernel/ftrace.c unsigned char *new, *old; new 240 arch/sh/kernel/ftrace.c new = ftrace_nop_replace(ip); new 242 arch/sh/kernel/ftrace.c return ftrace_modify_code(rec->ip, old, new); new 247 arch/sh/kernel/ftrace.c unsigned char *new, *old; new 251 arch/sh/kernel/ftrace.c new = ftrace_call_replace(ip, addr); new 253 arch/sh/kernel/ftrace.c return ftrace_modify_code(rec->ip, old, new); new 55 arch/sparc/include/asm/atomic_64.h static inline int atomic_xchg(atomic_t *v, int new) new 57 arch/sparc/include/asm/atomic_64.h return xchg(&v->counter, new); new 62 arch/sparc/include/asm/atomic_64.h #define atomic64_xchg(v, new) (xchg(&((v)->counter), new)) new 15 arch/sparc/include/asm/cmpxchg_32.h unsigned long __xchg_u32(volatile u32 *m, u32 new); new 66 arch/sparc/include/asm/cmpxchg_32.h u64 __cmpxchg_u64(u64 *ptr, u64 old, u64 new); new 67 arch/sparc/include/asm/cmpxchg_32.h #define cmpxchg64(ptr, old, new) __cmpxchg_u64(ptr, old, new) new 11 arch/sparc/include/asm/cmpxchg_64.h __cmpxchg_u32(volatile int *m, int old, int new) new 14 arch/sparc/include/asm/cmpxchg_64.h : "=&r" (new) new 15 arch/sparc/include/asm/cmpxchg_64.h : "0" (new), "r" (m), "r" (old) new 18 arch/sparc/include/asm/cmpxchg_64.h return new; new 115 arch/sparc/include/asm/cmpxchg_64.h __cmpxchg_u64(volatile long *m, unsigned long old, unsigned long new) new 118 arch/sparc/include/asm/cmpxchg_64.h : "=&r" (new) new 119 arch/sparc/include/asm/cmpxchg_64.h : "0" (new), "r" (m), "r" (old) new 122 arch/sparc/include/asm/cmpxchg_64.h return new; new 131 arch/sparc/include/asm/cmpxchg_64.h __cmpxchg_u8(volatile unsigned char *m, unsigned char old, unsigned char new) new 141 arch/sparc/include/asm/cmpxchg_64.h new32 = (load32 & ~mask) | (new << bit_shift); new 157 arch/sparc/include/asm/cmpxchg_64.h __cmpxchg(volatile void *ptr, unsigned long old, unsigned long new, int size) new 161 arch/sparc/include/asm/cmpxchg_64.h return __cmpxchg_u8(ptr, old, new); new 163 arch/sparc/include/asm/cmpxchg_64.h return __cmpxchg_u32(ptr, old, new); new 165 arch/sparc/include/asm/cmpxchg_64.h return __cmpxchg_u64(ptr, old, new); new 186 arch/sparc/include/asm/cmpxchg_64.h unsigned long new, int size) new 190 arch/sparc/include/asm/cmpxchg_64.h case 8: return __cmpxchg(ptr, old, new, size); new 192 arch/sparc/include/asm/cmpxchg_64.h return __cmpxchg_local_generic(ptr, old, new, size); new 26 arch/sparc/kernel/ftrace.c static int ftrace_modify_code(unsigned long ip, u32 old, u32 new) new 47 arch/sparc/kernel/ftrace.c : [new] "0" (new), [old] "r" (old), [ip] "r" (ip) new 50 arch/sparc/kernel/ftrace.c if (replaced != old && replaced != new) new 59 arch/sparc/kernel/ftrace.c u32 old, new; new 62 arch/sparc/kernel/ftrace.c new = ftrace_nop; new 63 arch/sparc/kernel/ftrace.c return ftrace_modify_code(ip, old, new); new 69 arch/sparc/kernel/ftrace.c u32 old, new; new 72 arch/sparc/kernel/ftrace.c new = ftrace_call_replace(ip, addr); new 73 arch/sparc/kernel/ftrace.c return ftrace_modify_code(ip, old, new); new 79 arch/sparc/kernel/ftrace.c u32 old, new; new 82 arch/sparc/kernel/ftrace.c new = ftrace_call_replace(ip, (unsigned long)func); new 83 arch/sparc/kernel/ftrace.c return ftrace_modify_code(ip, old, new); new 100 arch/sparc/kernel/ftrace.c u32 old, new; new 103 arch/sparc/kernel/ftrace.c new = ftrace_call_replace(ip, (unsigned long) &ftrace_graph_caller); new 104 arch/sparc/kernel/ftrace.c return ftrace_modify_code(ip, old, new); new 110 arch/sparc/kernel/ftrace.c u32 old, new; new 113 arch/sparc/kernel/ftrace.c new = ftrace_call_replace(ip, (unsigned long) &ftrace_stub); new 115 arch/sparc/kernel/ftrace.c return ftrace_modify_code(ip, old, new); new 860 arch/sparc/kernel/ldc.c unsigned long new; new 889 arch/sparc/kernel/ldc.c new = lp->rx_head; new 890 arch/sparc/kernel/ldc.c new += LDC_PACKET_SIZE; new 891 arch/sparc/kernel/ldc.c if (new == (lp->rx_num_entries * LDC_PACKET_SIZE)) new 892 arch/sparc/kernel/ldc.c new = 0; new 893 arch/sparc/kernel/ldc.c lp->rx_head = new; new 895 arch/sparc/kernel/ldc.c err = __set_rx_head(lp, new); new 1533 arch/sparc/kernel/ldc.c unsigned long hv_err, new; new 1556 arch/sparc/kernel/ldc.c new = rx_advance(lp, lp->rx_head); new 1557 arch/sparc/kernel/ldc.c lp->rx_head = new; new 1559 arch/sparc/kernel/ldc.c err = __set_rx_head(lp, new); new 1742 arch/sparc/kernel/ldc.c unsigned long hv_err, new; new 1761 arch/sparc/kernel/ldc.c new = lp->rx_head; new 1766 arch/sparc/kernel/ldc.c BUG_ON(new == lp->rx_tail); new 1767 arch/sparc/kernel/ldc.c p = lp->rx_base + (new / LDC_PACKET_SIZE); new 1800 arch/sparc/kernel/ldc.c new = rx_advance(lp, new); new 1809 arch/sparc/kernel/ldc.c new = rx_advance(lp, new); new 1810 arch/sparc/kernel/ldc.c err = rx_set_head(lp, new); new 1834 arch/sparc/kernel/ldc.c new = rx_advance(lp, new); new 1836 arch/sparc/kernel/ldc.c err = rx_set_head(lp, new); new 1864 arch/sparc/kernel/ldc.c new = rx_advance(lp, new); new 1876 arch/sparc/kernel/ldc.c if (new == lp->rx_tail) { new 1877 arch/sparc/kernel/ldc.c err = rx_data_wait(lp, new); new 1884 arch/sparc/kernel/ldc.c err = rx_set_head(lp, new); new 2065 arch/sparc/kernel/ldc.c unsigned long new = (base + PAGE_SIZE) & PAGE_MASK; new 2067 arch/sparc/kernel/ldc.c len -= (new - base); new 2068 arch/sparc/kernel/ldc.c base = new; new 2088 arch/sparc/kernel/ldc.c unsigned long tlen, new = pa + PAGE_SIZE; new 2116 arch/sparc/kernel/ldc.c pa = new; new 1628 arch/sparc/kernel/smp_64.c pud_t *new; new 1630 arch/sparc/kernel/smp_64.c new = memblock_alloc_from(PAGE_SIZE, PAGE_SIZE, PAGE_SIZE); new 1631 arch/sparc/kernel/smp_64.c if (!new) new 1633 arch/sparc/kernel/smp_64.c pgd_populate(&init_mm, pgd, new); new 1638 arch/sparc/kernel/smp_64.c pmd_t *new; new 1640 arch/sparc/kernel/smp_64.c new = memblock_alloc_from(PAGE_SIZE, PAGE_SIZE, PAGE_SIZE); new 1641 arch/sparc/kernel/smp_64.c if (!new) new 1643 arch/sparc/kernel/smp_64.c pud_populate(&init_mm, pud, new); new 1648 arch/sparc/kernel/smp_64.c pte_t *new; new 1650 arch/sparc/kernel/smp_64.c new = memblock_alloc_from(PAGE_SIZE, PAGE_SIZE, PAGE_SIZE); new 1651 arch/sparc/kernel/smp_64.c if (!new) new 1653 arch/sparc/kernel/smp_64.c pmd_populate_kernel(&init_mm, pmd, new); new 667 arch/sparc/kernel/time_64.c if ((val == CPUFREQ_PRECHANGE && freq->old < freq->new) || new 668 arch/sparc/kernel/time_64.c (val == CPUFREQ_POSTCHANGE && freq->old > freq->new)) { new 671 arch/sparc/kernel/time_64.c freq->new); new 70 arch/sparc/lib/atomic32.c int atomic_xchg(atomic_t *v, int new) new 77 arch/sparc/lib/atomic32.c v->counter = new; new 83 arch/sparc/lib/atomic32.c int atomic_cmpxchg(atomic_t *v, int old, int new) new 91 arch/sparc/lib/atomic32.c v->counter = new; new 162 arch/sparc/lib/atomic32.c unsigned long __cmpxchg_u32(volatile u32 *ptr, u32 old, u32 new) new 169 arch/sparc/lib/atomic32.c *ptr = new; new 176 arch/sparc/lib/atomic32.c u64 __cmpxchg_u64(u64 *ptr, u64 old, u64 new) new 183 arch/sparc/lib/atomic32.c *ptr = new; new 190 arch/sparc/lib/atomic32.c unsigned long __xchg_u32(volatile u32 *ptr, u32 new) new 197 arch/sparc/lib/atomic32.c *ptr = new; new 1808 arch/sparc/mm/init_64.c pud_t *new; new 1810 arch/sparc/mm/init_64.c new = memblock_alloc_from(PAGE_SIZE, PAGE_SIZE, new 1812 arch/sparc/mm/init_64.c if (!new) new 1815 arch/sparc/mm/init_64.c pgd_populate(&init_mm, pgd, new); new 1819 arch/sparc/mm/init_64.c pmd_t *new; new 1825 arch/sparc/mm/init_64.c new = memblock_alloc_from(PAGE_SIZE, PAGE_SIZE, new 1827 arch/sparc/mm/init_64.c if (!new) new 1830 arch/sparc/mm/init_64.c pud_populate(&init_mm, pud, new); new 1835 arch/sparc/mm/init_64.c pte_t *new; new 1841 arch/sparc/mm/init_64.c new = memblock_alloc_from(PAGE_SIZE, PAGE_SIZE, new 1843 arch/sparc/mm/init_64.c if (!new) new 1846 arch/sparc/mm/init_64.c pmd_populate_kernel(&init_mm, pmd, new); new 117 arch/sparc/vdso/vma.c const char *new) new 121 arch/sparc/vdso/vma.c Elf64_Sym *nsym = find_sym64(e, new); new 185 arch/sparc/vdso/vma.c const char *new) new 189 arch/sparc/vdso/vma.c Elf32_Sym *nsym = find_sym32(e, new); new 492 arch/um/drivers/chan_kern.c struct chan *new; new 509 arch/um/drivers/chan_kern.c new = parse_chan(line, in, device, opts, error_out); new 510 arch/um/drivers/chan_kern.c if (new == NULL) new 513 arch/um/drivers/chan_kern.c new->input = 1; new 514 arch/um/drivers/chan_kern.c list_add(&new->list, chans); new 515 arch/um/drivers/chan_kern.c line->chan_in = new; new 517 arch/um/drivers/chan_kern.c new = parse_chan(line, out, device, opts, error_out); new 518 arch/um/drivers/chan_kern.c if (new == NULL) new 521 arch/um/drivers/chan_kern.c list_add(&new->list, chans); new 522 arch/um/drivers/chan_kern.c new->output = 1; new 523 arch/um/drivers/chan_kern.c line->chan_out = new; new 526 arch/um/drivers/chan_kern.c new = parse_chan(line, str, device, opts, error_out); new 527 arch/um/drivers/chan_kern.c if (new == NULL) new 530 arch/um/drivers/chan_kern.c list_add(&new->list, chans); new 531 arch/um/drivers/chan_kern.c new->input = 1; new 532 arch/um/drivers/chan_kern.c new->output = 1; new 533 arch/um/drivers/chan_kern.c line->chan_in = line->chan_out = new; new 77 arch/um/drivers/chan_user.c struct termios save, new; new 89 arch/um/drivers/chan_user.c new = save; new 95 arch/um/drivers/chan_user.c new.c_oflag |= OPOST; new 96 arch/um/drivers/chan_user.c CATCH_EINTR(err = tcsetattr(fd, TCSAFLUSH, &new)); new 387 arch/um/drivers/line.c char *new = kstrdup(init, GFP_KERNEL); new 388 arch/um/drivers/line.c if (!new) { new 396 arch/um/drivers/line.c line->init_str = new; new 398 arch/um/drivers/line.c err = parse_chan_pair(new, line, n, opts, error_out); new 411 arch/um/drivers/line.c kfree(new); new 79 arch/um/drivers/mconsole_kern.c struct mconsole_entry *new; new 87 arch/um/drivers/mconsole_kern.c new = kmalloc(sizeof(*new), GFP_NOWAIT); new 88 arch/um/drivers/mconsole_kern.c if (new == NULL) new 91 arch/um/drivers/mconsole_kern.c new->request = req; new 92 arch/um/drivers/mconsole_kern.c new->request.regs = get_irq_regs()->regs; new 93 arch/um/drivers/mconsole_kern.c list_add(&new->list, &mc_requests); new 248 arch/um/drivers/mconsole_kern.c void mconsole_register_dev(struct mc_device *new) new 251 arch/um/drivers/mconsole_kern.c BUG_ON(!list_empty(&new->list)); new 252 arch/um/drivers/mconsole_kern.c list_add(&new->list, &mconsole_devices); new 42 arch/um/drivers/mconsole_kern.h extern void mconsole_register_dev(struct mc_device *new); new 46 arch/um/drivers/mconsole_kern.h static inline void mconsole_register_dev(struct mc_device *new) new 47 arch/um/drivers/net_kern.c struct sk_buff *new; new 57 arch/um/drivers/net_kern.c new = dev_alloc_skb(max); new 58 arch/um/drivers/net_kern.c if (new == NULL) new 61 arch/um/drivers/net_kern.c skb_put(new, max); new 64 arch/um/drivers/net_kern.c drop_skb = new; new 585 arch/um/drivers/net_kern.c void register_transport(struct transport *new) new 594 arch/um/drivers/net_kern.c BUG_ON(!list_empty(&new->list)); new 595 arch/um/drivers/net_kern.c list_add(&new->list, &transports); new 600 arch/um/drivers/net_kern.c match = check_transport(new, eth->init, eth->index, &init, new 605 arch/um/drivers/net_kern.c eth_configure(eth->index, init, mac, new, GFP_KERNEL); new 640 arch/um/drivers/net_kern.c struct eth_init *new; new 651 arch/um/drivers/net_kern.c new = memblock_alloc(sizeof(*new), SMP_CACHE_BYTES); new 652 arch/um/drivers/net_kern.c if (!new) new 654 arch/um/drivers/net_kern.c sizeof(*new)); new 656 arch/um/drivers/net_kern.c INIT_LIST_HEAD(&new->list); new 657 arch/um/drivers/net_kern.c new->index = n; new 658 arch/um/drivers/net_kern.c new->init = str; new 660 arch/um/drivers/net_kern.c list_add_tail(&new->list, ð_cmd_line); new 169 arch/um/drivers/port_user.c int new, err; new 174 arch/um/drivers/port_user.c new = accept(fd, NULL, 0); new 175 arch/um/drivers/port_user.c if (new < 0) new 183 arch/um/drivers/port_user.c { .sock_fd = new, new 191 arch/um/drivers/port_user.c return new; new 199 arch/um/drivers/port_user.c close(new); new 1600 arch/um/drivers/vector_kern.c struct vector_cmd_line_arg *new; new 1608 arch/um/drivers/vector_kern.c new = memblock_alloc(sizeof(*new), SMP_CACHE_BYTES); new 1609 arch/um/drivers/vector_kern.c if (!new) new 1611 arch/um/drivers/vector_kern.c sizeof(*new)); new 1612 arch/um/drivers/vector_kern.c INIT_LIST_HEAD(&new->list); new 1613 arch/um/drivers/vector_kern.c new->unit = n; new 1614 arch/um/drivers/vector_kern.c new->arguments = str; new 1615 arch/um/drivers/vector_kern.c list_add_tail(&new->list, &vec_cmd_line); new 88 arch/um/drivers/xterm.c int pid, fd, new, err; new 152 arch/um/drivers/xterm.c new = xterm_fd(fd, &data->helper_pid); new 153 arch/um/drivers/xterm.c if (new < 0) { new 154 arch/um/drivers/xterm.c err = new; new 160 arch/um/drivers/xterm.c err = os_set_fd_block(new, 0); new 167 arch/um/drivers/xterm.c CATCH_EINTR(err = tcgetattr(new, &data->tt)); new 169 arch/um/drivers/xterm.c new = err; new 174 arch/um/drivers/xterm.c err = raw(new); new 176 arch/um/drivers/xterm.c new = err; new 185 arch/um/drivers/xterm.c return new; new 188 arch/um/drivers/xterm.c close(new); new 48 arch/um/include/asm/mmu_context.h static inline void activate_mm(struct mm_struct *old, struct mm_struct *new) new 54 arch/um/include/asm/mmu_context.h __switch_mm(&new->context.id); new 55 arch/um/include/asm/mmu_context.h down_write_nested(&new->mmap_sem, 1); new 56 arch/um/include/asm/mmu_context.h uml_setup_stubs(new); new 57 arch/um/include/asm/mmu_context.h up_write(&new->mmap_sem); new 39 arch/um/include/asm/ptrace-generic.h extern int arch_set_tls(struct task_struct *new, unsigned long tls); new 66 arch/um/include/shared/net_kern.h extern void register_transport(struct transport *new); new 463 arch/um/os-Linux/file.c int new; new 465 arch/um/os-Linux/file.c new = accept(fd, NULL, 0); new 466 arch/um/os-Linux/file.c if (new < 0) new 468 arch/um/os-Linux/file.c return new; new 504 arch/um/os-Linux/file.c int new, n; new 505 arch/um/os-Linux/file.c char buf[CMSG_SPACE(sizeof(new))]; new 538 arch/um/os-Linux/file.c new = ((int *) CMSG_DATA(cmsg))[0]; new 539 arch/um/os-Linux/file.c return new; new 105 arch/um/os-Linux/sigio.c struct pollfd *new; new 110 arch/um/os-Linux/sigio.c new = uml_kmalloc(n * sizeof(struct pollfd), UM_GFP_ATOMIC); new 111 arch/um/os-Linux/sigio.c if (new == NULL) { new 117 arch/um/os-Linux/sigio.c memcpy(new, polls->poll, polls->used * sizeof(struct pollfd)); new 120 arch/um/os-Linux/sigio.c polls->poll = new; new 431 arch/um/os-Linux/sigio.c struct sigaction old, new; new 472 arch/um/os-Linux/sigio.c new = old; new 473 arch/um/os-Linux/sigio.c new.sa_handler = handler; new 474 arch/um/os-Linux/sigio.c if (sigaction(SIGIO, &new, NULL) < 0) { new 346 arch/um/os-Linux/start_up.c struct iomem_region *new; new 370 arch/um/os-Linux/start_up.c new = malloc(sizeof(*new)); new 371 arch/um/os-Linux/start_up.c if (new == NULL) { new 378 arch/um/os-Linux/start_up.c *new = ((struct iomem_region) { .next = iomem_regions, new 384 arch/um/os-Linux/start_up.c iomem_regions = new; new 385 arch/um/os-Linux/start_up.c iomem_size += new->size + UM_KERN_PAGE_SIZE; new 198 arch/x86/boot/compressed/eboot.c struct setup_data *data, *new; new 218 arch/x86/boot/compressed/eboot.c size + sizeof(struct setup_data), &new); new 225 arch/x86/boot/compressed/eboot.c new->data, &size); new 228 arch/x86/boot/compressed/eboot.c efi_call_early(free_pool, new); new 231 arch/x86/boot/compressed/eboot.c new->type = SETUP_APPLE_PROPERTIES; new 232 arch/x86/boot/compressed/eboot.c new->len = size; new 233 arch/x86/boot/compressed/eboot.c new->next = 0; new 237 arch/x86/boot/compressed/eboot.c boot_params->hdr.setup_data = (unsigned long)new; new 241 arch/x86/boot/compressed/eboot.c data->next = (unsigned long)new; new 75 arch/x86/boot/compressed/pgtable_64.c unsigned long new = bios_start; new 89 arch/x86/boot/compressed/pgtable_64.c new = entry->addr + entry->size; new 92 arch/x86/boot/compressed/pgtable_64.c new = round_down(new, PAGE_SIZE); new 95 arch/x86/boot/compressed/pgtable_64.c if (new - TRAMPOLINE_32BIT_SIZE < entry->addr) new 99 arch/x86/boot/compressed/pgtable_64.c if (new - TRAMPOLINE_32BIT_SIZE > bios_start) new 102 arch/x86/boot/compressed/pgtable_64.c bios_start = new; new 438 arch/x86/events/amd/core.c int idx, new = -1; new 457 arch/x86/events/amd/core.c if (new == -1 || hwc->idx == idx) new 470 arch/x86/events/amd/core.c if (new != -1) new 471 arch/x86/events/amd/core.c cmpxchg(nb->owners + new, event, NULL); new 472 arch/x86/events/amd/core.c new = idx; new 479 arch/x86/events/amd/core.c if (new == -1) new 482 arch/x86/events/amd/core.c return &nb->event_constraints[new]; new 85 arch/x86/events/amd/uncore.c u64 prev, new; new 94 arch/x86/events/amd/uncore.c rdpmcl(hwc->event_base_rdpmc, new); new 95 arch/x86/events/amd/uncore.c local64_set(&hwc->prev_count, new); new 96 arch/x86/events/amd/uncore.c delta = (new << COUNTER_SHIFT) - (prev << COUNTER_SHIFT); new 1672 arch/x86/events/intel/ds.c s64 new, old; new 1712 arch/x86/events/intel/ds.c new = ((s64)(new_raw_count << shift) >> shift); new 1714 arch/x86/events/intel/ds.c local64_add(new - old + count * period, &event->count); new 1716 arch/x86/events/intel/ds.c local64_set(&hwc->period_left, -new); new 246 arch/x86/include/asm/alternative.h : output : [old] "i" (oldfunc), [new] "i" (newfunc), ## input) new 192 arch/x86/include/asm/atomic.h static __always_inline int arch_atomic_cmpxchg(atomic_t *v, int old, int new) new 194 arch/x86/include/asm/atomic.h return arch_cmpxchg(&v->counter, old, new); new 198 arch/x86/include/asm/atomic.h static __always_inline bool arch_atomic_try_cmpxchg(atomic_t *v, int *old, int new) new 200 arch/x86/include/asm/atomic.h return try_cmpxchg(&v->counter, old, new); new 203 arch/x86/include/asm/atomic.h static inline int arch_atomic_xchg(atomic_t *v, int new) new 205 arch/x86/include/asm/atomic.h return arch_xchg(&v->counter, new); new 178 arch/x86/include/asm/atomic64_64.h static inline s64 arch_atomic64_cmpxchg(atomic64_t *v, s64 old, s64 new) new 180 arch/x86/include/asm/atomic64_64.h return arch_cmpxchg(&v->counter, old, new); new 184 arch/x86/include/asm/atomic64_64.h static __always_inline bool arch_atomic64_try_cmpxchg(atomic64_t *v, s64 *old, s64 new) new 186 arch/x86/include/asm/atomic64_64.h return try_cmpxchg(&v->counter, old, new); new 189 arch/x86/include/asm/atomic64_64.h static inline s64 arch_atomic64_xchg(atomic64_t *v, s64 new) new 191 arch/x86/include/asm/atomic64_64.h return arch_xchg(&v->counter, new); new 85 arch/x86/include/asm/cmpxchg.h #define __raw_cmpxchg(ptr, old, new, size, lock) \ new 89 arch/x86/include/asm/cmpxchg.h __typeof__(*(ptr)) __new = (new); \ new 133 arch/x86/include/asm/cmpxchg.h #define __cmpxchg(ptr, old, new, size) \ new 134 arch/x86/include/asm/cmpxchg.h __raw_cmpxchg((ptr), (old), (new), (size), LOCK_PREFIX) new 136 arch/x86/include/asm/cmpxchg.h #define __sync_cmpxchg(ptr, old, new, size) \ new 137 arch/x86/include/asm/cmpxchg.h __raw_cmpxchg((ptr), (old), (new), (size), "lock; ") new 139 arch/x86/include/asm/cmpxchg.h #define __cmpxchg_local(ptr, old, new, size) \ new 140 arch/x86/include/asm/cmpxchg.h __raw_cmpxchg((ptr), (old), (new), (size), "") new 148 arch/x86/include/asm/cmpxchg.h #define arch_cmpxchg(ptr, old, new) \ new 149 arch/x86/include/asm/cmpxchg.h __cmpxchg(ptr, old, new, sizeof(*(ptr))) new 151 arch/x86/include/asm/cmpxchg.h #define arch_sync_cmpxchg(ptr, old, new) \ new 152 arch/x86/include/asm/cmpxchg.h __sync_cmpxchg(ptr, old, new, sizeof(*(ptr))) new 154 arch/x86/include/asm/cmpxchg.h #define arch_cmpxchg_local(ptr, old, new) \ new 155 arch/x86/include/asm/cmpxchg.h __cmpxchg_local(ptr, old, new, sizeof(*(ptr))) new 173 arch/x86/include/asm/cmpxchg.h : [new] "q" (__new) \ new 185 arch/x86/include/asm/cmpxchg.h : [new] "r" (__new) \ new 197 arch/x86/include/asm/cmpxchg.h : [new] "r" (__new) \ new 209 arch/x86/include/asm/cmpxchg.h : [new] "r" (__new) \ new 221 arch/x86/include/asm/cmpxchg.h #define __try_cmpxchg(ptr, pold, new, size) \ new 222 arch/x86/include/asm/cmpxchg.h __raw_try_cmpxchg((ptr), (pold), (new), (size), LOCK_PREFIX) new 224 arch/x86/include/asm/cmpxchg.h #define try_cmpxchg(ptr, pold, new) \ new 225 arch/x86/include/asm/cmpxchg.h __try_cmpxchg((ptr), (pold), (new), sizeof(*(ptr))) new 47 arch/x86/include/asm/cmpxchg_32.h static inline u64 __cmpxchg64(volatile u64 *ptr, u64 old, u64 new) new 53 arch/x86/include/asm/cmpxchg_32.h : "b" ((u32)new), new 54 arch/x86/include/asm/cmpxchg_32.h "c" ((u32)(new >> 32)), new 60 arch/x86/include/asm/cmpxchg_32.h static inline u64 __cmpxchg64_local(volatile u64 *ptr, u64 old, u64 new) new 66 arch/x86/include/asm/cmpxchg_32.h : "b" ((u32)new), new 67 arch/x86/include/asm/cmpxchg_32.h "c" ((u32)(new >> 32)), new 182 arch/x86/include/asm/kvm_emulate.h const void *new, new 35 arch/x86/include/asm/kvm_page_track.h void (*track_write)(struct kvm_vcpu *vcpu, gpa_t gpa, const u8 *new, new 72 arch/x86/include/asm/kvm_page_track.h void kvm_page_track_write(struct kvm_vcpu *vcpu, gpa_t gpa, const u8 *new, new 43 arch/x86/include/asm/mc146818rtc.h unsigned long new; new 44 arch/x86/include/asm/mc146818rtc.h new = ((smp_processor_id() + 1) << 8) | reg; new 50 arch/x86/include/asm/mc146818rtc.h if (__cmpxchg(&cmos_lock, 0, new, sizeof(cmos_lock)) == 0) new 200 arch/x86/include/asm/pgtable-3level.h union split_pmd old, new, *ptr; new 204 arch/x86/include/asm/pgtable-3level.h new.pmd = pmd; new 207 arch/x86/include/asm/pgtable-3level.h old.pmd_low = xchg(&ptr->pmd_low, new.pmd_low); new 209 arch/x86/include/asm/pgtable-3level.h ptr->pmd_high = new.pmd_high; new 489 arch/x86/include/asm/pgtable_types.h pgprot_t new; new 491 arch/x86/include/asm/pgtable_types.h pgprot_val(new) = (val & ~(_PAGE_PAT | _PAGE_PAT_LARGE)) | new 493 arch/x86/include/asm/pgtable_types.h return new; new 498 arch/x86/include/asm/pgtable_types.h pgprot_t new; new 500 arch/x86/include/asm/pgtable_types.h pgprot_val(new) = (val & ~(_PAGE_PAT | _PAGE_PAT_LARGE)) | new 503 arch/x86/include/asm/pgtable_types.h return new; new 31 arch/x86/include/asm/preempt.h int old, new; new 35 arch/x86/include/asm/preempt.h new = (old & PREEMPT_NEED_RESCHED) | new 37 arch/x86/include/asm/preempt.h } while (raw_cpu_cmpxchg_4(__preempt_count, old, new) != old); new 590 arch/x86/include/asm/uaccess.h #define __user_atomic_cmpxchg_inatomic(uval, ptr, old, new, size) \ new 594 arch/x86/include/asm/uaccess.h __typeof__(*(ptr)) __new = (new); \ new 672 arch/x86/include/asm/uaccess.h #define user_atomic_cmpxchg_inatomic(uval, ptr, old, new) \ new 676 arch/x86/include/asm/uaccess.h (old), (new), sizeof(*(ptr))) : \ new 571 arch/x86/kernel/acpi/boot.c unsigned int old, new; new 581 arch/x86/kernel/acpi/boot.c new = acpi_noirq ? old : 0; new 589 arch/x86/kernel/acpi/boot.c new &= ~mask; new 592 arch/x86/kernel/acpi/boot.c new |= mask; new 596 arch/x86/kernel/acpi/boot.c if (old == new) new 599 arch/x86/kernel/acpi/boot.c printk(PREFIX "setting ELCR to %04x (from %04x)\n", new, old); new 600 arch/x86/kernel/acpi/boot.c outb(new, 0x4d0); new 601 arch/x86/kernel/acpi/boot.c outb(new >> 8, 0x4d1); new 1737 arch/x86/kernel/acpi/boot.c unsigned int old, new, val; new 1740 arch/x86/kernel/acpi/boot.c new = (((old & ~0x3) + 2) + ((old >> 1) & 0x1)); new 1741 arch/x86/kernel/acpi/boot.c val = cmpxchg(lock, old, new); new 1743 arch/x86/kernel/acpi/boot.c return ((new & 0x3) < 3) ? -1 : 0; new 1748 arch/x86/kernel/acpi/boot.c unsigned int old, new, val; new 1751 arch/x86/kernel/acpi/boot.c new = old & ~0x3; new 1752 arch/x86/kernel/acpi/boot.c val = cmpxchg(lock, old, new); new 353 arch/x86/kernel/apb_timer.c u64 old, new; new 383 arch/x86/kernel/apb_timer.c new = dw_apb_clocksource_read(clocksource_apbt); new 384 arch/x86/kernel/apb_timer.c } while (new < old); new 392 arch/x86/kernel/apic/apic.c static inline int eilvt_entry_is_changeable(unsigned int old, unsigned int new) new 395 arch/x86/kernel/apic/apic.c || (new == APIC_EILVT_MASKED) new 396 arch/x86/kernel/apic/apic.c || ((new & ~APIC_EILVT_MASKED) == old); new 399 arch/x86/kernel/apic/apic.c static unsigned int reserve_eilvt_offset(int offset, unsigned int new) new 409 arch/x86/kernel/apic/apic.c if (vector && !eilvt_entry_is_changeable(vector, new)) new 412 arch/x86/kernel/apic/apic.c rsvd = atomic_cmpxchg(&eilvt_offsets[offset], rsvd, new); new 413 arch/x86/kernel/apic/apic.c } while (rsvd != new); new 420 arch/x86/kernel/apic/apic.c return new; new 432 arch/x86/kernel/apic/apic.c unsigned int new, old, reserved; new 434 arch/x86/kernel/apic/apic.c new = (mask << 16) | (msg_type << 8) | vector; new 436 arch/x86/kernel/apic/apic.c reserved = reserve_eilvt_offset(offset, new); new 438 arch/x86/kernel/apic/apic.c if (reserved != new) { new 442 arch/x86/kernel/apic/apic.c smp_processor_id(), reg, offset, new, reserved); new 446 arch/x86/kernel/apic/apic.c if (!eilvt_entry_is_changeable(old, new)) { new 450 arch/x86/kernel/apic/apic.c smp_processor_id(), reg, offset, new, old); new 454 arch/x86/kernel/apic/apic.c apic_write(reg, new); new 434 arch/x86/kernel/cpu/mce/amd.c static int setup_APIC_mce_threshold(int reserved, int new) new 436 arch/x86/kernel/cpu/mce/amd.c if (reserved < 0 && !setup_APIC_eilvt(new, THRESHOLD_APIC_VECTOR, new 438 arch/x86/kernel/cpu/mce/amd.c return new; new 443 arch/x86/kernel/cpu/mce/amd.c static int setup_APIC_deferred_error(int reserved, int new) new 445 arch/x86/kernel/cpu/mce/amd.c if (reserved < 0 && !setup_APIC_eilvt(new, DEFERRED_ERROR_VECTOR, new 447 arch/x86/kernel/cpu/mce/amd.c return new; new 525 arch/x86/kernel/cpu/mce/amd.c int new; new 543 arch/x86/kernel/cpu/mce/amd.c new = (misc_high & MASK_LVTOFF_HI) >> 20; new 551 arch/x86/kernel/cpu/mce/amd.c new = (smca_low & SMCA_THR_LVT_OFF) >> 12; new 554 arch/x86/kernel/cpu/mce/amd.c offset = setup_APIC_mce_threshold(offset, new); new 555 arch/x86/kernel/cpu/mce/amd.c if (offset == new) new 1059 arch/x86/kernel/cpu/mce/amd.c unsigned long new; new 1064 arch/x86/kernel/cpu/mce/amd.c if (kstrtoul(buf, 0, &new) < 0) new 1067 arch/x86/kernel/cpu/mce/amd.c b->interrupt_enable = !!new; new 1081 arch/x86/kernel/cpu/mce/amd.c unsigned long new; new 1083 arch/x86/kernel/cpu/mce/amd.c if (kstrtoul(buf, 0, &new) < 0) new 1086 arch/x86/kernel/cpu/mce/amd.c if (new > THRESHOLD_MAX) new 1087 arch/x86/kernel/cpu/mce/amd.c new = THRESHOLD_MAX; new 1088 arch/x86/kernel/cpu/mce/amd.c if (new < 1) new 1089 arch/x86/kernel/cpu/mce/amd.c new = 1; new 1093 arch/x86/kernel/cpu/mce/amd.c b->threshold_limit = new; new 2139 arch/x86/kernel/cpu/mce/core.c u64 new; new 2141 arch/x86/kernel/cpu/mce/core.c if (kstrtou64(buf, 0, &new) < 0) new 2152 arch/x86/kernel/cpu/mce/core.c b->ctl = new; new 2162 arch/x86/kernel/cpu/mce/core.c u64 new; new 2164 arch/x86/kernel/cpu/mce/core.c if (kstrtou64(buf, 0, &new) < 0) new 2168 arch/x86/kernel/cpu/mce/core.c if (mca_cfg.ignore_ce ^ !!new) { new 2169 arch/x86/kernel/cpu/mce/core.c if (new) { new 2189 arch/x86/kernel/cpu/mce/core.c u64 new; new 2191 arch/x86/kernel/cpu/mce/core.c if (kstrtou64(buf, 0, &new) < 0) new 2195 arch/x86/kernel/cpu/mce/core.c if (mca_cfg.cmci_disabled ^ !!new) { new 2196 arch/x86/kernel/cpu/mce/core.c if (new) { new 155 arch/x86/kernel/ftrace.c unsigned const char *new, *old; new 159 arch/x86/kernel/ftrace.c new = ftrace_nop_replace(); new 170 arch/x86/kernel/ftrace.c return ftrace_modify_code_direct(rec->ip, old, new); new 181 arch/x86/kernel/ftrace.c unsigned const char *new, *old; new 185 arch/x86/kernel/ftrace.c new = ftrace_call_replace(ip, addr); new 188 arch/x86/kernel/ftrace.c return ftrace_modify_code_direct(rec->ip, old, new); new 247 arch/x86/kernel/ftrace.c static int update_ftrace_func(unsigned long ip, void *new) new 261 arch/x86/kernel/ftrace.c ret = ftrace_modify_code(ip, old, new); new 271 arch/x86/kernel/ftrace.c unsigned char *new; new 276 arch/x86/kernel/ftrace.c new = ftrace_call_replace(ip, (unsigned long)func); new 277 arch/x86/kernel/ftrace.c ret = update_ftrace_func(ip, new); new 282 arch/x86/kernel/ftrace.c new = ftrace_call_replace(ip, (unsigned long)func); new 283 arch/x86/kernel/ftrace.c ret = update_ftrace_func(ip, new); new 458 arch/x86/kernel/ftrace.c static int add_update_code(unsigned long ip, unsigned const char *new) new 462 arch/x86/kernel/ftrace.c new++; new 463 arch/x86/kernel/ftrace.c return ftrace_write(ip, new, MCOUNT_INSN_SIZE - 1); new 469 arch/x86/kernel/ftrace.c unsigned const char *new; new 471 arch/x86/kernel/ftrace.c new = ftrace_call_replace(ip, addr); new 472 arch/x86/kernel/ftrace.c return add_update_code(ip, new); new 478 arch/x86/kernel/ftrace.c unsigned const char *new; new 480 arch/x86/kernel/ftrace.c new = ftrace_nop_replace(); new 481 arch/x86/kernel/ftrace.c return add_update_code(ip, new); new 513 arch/x86/kernel/ftrace.c unsigned const char *new; new 515 arch/x86/kernel/ftrace.c new = ftrace_call_replace(ip, addr); new 517 arch/x86/kernel/ftrace.c return ftrace_write(ip, new, 1); new 523 arch/x86/kernel/ftrace.c unsigned const char *new; new 525 arch/x86/kernel/ftrace.c new = ftrace_nop_replace(); new 527 arch/x86/kernel/ftrace.c return ftrace_write(ip, new, 1); new 862 arch/x86/kernel/ftrace.c unsigned char *new; new 893 arch/x86/kernel/ftrace.c new = ftrace_call_replace(ip, (unsigned long)func); new 894 arch/x86/kernel/ftrace.c ret = update_ftrace_func(ip, new); new 991 arch/x86/kernel/ftrace.c unsigned char *new; new 994 arch/x86/kernel/ftrace.c new = ftrace_jmp_replace(ip, (unsigned long)func); new 996 arch/x86/kernel/ftrace.c return update_ftrace_func(ip, new); new 686 arch/x86/kernel/hpet.c union hpet_lock old, new; new 706 arch/x86/kernel/hpet.c new.value = hpet_readl(HPET_COUNTER); new 710 arch/x86/kernel/hpet.c WRITE_ONCE(hpet.value, new.value); new 713 arch/x86/kernel/hpet.c return (u64)new.value; new 732 arch/x86/kernel/hpet.c new.lockval = READ_ONCE(hpet.lockval); new 733 arch/x86/kernel/hpet.c } while ((new.value == old.value) && arch_spin_is_locked(&new.lock)); new 735 arch/x86/kernel/hpet.c return (u64)new.value; new 903 arch/x86/kernel/mpparse.c unsigned char old, new; new 908 arch/x86/kernel/mpparse.c new = mpf_checksum((unsigned char *)mpc, mpc->length); new 909 arch/x86/kernel/mpparse.c if (old == new) { new 338 arch/x86/kernel/smpboot.c int new; new 341 arch/x86/kernel/smpboot.c new = topology_phys_to_logical_pkg(pkg); new 342 arch/x86/kernel/smpboot.c if (new >= 0) new 345 arch/x86/kernel/smpboot.c new = logical_packages++; new 346 arch/x86/kernel/smpboot.c if (new != pkg) { new 348 arch/x86/kernel/smpboot.c cpu, pkg, new); new 351 arch/x86/kernel/smpboot.c cpu_data(cpu).logical_proc_id = new; new 361 arch/x86/kernel/smpboot.c int new; new 364 arch/x86/kernel/smpboot.c new = topology_phys_to_logical_die(die, cpu); new 365 arch/x86/kernel/smpboot.c if (new >= 0) new 368 arch/x86/kernel/smpboot.c new = logical_die++; new 369 arch/x86/kernel/smpboot.c if (new != die) { new 371 arch/x86/kernel/smpboot.c cpu, die, new); new 374 arch/x86/kernel/smpboot.c cpu_data(cpu).logical_die_id = new; new 991 arch/x86/kernel/tsc.c if ((val == CPUFREQ_PRECHANGE && freq->old < freq->new) || new 992 arch/x86/kernel/tsc.c (val == CPUFREQ_POSTCHANGE && freq->old > freq->new)) { new 994 arch/x86/kernel/tsc.c cpufreq_scale(loops_per_jiffy_ref, ref_freq, freq->new); new 996 arch/x86/kernel/tsc.c tsc_khz = cpufreq_scale(tsc_khz_ref, ref_freq, freq->new); new 90 arch/x86/kernel/vm86_32.c #define set_flags(X, new, mask) \ new 91 arch/x86/kernel/vm86_32.c ((X) = ((X) & ~(mask)) | ((new) & (mask))) new 172 arch/x86/kvm/lapic.c struct kvm_apic_map *new, *old = NULL; new 183 arch/x86/kvm/lapic.c new = kvzalloc(sizeof(struct kvm_apic_map) + new 187 arch/x86/kvm/lapic.c if (!new) new 190 arch/x86/kvm/lapic.c new->max_apic_id = max_id; new 208 arch/x86/kvm/lapic.c x2apic_id <= new->max_apic_id) new 209 arch/x86/kvm/lapic.c new->phys_map[x2apic_id] = apic; new 214 arch/x86/kvm/lapic.c if (!apic_x2apic_mode(apic) && !new->phys_map[xapic_id]) new 215 arch/x86/kvm/lapic.c new->phys_map[xapic_id] = apic; new 223 arch/x86/kvm/lapic.c new->mode |= KVM_APIC_MODE_X2APIC; new 227 arch/x86/kvm/lapic.c new->mode |= KVM_APIC_MODE_XAPIC_FLAT; new 229 arch/x86/kvm/lapic.c new->mode |= KVM_APIC_MODE_XAPIC_CLUSTER; new 232 arch/x86/kvm/lapic.c if (!kvm_apic_map_get_logical_dest(new, ldr, &cluster, &mask)) new 241 arch/x86/kvm/lapic.c rcu_assign_pointer(kvm->arch.apic_map, new); new 5305 arch/x86/kvm/mmu.c const void *new) new 5313 arch/x86/kvm/mmu.c vcpu->arch.mmu->update_pte(vcpu, sp, spte, new); new 5316 arch/x86/kvm/mmu.c static bool need_remote_flush(u64 old, u64 new) new 5320 arch/x86/kvm/mmu.c if (!is_shadow_present_pte(new)) new 5322 arch/x86/kvm/mmu.c if ((old ^ new) & PT64_BASE_ADDR_MASK) new 5325 arch/x86/kvm/mmu.c new ^= shadow_nx_mask; new 5326 arch/x86/kvm/mmu.c return (old & ~new & PT64_PERM_MASK) != 0; new 5432 arch/x86/kvm/mmu.c const u8 *new, int bytes, new 223 arch/x86/kvm/page_track.c void kvm_page_track_write(struct kvm_vcpu *vcpu, gpa_t gpa, const u8 *new, new 238 arch/x86/kvm/page_track.c n->track_write(vcpu, gpa, new, bytes, n); new 4664 arch/x86/kvm/svm.c u64 *old, *new; new 4672 arch/x86/kvm/svm.c new = avic_get_physical_id_entry(vcpu, id); new 4673 arch/x86/kvm/svm.c if (!new || !old) new 4677 arch/x86/kvm/svm.c *new = *old; new 4679 arch/x86/kvm/svm.c to_svm(vcpu)->avic_physical_id_cache = new; new 894 arch/x86/kvm/trace.h TP_PROTO(unsigned int vcpu_id, unsigned int new, unsigned int old), new 895 arch/x86/kvm/trace.h TP_ARGS(vcpu_id, new, old), new 899 arch/x86/kvm/trace.h __field( unsigned int, new ) new 905 arch/x86/kvm/trace.h __entry->new = new; new 910 arch/x86/kvm/trace.h __entry->vcpu_id, __entry->old, __entry->new, new 911 arch/x86/kvm/trace.h __entry->old < __entry->new ? "growed" : "shrinked") new 1236 arch/x86/kvm/vmx/vmx.c struct pi_desc old, new; new 1262 arch/x86/kvm/vmx/vmx.c old.control = new.control = pi_desc->control; new 1267 arch/x86/kvm/vmx/vmx.c new.ndst = dest; new 1269 arch/x86/kvm/vmx/vmx.c new.ndst = (dest << 8) & 0xFF00; new 1271 arch/x86/kvm/vmx/vmx.c new.sn = 0; new 1273 arch/x86/kvm/vmx/vmx.c new.control) != old.control); new 7321 arch/x86/kvm/vmx/vmx.c struct pi_desc old, new; new 7325 arch/x86/kvm/vmx/vmx.c old.control = new.control = pi_desc->control; new 7332 arch/x86/kvm/vmx/vmx.c new.ndst = dest; new 7334 arch/x86/kvm/vmx/vmx.c new.ndst = (dest << 8) & 0xFF00; new 7337 arch/x86/kvm/vmx/vmx.c new.nv = POSTED_INTR_VECTOR; new 7339 arch/x86/kvm/vmx/vmx.c new.control) != old.control); new 7365 arch/x86/kvm/vmx/vmx.c struct pi_desc old, new; new 7385 arch/x86/kvm/vmx/vmx.c old.control = new.control = pi_desc->control; new 7402 arch/x86/kvm/vmx/vmx.c new.ndst = dest; new 7404 arch/x86/kvm/vmx/vmx.c new.ndst = (dest << 8) & 0xFF00; new 7407 arch/x86/kvm/vmx/vmx.c new.nv = POSTED_INTR_WAKEUP_VECTOR; new 7409 arch/x86/kvm/vmx/vmx.c new.control) != old.control); new 5819 arch/x86/kvm/x86.c #define CMPXCHG_TYPE(t, ptr, old, new) \ new 5820 arch/x86/kvm/x86.c (cmpxchg((t *)(ptr), *(t *)(old), *(t *)(new)) == *(t *)(old)) new 5823 arch/x86/kvm/x86.c # define CMPXCHG64(ptr, old, new) CMPXCHG_TYPE(u64, ptr, old, new) new 5825 arch/x86/kvm/x86.c # define CMPXCHG64(ptr, old, new) \ new 5826 arch/x86/kvm/x86.c (cmpxchg64((u64 *)(ptr), *(u64 *)(old), *(u64 *)(new)) == *(u64 *)(old)) new 5832 arch/x86/kvm/x86.c const void *new, new 5862 arch/x86/kvm/x86.c exchanged = CMPXCHG_TYPE(u8, kaddr, old, new); new 5865 arch/x86/kvm/x86.c exchanged = CMPXCHG_TYPE(u16, kaddr, old, new); new 5868 arch/x86/kvm/x86.c exchanged = CMPXCHG_TYPE(u32, kaddr, old, new); new 5871 arch/x86/kvm/x86.c exchanged = CMPXCHG64(kaddr, old, new); new 5882 arch/x86/kvm/x86.c kvm_page_track_write(vcpu, gpa, new, bytes); new 5889 arch/x86/kvm/x86.c return emulator_write_emulated(ctxt, addr, new, bytes, exception); new 6981 arch/x86/kvm/x86.c khz = freq->new; new 7085 arch/x86/kvm/x86.c if (freq->old < freq->new && send_ipi) { new 7108 arch/x86/kvm/x86.c if (val == CPUFREQ_PRECHANGE && freq->old > freq->new) new 7110 arch/x86/kvm/x86.c if (val == CPUFREQ_POSTCHANGE && freq->old < freq->new) new 9839 arch/x86/kvm/x86.c struct kvm_memory_slot *new) new 9842 arch/x86/kvm/x86.c if (new->flags & KVM_MEM_READONLY) { new 9843 arch/x86/kvm/x86.c kvm_mmu_slot_remove_write_access(kvm, new); new 9877 arch/x86/kvm/x86.c if (new->flags & KVM_MEM_LOG_DIRTY_PAGES) { new 9879 arch/x86/kvm/x86.c kvm_x86_ops->slot_enable_log_dirty(kvm, new); new 9881 arch/x86/kvm/x86.c kvm_mmu_slot_remove_write_access(kvm, new); new 9884 arch/x86/kvm/x86.c kvm_x86_ops->slot_disable_log_dirty(kvm, new); new 9891 arch/x86/kvm/x86.c const struct kvm_memory_slot *new, new 9917 arch/x86/kvm/x86.c !(new->flags & KVM_MEM_LOG_DIRTY_PAGES)) new 9918 arch/x86/kvm/x86.c kvm_mmu_zap_collapsible_sptes(kvm, new); new 9931 arch/x86/kvm/x86.c kvm_mmu_slot_apply_flags(kvm, (struct kvm_memory_slot *) new); new 548 arch/x86/mm/pat.c struct memtype *new; new 596 arch/x86/mm/pat.c new = kzalloc(sizeof(struct memtype), GFP_KERNEL); new 597 arch/x86/mm/pat.c if (!new) new 600 arch/x86/mm/pat.c new->start = start; new 601 arch/x86/mm/pat.c new->end = end; new 602 arch/x86/mm/pat.c new->type = actual_type; new 606 arch/x86/mm/pat.c err = rbt_memtype_check_insert(new, new_type); new 610 arch/x86/mm/pat.c cattr_name(new->type), cattr_name(req_type)); new 611 arch/x86/mm/pat.c kfree(new); new 620 arch/x86/mm/pat.c start, end - 1, cattr_name(new->type), cattr_name(req_type), new 32 arch/x86/mm/pat_internal.h extern int rbt_memtype_check_insert(struct memtype *new, new 38 arch/x86/mm/pat_internal.h static inline int rbt_memtype_check_insert(struct memtype *new, new 188 arch/x86/mm/pat_rbtree.c int rbt_memtype_check_insert(struct memtype *new, new 193 arch/x86/mm/pat_rbtree.c err = memtype_rb_check_conflict(&memtype_rbroot, new->start, new->end, new 194 arch/x86/mm/pat_rbtree.c new->type, ret_type); new 198 arch/x86/mm/pat_rbtree.c new->type = *ret_type; new 200 arch/x86/mm/pat_rbtree.c new->subtree_max_end = new->end; new 201 arch/x86/mm/pat_rbtree.c memtype_rb_insert(&memtype_rbroot, new); new 49 arch/x86/pci/irq.c int new); new 53 arch/x86/pci/mmconfig-shared.c static void list_add_sorted(struct pci_mmcfg_region *new) new 59 arch/x86/pci/mmconfig-shared.c if (cfg->segment > new->segment || new 60 arch/x86/pci/mmconfig-shared.c (cfg->segment == new->segment && new 61 arch/x86/pci/mmconfig-shared.c cfg->start_bus >= new->start_bus)) { new 62 arch/x86/pci/mmconfig-shared.c list_add_tail_rcu(&new->list, &cfg->list); new 66 arch/x86/pci/mmconfig-shared.c list_add_tail_rcu(&new->list, &pci_mmcfg_list); new 72 arch/x86/pci/mmconfig-shared.c struct pci_mmcfg_region *new; new 78 arch/x86/pci/mmconfig-shared.c new = kzalloc(sizeof(*new), GFP_KERNEL); new 79 arch/x86/pci/mmconfig-shared.c if (!new) new 82 arch/x86/pci/mmconfig-shared.c new->address = addr; new 83 arch/x86/pci/mmconfig-shared.c new->segment = segment; new 84 arch/x86/pci/mmconfig-shared.c new->start_bus = start; new 85 arch/x86/pci/mmconfig-shared.c new->end_bus = end; new 87 arch/x86/pci/mmconfig-shared.c res = &new->res; new 91 arch/x86/pci/mmconfig-shared.c snprintf(new->name, PCI_MMCFG_RESOURCE_NAME_LEN, new 93 arch/x86/pci/mmconfig-shared.c res->name = new->name; new 95 arch/x86/pci/mmconfig-shared.c return new; new 101 arch/x86/pci/mmconfig-shared.c struct pci_mmcfg_region *new; new 103 arch/x86/pci/mmconfig-shared.c new = pci_mmconfig_alloc(segment, start, end, addr); new 104 arch/x86/pci/mmconfig-shared.c if (new) { new 106 arch/x86/pci/mmconfig-shared.c list_add_sorted(new); new 112 arch/x86/pci/mmconfig-shared.c segment, start, end, &new->res, (unsigned long)addr); new 115 arch/x86/pci/mmconfig-shared.c return new; new 250 arch/x86/platform/efi/quirks.c void *new; new 282 arch/x86/platform/efi/quirks.c new = early_memremap(new_phys, new_size); new 283 arch/x86/platform/efi/quirks.c if (!new) { new 288 arch/x86/platform/efi/quirks.c efi_memmap_insert(&efi.memmap, new, &mr); new 289 arch/x86/platform/efi/quirks.c early_memunmap(new, new_size); new 410 arch/x86/platform/efi/quirks.c void *new, *new_md; new 469 arch/x86/platform/efi/quirks.c new = memremap(new_phys, new_size, MEMREMAP_WB); new 470 arch/x86/platform/efi/quirks.c if (!new) { new 480 arch/x86/platform/efi/quirks.c new_md = new; new 491 arch/x86/platform/efi/quirks.c memunmap(new); new 153 arch/x86/platform/intel-mid/pwr.c static int __update_power_state(struct mid_pwr *pwr, int reg, int bit, int new) new 162 arch/x86/platform/intel-mid/pwr.c if (curstate == new) new 166 arch/x86/platform/intel-mid/pwr.c mid_pwr_set_state(pwr, reg, (power & ~(3 << bit)) | (new << bit)); new 176 arch/x86/platform/intel-mid/pwr.c if (curstate != new) new 24 arch/x86/um/bugs_32.c struct sigaction old, new; new 27 arch/x86/um/bugs_32.c new.sa_handler = cmov_sigill_test_handler; new 30 arch/x86/um/bugs_32.c new.sa_flags = SA_NODEFER; new 31 arch/x86/um/bugs_32.c sigemptyset(&new.sa_mask); new 32 arch/x86/um/bugs_32.c sigaction(SIGILL, &new, &old); new 41 arch/x86/um/bugs_32.c sigaction(SIGILL, &old, &new); new 218 arch/x86/um/tls_32.c int arch_set_tls(struct task_struct *new, unsigned long tls) new 232 arch/x86/um/tls_32.c ret = set_tls_entry(new, &info, idx, 0); new 259 arch/xtensa/include/asm/atomic.h #define atomic_xchg(v, new) (xchg(&((v)->counter), new)) new 24 arch/xtensa/include/asm/cmpxchg.h __cmpxchg_u32(volatile int *p, int old, int new) new 38 arch/xtensa/include/asm/cmpxchg.h : "a" (new), "a" (p), "a" (old) new 47 arch/xtensa/include/asm/cmpxchg.h : "+a" (new) new 52 arch/xtensa/include/asm/cmpxchg.h return new; new 63 arch/xtensa/include/asm/cmpxchg.h : "a" (p), "a" (old), "r" (new) new 74 arch/xtensa/include/asm/cmpxchg.h __cmpxchg(volatile void *ptr, unsigned long old, unsigned long new, int size) new 77 arch/xtensa/include/asm/cmpxchg.h case 4: return __cmpxchg_u32(ptr, old, new); new 94 arch/xtensa/include/asm/cmpxchg.h unsigned long new, int size) new 98 arch/xtensa/include/asm/cmpxchg.h return __cmpxchg_u32(ptr, old, new); new 100 arch/xtensa/include/asm/cmpxchg.h return __cmpxchg_local_generic(ptr, old, new, size); new 609 arch/xtensa/platforms/iss/network.c struct iss_net_init *new; new 644 arch/xtensa/platforms/iss/network.c new = memblock_alloc(sizeof(*new), SMP_CACHE_BYTES); new 645 arch/xtensa/platforms/iss/network.c if (new == NULL) { new 650 arch/xtensa/platforms/iss/network.c INIT_LIST_HEAD(&new->list); new 651 arch/xtensa/platforms/iss/network.c new->index = n; new 652 arch/xtensa/platforms/iss/network.c new->init = str + 1; new 654 arch/xtensa/platforms/iss/network.c list_add_tail(&new->list, ð_cmd_line); new 726 block/bfq-iosched.c struct rb_node **new = &(root->rb_root.rb_node), *parent = NULL; new 744 block/bfq-iosched.c while (*new) { new 745 block/bfq-iosched.c struct bfq_weight_counter *__counter = container_of(*new, new 748 block/bfq-iosched.c parent = *new; new 755 block/bfq-iosched.c new = &((*new)->rb_left); new 757 block/bfq-iosched.c new = &((*new)->rb_right); new 781 block/bfq-iosched.c rb_link_node(&bfqq->weight_counter->weights_node, parent, new); new 15 block/blk-lib.c struct bio *new = bio_alloc(gfp, nr_pages); new 18 block/blk-lib.c bio_chain(bio, new); new 22 block/blk-lib.c return new; new 506 block/blk-mq-tag.c struct blk_mq_tags *new; new 519 block/blk-mq-tag.c new = blk_mq_alloc_rq_map(set, hctx->queue_num, tdepth, new 521 block/blk-mq-tag.c if (!new) new 523 block/blk-mq-tag.c ret = blk_mq_alloc_rqs(set, new, hctx->queue_num, tdepth); new 525 block/blk-mq-tag.c blk_mq_free_rq_map(new); new 531 block/blk-mq-tag.c *tagsptr = new; new 1916 block/blk-throttle.c static void throtl_downgrade_state(struct throtl_data *td, int new) new 1926 block/blk-throttle.c td->limit_index = new; new 189 crypto/jitterentropy.c __u64 new = 0; new 202 crypto/jitterentropy.c new = ec->data; new 218 crypto/jitterentropy.c tmp ^= ((new >> 63) & 1); new 219 crypto/jitterentropy.c tmp ^= ((new >> 60) & 1); new 220 crypto/jitterentropy.c tmp ^= ((new >> 55) & 1); new 221 crypto/jitterentropy.c tmp ^= ((new >> 30) & 1); new 222 crypto/jitterentropy.c tmp ^= ((new >> 27) & 1); new 223 crypto/jitterentropy.c tmp ^= ((new >> 22) & 1); new 224 crypto/jitterentropy.c new <<= 1; new 225 crypto/jitterentropy.c new ^= tmp; new 228 crypto/jitterentropy.c ec->data = new; new 77 drivers/acpi/acpi_memhotplug.c struct acpi_memory_info *info, *new; new 95 drivers/acpi/acpi_memhotplug.c new = kzalloc(sizeof(struct acpi_memory_info), GFP_KERNEL); new 96 drivers/acpi/acpi_memhotplug.c if (!new) new 99 drivers/acpi/acpi_memhotplug.c INIT_LIST_HEAD(&new->list); new 100 drivers/acpi/acpi_memhotplug.c new->caching = address64.info.mem.caching; new 101 drivers/acpi/acpi_memhotplug.c new->write_protect = address64.info.mem.write_protect; new 102 drivers/acpi/acpi_memhotplug.c new->start_addr = address64.address.minimum; new 103 drivers/acpi/acpi_memhotplug.c new->length = address64.address.address_length; new 104 drivers/acpi/acpi_memhotplug.c list_add_tail(&new->list, &mem_device->res_list); new 962 drivers/ata/sata_mv.c static inline void mv_write_cached_reg(void __iomem *addr, u32 *old, u32 new) new 964 drivers/ata/sata_mv.c if (new != *old) { new 966 drivers/ata/sata_mv.c *old = new; new 980 drivers/ata/sata_mv.c writelfl(new, addr); /* read after write */ new 984 drivers/ata/sata_mv.c writel(new, addr); /* unaffected by the errata */ new 1510 drivers/ata/sata_mv.c u32 old, new; new 1515 drivers/ata/sata_mv.c new = old | (1 << 22); new 1517 drivers/ata/sata_mv.c new = old & ~(1 << 22); new 1518 drivers/ata/sata_mv.c if (new != old) new 1519 drivers/ata/sata_mv.c writel(new, hpriv->base + GPIO_PORT_CTL); new 1537 drivers/ata/sata_mv.c u32 new, *old = &pp->cached.unknown_rsvd; new 1540 drivers/ata/sata_mv.c new = *old | 1; new 1542 drivers/ata/sata_mv.c new = *old & ~1; new 1543 drivers/ata/sata_mv.c mv_write_cached_reg(mv_ap_base(ap) + EDMA_UNKNOWN_RSVD, old, new); new 1078 drivers/atm/lanai.c u32 new = reg_read(lanai, Status_Reg); new 1079 drivers/atm/lanai.c u32 changes = new ^ lanai->status; new 1080 drivers/atm/lanai.c lanai->status = new; new 1083 drivers/atm/lanai.c status_message(lanai->number, name, new & flag) new 317 drivers/base/component.c struct component_match_array *new; new 322 drivers/base/component.c new = kmalloc_array(num, sizeof(*new), GFP_KERNEL); new 323 drivers/base/component.c if (!new) new 327 drivers/base/component.c memcpy(new, match->compare, sizeof(*new) * new 331 drivers/base/component.c match->compare = new; new 1001 drivers/base/core.c unsigned long new; new 1003 drivers/base/core.c ret = kstrtoul(buf, 0, &new); new 1006 drivers/base/core.c *(unsigned long *)(ea->var) = new; new 1027 drivers/base/core.c long new; new 1029 drivers/base/core.c ret = kstrtol(buf, 0, &new); new 1033 drivers/base/core.c if (new > INT_MAX || new < INT_MIN) new 1035 drivers/base/core.c *(int *)(ea->var) = new; new 99 drivers/base/regmap/regcache-rbtree.c struct rb_node **new, *parent; new 105 drivers/base/regmap/regcache-rbtree.c new = &root->rb_node; new 106 drivers/base/regmap/regcache-rbtree.c while (*new) { new 107 drivers/base/regmap/regcache-rbtree.c rbnode_tmp = rb_entry(*new, struct regcache_rbtree_node, node); new 113 drivers/base/regmap/regcache-rbtree.c parent = *new; new 119 drivers/base/regmap/regcache-rbtree.c new = &((*new)->rb_right); new 121 drivers/base/regmap/regcache-rbtree.c new = &((*new)->rb_left); new 125 drivers/base/regmap/regcache-rbtree.c rb_link_node(&rbnode->node, parent, new); new 535 drivers/base/regmap/regmap.c struct rb_node **new = &(root->rb_node), *parent = NULL; new 537 drivers/base/regmap/regmap.c while (*new) { new 539 drivers/base/regmap/regmap.c rb_entry(*new, struct regmap_range_node, node); new 541 drivers/base/regmap/regmap.c parent = *new; new 543 drivers/base/regmap/regmap.c new = &((*new)->rb_left); new 545 drivers/base/regmap/regmap.c new = &((*new)->rb_right); new 550 drivers/base/regmap/regmap.c rb_link_node(&data->node, parent, new); new 1055 drivers/base/regmap/regmap.c struct regmap_range_node *new; new 1111 drivers/base/regmap/regmap.c new = kzalloc(sizeof(*new), GFP_KERNEL); new 1112 drivers/base/regmap/regmap.c if (new == NULL) { new 1117 drivers/base/regmap/regmap.c new->map = map; new 1118 drivers/base/regmap/regmap.c new->name = range_cfg->name; new 1119 drivers/base/regmap/regmap.c new->range_min = range_cfg->range_min; new 1120 drivers/base/regmap/regmap.c new->range_max = range_cfg->range_max; new 1121 drivers/base/regmap/regmap.c new->selector_reg = range_cfg->selector_reg; new 1122 drivers/base/regmap/regmap.c new->selector_mask = range_cfg->selector_mask; new 1123 drivers/base/regmap/regmap.c new->selector_shift = range_cfg->selector_shift; new 1124 drivers/base/regmap/regmap.c new->window_start = range_cfg->window_start; new 1125 drivers/base/regmap/regmap.c new->window_len = range_cfg->window_len; new 1127 drivers/base/regmap/regmap.c if (!_regmap_range_add(map, new)) { new 1129 drivers/base/regmap/regmap.c kfree(new); new 369 drivers/base/swnode.c const void *new; new 380 drivers/base/swnode.c new = dst->pointer.str; new 382 drivers/base/swnode.c new = kmemdup(pointer, src->length, GFP_KERNEL); new 383 drivers/base/swnode.c if (!new) new 387 drivers/base/swnode.c new = kstrdup(src->value.str, GFP_KERNEL); new 388 drivers/base/swnode.c if (!new && src->value.str) new 391 drivers/base/swnode.c new = pointer; new 398 drivers/base/swnode.c property_set_pointer(dst, new); new 428 drivers/block/brd.c static struct brd_device *brd_init_one(int i, bool *new) new 432 drivers/block/brd.c *new = false; new 444 drivers/block/brd.c *new = true; new 460 drivers/block/brd.c bool new; new 463 drivers/block/brd.c brd = brd_init_one(MINOR(dev) / max_part, &new); new 467 drivers/block/brd.c if (new) new 27 drivers/block/drbd/drbd_interval.c struct rb_node **new = &root->rb_node, *parent = NULL; new 32 drivers/block/drbd/drbd_interval.c while (*new) { new 34 drivers/block/drbd/drbd_interval.c rb_entry(*new, struct drbd_interval, rb); new 36 drivers/block/drbd/drbd_interval.c parent = *new; new 40 drivers/block/drbd/drbd_interval.c new = &(*new)->rb_left; new 42 drivers/block/drbd/drbd_interval.c new = &(*new)->rb_right; new 44 drivers/block/drbd/drbd_interval.c new = &(*new)->rb_left; new 46 drivers/block/drbd/drbd_interval.c new = &(*new)->rb_right; new 52 drivers/block/drbd/drbd_interval.c rb_link_node(&this->rb, parent, new); new 1382 drivers/block/drbd/drbd_nl.c unsigned int now, new, local, peer; new 1415 drivers/block/drbd/drbd_nl.c new = min(local, peer); new 1417 drivers/block/drbd/drbd_nl.c if (device->state.role == R_PRIMARY && new < now) new 1418 drivers/block/drbd/drbd_nl.c drbd_err(device, "ASSERT FAILED new < now; (%u < %u)\n", new, now); new 1420 drivers/block/drbd/drbd_nl.c if (new != now) new 1421 drivers/block/drbd/drbd_nl.c drbd_info(device, "max BIO size = %u\n", new); new 1423 drivers/block/drbd/drbd_nl.c drbd_setup_queue_param(device, bdev, new, o); new 2040 drivers/block/drbd/drbd_worker.c unsigned long old, new; new 2043 drivers/block/drbd/drbd_worker.c new = old & ~DRBD_DEVICE_WORK_MASK; new 2044 drivers/block/drbd/drbd_worker.c } while (cmpxchg(flags, old, new) != old); new 233 drivers/block/xen-blkback/blkback.c struct rb_node **new = NULL, *parent = NULL; new 243 drivers/block/xen-blkback/blkback.c new = &ring->persistent_gnts.rb_node; new 244 drivers/block/xen-blkback/blkback.c while (*new) { new 245 drivers/block/xen-blkback/blkback.c this = container_of(*new, struct persistent_gnt, node); new 247 drivers/block/xen-blkback/blkback.c parent = *new; new 249 drivers/block/xen-blkback/blkback.c new = &((*new)->rb_left); new 251 drivers/block/xen-blkback/blkback.c new = &((*new)->rb_right); new 260 drivers/block/xen-blkback/blkback.c rb_link_node(&(persistent_gnt->node), parent, new); new 99 drivers/char/agp/generic.c struct agp_memory *new; new 105 drivers/char/agp/generic.c new = kzalloc(sizeof(struct agp_memory), GFP_KERNEL); new 106 drivers/char/agp/generic.c if (new == NULL) new 109 drivers/char/agp/generic.c new->key = agp_get_key(); new 111 drivers/char/agp/generic.c if (new->key < 0) { new 112 drivers/char/agp/generic.c kfree(new); new 116 drivers/char/agp/generic.c agp_alloc_page_array(alloc_size, new); new 118 drivers/char/agp/generic.c if (new->pages == NULL) { new 119 drivers/char/agp/generic.c agp_free_key(new->key); new 120 drivers/char/agp/generic.c kfree(new); new 123 drivers/char/agp/generic.c new->num_scratch_pages = 0; new 124 drivers/char/agp/generic.c return new; new 129 drivers/char/agp/generic.c struct agp_memory *new; new 131 drivers/char/agp/generic.c new = kzalloc(sizeof(struct agp_memory), GFP_KERNEL); new 132 drivers/char/agp/generic.c if (new == NULL) new 135 drivers/char/agp/generic.c new->key = agp_get_key(); new 137 drivers/char/agp/generic.c if (new->key < 0) { new 138 drivers/char/agp/generic.c kfree(new); new 142 drivers/char/agp/generic.c agp_alloc_page_array(PAGE_SIZE * scratch_pages, new); new 144 drivers/char/agp/generic.c if (new->pages == NULL) { new 145 drivers/char/agp/generic.c agp_free_key(new->key); new 146 drivers/char/agp/generic.c kfree(new); new 149 drivers/char/agp/generic.c new->num_scratch_pages = scratch_pages; new 150 drivers/char/agp/generic.c new->type = AGP_NORMAL_MEMORY; new 151 drivers/char/agp/generic.c return new; new 222 drivers/char/agp/generic.c struct agp_memory *new; new 235 drivers/char/agp/generic.c new = agp_generic_alloc_user(page_count, type); new 236 drivers/char/agp/generic.c if (new) new 237 drivers/char/agp/generic.c new->bridge = bridge; new 238 drivers/char/agp/generic.c return new; new 242 drivers/char/agp/generic.c new = bridge->driver->alloc_by_type(page_count, type); new 243 drivers/char/agp/generic.c if (new) new 244 drivers/char/agp/generic.c new->bridge = bridge; new 245 drivers/char/agp/generic.c return new; new 250 drivers/char/agp/generic.c new = agp_create_memory(scratch_pages); new 252 drivers/char/agp/generic.c if (new == NULL) new 256 drivers/char/agp/generic.c if (bridge->driver->agp_alloc_pages(bridge, new, page_count)) { new 257 drivers/char/agp/generic.c agp_free_memory(new); new 260 drivers/char/agp/generic.c new->bridge = bridge; new 261 drivers/char/agp/generic.c return new; new 268 drivers/char/agp/generic.c agp_free_memory(new); new 271 drivers/char/agp/generic.c new->pages[i] = page; new 272 drivers/char/agp/generic.c new->page_count++; new 274 drivers/char/agp/generic.c new->bridge = bridge; new 276 drivers/char/agp/generic.c return new; new 1165 drivers/char/agp/generic.c struct agp_memory *new; new 1170 drivers/char/agp/generic.c new = agp_create_user_memory(page_count); new 1171 drivers/char/agp/generic.c if (new == NULL) new 1175 drivers/char/agp/generic.c new->pages[i] = NULL; new 1176 drivers/char/agp/generic.c new->page_count = 0; new 1177 drivers/char/agp/generic.c new->type = type; new 1178 drivers/char/agp/generic.c new->num_scratch_pages = pages; new 1180 drivers/char/agp/generic.c return new; new 241 drivers/char/agp/intel-gtt.c struct agp_memory *new; new 258 drivers/char/agp/intel-gtt.c new = agp_create_memory(pg_count); new 259 drivers/char/agp/intel-gtt.c if (new == NULL) new 262 drivers/char/agp/intel-gtt.c new->pages[0] = page; new 265 drivers/char/agp/intel-gtt.c new->pages[1] = new->pages[0] + 1; new 266 drivers/char/agp/intel-gtt.c new->pages[2] = new->pages[1] + 1; new 267 drivers/char/agp/intel-gtt.c new->pages[3] = new->pages[2] + 1; new 269 drivers/char/agp/intel-gtt.c new->page_count = pg_count; new 270 drivers/char/agp/intel-gtt.c new->num_scratch_pages = pg_count; new 271 drivers/char/agp/intel-gtt.c new->type = AGP_PHYS_MEMORY; new 272 drivers/char/agp/intel-gtt.c new->physical = page_to_phys(new->pages[0]); new 273 drivers/char/agp/intel-gtt.c return new; new 982 drivers/char/agp/intel-gtt.c struct agp_memory *new; new 988 drivers/char/agp/intel-gtt.c new = agp_create_memory(1); new 989 drivers/char/agp/intel-gtt.c if (new == NULL) new 992 drivers/char/agp/intel-gtt.c new->type = AGP_DCACHE_MEMORY; new 993 drivers/char/agp/intel-gtt.c new->page_count = pg_count; new 994 drivers/char/agp/intel-gtt.c new->num_scratch_pages = 0; new 995 drivers/char/agp/intel-gtt.c agp_free_page_array(new); new 996 drivers/char/agp/intel-gtt.c return new; new 23 drivers/char/agp/isoch.c static void agp_3_5_dev_list_insert(struct list_head *head, struct list_head *new) new 25 drivers/char/agp/isoch.c struct agp_3_5_dev *cur, *n = list_entry(new, struct agp_3_5_dev, list); new 33 drivers/char/agp/isoch.c list_add_tail(new, pos); new 56 drivers/clk/clk-multiplier.c static bool __is_best_rate(unsigned long rate, unsigned long new, new 60 drivers/clk/clk-multiplier.c return abs(rate - new) < abs(rate - best); new 62 drivers/clk/clk-multiplier.c return new >= rate && new < best; new 83 drivers/clk/qcom/clk-regmap-mux-div.c unsigned long new) new 85 drivers/clk/qcom/clk-regmap-mux-div.c return (req <= new && new < best) || (best < req && best < new); new 319 drivers/clocksource/arm_arch_timer.c u64 old, new; new 322 drivers/clocksource/arm_arch_timer.c new = read_sysreg(cntpct_el0); new 323 drivers/clocksource/arm_arch_timer.c return (((old ^ new) >> 32) & 1) ? old : new; new 328 drivers/clocksource/arm_arch_timer.c u64 old, new; new 331 drivers/clocksource/arm_arch_timer.c new = read_sysreg(cntvct_el0); new 332 drivers/clocksource/arm_arch_timer.c return (((old ^ new) >> 32) & 1) ? old : new; new 207 drivers/cpufreq/cppc_cpufreq.c freqs.new = target_freq; new 263 drivers/cpufreq/cpufreq-nforce2.c freqs.new = target_fsb * fid * 100; new 265 drivers/cpufreq/cpufreq-nforce2.c if (freqs.old == freqs.new) new 269 drivers/cpufreq/cpufreq-nforce2.c freqs.old, freqs.new); new 319 drivers/cpufreq/cpufreq.c if (val == CPUFREQ_POSTCHANGE && ci->old != ci->new) { new 321 drivers/cpufreq/cpufreq.c ci->new); new 323 drivers/cpufreq/cpufreq.c loops_per_jiffy, ci->new); new 352 drivers/cpufreq/cpufreq.c state, freqs->new); new 375 drivers/cpufreq/cpufreq.c pr_debug("FREQ: %u - CPUs: %*pbl\n", freqs->new, new 379 drivers/cpufreq/cpufreq.c trace_cpu_frequency(freqs->new, cpu); new 384 drivers/cpufreq/cpufreq.c cpufreq_stats_record_transition(policy, freqs->new); new 385 drivers/cpufreq/cpufreq.c policy->cur = freqs->new; new 397 drivers/cpufreq/cpufreq.c swap(freqs->old, freqs->new); new 1648 drivers/cpufreq/cpufreq.c freqs.new = new_freq; new 2036 drivers/cpufreq/cpufreq.c freqs->new = cpufreq_driver->get_intermediate(policy, index); new 2039 drivers/cpufreq/cpufreq.c if (!freqs->new) new 2043 drivers/cpufreq/cpufreq.c __func__, policy->cpu, freqs->old, freqs->new); new 2075 drivers/cpufreq/cpufreq.c intermediate_freq = freqs.new; new 2078 drivers/cpufreq/cpufreq.c freqs.old = freqs.new; new 2081 drivers/cpufreq/cpufreq.c freqs.new = newfreq; new 2083 drivers/cpufreq/cpufreq.c __func__, policy->cpu, freqs.old, freqs.new); new 2104 drivers/cpufreq/cpufreq.c freqs.new = policy->restore_freq; new 262 drivers/cpufreq/gx-suspmod.c freqs.new = new_khz; new 317 drivers/cpufreq/gx-suspmod.c pr_debug("suspend modulation w/ clock speed: %d kHz.\n", freqs.new); new 2331 drivers/cpufreq/intel_pstate.c freqs.new = target_freq; new 2336 drivers/cpufreq/intel_pstate.c target_pstate = DIV_ROUND_UP(freqs.new, cpu->pstate.scaling); new 2339 drivers/cpufreq/intel_pstate.c target_pstate = freqs.new / cpu->pstate.scaling; new 2342 drivers/cpufreq/intel_pstate.c target_pstate = DIV_ROUND_CLOSEST(freqs.new, cpu->pstate.scaling); new 2352 drivers/cpufreq/intel_pstate.c freqs.new = target_pstate * cpu->pstate.scaling; new 272 drivers/cpufreq/longhaul.c freqs.new = speed; new 347 drivers/cpufreq/longhaul.c freqs.new = calc_speed(longhaul_get_cpu_mult()); new 349 drivers/cpufreq/longhaul.c if (unlikely(freqs.new != speed)) { new 964 drivers/cpufreq/longhaul.c freqs.new = longhaul_table[i].frequency; new 216 drivers/cpufreq/pcc-cpufreq.c freqs.new = target_freq; new 231 drivers/cpufreq/powernow-k6.c freqs.new = clock_ratio[i].frequency; new 267 drivers/cpufreq/powernow-k7.c freqs.new = powernow_table[index].frequency; new 274 drivers/cpufreq/powernow-k7.c if (freqs.old > freqs.new) { new 913 drivers/cpufreq/powernow-k8.c freqs.new = find_khz_freq_from_fid(fid); new 199 drivers/cpufreq/s3c24xx-cpufreq.c freqs.new = cpu_new.freq; new 202 drivers/cpufreq/s3c24xx-cpufreq.c freqs.freqs.new = cpu_new.freq.armclk / 1000; new 67 drivers/cpufreq/sh-cpufreq.c freqs.new = (freq + 500) / 1000; new 42 drivers/cpufreq/unicore2-cpufreq.c freqs.new = target_freq; new 186 drivers/crypto/qat/qat_common/adf_cfg.c static void adf_cfg_keyval_add(struct adf_cfg_key_val *new, new 189 drivers/crypto/qat/qat_common/adf_cfg.c list_add_tail(&new->list, &sec->param_head); new 302 drivers/devfreq/devfreq.c freqs.new = new_freq; new 307 drivers/devfreq/devfreq.c freqs.new = cur_freq; new 312 drivers/devfreq/devfreq.c freqs.new = new_freq; new 133 drivers/devfreq/governor_passive.c unsigned long freq = freqs->new; new 150 drivers/dma-buf/dma-resv.c struct dma_resv_list *old, *new; new 167 drivers/dma-buf/dma-resv.c new = dma_resv_list_alloc(max); new 168 drivers/dma-buf/dma-resv.c if (!new) new 183 drivers/dma-buf/dma-resv.c RCU_INIT_POINTER(new->shared[--k], fence); new 185 drivers/dma-buf/dma-resv.c RCU_INIT_POINTER(new->shared[j++], fence); new 187 drivers/dma-buf/dma-resv.c new->shared_count = j; new 197 drivers/dma-buf/dma-resv.c rcu_assign_pointer(obj->fence, new); new 206 drivers/dma-buf/dma-resv.c fence = rcu_dereference_protected(new->shared[i], new 314 drivers/dma-buf/dma-resv.c struct dma_fence *old, *new; new 365 drivers/dma-buf/dma-resv.c new = dma_fence_get_rcu_safe(&src->fence_excl); new 374 drivers/dma-buf/dma-resv.c RCU_INIT_POINTER(dst->fence_excl, new); new 304 drivers/dma/altera-msgdma.c struct msgdma_sw_desc *new; new 308 drivers/dma/altera-msgdma.c new = tx_to_desc(tx); new 312 drivers/dma/altera-msgdma.c list_add_tail(&new->node, &mdev->pending_list); new 333 drivers/dma/altera-msgdma.c struct msgdma_sw_desc *new, *first = NULL; new 352 drivers/dma/altera-msgdma.c new = msgdma_get_descriptor(mdev); new 355 drivers/dma/altera-msgdma.c desc = &new->hw_desc; new 362 drivers/dma/altera-msgdma.c first = new; new 364 drivers/dma/altera-msgdma.c list_add_tail(&new->node, &first->tx_list); new 392 drivers/dma/altera-msgdma.c struct msgdma_sw_desc *new, *first = NULL; new 418 drivers/dma/altera-msgdma.c new = msgdma_get_descriptor(mdev); new 420 drivers/dma/altera-msgdma.c desc = &new->hw_desc; new 436 drivers/dma/altera-msgdma.c first = new; new 438 drivers/dma/altera-msgdma.c list_add_tail(&new->node, &first->tx_list); new 726 drivers/dma/ep93xx_dma.c struct ep93xx_dma_desc *new; new 736 drivers/dma/ep93xx_dma.c new = list_first_entry(&edmac->queue, struct ep93xx_dma_desc, node); new 737 drivers/dma/ep93xx_dma.c list_del_init(&new->node); new 739 drivers/dma/ep93xx_dma.c ep93xx_dma_set_active(edmac, new); new 764 drivers/dma/fsldma.c struct fsl_desc_sw *first = NULL, *prev = NULL, *new; new 778 drivers/dma/fsldma.c new = fsl_dma_alloc_descriptor(chan); new 779 drivers/dma/fsldma.c if (!new) { new 786 drivers/dma/fsldma.c set_desc_cnt(chan, &new->hw, copy); new 787 drivers/dma/fsldma.c set_desc_src(chan, &new->hw, dma_src); new 788 drivers/dma/fsldma.c set_desc_dst(chan, &new->hw, dma_dst); new 791 drivers/dma/fsldma.c first = new; new 793 drivers/dma/fsldma.c set_desc_next(chan, &prev->hw, new->async_tx.phys); new 795 drivers/dma/fsldma.c new->async_tx.cookie = 0; new 796 drivers/dma/fsldma.c async_tx_ack(&new->async_tx); new 798 drivers/dma/fsldma.c prev = new; new 804 drivers/dma/fsldma.c list_add_tail(&new->node, &first->tx_list); new 807 drivers/dma/fsldma.c new->async_tx.flags = flags; /* client is in control of this ack */ new 808 drivers/dma/fsldma.c new->async_tx.cookie = -EBUSY; new 811 drivers/dma/fsldma.c set_ld_eol(chan, new); new 450 drivers/dma/mmp_pdma.c struct mmp_pdma_desc_sw *first = NULL, *prev = NULL, *new; new 470 drivers/dma/mmp_pdma.c new = mmp_pdma_alloc_descriptor(chan); new 471 drivers/dma/mmp_pdma.c if (!new) { new 480 drivers/dma/mmp_pdma.c new->desc.dcmd = chan->dcmd | (DCMD_LENGTH & copy); new 481 drivers/dma/mmp_pdma.c new->desc.dsadr = dma_src; new 482 drivers/dma/mmp_pdma.c new->desc.dtadr = dma_dst; new 485 drivers/dma/mmp_pdma.c first = new; new 487 drivers/dma/mmp_pdma.c prev->desc.ddadr = new->async_tx.phys; new 489 drivers/dma/mmp_pdma.c new->async_tx.cookie = 0; new 490 drivers/dma/mmp_pdma.c async_tx_ack(&new->async_tx); new 492 drivers/dma/mmp_pdma.c prev = new; new 505 drivers/dma/mmp_pdma.c list_add_tail(&new->node, &first->tx_list); new 512 drivers/dma/mmp_pdma.c new->desc.ddadr = DDADR_STOP; new 513 drivers/dma/mmp_pdma.c new->desc.dcmd |= DCMD_ENDIRQEN; new 531 drivers/dma/mmp_pdma.c struct mmp_pdma_desc_sw *first = NULL, *prev = NULL, *new = NULL; new 554 drivers/dma/mmp_pdma.c new = mmp_pdma_alloc_descriptor(chan); new 555 drivers/dma/mmp_pdma.c if (!new) { new 560 drivers/dma/mmp_pdma.c new->desc.dcmd = chan->dcmd | (DCMD_LENGTH & len); new 562 drivers/dma/mmp_pdma.c new->desc.dsadr = addr; new 563 drivers/dma/mmp_pdma.c new->desc.dtadr = chan->dev_addr; new 565 drivers/dma/mmp_pdma.c new->desc.dsadr = chan->dev_addr; new 566 drivers/dma/mmp_pdma.c new->desc.dtadr = addr; new 570 drivers/dma/mmp_pdma.c first = new; new 572 drivers/dma/mmp_pdma.c prev->desc.ddadr = new->async_tx.phys; new 574 drivers/dma/mmp_pdma.c new->async_tx.cookie = 0; new 575 drivers/dma/mmp_pdma.c async_tx_ack(&new->async_tx); new 576 drivers/dma/mmp_pdma.c prev = new; new 579 drivers/dma/mmp_pdma.c list_add_tail(&new->node, &first->tx_list); new 591 drivers/dma/mmp_pdma.c new->desc.ddadr = DDADR_STOP; new 592 drivers/dma/mmp_pdma.c new->desc.dcmd |= DCMD_ENDIRQEN; new 612 drivers/dma/mmp_pdma.c struct mmp_pdma_desc_sw *first = NULL, *prev = NULL, *new; new 646 drivers/dma/mmp_pdma.c new = mmp_pdma_alloc_descriptor(chan); new 647 drivers/dma/mmp_pdma.c if (!new) { new 652 drivers/dma/mmp_pdma.c new->desc.dcmd = (chan->dcmd | DCMD_ENDIRQEN | new 654 drivers/dma/mmp_pdma.c new->desc.dsadr = dma_src; new 655 drivers/dma/mmp_pdma.c new->desc.dtadr = dma_dst; new 658 drivers/dma/mmp_pdma.c first = new; new 660 drivers/dma/mmp_pdma.c prev->desc.ddadr = new->async_tx.phys; new 662 drivers/dma/mmp_pdma.c new->async_tx.cookie = 0; new 663 drivers/dma/mmp_pdma.c async_tx_ack(&new->async_tx); new 665 drivers/dma/mmp_pdma.c prev = new; new 674 drivers/dma/mmp_pdma.c list_add_tail(&new->node, &first->tx_list); new 681 drivers/dma/mmp_pdma.c new->desc.ddadr = first->async_tx.phys; new 507 drivers/dma/sh/shdma-base.c struct shdma_desc *new; new 514 drivers/dma/sh/shdma-base.c new = shdma_get_desc(schan); new 515 drivers/dma/sh/shdma-base.c if (!new) { new 520 drivers/dma/sh/shdma-base.c ops->desc_setup(schan, new, *src, *dst, ©_size); new 524 drivers/dma/sh/shdma-base.c new->async_tx.cookie = -EBUSY; new 525 drivers/dma/sh/shdma-base.c *first = new; new 528 drivers/dma/sh/shdma-base.c new->async_tx.cookie = -EINVAL; new 533 drivers/dma/sh/shdma-base.c copy_size, *len, src, dst, &new->async_tx, new 534 drivers/dma/sh/shdma-base.c new->async_tx.cookie); new 536 drivers/dma/sh/shdma-base.c new->mark = DESC_PREPARED; new 537 drivers/dma/sh/shdma-base.c new->async_tx.flags = flags; new 538 drivers/dma/sh/shdma-base.c new->direction = direction; new 539 drivers/dma/sh/shdma-base.c new->partial = 0; new 547 drivers/dma/sh/shdma-base.c return new; new 565 drivers/dma/sh/shdma-base.c struct shdma_desc *first = NULL, *new = NULL /* compiler... */; new 600 drivers/dma/sh/shdma-base.c new = shdma_add_desc(schan, flags, new 604 drivers/dma/sh/shdma-base.c new = shdma_add_desc(schan, flags, new 607 drivers/dma/sh/shdma-base.c if (!new) new 610 drivers/dma/sh/shdma-base.c new->cyclic = cyclic; new 612 drivers/dma/sh/shdma-base.c new->chunks = 1; new 614 drivers/dma/sh/shdma-base.c new->chunks = chunks--; new 615 drivers/dma/sh/shdma-base.c list_add_tail(&new->node, &tx_list); new 619 drivers/dma/sh/shdma-base.c if (new != first) new 620 drivers/dma/sh/shdma-base.c new->async_tx.cookie = -ENOSPC; new 630 drivers/dma/sh/shdma-base.c list_for_each_entry(new, &tx_list, node) new 631 drivers/dma/sh/shdma-base.c new->mark = DESC_IDLE; new 839 drivers/dma/xgene-dma.c struct xgene_dma_desc_sw *first = NULL, *new; new 851 drivers/dma/xgene-dma.c new = xgene_dma_alloc_descriptor(chan); new 852 drivers/dma/xgene-dma.c if (!new) new 856 drivers/dma/xgene-dma.c xgene_dma_prep_xor_desc(chan, new, &dst, src, new 860 drivers/dma/xgene-dma.c first = new; new 862 drivers/dma/xgene-dma.c new->tx.cookie = 0; new 863 drivers/dma/xgene-dma.c async_tx_ack(&new->tx); new 866 drivers/dma/xgene-dma.c list_add_tail(&new->node, &first->tx_list); new 869 drivers/dma/xgene-dma.c new->tx.flags = flags; /* client is in control of this ack */ new 870 drivers/dma/xgene-dma.c new->tx.cookie = -EBUSY; new 871 drivers/dma/xgene-dma.c list_splice(&first->tx_list, &new->tx_list); new 873 drivers/dma/xgene-dma.c return &new->tx; new 887 drivers/dma/xgene-dma.c struct xgene_dma_desc_sw *first = NULL, *new; new 913 drivers/dma/xgene-dma.c new = xgene_dma_alloc_descriptor(chan); new 914 drivers/dma/xgene-dma.c if (!new) new 918 drivers/dma/xgene-dma.c first = new; new 920 drivers/dma/xgene-dma.c new->tx.cookie = 0; new 921 drivers/dma/xgene-dma.c async_tx_ack(&new->tx); new 924 drivers/dma/xgene-dma.c list_add_tail(&new->node, &first->tx_list); new 931 drivers/dma/xgene-dma.c xgene_dma_prep_xor_desc(chan, new, &dst[0], src, new 941 drivers/dma/xgene-dma.c xgene_dma_prep_xor_desc(chan, new, &dst[1], _src, new 946 drivers/dma/xgene-dma.c new->tx.flags = flags; /* client is in control of this ack */ new 947 drivers/dma/xgene-dma.c new->tx.cookie = -EBUSY; new 948 drivers/dma/xgene-dma.c list_splice(&first->tx_list, &new->tx_list); new 950 drivers/dma/xgene-dma.c return &new->tx; new 372 drivers/dma/xilinx/zynqmp_dma.c struct zynqmp_dma_desc_sw *desc, *new; new 376 drivers/dma/xilinx/zynqmp_dma.c new = tx_to_desc(tx); new 386 drivers/dma/xilinx/zynqmp_dma.c desc->src_v->nxtdscraddr = new->src_p; new 388 drivers/dma/xilinx/zynqmp_dma.c desc->dst_v->nxtdscraddr = new->dst_p; new 392 drivers/dma/xilinx/zynqmp_dma.c list_add_tail(&new->node, &chan->pending_list); new 805 drivers/dma/xilinx/zynqmp_dma.c struct zynqmp_dma_desc_sw *new, *first = NULL; new 826 drivers/dma/xilinx/zynqmp_dma.c new = zynqmp_dma_get_descriptor(chan); new 829 drivers/dma/xilinx/zynqmp_dma.c desc = (struct zynqmp_dma_desc_ll *)new->src_v; new 837 drivers/dma/xilinx/zynqmp_dma.c first = new; new 839 drivers/dma/xilinx/zynqmp_dma.c list_add_tail(&new->node, &first->tx_list); new 64 drivers/edac/edac_mc.c void edac_set_report_status(int new) new 66 drivers/edac/edac_mc.c if (new == EDAC_REPORTING_ENABLED || new 67 drivers/edac/edac_mc.c new == EDAC_REPORTING_DISABLED || new 68 drivers/edac/edac_mc.c new == EDAC_REPORTING_FORCE) new 69 drivers/edac/edac_mc.c edac_report = new; new 880 drivers/firewire/core-device.c struct fw_device *new = data; new 881 drivers/firewire/core-device.c struct fw_card *card = new->card; new 890 drivers/firewire/core-device.c if (memcmp(old->config_rom, new->config_rom, 6 * 4) == 0 && new 894 drivers/firewire/core-device.c struct fw_node *current_node = new->node; new 897 drivers/firewire/core-device.c new->node = obsolete_node; new 898 drivers/firewire/core-device.c new->node->data = new; new 902 drivers/firewire/core-device.c old->max_speed = new->max_speed; new 214 drivers/firewire/core-iso.c int try, new, old = allocate ? BANDWIDTH_AVAILABLE_INITIAL : 0; new 223 drivers/firewire/core-iso.c new = allocate ? old - bandwidth : old + bandwidth; new 224 drivers/firewire/core-iso.c if (new < 0 || new > BANDWIDTH_AVAILABLE_INITIAL) new 228 drivers/firewire/core-iso.c data[1] = cpu_to_be32(new); new 301 drivers/firewire/net.c struct fwnet_fragment_info *fi, *fi2, *new; new 350 drivers/firewire/net.c new = kmalloc(sizeof(*new), GFP_ATOMIC); new 351 drivers/firewire/net.c if (!new) new 354 drivers/firewire/net.c new->offset = offset; new 355 drivers/firewire/net.c new->len = len; new 356 drivers/firewire/net.c list_add(&new->fi_link, list); new 358 drivers/firewire/net.c return new; new 365 drivers/firewire/net.c struct fwnet_partial_datagram *new; new 368 drivers/firewire/net.c new = kmalloc(sizeof(*new), GFP_ATOMIC); new 369 drivers/firewire/net.c if (!new) new 372 drivers/firewire/net.c INIT_LIST_HEAD(&new->fi_list); new 373 drivers/firewire/net.c fi = fwnet_frag_new(new, frag_off, frag_len); new 377 drivers/firewire/net.c new->datagram_label = datagram_label; new 378 drivers/firewire/net.c new->datagram_size = dg_size; new 379 drivers/firewire/net.c new->skb = dev_alloc_skb(dg_size + LL_RESERVED_SPACE(net)); new 380 drivers/firewire/net.c if (new->skb == NULL) new 383 drivers/firewire/net.c skb_reserve(new->skb, LL_RESERVED_SPACE(net)); new 384 drivers/firewire/net.c new->pbuf = skb_put(new->skb, dg_size); new 385 drivers/firewire/net.c memcpy(new->pbuf + frag_off, frag_buf, frag_len); new 386 drivers/firewire/net.c list_add_tail(&new->pd_link, &peer->pd_list); new 388 drivers/firewire/net.c return new; new 393 drivers/firewire/net.c kfree(new); new 263 drivers/firmware/efi/memmap.c void *old, *new; new 281 drivers/firmware/efi/memmap.c for (old = old_memmap->map, new = buf; new 283 drivers/firmware/efi/memmap.c old += old_memmap->desc_size, new += old_memmap->desc_size) { new 286 drivers/firmware/efi/memmap.c memcpy(new, old, old_memmap->desc_size); new 287 drivers/firmware/efi/memmap.c md = new; new 301 drivers/firmware/efi/memmap.c new += old_memmap->desc_size; new 302 drivers/firmware/efi/memmap.c memcpy(new, old, old_memmap->desc_size); new 303 drivers/firmware/efi/memmap.c md = new; new 314 drivers/firmware/efi/memmap.c new += old_memmap->desc_size; new 315 drivers/firmware/efi/memmap.c memcpy(new, old, old_memmap->desc_size); new 316 drivers/firmware/efi/memmap.c md = new; new 322 drivers/firmware/efi/memmap.c new += old_memmap->desc_size; new 323 drivers/firmware/efi/memmap.c memcpy(new, old, old_memmap->desc_size); new 324 drivers/firmware/efi/memmap.c md = new; new 336 drivers/firmware/efi/memmap.c new += old_memmap->desc_size; new 337 drivers/firmware/efi/memmap.c memcpy(new, old, old_memmap->desc_size); new 338 drivers/firmware/efi/memmap.c md = new; new 156 drivers/fpga/dfl-afu-dma-region.c struct rb_node **new, *parent = NULL; new 161 drivers/fpga/dfl-afu-dma-region.c new = &afu->dma_regions.rb_node; new 163 drivers/fpga/dfl-afu-dma-region.c while (*new) { new 166 drivers/fpga/dfl-afu-dma-region.c this = container_of(*new, struct dfl_afu_dma_region, node); new 168 drivers/fpga/dfl-afu-dma-region.c parent = *new; new 174 drivers/fpga/dfl-afu-dma-region.c new = &((*new)->rb_left); new 176 drivers/fpga/dfl-afu-dma-region.c new = &((*new)->rb_right); new 181 drivers/fpga/dfl-afu-dma-region.c rb_link_node(®ion->node, parent, new); new 212 drivers/gpio/gpio-stmpe.c u8 new = stmpe_gpio->regs[i][j]; new 214 drivers/gpio/gpio-stmpe.c if (new == old) new 217 drivers/gpio/gpio-stmpe.c stmpe_gpio->oldregs[i][j] = new; new 218 drivers/gpio/gpio-stmpe.c stmpe_reg_write(stmpe, stmpe->regs[regmap[i][j]], new); new 206 drivers/gpio/gpio-tc3589x.c u8 new = tc3589x_gpio->regs[i][j]; new 208 drivers/gpio/gpio-tc3589x.c if (new == old) new 211 drivers/gpio/gpio-tc3589x.c tc3589x_gpio->oldregs[i][j] = new; new 212 drivers/gpio/gpio-tc3589x.c tc3589x_reg_write(tc3589x, regmap[i] + j * 8, new); new 222 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c struct dma_resv_list *old, *new; new 232 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c new = kmalloc(offsetof(typeof(*new), shared[old->shared_max]), new 234 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c if (!new) new 247 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c RCU_INIT_POINTER(new->shared[--j], f); new 249 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c RCU_INIT_POINTER(new->shared[k++], f); new 251 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c new->shared_max = old->shared_max; new 252 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c new->shared_count = k; new 257 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c RCU_INIT_POINTER(resv->fence, new); new 265 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c f = rcu_dereference_protected(new->shared[i], new 130 drivers/gpu/drm/amd/display/dc/core/dc.c static inline void elevate_update_type(enum surface_update_type *original, enum surface_update_type new) new 132 drivers/gpu/drm/amd/display/dc/core/dc.c if (new > *original) new 133 drivers/gpu/drm/amd/display/dc/core/dc.c *original = new; new 453 drivers/gpu/drm/arm/display/komeda/komeda_crtc.c struct komeda_crtc_state *new; new 455 drivers/gpu/drm/arm/display/komeda/komeda_crtc.c new = kzalloc(sizeof(*new), GFP_KERNEL); new 456 drivers/gpu/drm/arm/display/komeda/komeda_crtc.c if (!new) new 459 drivers/gpu/drm/arm/display/komeda/komeda_crtc.c __drm_atomic_helper_crtc_duplicate_state(crtc, &new->base); new 461 drivers/gpu/drm/arm/display/komeda/komeda_crtc.c new->affected_pipes = old->active_pipes; new 462 drivers/gpu/drm/arm/display/komeda/komeda_crtc.c new->clock_ratio = old->clock_ratio; new 463 drivers/gpu/drm/arm/display/komeda/komeda_crtc.c new->max_slave_zorder = old->max_slave_zorder; new 465 drivers/gpu/drm/arm/display/komeda/komeda_crtc.c return &new->base; new 104 drivers/gpu/drm/arm/display/komeda/komeda_kms.c struct komeda_plane_state *new = to_kplane_st(plane_st); new 114 drivers/gpu/drm/arm/display/komeda/komeda_kms.c if (!last || (new->base.zpos > last->base.zpos)) { new 115 drivers/gpu/drm/arm/display/komeda/komeda_kms.c list_add_tail(&new->zlist_node, zorder_list); new 121 drivers/gpu/drm/arm/display/komeda/komeda_kms.c if (new->base.zpos < node->base.zpos) { new 122 drivers/gpu/drm/arm/display/komeda/komeda_kms.c list_add_tail(&new->zlist_node, &node->zlist_node); new 124 drivers/gpu/drm/arm/display/komeda/komeda_kms.c } else if (node->base.zpos == new->base.zpos) { new 126 drivers/gpu/drm/arm/display/komeda/komeda_kms.c struct drm_plane *b = new->base.plane; new 15 drivers/gpu/drm/arm/display/komeda/komeda_pipeline_state.c static inline bool is_switching_user(void *old, void *new) new 17 drivers/gpu/drm/arm/display/komeda/komeda_pipeline_state.c if (!old || !new) new 20 drivers/gpu/drm/arm/display/komeda/komeda_pipeline_state.c return old != new; new 1176 drivers/gpu/drm/arm/display/komeda/komeda_pipeline_state.c struct komeda_pipeline_state *new) new 1178 drivers/gpu/drm/arm/display/komeda/komeda_pipeline_state.c struct drm_atomic_state *drm_st = new->obj.state; new 1186 drivers/gpu/drm/arm/display/komeda/komeda_pipeline_state.c disabling_comps = (~new->active_comps) & old->active_comps; new 1192 drivers/gpu/drm/arm/display/komeda/komeda_pipeline_state.c drm_st, NULL, new->crtc); new 1256 drivers/gpu/drm/arm/display/komeda/komeda_pipeline_state.c struct komeda_pipeline_state *new = priv_to_pipe_st(pipe->obj.state); new 1263 drivers/gpu/drm/arm/display/komeda/komeda_pipeline_state.c changed_comps = new->active_comps | old->active_comps; new 1266 drivers/gpu/drm/arm/display/komeda/komeda_pipeline_state.c pipe->id, new->active_comps, changed_comps); new 1271 drivers/gpu/drm/arm/display/komeda/komeda_pipeline_state.c if (new->active_comps & BIT(c->id)) new 161 drivers/gpu/drm/arm/display/komeda/komeda_plane.c struct komeda_plane_state *new; new 166 drivers/gpu/drm/arm/display/komeda/komeda_plane.c new = kzalloc(sizeof(*new), GFP_KERNEL); new 167 drivers/gpu/drm/arm/display/komeda/komeda_plane.c if (!new) new 170 drivers/gpu/drm/arm/display/komeda/komeda_plane.c __drm_atomic_helper_plane_duplicate_state(plane, &new->base); new 172 drivers/gpu/drm/arm/display/komeda/komeda_plane.c return &new->base; new 276 drivers/gpu/drm/drm_context.c static int drm_context_switch(struct drm_device * dev, int old, int new) new 283 drivers/gpu/drm/drm_context.c DRM_DEBUG("Context switch from %d to %d\n", old, new); new 285 drivers/gpu/drm/drm_context.c if (new == dev->last_context) { new 305 drivers/gpu/drm/drm_context.c struct drm_file *file_priv, int new) new 307 drivers/gpu/drm/drm_context.c dev->last_context = new; /* PRE/POST: This is the _only_ writer. */ new 1651 drivers/gpu/drm/drm_edid.c u8 *edid, *new; new 1681 drivers/gpu/drm/drm_edid.c new = krealloc(edid, (valid_extensions + 1) * EDID_LENGTH, GFP_KERNEL); new 1682 drivers/gpu/drm/drm_edid.c if (!new) new 1684 drivers/gpu/drm/drm_edid.c edid = new; new 1708 drivers/gpu/drm/drm_edid.c new = kmalloc_array(valid_extensions + 1, EDID_LENGTH, new 1710 drivers/gpu/drm/drm_edid.c if (!new) new 1713 drivers/gpu/drm/drm_edid.c base = new; new 1725 drivers/gpu/drm/drm_edid.c edid = new; new 62 drivers/gpu/drm/drm_lock.c unsigned int old, new, prev; new 69 drivers/gpu/drm/drm_lock.c new = old | _DRM_LOCK_CONT; new 71 drivers/gpu/drm/drm_lock.c new = context | _DRM_LOCK_HELD | new 75 drivers/gpu/drm/drm_lock.c prev = cmpxchg(lock, old, new); new 89 drivers/gpu/drm/drm_lock.c if ((_DRM_LOCKING_CONTEXT(new)) == context && (new & _DRM_LOCK_HELD)) { new 111 drivers/gpu/drm/drm_lock.c unsigned int old, new, prev; new 117 drivers/gpu/drm/drm_lock.c new = context | _DRM_LOCK_HELD; new 118 drivers/gpu/drm/drm_lock.c prev = cmpxchg(lock, old, new); new 126 drivers/gpu/drm/drm_lock.c unsigned int old, new, prev; new 140 drivers/gpu/drm/drm_lock.c new = _DRM_LOCKING_CONTEXT(old); new 141 drivers/gpu/drm/drm_lock.c prev = cmpxchg(lock, old, new); new 604 drivers/gpu/drm/drm_mm.c void drm_mm_replace_node(struct drm_mm_node *old, struct drm_mm_node *new) new 610 drivers/gpu/drm/drm_mm.c *new = *old; new 612 drivers/gpu/drm/drm_mm.c list_replace(&old->node_list, &new->node_list); new 613 drivers/gpu/drm/drm_mm.c rb_replace_node_cached(&old->rb, &new->rb, &mm->interval_tree); new 616 drivers/gpu/drm/drm_mm.c list_replace(&old->hole_stack, &new->hole_stack); new 618 drivers/gpu/drm/drm_mm.c &new->rb_hole_size, new 621 drivers/gpu/drm/drm_mm.c &new->rb_hole_addr, new 626 drivers/gpu/drm/drm_mm.c new->allocated = true; new 622 drivers/gpu/drm/drm_probe_helper.c const char *old, *new; new 643 drivers/gpu/drm/drm_probe_helper.c new = drm_get_connector_status_name(connector->status); new 649 drivers/gpu/drm/drm_probe_helper.c old, new); new 267 drivers/gpu/drm/drm_vma_manager.c struct drm_vma_offset_file *new, *entry; new 274 drivers/gpu/drm/drm_vma_manager.c new = kmalloc(sizeof(*entry), GFP_KERNEL); new 294 drivers/gpu/drm/drm_vma_manager.c if (!new) { new 299 drivers/gpu/drm/drm_vma_manager.c new->vm_tag = tag; new 300 drivers/gpu/drm/drm_vma_manager.c new->vm_count = 1; new 301 drivers/gpu/drm/drm_vma_manager.c rb_link_node(&new->vm_rb, parent, iter); new 302 drivers/gpu/drm/drm_vma_manager.c rb_insert_color(&new->vm_rb, &node->vm_files); new 303 drivers/gpu/drm/drm_vma_manager.c new = NULL; new 307 drivers/gpu/drm/drm_vma_manager.c kfree(new); new 11471 drivers/gpu/drm/i915/display/intel_display.c struct intel_plane_state *new) new 11474 drivers/gpu/drm/i915/display/intel_display.c if (new->base.visible != cur->base.visible) new 11477 drivers/gpu/drm/i915/display/intel_display.c if (!cur->base.fb || !new->base.fb) new 11480 drivers/gpu/drm/i915/display/intel_display.c if (cur->base.fb->modifier != new->base.fb->modifier || new 11481 drivers/gpu/drm/i915/display/intel_display.c cur->base.rotation != new->base.rotation || new 11482 drivers/gpu/drm/i915/display/intel_display.c drm_rect_width(&new->base.src) != drm_rect_width(&cur->base.src) || new 11483 drivers/gpu/drm/i915/display/intel_display.c drm_rect_height(&new->base.src) != drm_rect_height(&cur->base.src) || new 11484 drivers/gpu/drm/i915/display/intel_display.c drm_rect_width(&new->base.dst) != drm_rect_width(&cur->base.dst) || new 11485 drivers/gpu/drm/i915/display/intel_display.c drm_rect_height(&new->base.dst) != drm_rect_height(&cur->base.dst)) new 286 drivers/gpu/drm/i915/display/intel_frontbuffer.c struct intel_frontbuffer *new, new 304 drivers/gpu/drm/i915/display/intel_frontbuffer.c if (new) { new 305 drivers/gpu/drm/i915/display/intel_frontbuffer.c WARN_ON(atomic_read(&new->bits) & frontbuffer_bits); new 306 drivers/gpu/drm/i915/display/intel_frontbuffer.c atomic_or(frontbuffer_bits, &new->bits); new 119 drivers/gpu/drm/i915/display/intel_frontbuffer.h struct intel_frontbuffer *new, new 742 drivers/gpu/drm/i915/gt/selftest_hangcheck.c struct i915_request *new; new 745 drivers/gpu/drm/i915/gt/selftest_hangcheck.c new = igt_request_alloc(ctx[idx], engine); new 746 drivers/gpu/drm/i915/gt/selftest_hangcheck.c if (IS_ERR(new)) { new 748 drivers/gpu/drm/i915/gt/selftest_hangcheck.c err = PTR_ERR(new); new 756 drivers/gpu/drm/i915/gt/selftest_hangcheck.c rq[idx] = i915_request_get(new); new 757 drivers/gpu/drm/i915/gt/selftest_hangcheck.c i915_request_add(new); new 382 drivers/gpu/drm/i915/gt/selftest_workarounds.c static u32 reg_write(u32 old, u32 new, u32 rsvd) new 385 drivers/gpu/drm/i915/gt/selftest_workarounds.c old &= ~(new >> 16); new 386 drivers/gpu/drm/i915/gt/selftest_workarounds.c old |= new & (new >> 16); new 389 drivers/gpu/drm/i915/gt/selftest_workarounds.c old |= new & rsvd; new 72 drivers/gpu/drm/i915/gvt/cfg_space.c u8 mask, new, old; new 78 drivers/gpu/drm/i915/gvt/cfg_space.c new = src[i] & mask; new 86 drivers/gpu/drm/i915/gvt/cfg_space.c new = (~new & old) & mask; new 88 drivers/gpu/drm/i915/gvt/cfg_space.c cfg_base[off + i] = (old & ~mask) | new; new 179 drivers/gpu/drm/i915/gvt/cfg_space.c u8 new = *(u8 *)p_data; new 180 drivers/gpu/drm/i915/gvt/cfg_space.c u8 changed = old ^ new; new 210 drivers/gpu/drm/i915/gvt/cfg_space.c u32 new = *(u32 *)(p_data); new 212 drivers/gpu/drm/i915/gvt/cfg_space.c if ((new & PCI_ROM_ADDRESS_MASK) == PCI_ROM_ADDRESS_MASK) new 223 drivers/gpu/drm/i915/gvt/cfg_space.c u32 new = *(u32 *)(p_data); new 238 drivers/gpu/drm/i915/gvt/cfg_space.c if (new == 0xffffffff) { new 271 drivers/gpu/drm/i915/gvt/cfg_space.c intel_vgpu_write_pci_bar(vgpu, offset, new, lo); new 277 drivers/gpu/drm/i915/gvt/cfg_space.c intel_vgpu_write_pci_bar(vgpu, offset, new, lo); new 281 drivers/gpu/drm/i915/gvt/cfg_space.c intel_vgpu_write_pci_bar(vgpu, offset, new, lo); new 1428 drivers/gpu/drm/i915/gvt/gtt.c struct intel_gvt_gtt_entry old, new; new 1435 drivers/gpu/drm/i915/gvt/gtt.c old.type = new.type = get_entry_type(spt->guest_page.type); new 1436 drivers/gpu/drm/i915/gvt/gtt.c old.val64 = new.val64 = 0; new 1441 drivers/gpu/drm/i915/gvt/gtt.c ops->get_entry(NULL, &new, index, true, new 1444 drivers/gpu/drm/i915/gvt/gtt.c if (old.val64 == new.val64 new 1450 drivers/gpu/drm/i915/gvt/gtt.c new.val64, index); new 1452 drivers/gpu/drm/i915/gvt/gtt.c ret = ppgtt_populate_shadow_entry(vgpu, spt, index, &new); new 1456 drivers/gpu/drm/i915/gvt/gtt.c ops->set_entry(oos_page->mem, &new, index, false, 0, vgpu); new 272 drivers/gpu/drm/i915/gvt/handlers.c #define CALC_MODE_MASK_REG(old, new) \ new 273 drivers/gpu/drm/i915/gvt/handlers.c (((new) & GENMASK(31, 16)) \ new 274 drivers/gpu/drm/i915/gvt/handlers.c | ((((old) & GENMASK(15, 0)) & ~((new) >> 16)) \ new 275 drivers/gpu/drm/i915/gvt/handlers.c | ((new) & ((new) >> 16)))) new 280 drivers/gpu/drm/i915/gvt/handlers.c u32 old, new; new 284 drivers/gpu/drm/i915/gvt/handlers.c new = CALC_MODE_MASK_REG(old, *(u32 *)p_data); new 306 drivers/gpu/drm/i915/gvt/handlers.c vgpu_vreg(vgpu, offset) = new; new 307 drivers/gpu/drm/i915/gvt/handlers.c vgpu_vreg(vgpu, ack_reg_offset) = (new & GENMASK(15, 0)); new 259 drivers/gpu/drm/i915/gvt/kvmgt.c struct gvt_dma *new, *itr; new 262 drivers/gpu/drm/i915/gvt/kvmgt.c new = kzalloc(sizeof(struct gvt_dma), GFP_KERNEL); new 263 drivers/gpu/drm/i915/gvt/kvmgt.c if (!new) new 266 drivers/gpu/drm/i915/gvt/kvmgt.c new->vgpu = vgpu; new 267 drivers/gpu/drm/i915/gvt/kvmgt.c new->gfn = gfn; new 268 drivers/gpu/drm/i915/gvt/kvmgt.c new->dma_addr = dma_addr; new 269 drivers/gpu/drm/i915/gvt/kvmgt.c new->size = size; new 270 drivers/gpu/drm/i915/gvt/kvmgt.c kref_init(&new->ref); new 283 drivers/gpu/drm/i915/gvt/kvmgt.c rb_link_node(&new->gfn_node, parent, link); new 284 drivers/gpu/drm/i915/gvt/kvmgt.c rb_insert_color(&new->gfn_node, &vgpu->vdev.gfn_cache); new 298 drivers/gpu/drm/i915/gvt/kvmgt.c rb_link_node(&new->dma_addr_node, parent, link); new 299 drivers/gpu/drm/i915/gvt/kvmgt.c rb_insert_color(&new->dma_addr_node, &vgpu->vdev.dma_addr_cache); new 3428 drivers/gpu/drm/i915/i915_debugfs.c u16 new[8] = { 0 }; new 3452 drivers/gpu/drm/i915/i915_debugfs.c &new[0], &new[1], &new[2], &new[3], new 3453 drivers/gpu/drm/i915/i915_debugfs.c &new[4], &new[5], &new[6], &new[7]); new 3460 drivers/gpu/drm/i915/i915_debugfs.c wm[level] = new[level]; new 577 drivers/gpu/drm/i915/i915_pmu.c u64 prev, new; new 581 drivers/gpu/drm/i915/i915_pmu.c new = __i915_pmu_event_read(event); new 583 drivers/gpu/drm/i915/i915_pmu.c if (local64_cmpxchg(&hwc->prev_count, prev, new) != prev) new 586 drivers/gpu/drm/i915/i915_pmu.c local64_add(new - prev, &event->count); new 52 drivers/gpu/drm/i915/i915_sw_fence.c int old, int new) new 54 drivers/gpu/drm/i915/i915_sw_fence.c debug_object_active_state(fence, &i915_sw_fence_debug_descr, old, new); new 93 drivers/gpu/drm/i915/i915_sw_fence.c int old, int new) new 146 drivers/gpu/drm/i915/i915_trace.h TP_PROTO(struct drm_i915_private *dev_priv, bool old, bool new), new 147 drivers/gpu/drm/i915/i915_trace.h TP_ARGS(dev_priv, old, new), new 153 drivers/gpu/drm/i915/i915_trace.h __field(bool, new) new 163 drivers/gpu/drm/i915/i915_trace.h __entry->new = new; new 167 drivers/gpu/drm/i915/i915_trace.h onoff(__entry->old), onoff(__entry->new), new 808 drivers/gpu/drm/i915/intel_pm.c static bool is_disabling(int old, int new, int threshold) new 810 drivers/gpu/drm/i915/intel_pm.c return old >= threshold && new < threshold; new 813 drivers/gpu/drm/i915/intel_pm.c static bool is_enabling(int old, int new, int threshold) new 815 drivers/gpu/drm/i915/intel_pm.c return old < threshold && new >= threshold; new 3465 drivers/gpu/drm/i915/intel_pm.c const struct ilk_wm_values *new) new 3472 drivers/gpu/drm/i915/intel_pm.c if (old->wm_linetime[pipe] != new->wm_linetime[pipe]) { new 3478 drivers/gpu/drm/i915/intel_pm.c if (old->wm_pipe[pipe] != new->wm_pipe[pipe]) { new 3485 drivers/gpu/drm/i915/intel_pm.c if (old->enable_fbc_wm != new->enable_fbc_wm) { new 3491 drivers/gpu/drm/i915/intel_pm.c if (old->partitioning != new->partitioning) { new 3503 drivers/gpu/drm/i915/intel_pm.c if (old->wm_lp[wm_lp - 1] != new->wm_lp[wm_lp - 1] || new 3504 drivers/gpu/drm/i915/intel_pm.c old->wm_lp_spr[wm_lp - 1] != new->wm_lp_spr[wm_lp - 1]) new 5362 drivers/gpu/drm/i915/intel_pm.c const struct skl_ddb_entry *old, *new; new 5365 drivers/gpu/drm/i915/intel_pm.c new = &new_crtc_state->wm.skl.plane_ddb_y[plane_id]; new 5367 drivers/gpu/drm/i915/intel_pm.c if (skl_ddb_entry_equal(old, new)) new 5372 drivers/gpu/drm/i915/intel_pm.c old->start, old->end, new->start, new->end, new 5373 drivers/gpu/drm/i915/intel_pm.c skl_ddb_entry_size(old), skl_ddb_entry_size(new)); new 13 drivers/gpu/drm/msm/adreno/a5xx_preempt.c enum preempt_state old, enum preempt_state new) new 16 drivers/gpu/drm/msm/adreno/a5xx_preempt.c old, new); new 26 drivers/gpu/drm/msm/adreno/a5xx_preempt.c enum preempt_state new) new 34 drivers/gpu/drm/msm/adreno/a5xx_preempt.c atomic_set(&gpu->preempt_state, new); new 285 drivers/gpu/drm/msm/disp/dpu1/dpu_core_perf.c struct dpu_core_perf_params *new, *old; new 313 drivers/gpu/drm/msm/disp/dpu1/dpu_core_perf.c new = &dpu_cstate->new_perf; new 323 drivers/gpu/drm/msm/disp/dpu1/dpu_core_perf.c if ((params_changed && ((new->bw_ctl > old->bw_ctl) || new 324 drivers/gpu/drm/msm/disp/dpu1/dpu_core_perf.c (new->max_per_pipe_ib > old->max_per_pipe_ib))) || new 325 drivers/gpu/drm/msm/disp/dpu1/dpu_core_perf.c (!params_changed && ((new->bw_ctl < old->bw_ctl) || new 326 drivers/gpu/drm/msm/disp/dpu1/dpu_core_perf.c (new->max_per_pipe_ib < old->max_per_pipe_ib)))) { new 329 drivers/gpu/drm/msm/disp/dpu1/dpu_core_perf.c new->bw_ctl, old->bw_ctl); new 330 drivers/gpu/drm/msm/disp/dpu1/dpu_core_perf.c old->bw_ctl = new->bw_ctl; new 331 drivers/gpu/drm/msm/disp/dpu1/dpu_core_perf.c old->max_per_pipe_ib = new->max_per_pipe_ib; new 336 drivers/gpu/drm/msm/disp/dpu1/dpu_core_perf.c (new->core_clk_rate > old->core_clk_rate)) || new 338 drivers/gpu/drm/msm/disp/dpu1/dpu_core_perf.c (new->core_clk_rate < old->core_clk_rate))) { new 339 drivers/gpu/drm/msm/disp/dpu1/dpu_core_perf.c old->core_clk_rate = new->core_clk_rate; new 345 drivers/gpu/drm/msm/disp/dpu1/dpu_core_perf.c memset(new, 0, sizeof(*new)); new 350 drivers/gpu/drm/msm/disp/dpu1/dpu_core_perf.c trace_dpu_perf_crtc_update(crtc->base.id, new->bw_ctl, new 351 drivers/gpu/drm/msm/disp/dpu1/dpu_core_perf.c new->core_clk_rate, stop_req, update_bus, update_clk); new 32 drivers/gpu/drm/nouveau/dispnv50/base.c int (*new)(struct nouveau_drm *, int, s32, struct nv50_wndw **); new 52 drivers/gpu/drm/nouveau/dispnv50/base.c return bases[cid].new(drm, head, bases[cid].oclass, pwndw); new 43 drivers/gpu/drm/nouveau/dispnv50/core.c int (*new)(struct nouveau_drm *, s32, struct nv50_core **); new 70 drivers/gpu/drm/nouveau/dispnv50/core.c return cores[cid].new(drm, cores[cid].oclass, pcore); new 32 drivers/gpu/drm/nouveau/dispnv50/curs.c int (*new)(struct nouveau_drm *, int, s32, struct nv50_wndw **); new 52 drivers/gpu/drm/nouveau/dispnv50/curs.c return curses[cid].new(drm, head, curses[cid].oclass, pwndw); new 33 drivers/gpu/drm/nouveau/dispnv50/ovly.c int (*new)(struct nouveau_drm *, int, s32, struct nv50_wndw **); new 52 drivers/gpu/drm/nouveau/dispnv50/ovly.c ret = ovlys[cid].new(drm, head, ovlys[cid].oclass, pwndw); new 705 drivers/gpu/drm/nouveau/dispnv50/wndw.c int (*new)(struct nouveau_drm *, enum drm_plane_type, new 721 drivers/gpu/drm/nouveau/dispnv50/wndw.c ret = wndws[cid].new(drm, type, index, wndws[cid].oclass, pwndw); new 268 drivers/gpu/drm/nouveau/nvif/object.c struct nvif_ioctl_new_v0 new; new 286 drivers/gpu/drm/nouveau/nvif/object.c args->new.version = 0; new 287 drivers/gpu/drm/nouveau/nvif/object.c args->new.route = parent->client->route; new 288 drivers/gpu/drm/nouveau/nvif/object.c args->new.token = nvif_handle(object); new 289 drivers/gpu/drm/nouveau/nvif/object.c args->new.object = nvif_handle(object); new 290 drivers/gpu/drm/nouveau/nvif/object.c args->new.handle = handle; new 291 drivers/gpu/drm/nouveau/nvif/object.c args->new.oclass = oclass; new 293 drivers/gpu/drm/nouveau/nvif/object.c memcpy(args->new.data, data, size); new 296 drivers/gpu/drm/nouveau/nvif/object.c memcpy(data, args->new.data, size); new 37 drivers/gpu/drm/nouveau/nvkm/engine/disp/g84.c .head = { .cnt = nv50_head_cnt, .new = nv50_head_new }, new 38 drivers/gpu/drm/nouveau/nvkm/engine/disp/g84.c .dac = { .cnt = nv50_dac_cnt, .new = nv50_dac_new }, new 39 drivers/gpu/drm/nouveau/nvkm/engine/disp/g84.c .sor = { .cnt = nv50_sor_cnt, .new = g84_sor_new }, new 40 drivers/gpu/drm/nouveau/nvkm/engine/disp/g84.c .pior = { .cnt = nv50_pior_cnt, .new = nv50_pior_new }, new 37 drivers/gpu/drm/nouveau/nvkm/engine/disp/g94.c .head = { .cnt = nv50_head_cnt, .new = nv50_head_new }, new 38 drivers/gpu/drm/nouveau/nvkm/engine/disp/g94.c .dac = { .cnt = nv50_dac_cnt, .new = nv50_dac_new }, new 39 drivers/gpu/drm/nouveau/nvkm/engine/disp/g94.c .sor = { .cnt = g94_sor_cnt, .new = g94_sor_new }, new 40 drivers/gpu/drm/nouveau/nvkm/engine/disp/g94.c .pior = { .cnt = nv50_pior_cnt, .new = nv50_pior_new }, new 263 drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c .head = { .cnt = gf119_head_cnt, .new = gf119_head_new }, new 264 drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c .dac = { .cnt = gf119_dac_cnt, .new = gf119_dac_new }, new 265 drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c .sor = { .cnt = gf119_sor_cnt, .new = gf119_sor_new }, new 38 drivers/gpu/drm/nouveau/nvkm/engine/disp/gk104.c .head = { .cnt = gf119_head_cnt, .new = gf119_head_new }, new 39 drivers/gpu/drm/nouveau/nvkm/engine/disp/gk104.c .dac = { .cnt = gf119_dac_cnt, .new = gf119_dac_new }, new 40 drivers/gpu/drm/nouveau/nvkm/engine/disp/gk104.c .sor = { .cnt = gf119_sor_cnt, .new = gk104_sor_new }, new 38 drivers/gpu/drm/nouveau/nvkm/engine/disp/gk110.c .head = { .cnt = gf119_head_cnt, .new = gf119_head_new }, new 39 drivers/gpu/drm/nouveau/nvkm/engine/disp/gk110.c .dac = { .cnt = gf119_dac_cnt, .new = gf119_dac_new }, new 40 drivers/gpu/drm/nouveau/nvkm/engine/disp/gk110.c .sor = { .cnt = gf119_sor_cnt, .new = gk104_sor_new }, new 38 drivers/gpu/drm/nouveau/nvkm/engine/disp/gm107.c .head = { .cnt = gf119_head_cnt, .new = gf119_head_new }, new 39 drivers/gpu/drm/nouveau/nvkm/engine/disp/gm107.c .dac = { .cnt = gf119_dac_cnt, .new = gf119_dac_new }, new 40 drivers/gpu/drm/nouveau/nvkm/engine/disp/gm107.c .sor = { .cnt = gf119_sor_cnt, .new = gm107_sor_new }, new 38 drivers/gpu/drm/nouveau/nvkm/engine/disp/gm200.c .head = { .cnt = gf119_head_cnt, .new = gf119_head_new }, new 39 drivers/gpu/drm/nouveau/nvkm/engine/disp/gm200.c .dac = { .cnt = gf119_dac_cnt, .new = gf119_dac_new }, new 40 drivers/gpu/drm/nouveau/nvkm/engine/disp/gm200.c .sor = { .cnt = gf119_sor_cnt, .new = gm200_sor_new }, new 38 drivers/gpu/drm/nouveau/nvkm/engine/disp/gp100.c .head = { .cnt = gf119_head_cnt, .new = gf119_head_new }, new 39 drivers/gpu/drm/nouveau/nvkm/engine/disp/gp100.c .sor = { .cnt = gf119_sor_cnt, .new = gm200_sor_new }, new 65 drivers/gpu/drm/nouveau/nvkm/engine/disp/gp102.c .head = { .cnt = gf119_head_cnt, .new = gf119_head_new }, new 66 drivers/gpu/drm/nouveau/nvkm/engine/disp/gp102.c .sor = { .cnt = gf119_sor_cnt, .new = gm200_sor_new }, new 37 drivers/gpu/drm/nouveau/nvkm/engine/disp/gt200.c .head = { .cnt = nv50_head_cnt, .new = nv50_head_new }, new 38 drivers/gpu/drm/nouveau/nvkm/engine/disp/gt200.c .dac = { .cnt = nv50_dac_cnt, .new = nv50_dac_new }, new 39 drivers/gpu/drm/nouveau/nvkm/engine/disp/gt200.c .sor = { .cnt = nv50_sor_cnt, .new = g84_sor_new }, new 40 drivers/gpu/drm/nouveau/nvkm/engine/disp/gt200.c .pior = { .cnt = nv50_pior_cnt, .new = nv50_pior_new }, new 37 drivers/gpu/drm/nouveau/nvkm/engine/disp/gt215.c .head = { .cnt = nv50_head_cnt, .new = nv50_head_new }, new 38 drivers/gpu/drm/nouveau/nvkm/engine/disp/gt215.c .dac = { .cnt = nv50_dac_cnt, .new = nv50_dac_new }, new 39 drivers/gpu/drm/nouveau/nvkm/engine/disp/gt215.c .sor = { .cnt = g94_sor_cnt, .new = gt215_sor_new }, new 40 drivers/gpu/drm/nouveau/nvkm/engine/disp/gt215.c .pior = { .cnt = nv50_pior_cnt, .new = nv50_pior_new }, new 421 drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c .head = { .cnt = gv100_head_cnt, .new = gv100_head_new }, new 422 drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c .sor = { .cnt = gv100_sor_cnt, .new = gv100_sor_new }, new 35 drivers/gpu/drm/nouveau/nvkm/engine/disp/mcp77.c .head = { .cnt = nv50_head_cnt, .new = nv50_head_new }, new 36 drivers/gpu/drm/nouveau/nvkm/engine/disp/mcp77.c .dac = { .cnt = nv50_dac_cnt, .new = nv50_dac_new }, new 37 drivers/gpu/drm/nouveau/nvkm/engine/disp/mcp77.c .sor = { .cnt = g94_sor_cnt, .new = mcp77_sor_new }, new 38 drivers/gpu/drm/nouveau/nvkm/engine/disp/mcp77.c .pior = { .cnt = nv50_pior_cnt, .new = nv50_pior_new }, new 35 drivers/gpu/drm/nouveau/nvkm/engine/disp/mcp89.c .head = { .cnt = nv50_head_cnt, .new = nv50_head_new }, new 36 drivers/gpu/drm/nouveau/nvkm/engine/disp/mcp89.c .dac = { .cnt = nv50_dac_cnt, .new = nv50_dac_new }, new 37 drivers/gpu/drm/nouveau/nvkm/engine/disp/mcp89.c .sor = { .cnt = g94_sor_cnt, .new = mcp89_sor_new }, new 38 drivers/gpu/drm/nouveau/nvkm/engine/disp/mcp89.c .pior = { .cnt = nv50_pior_cnt, .new = nv50_pior_new }, new 100 drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c ret = func->head.new(&disp->base, i); new 110 drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c ret = func->dac.new(&disp->base, i); new 121 drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c ret = func->pior.new(&disp->base, i); new 131 drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c ret = func->sor.new(&disp->base, i); new 765 drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c .head = { .cnt = nv50_head_cnt, .new = nv50_head_new }, new 766 drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c .dac = { .cnt = nv50_dac_cnt, .new = nv50_dac_new }, new 767 drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c .sor = { .cnt = nv50_sor_cnt, .new = nv50_sor_new }, new 768 drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c .pior = { .cnt = nv50_pior_cnt, .new = nv50_pior_new }, new 66 drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.h int (*new)(struct nvkm_disp *, int id); new 143 drivers/gpu/drm/nouveau/nvkm/engine/disp/tu102.c .head = { .cnt = gv100_head_cnt, .new = gv100_head_new }, new 144 drivers/gpu/drm/nouveau/nvkm/engine/disp/tu102.c .sor = { .cnt = gv100_sor_cnt, .new = tu102_sor_new }, new 390 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gm20b.c u32 new = pll->pl; new 398 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gm20b.c if ((old & new) == 0) { new 399 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gm20b.c cur_pll.base.pl = min(old | BIT(ffs(new) - 1), new 400 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gm20b.c new | BIT(ffs(old) - 1)); new 404 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gm20b.c cur_pll.base.pl = new; new 763 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c struct nvkm_vma *new; new 767 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c if (!(new = nvkm_vma_new(vma->addr + (vma->size - tail), tail))) new 771 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c new->mapref = vma->mapref; new 772 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c new->sparse = vma->sparse; new 773 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c new->page = vma->page; new 774 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c new->refd = vma->refd; new 775 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c new->used = vma->used; new 776 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c new->part = vma->part; new 777 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c new->user = vma->user; new 778 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c new->busy = vma->busy; new 779 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c new->mapped = vma->mapped; new 780 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c list_add(&new->head, &vma->head); new 781 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c return new; new 60 drivers/gpu/drm/radeon/mkregtable.c static inline void __list_add(struct list_head *new, new 63 drivers/gpu/drm/radeon/mkregtable.c next->prev = new; new 64 drivers/gpu/drm/radeon/mkregtable.c new->next = next; new 65 drivers/gpu/drm/radeon/mkregtable.c new->prev = prev; new 66 drivers/gpu/drm/radeon/mkregtable.c prev->next = new; new 69 drivers/gpu/drm/radeon/mkregtable.c extern void __list_add(struct list_head *new, new 81 drivers/gpu/drm/radeon/mkregtable.c static inline void list_add_tail(struct list_head *new, struct list_head *head) new 83 drivers/gpu/drm/radeon/mkregtable.c __list_add(new, head->prev, head); new 231 drivers/gpu/drm/tegra/hub.c struct tegra_dc *new) new 235 drivers/gpu/drm/tegra/hub.c struct tegra_dc *old = plane->dc, *dc = new ? new : old; new 236 drivers/gpu/drm/tegra/hub.c struct device *dev = new ? new->dev : old->dev; new 243 drivers/gpu/drm/tegra/hub.c if (new && (owner != OWNER_MASK && owner != new->pipe)) { new 259 drivers/gpu/drm/tegra/hub.c if (new) new 260 drivers/gpu/drm/tegra/hub.c value |= OWNER(new->pipe); new 266 drivers/gpu/drm/tegra/hub.c plane->dc = new; new 625 drivers/gpu/drm/tegra/hub.c struct drm_crtc_state *old, *new; new 644 drivers/gpu/drm/tegra/hub.c for_each_oldnew_crtc_in_state(state, crtc, old, new, i) { new 645 drivers/gpu/drm/tegra/hub.c struct tegra_dc_state *dc = to_dc_state(new); new 647 drivers/gpu/drm/tegra/hub.c if (new->active) { new 389 drivers/gpu/drm/tegra/plane.c struct drm_plane_state *new; new 393 drivers/gpu/drm/tegra/plane.c for_each_new_plane_in_state(state->base.state, plane, new, i) { new 403 drivers/gpu/drm/tegra/plane.c if (new->fb && __drm_format_has_alpha(new->fb->format->format)) new 408 drivers/gpu/drm/tegra/plane.c if (new->normalized_zpos > state->base.normalized_zpos) new 419 drivers/gpu/drm/tegra/plane.c if (!new->fb) new 428 drivers/gpu/drm/tegra/plane.c struct drm_plane_state *new; new 452 drivers/gpu/drm/tegra/plane.c new = drm_atomic_get_new_plane_state(state->base.state, plane); new 453 drivers/gpu/drm/tegra/plane.c tegra_state = to_tegra_plane_state(new); new 459 drivers/gpu/drm/tegra/plane.c if (new->fb) new 89 drivers/hid/usbhid/usbkbd.c unsigned char *new; new 118 drivers/hid/usbhid/usbkbd.c input_report_key(kbd->dev, usb_kbd_keycode[i + 224], (kbd->new[0] >> i) & 1); new 122 drivers/hid/usbhid/usbkbd.c if (kbd->old[i] > 3 && memscan(kbd->new + 2, kbd->old[i], 6) == kbd->new + 8) { new 131 drivers/hid/usbhid/usbkbd.c if (kbd->new[i] > 3 && memscan(kbd->old + 2, kbd->new[i], 6) == kbd->old + 8) { new 132 drivers/hid/usbhid/usbkbd.c if (usb_kbd_keycode[kbd->new[i]]) new 133 drivers/hid/usbhid/usbkbd.c input_report_key(kbd->dev, usb_kbd_keycode[kbd->new[i]], 1); new 137 drivers/hid/usbhid/usbkbd.c kbd->new[i]); new 143 drivers/hid/usbhid/usbkbd.c memcpy(kbd->old, kbd->new, 8); new 242 drivers/hid/usbhid/usbkbd.c if (!(kbd->new = usb_alloc_coherent(dev, 8, GFP_ATOMIC, &kbd->new_dma))) new 256 drivers/hid/usbhid/usbkbd.c usb_free_coherent(dev, 8, kbd->new, kbd->new_dma); new 336 drivers/hid/usbhid/usbkbd.c kbd->new, (maxp > 8 ? 8 : maxp), new 2051 drivers/hv/vmbus_drv.c int vmbus_allocate_mmio(struct resource **new, struct hv_device *device_obj, new 2076 drivers/hv/vmbus_drv.c *new = request_mem_region_exclusive(start, size, dev_n); new 2077 drivers/hv/vmbus_drv.c if (*new) { new 2097 drivers/hv/vmbus_drv.c *new = request_mem_region_exclusive(start, size, dev_n); new 2098 drivers/hv/vmbus_drv.c if (*new) { new 2099 drivers/hv/vmbus_drv.c shadow->name = (char *)*new; new 486 drivers/hwmon/lm93.c u8 new = LM93_TEMP_OFFSET_TO_REG(off, mode); new 490 drivers/hwmon/lm93.c return (old & 0xf0) | (new & 0x0f); new 494 drivers/hwmon/lm93.c return (new << 4 & 0xf0) | (old & 0x0f); new 707 drivers/hwtracing/intel_th/msu.c enum lockout_state new) new 724 drivers/hwtracing/intel_th/msu.c win->lockout = new; new 726 drivers/hwtracing/intel_th/msu.c if (old == expect && new == WIN_LOCKED) new 739 drivers/hwtracing/intel_th/msu.c if (expect == WIN_LOCKED && old == new) new 127 drivers/ide/ns87415.c unsigned int bit, other, new, *old = (unsigned int *) hwif->select_data; new 131 drivers/ide/ns87415.c new = *old; new 137 drivers/ide/ns87415.c new &= ~bit; new 139 drivers/ide/ns87415.c new |= bit; new 144 drivers/ide/ns87415.c new = use_dma ? ((new & ~other) | bit) : (new & ~bit); new 146 drivers/ide/ns87415.c if (new != *old) { new 159 drivers/ide/ns87415.c *old = new; new 160 drivers/ide/ns87415.c (void) pci_write_config_dword(dev, 0x40, new); new 281 drivers/ide/trm290.c u16 new, old, compat = hwif->channel ? 0x374 : 0x3f4; new 294 drivers/ide/trm290.c new = inw(hwif->config_data); new 297 drivers/ide/trm290.c hwif->name, old, new & ~1); new 202 drivers/iio/multiplexer/iio-mux.c char *new; new 212 drivers/iio/multiplexer/iio-mux.c new = devm_kmemdup(dev, buf, len + 1, GFP_KERNEL); new 213 drivers/iio/multiplexer/iio-mux.c if (!new) { new 218 drivers/iio/multiplexer/iio-mux.c new[len] = 0; new 225 drivers/iio/multiplexer/iio-mux.c devm_kfree(dev, new); new 230 drivers/iio/multiplexer/iio-mux.c mux->child[idx].ext_info_cache[private].data = new; new 376 drivers/infiniband/core/verbs.c const struct rdma_ah_attr *new) new 379 drivers/infiniband/core/verbs.c *old = *new; new 248 drivers/infiniband/hw/cxgb3/iwch_cm.c static void __state_set(struct iwch_ep_common *epc, enum iwch_ep_state new) new 250 drivers/infiniband/hw/cxgb3/iwch_cm.c epc->state = new; new 253 drivers/infiniband/hw/cxgb3/iwch_cm.c static void state_set(struct iwch_ep_common *epc, enum iwch_ep_state new) new 258 drivers/infiniband/hw/cxgb3/iwch_cm.c pr_debug("%s - %s -> %s\n", __func__, states[epc->state], states[new]); new 259 drivers/infiniband/hw/cxgb3/iwch_cm.c __state_set(epc, new); new 2126 drivers/infiniband/hw/cxgb3/iwch_cm.c int iwch_ep_redirect(void *ctx, struct dst_entry *old, struct dst_entry *new, new 2134 drivers/infiniband/hw/cxgb3/iwch_cm.c pr_debug("%s ep %p redirect to dst %p l2t %p\n", __func__, ep, new, new 2136 drivers/infiniband/hw/cxgb3/iwch_cm.c dst_hold(new); new 2140 drivers/infiniband/hw/cxgb3/iwch_cm.c ep->dst = new; new 227 drivers/infiniband/hw/cxgb3/iwch_cm.h int iwch_ep_redirect(void *ctx, struct dst_entry *old, struct dst_entry *new, struct l2t_entry *l2t); new 276 drivers/infiniband/hw/cxgb4/cm.c static void __state_set(struct c4iw_ep_common *epc, enum c4iw_ep_state new) new 278 drivers/infiniband/hw/cxgb4/cm.c epc->state = new; new 281 drivers/infiniband/hw/cxgb4/cm.c static void state_set(struct c4iw_ep_common *epc, enum c4iw_ep_state new) new 284 drivers/infiniband/hw/cxgb4/cm.c pr_debug("%s -> %s\n", states[epc->state], states[new]); new 285 drivers/infiniband/hw/cxgb4/cm.c __state_set(epc, new); new 945 drivers/infiniband/hw/cxgb4/iw_cxgb4.h int c4iw_ep_redirect(void *ctx, struct dst_entry *old, struct dst_entry *new, new 1150 drivers/infiniband/hw/hfi1/mad.c static int logical_transition_allowed(int old, int new) new 1153 drivers/infiniband/hw/hfi1/mad.c new < IB_PORT_NOP || new > IB_PORT_ACTIVE_DEFER) { new 1155 drivers/infiniband/hw/hfi1/mad.c old, new); new 1159 drivers/infiniband/hw/hfi1/mad.c if (new == IB_PORT_NOP) new 1164 drivers/infiniband/hw/hfi1/mad.c new -= IB_PORT_DOWN; new 1166 drivers/infiniband/hw/hfi1/mad.c if (old < 0 || new < 0) new 1168 drivers/infiniband/hw/hfi1/mad.c return logical_state_transitions.allowed[old][new]; new 1171 drivers/infiniband/hw/hfi1/mad.c static int physical_transition_allowed(int old, int new) new 1174 drivers/infiniband/hw/hfi1/mad.c new < IB_PORTPHYSSTATE_NOP || new > OPA_PORTPHYSSTATE_MAX) { new 1176 drivers/infiniband/hw/hfi1/mad.c old, new); new 1180 drivers/infiniband/hw/hfi1/mad.c if (new == IB_PORTPHYSSTATE_NOP) new 1185 drivers/infiniband/hw/hfi1/mad.c new -= IB_PORTPHYSSTATE_POLLING; new 1187 drivers/infiniband/hw/hfi1/mad.c if (old < 0 || new < 0) new 1189 drivers/infiniband/hw/hfi1/mad.c return physical_state_transitions.allowed[old][new]; new 262 drivers/infiniband/hw/hfi1/pio_copy.c u64 new, temp; new 264 drivers/infiniband/hw/hfi1/pio_copy.c new = *(u64 *)src; new 265 drivers/infiniband/hw/hfi1/pio_copy.c temp = pbuf->carry.val64 | (new << mshift(pbuf->carry_bytes)); new 267 drivers/infiniband/hw/hfi1/pio_copy.c pbuf->carry.val64 = new >> zshift(pbuf->carry_bytes); new 189 drivers/infiniband/hw/mlx4/cm.c static void sl_id_map_add(struct ib_device *ibdev, struct id_map_entry *new) new 194 drivers/infiniband/hw/mlx4/cm.c int slave_id = new->slave_id; new 195 drivers/infiniband/hw/mlx4/cm.c int sl_cm_id = new->sl_cm_id; new 202 drivers/infiniband/hw/mlx4/cm.c rb_replace_node(&ent->node, &new->node, sl_id_map); new 217 drivers/infiniband/hw/mlx4/cm.c rb_link_node(&new->node, parent, link); new 218 drivers/infiniband/hw/mlx4/cm.c rb_insert_color(&new->node, sl_id_map); new 1220 drivers/infiniband/hw/mlx5/mlx5_ib.h void mlx5_ib_copy_pas(u64 *old, u64 *new, int step, int num); new 160 drivers/infiniband/hw/qib/qib_user_sdma.c qib_user_sdma_rb_insert(struct rb_root *root, struct qib_user_sdma_rb_node *new) new 169 drivers/infiniband/hw/qib/qib_user_sdma.c if (new->pid < got->pid) new 171 drivers/infiniband/hw/qib/qib_user_sdma.c else if (new->pid > got->pid) new 177 drivers/infiniband/hw/qib/qib_user_sdma.c rb_link_node(&new->node, parent, node); new 178 drivers/infiniband/hw/qib/qib_user_sdma.c rb_insert_color(&new->node, root); new 283 drivers/infiniband/sw/rxe/rxe_pool.c static void insert_index(struct rxe_pool *pool, struct rxe_pool_entry *new) new 293 drivers/infiniband/sw/rxe/rxe_pool.c if (elem->index == new->index) { new 298 drivers/infiniband/sw/rxe/rxe_pool.c if (elem->index > new->index) new 304 drivers/infiniband/sw/rxe/rxe_pool.c rb_link_node(&new->node, parent, link); new 305 drivers/infiniband/sw/rxe/rxe_pool.c rb_insert_color(&new->node, &pool->tree); new 310 drivers/infiniband/sw/rxe/rxe_pool.c static void insert_key(struct rxe_pool *pool, struct rxe_pool_entry *new) new 322 drivers/infiniband/sw/rxe/rxe_pool.c (u8 *)new + pool->key_offset, pool->key_size); new 335 drivers/infiniband/sw/rxe/rxe_pool.c rb_link_node(&new->node, parent, link); new 336 drivers/infiniband/sw/rxe/rxe_pool.c rb_insert_color(&new->node, &pool->tree); new 226 drivers/infiniband/sw/siw/siw_mem.c struct siw_mem *new = NULL; new 234 drivers/infiniband/sw/siw/siw_mem.c new = siw_mem_id2obj(sdev, sge->lkey >> 8); new 235 drivers/infiniband/sw/siw/siw_mem.c if (unlikely(!new)) { new 240 drivers/infiniband/sw/siw/siw_mem.c *mem = new; new 255 drivers/infiniband/sw/siw/siw_mem.c if (new) { new 257 drivers/infiniband/sw/siw/siw_mem.c siw_mem_put(new); new 108 drivers/infiniband/ulp/srpt/ib_srpt.c static bool srpt_set_ch_state(struct srpt_rdma_ch *ch, enum rdma_ch_state new) new 116 drivers/infiniband/ulp/srpt/ib_srpt.c if (new > prev) { new 117 drivers/infiniband/ulp/srpt/ib_srpt.c ch->state = new; new 779 drivers/infiniband/ulp/srpt/ib_srpt.c enum srpt_command_state new) new 785 drivers/infiniband/ulp/srpt/ib_srpt.c ioctx->state = new; new 800 drivers/infiniband/ulp/srpt/ib_srpt.c enum srpt_command_state new) new 806 drivers/infiniband/ulp/srpt/ib_srpt.c WARN_ON(new == SRPT_STATE_NEW); new 810 drivers/infiniband/ulp/srpt/ib_srpt.c ioctx->state = new; new 247 drivers/input/ff-memless.c struct ff_effect *new = state->effect; new 252 drivers/input/ff-memless.c switch (new->type) { new 254 drivers/input/ff-memless.c i = new->direction * 360 / 0xffff; new 256 drivers/input/ff-memless.c new->u.constant.level, new 257 drivers/input/ff-memless.c &new->u.constant.envelope)); new 272 drivers/input/ff-memless.c strong = (u32)new->u.rumble.strong_magnitude * gain / 0xffff; new 273 drivers/input/ff-memless.c weak = (u32)new->u.rumble.weak_magnitude * gain / 0xffff; new 279 drivers/input/ff-memless.c new->direction, strong); new 284 drivers/input/ff-memless.c new->direction, weak); new 295 drivers/input/ff-memless.c i = apply_envelope(state, abs(new->u.periodic.magnitude), new 296 drivers/input/ff-memless.c &new->u.periodic.envelope); new 305 drivers/input/ff-memless.c new->direction, i); new 183 drivers/input/joystick/iforce/iforce-ff.c struct ff_effect *new) new 188 drivers/input/joystick/iforce/iforce-ff.c if (new->type != FF_SPRING && new->type != FF_FRICTION) { new 195 drivers/input/joystick/iforce/iforce-ff.c ret |= old->u.condition[i].right_saturation != new->u.condition[i].right_saturation new 196 drivers/input/joystick/iforce/iforce-ff.c || old->u.condition[i].left_saturation != new->u.condition[i].left_saturation new 197 drivers/input/joystick/iforce/iforce-ff.c || old->u.condition[i].right_coeff != new->u.condition[i].right_coeff new 198 drivers/input/joystick/iforce/iforce-ff.c || old->u.condition[i].left_coeff != new->u.condition[i].left_coeff new 199 drivers/input/joystick/iforce/iforce-ff.c || old->u.condition[i].deadband != new->u.condition[i].deadband new 200 drivers/input/joystick/iforce/iforce-ff.c || old->u.condition[i].center != new->u.condition[i].center; new 259 drivers/input/joystick/iforce/iforce-ff.c struct ff_effect *new) new 261 drivers/input/joystick/iforce/iforce-ff.c if (new->type != FF_PERIODIC) { new 266 drivers/input/joystick/iforce/iforce-ff.c return (old->u.periodic.period != new->u.periodic.period new 267 drivers/input/joystick/iforce/iforce-ff.c || old->u.periodic.magnitude != new->u.periodic.magnitude new 268 drivers/input/joystick/iforce/iforce-ff.c || old->u.periodic.offset != new->u.periodic.offset new 269 drivers/input/joystick/iforce/iforce-ff.c || old->u.periodic.phase != new->u.periodic.phase); new 276 drivers/input/joystick/iforce/iforce-ff.c static int need_core(struct ff_effect *old, struct ff_effect *new) new 278 drivers/input/joystick/iforce/iforce-ff.c if (old->direction != new->direction new 279 drivers/input/joystick/iforce/iforce-ff.c || old->trigger.button != new->trigger.button new 280 drivers/input/joystick/iforce/iforce-ff.c || old->trigger.interval != new->trigger.interval new 281 drivers/input/joystick/iforce/iforce-ff.c || old->replay.length != new->replay.length new 282 drivers/input/joystick/iforce/iforce-ff.c || old->replay.delay != new->replay.delay) new 82 drivers/input/keyboard/jornada680_kbd.c unsigned char new = jornadakbd->new_scan[i]; new 84 drivers/input/keyboard/jornada680_kbd.c unsigned int xor = new ^ old; new 97 drivers/input/keyboard/jornada680_kbd.c !(new & bit)); new 29 drivers/input/keyboard/maple_keyb.c unsigned char new[8]; new 92 drivers/input/keyboard/maple_keyb.c input_report_key(dev, keycode, (kbd->new[0] >> i) & 1); new 96 drivers/input/keyboard/maple_keyb.c ptr = memchr(kbd->new + 2, kbd->old[i], 6); new 108 drivers/input/keyboard/maple_keyb.c ptr = memchr(kbd->old + 2, kbd->new[i], 6); new 109 drivers/input/keyboard/maple_keyb.c code = kbd->new[i]; new 122 drivers/input/keyboard/maple_keyb.c memcpy(kbd->old, kbd->new, 8); new 138 drivers/input/keyboard/maple_keyb.c memcpy(kbd->new, buf + 2, 8); new 635 drivers/interconnect/core.c struct icc_node **new; new 653 drivers/interconnect/core.c new = krealloc(node->links, new 656 drivers/interconnect/core.c if (!new) { new 661 drivers/interconnect/core.c node->links = new; new 680 drivers/interconnect/core.c struct icc_node **new; new 703 drivers/interconnect/core.c new = krealloc(src->links, src->num_links * sizeof(*src->links), new 705 drivers/interconnect/core.c if (new) new 706 drivers/interconnect/core.c src->links = new; new 454 drivers/iommu/io-pgtable-arm-v7s.c arm_v7s_iopte old, new; new 456 drivers/iommu/io-pgtable-arm-v7s.c new = virt_to_phys(table) | ARM_V7S_PTE_TYPE_TABLE; new 458 drivers/iommu/io-pgtable-arm-v7s.c new |= ARM_V7S_ATTR_NS_TABLE; new 467 drivers/iommu/io-pgtable-arm-v7s.c old = cmpxchg_relaxed(ptep, curr, new); new 356 drivers/iommu/io-pgtable-arm.c arm_lpae_iopte old, new; new 358 drivers/iommu/io-pgtable-arm.c new = __pa(table) | ARM_LPAE_PTE_TYPE_TABLE; new 360 drivers/iommu/io-pgtable-arm.c new |= ARM_LPAE_PTE_NSTABLE; new 369 drivers/iommu/io-pgtable-arm.c old = cmpxchg64_relaxed(ptep, curr, new); new 377 drivers/iommu/io-pgtable-arm.c WRITE_ONCE(*ptep, new | ARM_LPAE_PTE_SW_SYNC); new 292 drivers/iommu/iommu.c int iommu_insert_resv_region(struct iommu_resv_region *new, new 298 drivers/iommu/iommu.c nr = iommu_alloc_resv_region(new->start, new->length, new 299 drivers/iommu/iommu.c new->prot, new->type); new 316 drivers/iommu/iommu.c if (iter->type != new->type) { new 126 drivers/iommu/iova.c __cached_rbnode_insert_update(struct iova_domain *iovad, struct iova *new) new 128 drivers/iommu/iova.c if (new->pfn_hi < iovad->dma_32bit_pfn) new 129 drivers/iommu/iova.c iovad->cached32_node = &new->node; new 131 drivers/iommu/iova.c iovad->cached_node = &new->node; new 157 drivers/iommu/iova.c struct rb_node **new, *parent = NULL; new 159 drivers/iommu/iova.c new = (start) ? &start : &(root->rb_node); new 161 drivers/iommu/iova.c while (*new) { new 162 drivers/iommu/iova.c struct iova *this = rb_entry(*new, struct iova, node); new 164 drivers/iommu/iova.c parent = *new; new 167 drivers/iommu/iova.c new = &((*new)->rb_left); new 169 drivers/iommu/iova.c new = &((*new)->rb_right); new 176 drivers/iommu/iova.c rb_link_node(&iova->node, parent, new); new 182 drivers/iommu/iova.c struct iova *new, bool size_aligned) new 215 drivers/iommu/iova.c new->pfn_lo = new_pfn; new 216 drivers/iommu/iova.c new->pfn_hi = new->pfn_lo + size - 1; new 219 drivers/iommu/iova.c iova_insert_rbtree(&iovad->rbroot, new, prev); new 220 drivers/iommu/iova.c __cached_rbnode_insert_update(iovad, new); new 1536 drivers/irqchip/irq-gic-v3-its.c struct lpi_range *new, *old; new 1538 drivers/irqchip/irq-gic-v3-its.c new = mk_lpi_range(base, nr_lpis); new 1539 drivers/irqchip/irq-gic-v3-its.c if (!new) new 1555 drivers/irqchip/irq-gic-v3-its.c list_add(&new->entry, &old->entry); new 1560 drivers/irqchip/irq-gic-v3-its.c merge_lpi_ranges(old, new); new 1561 drivers/irqchip/irq-gic-v3-its.c merge_lpi_ranges(new, list_next_entry(new, entry)); new 25 drivers/isdn/mISDN/dsp_hwec.c .new = NULL, new 267 drivers/isdn/mISDN/dsp_pipeline.c pipeline_entry->p = elem->new(args); new 293 drivers/leds/leds-tca6507.c static void set_code(struct tca6507_chip *tca, int reg, int bank, int new) new 299 drivers/leds/leds-tca6507.c new <<= 4; new 302 drivers/leds/leds-tca6507.c n |= new; new 1546 drivers/lightnvm/pblk-core.c struct pblk_line *cur, *new = NULL; new 1549 drivers/lightnvm/pblk-core.c new = l_mg->data_next; new 1550 drivers/lightnvm/pblk-core.c if (!new) new 1555 drivers/lightnvm/pblk-core.c l_mg->data_line = new; new 1557 drivers/lightnvm/pblk-core.c pblk_line_setup_metadata(new, l_mg, &pblk->lm); new 1561 drivers/lightnvm/pblk-core.c left_seblks = atomic_read(&new->left_seblks); new 1564 drivers/lightnvm/pblk-core.c if (atomic_read(&new->left_eblks)) { new 1565 drivers/lightnvm/pblk-core.c if (pblk_line_erase(pblk, new)) new 1573 drivers/lightnvm/pblk-core.c if (pblk_line_alloc_bitmaps(pblk, new)) new 1577 drivers/lightnvm/pblk-core.c if (!pblk_line_init_metadata(pblk, new, cur)) { new 1578 drivers/lightnvm/pblk-core.c new = pblk_line_retry(pblk, new); new 1579 drivers/lightnvm/pblk-core.c if (!new) new 1585 drivers/lightnvm/pblk-core.c if (!pblk_line_init_bb(pblk, new, 1)) { new 1586 drivers/lightnvm/pblk-core.c new = pblk_line_retry(pblk, new); new 1587 drivers/lightnvm/pblk-core.c if (!new) new 1593 drivers/lightnvm/pblk-core.c pblk_rl_free_lines_dec(&pblk->rl, new, true); new 1603 drivers/lightnvm/pblk-core.c pblk_stop_writes(pblk, new); new 1612 drivers/lightnvm/pblk-core.c return new; new 1325 drivers/md/bcache/bset.c void bch_btree_sort_into(struct btree_keys *b, struct btree_keys *new, new 1333 drivers/md/bcache/bset.c btree_mergesort(b, new->set->data, &iter, false, true); new 1337 drivers/md/bcache/bset.c new->set->size = 0; // XXX: why? new 376 drivers/md/bcache/bset.h void bch_btree_sort_into(struct btree_keys *b, struct btree_keys *new, new 97 drivers/md/bcache/io.c unsigned int new = count - ca->set->error_decay; new 104 drivers/md/bcache/io.c count = atomic_cmpxchg(&ca->io_count, old, new); new 107 drivers/md/bcache/io.c count = new; new 112 drivers/md/bcache/io.c new = ((uint64_t) errors * 127) / 128; new 114 drivers/md/bcache/io.c old, new); new 470 drivers/md/bcache/util.h #define RB_INSERT(root, new, member, cmp) \ new 474 drivers/md/bcache/util.h typeof(new) this; \ new 479 drivers/md/bcache/util.h this = container_of(*n, typeof(*(new)), member); \ new 480 drivers/md/bcache/util.h res = cmp(new, this); \ new 488 drivers/md/bcache/util.h rb_link_node(&(new)->member, parent, n); \ new 489 drivers/md/bcache/util.h rb_insert_color(&(new)->member, root); \ new 116 drivers/md/dm-bio-prison-v1.c struct rb_node **new = &prison->cells.rb_node, *parent = NULL; new 118 drivers/md/dm-bio-prison-v1.c while (*new) { new 120 drivers/md/dm-bio-prison-v1.c rb_entry(*new, struct dm_bio_prison_cell, node); new 124 drivers/md/dm-bio-prison-v1.c parent = *new; new 126 drivers/md/dm-bio-prison-v1.c new = &((*new)->rb_left); new 128 drivers/md/dm-bio-prison-v1.c new = &((*new)->rb_right); new 140 drivers/md/dm-bio-prison-v1.c rb_link_node(&cell_prealloc->node, parent, new); new 120 drivers/md/dm-bio-prison-v2.c struct rb_node **new = &prison->cells.rb_node, *parent = NULL; new 122 drivers/md/dm-bio-prison-v2.c while (*new) { new 124 drivers/md/dm-bio-prison-v2.c rb_entry(*new, struct dm_bio_prison_cell_v2, node); new 128 drivers/md/dm-bio-prison-v2.c parent = *new; new 130 drivers/md/dm-bio-prison-v2.c new = &((*new)->rb_left); new 133 drivers/md/dm-bio-prison-v2.c new = &((*new)->rb_right); new 143 drivers/md/dm-bio-prison-v2.c rb_link_node(&cell_prealloc->node, parent, new); new 267 drivers/md/dm-bufio.c struct rb_node **new = &c->buffer_tree.rb_node, *parent = NULL; new 270 drivers/md/dm-bufio.c while (*new) { new 271 drivers/md/dm-bufio.c found = container_of(*new, struct dm_buffer, node); new 278 drivers/md/dm-bufio.c parent = *new; new 279 drivers/md/dm-bufio.c new = (found->block < b->block) ? new 280 drivers/md/dm-bufio.c &((*new)->rb_left) : &((*new)->rb_right); new 283 drivers/md/dm-bufio.c rb_link_node(&b->node, parent, new); new 1343 drivers/md/dm-bufio.c struct dm_buffer *new; new 1350 drivers/md/dm-bufio.c new = __find(c, new_block); new 1351 drivers/md/dm-bufio.c if (new) { new 1352 drivers/md/dm-bufio.c if (new->hold_count) { new 1361 drivers/md/dm-bufio.c __make_buffer_clean(new); new 1362 drivers/md/dm-bufio.c __unlink_buffer(new); new 1363 drivers/md/dm-bufio.c __free_buffer_wake(new); new 84 drivers/md/dm-cache-background-tracker.c struct rb_node **new = &b->pending.rb_node, *parent = NULL; new 86 drivers/md/dm-cache-background-tracker.c while (*new) { new 87 drivers/md/dm-cache-background-tracker.c w = container_of(*new, struct bt_work, node); new 89 drivers/md/dm-cache-background-tracker.c parent = *new; new 92 drivers/md/dm-cache-background-tracker.c new = &((*new)->rb_left); new 95 drivers/md/dm-cache-background-tracker.c new = &((*new)->rb_right); new 102 drivers/md/dm-cache-background-tracker.c rb_link_node(&nw->node, parent, new); new 113 drivers/md/dm-cache-background-tracker.c struct rb_node **new = &b->pending.rb_node; new 115 drivers/md/dm-cache-background-tracker.c while (*new) { new 116 drivers/md/dm-cache-background-tracker.c w = container_of(*new, struct bt_work, node); new 120 drivers/md/dm-cache-background-tracker.c new = &((*new)->rb_left); new 123 drivers/md/dm-cache-background-tracker.c new = &((*new)->rb_right); new 129 drivers/md/dm-cache-background-tracker.c return *new ? w : NULL; new 53 drivers/md/dm-dust.c static bool dust_rb_insert(struct rb_root *root, struct badblock *new) new 57 drivers/md/dm-dust.c sector_t value = new->bb; new 71 drivers/md/dm-dust.c rb_link_node(&new->node, parent, link); new 72 drivers/md/dm-dust.c rb_insert_color(&new->node, root); new 60 drivers/md/dm-exception-store.h chunk_t old, chunk_t new), new 370 drivers/md/dm-ioctl.c const char *new) new 382 drivers/md/dm-ioctl.c new_data = kstrdup(new, GFP_KERNEL); new 392 drivers/md/dm-ioctl.c hc = __get_uuid_cell(new); new 394 drivers/md/dm-ioctl.c hc = __get_name_cell(new); new 400 drivers/md/dm-ioctl.c param->name, new); new 413 drivers/md/dm-ioctl.c param->name, change_uuid ? "uuid " : "", new); new 425 drivers/md/dm-ioctl.c param->name, new, hc->uuid); new 1320 drivers/md/dm-ioctl.c static bool is_valid_type(enum dm_queue_mode cur, enum dm_queue_mode new) new 1322 drivers/md/dm-ioctl.c if (cur == new || new 1323 drivers/md/dm-ioctl.c (cur == DM_TYPE_BIO_BASED && new == DM_TYPE_DAX_BIO_BASED)) new 214 drivers/md/dm-raid1.c struct mirror *new; new 244 drivers/md/dm-raid1.c new = get_valid_mirror(ms); new 245 drivers/md/dm-raid1.c if (new) new 246 drivers/md/dm-raid1.c set_default_mirror(new); new 450 drivers/md/dm-snap-persistent.c chunk_t old, chunk_t new), new 495 drivers/md/dm-snap-persistent.c chunk_t new), new 613 drivers/md/dm-snap-persistent.c chunk_t old, chunk_t new), new 33 drivers/md/dm-snap-transient.c chunk_t old, chunk_t new), new 815 drivers/md/dm-snap.c static int dm_add_exception(void *context, chunk_t old, chunk_t new) new 828 drivers/md/dm-snap.c e->new_chunk = new; new 329 drivers/md/dm-zoned-metadata.c struct rb_node **new = &(root->rb_node), *parent = NULL; new 333 drivers/md/dm-zoned-metadata.c while (*new) { new 334 drivers/md/dm-zoned-metadata.c b = container_of(*new, struct dmz_mblock, node); new 335 drivers/md/dm-zoned-metadata.c parent = *new; new 336 drivers/md/dm-zoned-metadata.c new = (b->no < mblk->no) ? &((*new)->rb_left) : &((*new)->rb_right); new 340 drivers/md/dm-zoned-metadata.c rb_link_node(&mblk->node, parent, new); new 649 drivers/md/md.c struct mddev *mddev, *new = NULL; new 662 drivers/md/md.c kfree(new); new 666 drivers/md/md.c if (new) { new 667 drivers/md/md.c list_add(&new->all_mddevs, &all_mddevs); new 669 drivers/md/md.c new->hold_active = UNTIL_IOCTL; new 670 drivers/md/md.c return new; new 672 drivers/md/md.c } else if (new) { new 686 drivers/md/md.c kfree(new); new 697 drivers/md/md.c new->unit = dev; new 698 drivers/md/md.c new->md_minor = MINOR(dev); new 699 drivers/md/md.c new->hold_active = UNTIL_STOP; new 700 drivers/md/md.c list_add(&new->all_mddevs, &all_mddevs); new 702 drivers/md/md.c return new; new 706 drivers/md/md.c new = kzalloc(sizeof(*new), GFP_KERNEL); new 707 drivers/md/md.c if (!new) new 710 drivers/md/md.c new->unit = unit; new 712 drivers/md/md.c new->md_minor = MINOR(unit); new 714 drivers/md/md.c new->md_minor = MINOR(unit) >> MdpMinorShift; new 716 drivers/md/md.c mddev_init(new); new 3231 drivers/md/md.c sector_t new; new 3239 drivers/md/md.c new = blocks * 2; new 3240 drivers/md/md.c if (new != blocks * 2) new 3243 drivers/md/md.c *sectors = new; new 4997 drivers/md/md.c unsigned long long new; new 5000 drivers/md/md.c err = kstrtoull(buf, 10, &new); new 5003 drivers/md/md.c if (new != (sector_t)new) new 5014 drivers/md/md.c mddev->suspend_lo = new; new 5034 drivers/md/md.c unsigned long long new; new 5037 drivers/md/md.c err = kstrtoull(buf, 10, &new); new 5040 drivers/md/md.c if (new != (sector_t)new) new 5051 drivers/md/md.c mddev->suspend_hi = new; new 5076 drivers/md/md.c unsigned long long new; new 5079 drivers/md/md.c err = kstrtoull(buf, 10, &new); new 5082 drivers/md/md.c if (new != (sector_t)new) new 5090 drivers/md/md.c mddev->reshape_position = new; new 411 drivers/md/persistent-data/dm-space-map-common.c int (*mutator)(void *context, uint32_t old, uint32_t *new), new 498 drivers/md/persistent-data/dm-space-map-common.c static int set_ref_count(void *context, uint32_t old, uint32_t *new) new 500 drivers/md/persistent-data/dm-space-map-common.c *new = *((uint32_t *) context); new 510 drivers/md/persistent-data/dm-space-map-common.c static int inc_ref_count(void *context, uint32_t old, uint32_t *new) new 512 drivers/md/persistent-data/dm-space-map-common.c *new = old + 1; new 521 drivers/md/persistent-data/dm-space-map-common.c static int dec_ref_count(void *context, uint32_t old, uint32_t *new) new 528 drivers/md/persistent-data/dm-space-map-common.c *new = old - 1; new 268 drivers/md/persistent-data/dm-transaction-manager.c dm_block_t new; new 271 drivers/md/persistent-data/dm-transaction-manager.c r = dm_sm_new_block(tm->sm, &new); new 290 drivers/md/persistent-data/dm-transaction-manager.c r = dm_bm_write_lock_zero(tm->bm, new, v, result); new 3575 drivers/md/raid10.c static int setup_geo(struct geom *geo, struct mddev *mddev, enum geo_type new) new 3579 drivers/md/raid10.c switch (new) { new 4190 drivers/md/raid10.c struct geom new; new 4199 drivers/md/raid10.c if (setup_geo(&new, mddev, geo_start) != conf->copies) new 1566 drivers/md/raid5-cache.c unsigned long new = (unsigned long)space; /* overflow in theory */ new 1572 drivers/md/raid5-cache.c if (new < target) new 1574 drivers/md/raid5-cache.c } while (cmpxchg(&log->reclaim_target, target, new) != target); new 1526 drivers/md/raid5-ppl.c unsigned short new; new 1530 drivers/md/raid5-ppl.c if (kstrtou16(page, 10, &new)) new 1545 drivers/md/raid5-ppl.c ppl_conf->write_hint = new; new 6408 drivers/md/raid5.c unsigned long new; new 6413 drivers/md/raid5.c if (kstrtoul(page, 10, &new)) new 6422 drivers/md/raid5.c err = raid5_set_cache_size(mddev, new); new 6447 drivers/md/raid5.c unsigned long new; new 6455 drivers/md/raid5.c if (kstrtoul(page, 10, &new)) new 6458 drivers/md/raid5.c if (new != PARITY_DISABLE_RMW && !raid6_call.xor_syndrome) new 6461 drivers/md/raid5.c if (new != PARITY_DISABLE_RMW && new 6462 drivers/md/raid5.c new != PARITY_ENABLE_RMW && new 6463 drivers/md/raid5.c new != PARITY_PREFER_RMW) new 6466 drivers/md/raid5.c conf->rmw_level = new; new 6493 drivers/md/raid5.c unsigned long new; new 6498 drivers/md/raid5.c if (kstrtoul(page, 10, &new)) new 6507 drivers/md/raid5.c else if (new > conf->min_nr_stripes) new 6510 drivers/md/raid5.c conf->bypass_threshold = new; new 6538 drivers/md/raid5.c unsigned long new; new 6543 drivers/md/raid5.c if (kstrtoul(page, 10, &new)) new 6545 drivers/md/raid5.c new = !!new; new 6553 drivers/md/raid5.c else if (new != conf->skip_copy) { new 6555 drivers/md/raid5.c conf->skip_copy = new; new 6556 drivers/md/raid5.c if (new) new 6607 drivers/md/raid5.c unsigned int new; new 6614 drivers/md/raid5.c if (kstrtouint(page, 10, &new)) new 6617 drivers/md/raid5.c if (new > 8192) new 6626 drivers/md/raid5.c else if (new != conf->worker_cnt_per_group) { new 6633 drivers/md/raid5.c err = alloc_thread_groups(conf, new, new 1026 drivers/media/common/cx2341x.c const struct cx2341x_mpeg_params *new) new 1038 drivers/media/common/cx2341x.c cx2341x_api(priv, func, CX2341X_ENC_SET_OUTPUT_PORT, 2, new->port, 0); new 1041 drivers/media/common/cx2341x.c CMP_FIELD(old, new, is_50hz)) { new 1043 drivers/media/common/cx2341x.c new->is_50hz); new 1049 drivers/media/common/cx2341x.c CMP_FIELD(old, new, width) || new 1050 drivers/media/common/cx2341x.c CMP_FIELD(old, new, height) || new 1051 drivers/media/common/cx2341x.c CMP_FIELD(old, new, video_encoding)) { new 1052 drivers/media/common/cx2341x.c u16 w = new->width; new 1053 drivers/media/common/cx2341x.c u16 h = new->height; new 1055 drivers/media/common/cx2341x.c if (new->video_encoding == V4L2_MPEG_VIDEO_ENCODING_MPEG_1) { new 1065 drivers/media/common/cx2341x.c CMP_FIELD(old, new, stream_type)) { new 1067 drivers/media/common/cx2341x.c mpeg_stream_type[new->stream_type]); new 1072 drivers/media/common/cx2341x.c CMP_FIELD(old, new, video_aspect)) { new 1074 drivers/media/common/cx2341x.c 1 + new->video_aspect); new 1079 drivers/media/common/cx2341x.c CMP_FIELD(old, new, video_b_frames) || new 1080 drivers/media/common/cx2341x.c CMP_FIELD(old, new, video_gop_size)) { new 1082 drivers/media/common/cx2341x.c new->video_gop_size, new->video_b_frames + 1); new 1087 drivers/media/common/cx2341x.c CMP_FIELD(old, new, video_gop_closure)) { new 1089 drivers/media/common/cx2341x.c new->video_gop_closure); new 1094 drivers/media/common/cx2341x.c CMP_FIELD(old, new, audio_properties)) { new 1096 drivers/media/common/cx2341x.c 1, new->audio_properties); new 1101 drivers/media/common/cx2341x.c CMP_FIELD(old, new, audio_mute)) { new 1103 drivers/media/common/cx2341x.c new->audio_mute); new 1108 drivers/media/common/cx2341x.c CMP_FIELD(old, new, video_bitrate_mode) || new 1109 drivers/media/common/cx2341x.c CMP_FIELD(old, new, video_bitrate) || new 1110 drivers/media/common/cx2341x.c CMP_FIELD(old, new, video_bitrate_peak)) { new 1112 drivers/media/common/cx2341x.c new->video_bitrate_mode, new->video_bitrate, new 1113 drivers/media/common/cx2341x.c new->video_bitrate_peak / 400, 0, 0); new 1118 drivers/media/common/cx2341x.c CMP_FIELD(old, new, video_spatial_filter_mode) || new 1119 drivers/media/common/cx2341x.c CMP_FIELD(old, new, video_temporal_filter_mode) || new 1120 drivers/media/common/cx2341x.c CMP_FIELD(old, new, video_median_filter_type)) { new 1123 drivers/media/common/cx2341x.c new->video_spatial_filter_mode | new 1124 drivers/media/common/cx2341x.c (new->video_temporal_filter_mode << 1), new 1125 drivers/media/common/cx2341x.c new->video_median_filter_type); new 1130 drivers/media/common/cx2341x.c CMP_FIELD(old, new, video_luma_median_filter_bottom) || new 1131 drivers/media/common/cx2341x.c CMP_FIELD(old, new, video_luma_median_filter_top) || new 1132 drivers/media/common/cx2341x.c CMP_FIELD(old, new, video_chroma_median_filter_bottom) || new 1133 drivers/media/common/cx2341x.c CMP_FIELD(old, new, video_chroma_median_filter_top)) { new 1135 drivers/media/common/cx2341x.c new->video_luma_median_filter_bottom, new 1136 drivers/media/common/cx2341x.c new->video_luma_median_filter_top, new 1137 drivers/media/common/cx2341x.c new->video_chroma_median_filter_bottom, new 1138 drivers/media/common/cx2341x.c new->video_chroma_median_filter_top); new 1143 drivers/media/common/cx2341x.c CMP_FIELD(old, new, video_luma_spatial_filter_type) || new 1144 drivers/media/common/cx2341x.c CMP_FIELD(old, new, video_chroma_spatial_filter_type)) { new 1147 drivers/media/common/cx2341x.c 2, new->video_luma_spatial_filter_type, new 1148 drivers/media/common/cx2341x.c new->video_chroma_spatial_filter_type); new 1153 drivers/media/common/cx2341x.c CMP_FIELD(old, new, video_spatial_filter) || new 1154 drivers/media/common/cx2341x.c CMP_FIELD(old, new, video_temporal_filter)) { new 1156 drivers/media/common/cx2341x.c 2, new->video_spatial_filter, new 1157 drivers/media/common/cx2341x.c new->video_temporal_filter); new 1162 drivers/media/common/cx2341x.c CMP_FIELD(old, new, video_temporal_decimation)) { new 1164 drivers/media/common/cx2341x.c 1, new->video_temporal_decimation); new 1169 drivers/media/common/cx2341x.c CMP_FIELD(old, new, video_mute) || new 1170 drivers/media/common/cx2341x.c (new->video_mute && CMP_FIELD(old, new, video_mute_yuv))) { new 1172 drivers/media/common/cx2341x.c new->video_mute | (new->video_mute_yuv << 8)); new 1177 drivers/media/common/cx2341x.c CMP_FIELD(old, new, stream_insert_nav_packets)) { new 1179 drivers/media/common/cx2341x.c 7, new->stream_insert_nav_packets); new 504 drivers/media/common/siano/smscoreapi.c static void list_add_locked(struct list_head *new, struct list_head *head, new 511 drivers/media/common/siano/smscoreapi.c list_add(new, head); new 201 drivers/media/dvb-frontends/stv0910.c u8 shift, mask, old, new; new 208 drivers/media/dvb-frontends/stv0910.c new = ((val << shift) & mask) | (old & ~mask); new 209 drivers/media/dvb-frontends/stv0910.c if (new == old) new 211 drivers/media/dvb-frontends/stv0910.c return write_reg(state, field >> 16, new); new 557 drivers/media/i2c/et8ek8/et8ek8_driver.c struct et8ek8_gain new; new 560 drivers/media/i2c/et8ek8/et8ek8_driver.c new = et8ek8_gain_table[gain]; new 564 drivers/media/i2c/et8ek8/et8ek8_driver.c 0x124a, new.analog >> 8); new 568 drivers/media/i2c/et8ek8/et8ek8_driver.c 0x1249, new.analog & 0xff); new 573 drivers/media/i2c/et8ek8/et8ek8_driver.c 0x124d, new.digital >> 8); new 577 drivers/media/i2c/et8ek8/et8ek8_driver.c 0x124c, new.digital & 0xff); new 626 drivers/media/mc/mc-device.c struct media_graph new = { .top = 0 }; new 633 drivers/media/mc/mc-device.c ret = media_graph_walk_init(&new, mdev); new 639 drivers/media/mc/mc-device.c mdev->pm_count_walk = new; new 1499 drivers/media/pci/bt8xx/bttv-driver.c struct bttv_buffer *new) new 1504 drivers/media/pci/bt8xx/bttv-driver.c dprintk("switch_overlay: enter [new=%p]\n", new); new 1505 drivers/media/pci/bt8xx/bttv-driver.c if (new) new 1506 drivers/media/pci/bt8xx/bttv-driver.c new->vb.state = VIDEOBUF_DONE; new 1509 drivers/media/pci/bt8xx/bttv-driver.c btv->screen = new; new 1519 drivers/media/pci/bt8xx/bttv-driver.c if (NULL == new) new 2190 drivers/media/pci/bt8xx/bttv-driver.c struct bttv_buffer *new; new 2192 drivers/media/pci/bt8xx/bttv-driver.c new = videobuf_sg_alloc(sizeof(*new)); new 2193 drivers/media/pci/bt8xx/bttv-driver.c new->crop = btv->crop[!!fh->do_crop].rect; new 2194 drivers/media/pci/bt8xx/bttv-driver.c bttv_overlay_risc(btv, &fh->ov, fh->ovfmt, new); new 2195 drivers/media/pci/bt8xx/bttv-driver.c retval = bttv_switch_overlay(btv,fh,new); new 2539 drivers/media/pci/bt8xx/bttv-driver.c struct bttv_buffer *new; new 2560 drivers/media/pci/bt8xx/bttv-driver.c new = videobuf_sg_alloc(sizeof(*new)); new 2561 drivers/media/pci/bt8xx/bttv-driver.c new->crop = btv->crop[!!fh->do_crop].rect; new 2562 drivers/media/pci/bt8xx/bttv-driver.c bttv_overlay_risc(btv, &fh->ov, fh->ovfmt, new); new 2564 drivers/media/pci/bt8xx/bttv-driver.c new = NULL; new 2568 drivers/media/pci/bt8xx/bttv-driver.c retval = bttv_switch_overlay(btv, fh, new); new 2631 drivers/media/pci/bt8xx/bttv-driver.c struct bttv_buffer *new; new 2633 drivers/media/pci/bt8xx/bttv-driver.c new = videobuf_sg_alloc(sizeof(*new)); new 2634 drivers/media/pci/bt8xx/bttv-driver.c new->crop = btv->crop[!!fh->do_crop].rect; new 2635 drivers/media/pci/bt8xx/bttv-driver.c bttv_overlay_risc(btv, &fh->ov, fh->ovfmt, new); new 2636 drivers/media/pci/bt8xx/bttv-driver.c retval = bttv_switch_overlay(btv, fh, new); new 3600 drivers/media/pci/bt8xx/bttv-driver.c struct bttv_buffer_set old,new; new 3616 drivers/media/pci/bt8xx/bttv-driver.c memset(&new,0,sizeof(new)); new 3619 drivers/media/pci/bt8xx/bttv-driver.c btv->curr = new; new 3622 drivers/media/pci/bt8xx/bttv-driver.c bttv_buffer_activate_video(btv, &new); new 3627 drivers/media/pci/bt8xx/bttv-driver.c bttv_irq_wakeup_video(btv, &old, &new, VIDEOBUF_ERROR); new 3680 drivers/media/pci/bt8xx/bttv-driver.c struct bttv_buffer_set new; new 3687 drivers/media/pci/bt8xx/bttv-driver.c bttv_irq_next_video(btv, &new); new 3700 drivers/media/pci/bt8xx/bttv-driver.c btv->curr = new; new 3702 drivers/media/pci/bt8xx/bttv-driver.c bttv_buffer_activate_video(btv, &new); new 3712 drivers/media/pci/bt8xx/bttv-driver.c bttv_irq_wakeup_video(btv, &old, &new, VIDEOBUF_DONE); new 3719 drivers/media/pci/bt8xx/bttv-driver.c struct bttv_buffer *new = NULL; new 3726 drivers/media/pci/bt8xx/bttv-driver.c new = list_entry(btv->vcapture.next, struct bttv_buffer, vb.queue); new 3740 drivers/media/pci/bt8xx/bttv-driver.c btv->cvbi = new; new 3742 drivers/media/pci/bt8xx/bttv-driver.c bttv_buffer_activate_vbi(btv, new); new 1074 drivers/media/platform/sh_vou.c struct sh_vou_buffer *new = list_entry(vou_dev->active->list.next, new 1076 drivers/media/platform/sh_vou.c sh_vou_schedule_next(vou_dev, &new->vb); new 604 drivers/media/platform/vicodec/codec-fwht.c static int var_inter(const s16 *old, const s16 *new) new 609 drivers/media/platform/vicodec/codec-fwht.c for (i = 0; i < 8 * 8; i++, old++, new++) new 610 drivers/media/platform/vicodec/codec-fwht.c ret += (*old - *new) < 0 ? -(*old - *new) : (*old - *new); new 121 drivers/media/usb/au0828/au0828-core.c static void au0828_media_graph_notify(struct media_entity *new, new 211 drivers/media/usb/au0828/au0828-core.c static void au0828_media_graph_notify(struct media_entity *new, new 218 drivers/media/usb/au0828/au0828-core.c if (!new) { new 233 drivers/media/usb/au0828/au0828-core.c switch (new->function) { new 235 drivers/media/usb/au0828/au0828-core.c mixer = new; new 245 drivers/media/usb/au0828/au0828-core.c decoder = new; new 4360 drivers/media/v4l2-core/v4l2-ctrls.c void v4l2_ctrl_replace(struct v4l2_event *old, const struct v4l2_event *new) new 4364 drivers/media/v4l2-core/v4l2-ctrls.c old->u.ctrl = new->u.ctrl; new 4369 drivers/media/v4l2-core/v4l2-ctrls.c void v4l2_ctrl_merge(const struct v4l2_event *old, struct v4l2_event *new) new 4371 drivers/media/v4l2-core/v4l2-ctrls.c new->u.ctrl.changes |= old->u.ctrl.changes; new 257 drivers/media/v4l2-core/v4l2-dev.c enum v4l2_priority new) new 259 drivers/media/v4l2-core/v4l2-dev.c if (!prio_is_valid(new)) new 261 drivers/media/v4l2-core/v4l2-dev.c if (*local == new) new 264 drivers/media/v4l2-core/v4l2-dev.c atomic_inc(&global->prios[new]); new 267 drivers/media/v4l2-core/v4l2-dev.c *local = new; new 320 drivers/media/v4l2-core/v4l2-event.c const struct v4l2_event *new) new 324 drivers/media/v4l2-core/v4l2-event.c old->u.src_change = new->u.src_change; new 329 drivers/media/v4l2-core/v4l2-event.c struct v4l2_event *new) new 331 drivers/media/v4l2-core/v4l2-event.c new->u.src_change.changes |= old->u.src_change.changes; new 685 drivers/message/fusion/mptfc.c char *old, *new; new 701 drivers/message/fusion/mptfc.c new = new_speed == MPI_FCPORTPAGE0_CURRENT_SPEED_1GBIT ? "1 Gbps" : new 708 drivers/message/fusion/mptfc.c ioc->name, new); new 712 drivers/message/fusion/mptfc.c ioc->name, old, new); new 358 drivers/mfd/ab8500-core.c u8 new = ab8500->mask[i]; new 361 drivers/mfd/ab8500-core.c if (new == old) new 375 drivers/mfd/ab8500-core.c ab8500->oldmask[i] = new; new 378 drivers/mfd/ab8500-core.c set_register_interruptible(ab8500, AB8500_INTERRUPT, reg, new); new 1149 drivers/mfd/stmpe.c u8 new = stmpe->ier[i]; new 1152 drivers/mfd/stmpe.c if (new == old) new 1155 drivers/mfd/stmpe.c stmpe->oldier[i] = new; new 1156 drivers/mfd/stmpe.c stmpe_reg_write(stmpe, stmpe->regs[STMPE_IDX_IER_LSB + i], new); new 328 drivers/misc/echo/echo.c int new, old; new 333 drivers/misc/echo/echo.c new = (int)tx * (int)tx; new 337 drivers/misc/echo/echo.c ((new - old) + (1 << (ec->log2taps - 1))) >> ec->log2taps; new 271 drivers/misc/genwqe/card_ddcb.c __be32 old, new, icrc_hsi_shi; new 298 drivers/misc/genwqe/card_ddcb.c new = (old | DDCB_NEXT_BE32); new 301 drivers/misc/genwqe/card_ddcb.c icrc_hsi_shi = cmpxchg(&prev_ddcb->icrc_hsi_shi_32, old, new); new 624 drivers/misc/genwqe/card_ddcb.c __be32 old, new; new 647 drivers/misc/genwqe/card_ddcb.c new = (old | DDCB_PURGE_BE32); new 649 drivers/misc/genwqe/card_ddcb.c old, new); new 28 drivers/misc/ocxl/pasid.c struct id_range *cur, *new; new 31 drivers/misc/ocxl/pasid.c new = kmalloc(sizeof(struct id_range), GFP_KERNEL); new 32 drivers/misc/ocxl/pasid.c if (!new) new 44 drivers/misc/ocxl/pasid.c new->start = last_end + 1; new 45 drivers/misc/ocxl/pasid.c new->end = new->start + size - 1; new 47 drivers/misc/ocxl/pasid.c if (new->end > max_id) { new 48 drivers/misc/ocxl/pasid.c kfree(new); new 51 drivers/misc/ocxl/pasid.c list_add(&new->list, pos); new 52 drivers/misc/ocxl/pasid.c rc = new->start; new 93 drivers/mtd/devices/phram.c struct phram_mtd_list *new; new 96 drivers/mtd/devices/phram.c new = kzalloc(sizeof(*new), GFP_KERNEL); new 97 drivers/mtd/devices/phram.c if (!new) new 101 drivers/mtd/devices/phram.c new->mtd.priv = ioremap(start, len); new 102 drivers/mtd/devices/phram.c if (!new->mtd.priv) { new 108 drivers/mtd/devices/phram.c new->mtd.name = name; new 109 drivers/mtd/devices/phram.c new->mtd.size = len; new 110 drivers/mtd/devices/phram.c new->mtd.flags = MTD_CAP_RAM; new 111 drivers/mtd/devices/phram.c new->mtd._erase = phram_erase; new 112 drivers/mtd/devices/phram.c new->mtd._point = phram_point; new 113 drivers/mtd/devices/phram.c new->mtd._unpoint = phram_unpoint; new 114 drivers/mtd/devices/phram.c new->mtd._read = phram_read; new 115 drivers/mtd/devices/phram.c new->mtd._write = phram_write; new 116 drivers/mtd/devices/phram.c new->mtd.owner = THIS_MODULE; new 117 drivers/mtd/devices/phram.c new->mtd.type = MTD_RAM; new 118 drivers/mtd/devices/phram.c new->mtd.erasesize = PAGE_SIZE; new 119 drivers/mtd/devices/phram.c new->mtd.writesize = 1; new 122 drivers/mtd/devices/phram.c if (mtd_device_register(&new->mtd, NULL, 0)) { new 127 drivers/mtd/devices/phram.c list_add_tail(&new->list, &phram_list); new 131 drivers/mtd/devices/phram.c iounmap(new->mtd.priv); new 133 drivers/mtd/devices/phram.c kfree(new); new 340 drivers/mtd/mtd_blkdevs.c int add_mtd_blktrans_dev(struct mtd_blktrans_dev *new) new 342 drivers/mtd/mtd_blkdevs.c struct mtd_blktrans_ops *tr = new->tr; new 355 drivers/mtd/mtd_blkdevs.c if (new->devnum == -1) { new 359 drivers/mtd/mtd_blkdevs.c new->devnum = last_devnum+1; new 360 drivers/mtd/mtd_blkdevs.c list_add_tail(&new->list, &d->list); new 363 drivers/mtd/mtd_blkdevs.c } else if (d->devnum == new->devnum) { new 367 drivers/mtd/mtd_blkdevs.c } else if (d->devnum > new->devnum) { new 369 drivers/mtd/mtd_blkdevs.c list_add_tail(&new->list, &d->list); new 376 drivers/mtd/mtd_blkdevs.c if (new->devnum == -1) new 377 drivers/mtd/mtd_blkdevs.c new->devnum = last_devnum+1; new 382 drivers/mtd/mtd_blkdevs.c if (new->devnum > (MINORMASK >> tr->part_bits) || new 383 drivers/mtd/mtd_blkdevs.c (tr->part_bits && new->devnum >= 27 * 26)) { new 388 drivers/mtd/mtd_blkdevs.c list_add_tail(&new->list, &tr->devs); new 392 drivers/mtd/mtd_blkdevs.c mutex_init(&new->lock); new 393 drivers/mtd/mtd_blkdevs.c kref_init(&new->ref); new 395 drivers/mtd/mtd_blkdevs.c new->readonly = 1; new 404 drivers/mtd/mtd_blkdevs.c new->disk = gd; new 405 drivers/mtd/mtd_blkdevs.c gd->private_data = new; new 407 drivers/mtd/mtd_blkdevs.c gd->first_minor = (new->devnum) << tr->part_bits; new 411 drivers/mtd/mtd_blkdevs.c if (new->devnum < 26) new 413 drivers/mtd/mtd_blkdevs.c "%s%c", tr->name, 'a' + new->devnum); new 417 drivers/mtd/mtd_blkdevs.c 'a' - 1 + new->devnum / 26, new 418 drivers/mtd/mtd_blkdevs.c 'a' + new->devnum % 26); new 421 drivers/mtd/mtd_blkdevs.c "%s%d", tr->name, new->devnum); new 423 drivers/mtd/mtd_blkdevs.c set_capacity(gd, ((u64)new->size * tr->blksize) >> 9); new 426 drivers/mtd/mtd_blkdevs.c spin_lock_init(&new->queue_lock); new 427 drivers/mtd/mtd_blkdevs.c INIT_LIST_HEAD(&new->rq_list); new 429 drivers/mtd/mtd_blkdevs.c new->tag_set = kzalloc(sizeof(*new->tag_set), GFP_KERNEL); new 430 drivers/mtd/mtd_blkdevs.c if (!new->tag_set) new 433 drivers/mtd/mtd_blkdevs.c new->rq = blk_mq_init_sq_queue(new->tag_set, &mtd_mq_ops, 2, new 435 drivers/mtd/mtd_blkdevs.c if (IS_ERR(new->rq)) { new 436 drivers/mtd/mtd_blkdevs.c ret = PTR_ERR(new->rq); new 437 drivers/mtd/mtd_blkdevs.c new->rq = NULL; new 442 drivers/mtd/mtd_blkdevs.c blk_queue_write_cache(new->rq, true, false); new 444 drivers/mtd/mtd_blkdevs.c new->rq->queuedata = new; new 445 drivers/mtd/mtd_blkdevs.c blk_queue_logical_block_size(new->rq, tr->blksize); new 447 drivers/mtd/mtd_blkdevs.c blk_queue_flag_set(QUEUE_FLAG_NONROT, new->rq); new 448 drivers/mtd/mtd_blkdevs.c blk_queue_flag_clear(QUEUE_FLAG_ADD_RANDOM, new->rq); new 451 drivers/mtd/mtd_blkdevs.c blk_queue_flag_set(QUEUE_FLAG_DISCARD, new->rq); new 452 drivers/mtd/mtd_blkdevs.c blk_queue_max_discard_sectors(new->rq, UINT_MAX); new 455 drivers/mtd/mtd_blkdevs.c gd->queue = new->rq; new 457 drivers/mtd/mtd_blkdevs.c if (new->readonly) new 460 drivers/mtd/mtd_blkdevs.c device_add_disk(&new->mtd->dev, gd, NULL); new 462 drivers/mtd/mtd_blkdevs.c if (new->disk_attributes) { new 464 drivers/mtd/mtd_blkdevs.c new->disk_attributes); new 469 drivers/mtd/mtd_blkdevs.c kfree(new->tag_set); new 471 drivers/mtd/mtd_blkdevs.c put_disk(new->disk); new 473 drivers/mtd/mtd_blkdevs.c list_del(&new->list); new 895 drivers/mtd/mtdcore.c void register_mtd_user (struct mtd_notifier *new) new 901 drivers/mtd/mtdcore.c list_add(&new->list, &mtd_notifiers); new 906 drivers/mtd/mtdcore.c new->add(mtd); new 571 drivers/mtd/mtdpart.c static int mtd_add_partition_attrs(struct mtd_part *new) new 573 drivers/mtd/mtdpart.c int ret = sysfs_create_files(&new->mtd.dev.kobj, mtd_partition_attrs); new 584 drivers/mtd/mtdpart.c struct mtd_part *new; new 603 drivers/mtd/mtdpart.c new = allocate_partition(parent, &part, -1, offset); new 604 drivers/mtd/mtdpart.c if (IS_ERR(new)) new 605 drivers/mtd/mtdpart.c return PTR_ERR(new); new 608 drivers/mtd/mtdpart.c list_add(&new->list, &mtd_partitions); new 611 drivers/mtd/mtdpart.c ret = add_mtd_device(&new->mtd); new 615 drivers/mtd/mtdpart.c mtd_add_partition_attrs(new); new 621 drivers/mtd/mtdpart.c list_del(&new->list); new 624 drivers/mtd/mtdpart.c free_partition(new); new 3564 drivers/mtd/nand/onenand/onenand_base.c int ret, density, blksperdie, old, new, thisboundary; new 3587 drivers/mtd/nand/onenand/onenand_base.c new = boundary + (die * this->density_mask); new 3588 drivers/mtd/nand/onenand/onenand_base.c ret = flexonenand_check_blocks_erased(mtd, min(old, new) + 1, max(old, new)); new 147 drivers/mtd/nftlmount.c The new DiskOnChip driver scans the MediaHeader itself, and presents a virtual new 219 drivers/mtd/nftlmount.c The new DiskOnChip driver already scanned the bad block table. Just query it. new 106 drivers/mtd/ubi/fastmap.c struct ubi_vid_io_buf *new; new 109 drivers/mtd/ubi/fastmap.c new = ubi_alloc_vid_buf(ubi, GFP_KERNEL); new 110 drivers/mtd/ubi/fastmap.c if (!new) new 113 drivers/mtd/ubi/fastmap.c vh = ubi_get_vid_hdr(new); new 123 drivers/mtd/ubi/fastmap.c return new; new 824 drivers/mtd/ubi/fastmap.c struct ubi_ainf_peb *new; new 826 drivers/mtd/ubi/fastmap.c new = ubi_alloc_aeb(ai, old->pnum, old->ec); new 827 drivers/mtd/ubi/fastmap.c if (!new) new 830 drivers/mtd/ubi/fastmap.c new->vol_id = old->vol_id; new 831 drivers/mtd/ubi/fastmap.c new->sqnum = old->sqnum; new 832 drivers/mtd/ubi/fastmap.c new->lnum = old->lnum; new 833 drivers/mtd/ubi/fastmap.c new->scrub = old->scrub; new 834 drivers/mtd/ubi/fastmap.c new->copy_flag = old->copy_flag; new 836 drivers/mtd/ubi/fastmap.c return new; new 870 drivers/mtd/ubi/fastmap.c struct ubi_ainf_peb *new; new 872 drivers/mtd/ubi/fastmap.c new = clone_aeb(ai, aeb); new 873 drivers/mtd/ubi/fastmap.c if (!new) new 876 drivers/mtd/ubi/fastmap.c list_add(&new->u.list, &ai->fastmap); new 3417 drivers/net/bonding/bond_main.c const u64 *new = (const u64 *)_new; new 3423 drivers/net/bonding/bond_main.c u64 nv = new[i]; new 3499 drivers/net/bonding/bond_main.c const struct rtnl_link_stats64 *new = new 3502 drivers/net/bonding/bond_main.c bond_fold_stats(stats, new, &slave->slave_stats); new 3505 drivers/net/bonding/bond_main.c memcpy(&slave->slave_stats, new, sizeof(*new)); new 68 drivers/net/can/rx-offload.c static inline void __skb_queue_add_sort(struct sk_buff_head *head, struct sk_buff *new, new 77 drivers/net/can/rx-offload.c cb_new = can_rx_offload_get_cb(new); new 79 drivers/net/can/rx-offload.c netdev_dbg(new->dev, new 86 drivers/net/can/rx-offload.c if (compare(pos, new) < 0) new 92 drivers/net/can/rx-offload.c __skb_queue_head(head, new); new 94 drivers/net/can/rx-offload.c __skb_queue_after(head, insert, new); new 548 drivers/net/ethernet/amd/pcnet32.c int new, overlap; new 568 drivers/net/ethernet/amd/pcnet32.c for (new = 0; new < overlap; new++) { new 569 drivers/net/ethernet/amd/pcnet32.c new_rx_ring[new] = lp->rx_ring[new]; new 570 drivers/net/ethernet/amd/pcnet32.c new_dma_addr_list[new] = lp->rx_dma_addr[new]; new 571 drivers/net/ethernet/amd/pcnet32.c new_skb_list[new] = lp->rx_skbuff[new]; new 574 drivers/net/ethernet/amd/pcnet32.c for (; new < entries; new++) { new 576 drivers/net/ethernet/amd/pcnet32.c new_skb_list[new] = netdev_alloc_skb(dev, PKT_BUF_SKB); new 577 drivers/net/ethernet/amd/pcnet32.c rx_skbuff = new_skb_list[new]; new 586 drivers/net/ethernet/amd/pcnet32.c new_dma_addr_list[new] = new 590 drivers/net/ethernet/amd/pcnet32.c new_dma_addr_list[new])) { new 593 drivers/net/ethernet/amd/pcnet32.c dev_kfree_skb(new_skb_list[new]); new 596 drivers/net/ethernet/amd/pcnet32.c new_rx_ring[new].base = cpu_to_le32(new_dma_addr_list[new]); new 597 drivers/net/ethernet/amd/pcnet32.c new_rx_ring[new].buf_length = cpu_to_le16(NEG_BUF_SIZE); new 598 drivers/net/ethernet/amd/pcnet32.c new_rx_ring[new].status = cpu_to_le16(0x8000); new 601 drivers/net/ethernet/amd/pcnet32.c for (; new < lp->rx_ring_size; new++) { new 602 drivers/net/ethernet/amd/pcnet32.c if (lp->rx_skbuff[new]) { new 604 drivers/net/ethernet/amd/pcnet32.c lp->rx_dma_addr[new])) new 606 drivers/net/ethernet/amd/pcnet32.c lp->rx_dma_addr[new], new 609 drivers/net/ethernet/amd/pcnet32.c dev_kfree_skb(lp->rx_skbuff[new]); new 630 drivers/net/ethernet/amd/pcnet32.c while (--new >= lp->rx_ring_size) { new 631 drivers/net/ethernet/amd/pcnet32.c if (new_skb_list[new]) { new 633 drivers/net/ethernet/amd/pcnet32.c new_dma_addr_list[new])) new 635 drivers/net/ethernet/amd/pcnet32.c new_dma_addr_list[new], new 638 drivers/net/ethernet/amd/pcnet32.c dev_kfree_skb(new_skb_list[new]); new 64 drivers/net/ethernet/aurora/nb8800.c u32 new = (old & ~mask) | (val & mask); new 66 drivers/net/ethernet/aurora/nb8800.c if (new != old) new 67 drivers/net/ethernet/aurora/nb8800.c nb8800_writeb(priv, reg, new); new 74 drivers/net/ethernet/aurora/nb8800.c u32 new = (old & ~mask) | (val & mask); new 76 drivers/net/ethernet/aurora/nb8800.c if (new != old) new 77 drivers/net/ethernet/aurora/nb8800.c nb8800_writel(priv, reg, new); new 559 drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.c struct bmac1_stats *new = bnx2x_sp(bp, mac_stats.bmac1_stats); new 594 drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.c struct bmac2_stats *new = bnx2x_sp(bp, mac_stats.bmac2_stats); new 628 drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.c pstats->pfc_frames_tx_hi = new->tx_stat_gtpp_hi; new 629 drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.c pstats->pfc_frames_tx_lo = new->tx_stat_gtpp_lo; new 631 drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.c pstats->pfc_frames_rx_hi = new->rx_stat_grpp_hi; new 632 drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.c pstats->pfc_frames_rx_lo = new->rx_stat_grpp_lo; new 660 drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.c struct mstat_stats *new = bnx2x_sp(bp, mac_stats.mstat_stats); new 674 drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.c ADD_64(pstats->pfc_frames_tx_hi, new->stats_tx.tx_gtxpp_hi, new 675 drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.c pstats->pfc_frames_tx_lo, new->stats_tx.tx_gtxpp_lo); new 676 drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.c ADD_64(pstats->pfc_frames_rx_hi, new->stats_rx.rx_grxpp_hi, new 677 drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.c pstats->pfc_frames_rx_lo, new->stats_rx.rx_grxpp_lo); new 747 drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.c struct emac_stats *new = bnx2x_sp(bp, mac_stats.emac_stats); new 804 drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.c struct nig_stats *new = bnx2x_sp(bp, nig_stats); new 837 drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.c new->brb_discard - old->brb_discard); new 839 drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.c new->brb_truncate - old->brb_truncate); new 848 drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.c memcpy(old, new, sizeof(struct nig_stats)); new 390 drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.h DIFF_64(diff.hi, new->s##_hi, pstats->mac_stx[0].t##_hi, \ new 391 drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.h diff.lo, new->s##_lo, pstats->mac_stx[0].t##_lo); \ new 392 drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.h pstats->mac_stx[0].t##_hi = new->s##_hi; \ new 393 drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.h pstats->mac_stx[0].t##_lo = new->s##_lo; \ new 400 drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.h DIFF_64(diff.hi, new->s##_hi, old->s##_hi, \ new 401 drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.h diff.lo, new->s##_lo, old->s##_lo); \ new 415 drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.h ADD_64(pstats->mac_stx[1].t##_hi, new->diff##_hi, \ new 416 drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.h pstats->mac_stx[1].t##_lo, new->diff##_lo); \ new 423 drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.h new->s); \ new 67 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c static void cxgb_redirect(struct dst_entry *old, struct dst_entry *new, new 972 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c cxgb_redirect(nr->old, nr->new, nr->neigh, new 1110 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c static void cxgb_redirect(struct dst_entry *old, struct dst_entry *new, new 1130 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c e = t3_l2t_get(tdev, new, dev, daddr); new 1142 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c update_tcb = te->client->redirect(te->ctx, old, new, e); new 88 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.h struct dst_entry *new, struct l2t_entry *l2t); new 69 drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_flower.c struct ch_tc_flower_entry *new = kzalloc(sizeof(*new), GFP_KERNEL); new 70 drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_flower.c if (new) new 71 drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_flower.c spin_lock_init(&new->lock); new 72 drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_flower.c return new; new 735 drivers/net/ethernet/dec/tulip/winbond-840.c static inline void update_csr6(struct net_device *dev, int new) new 742 drivers/net/ethernet/dec/tulip/winbond-840.c new = 0; new 743 drivers/net/ethernet/dec/tulip/winbond-840.c if (new==np->csr6) new 768 drivers/net/ethernet/dec/tulip/winbond-840.c np->csr6 = new; new 771 drivers/net/ethernet/dec/tulip/winbond-840.c if (new & 0x200) new 1294 drivers/net/ethernet/dec/tulip/winbond-840.c int new; new 1300 drivers/net/ethernet/dec/tulip/winbond-840.c new = np->csr6 + 0x4000; new 1302 drivers/net/ethernet/dec/tulip/winbond-840.c new = (np->csr6 >> 14)&0x7f; new 1303 drivers/net/ethernet/dec/tulip/winbond-840.c if (new < 64) new 1304 drivers/net/ethernet/dec/tulip/winbond-840.c new *= 2; new 1306 drivers/net/ethernet/dec/tulip/winbond-840.c new = 127; /* load full packet before starting */ new 1307 drivers/net/ethernet/dec/tulip/winbond-840.c new = (np->csr6 & ~(0x7F << 14)) | (new<<14); new 1309 drivers/net/ethernet/dec/tulip/winbond-840.c netdev_dbg(dev, "Tx underflow, new csr6 %08x\n", new); new 1310 drivers/net/ethernet/dec/tulip/winbond-840.c update_csr6(dev, new); new 545 drivers/net/ethernet/freescale/enetc/enetc.c struct enetc_rx_swbd *new; new 547 drivers/net/ethernet/freescale/enetc/enetc.c new = &rx_ring->rx_swbd[rx_ring->next_to_alloc]; new 553 drivers/net/ethernet/freescale/enetc/enetc.c *new = *old; new 63 drivers/net/ethernet/intel/i40e/i40e_ddp.c static bool i40e_ddp_profiles_overlap(struct i40e_profile_info *new, new 67 drivers/net/ethernet/intel/i40e/i40e_ddp.c unsigned int group_id_new = (u8)((new->track_id & 0x00FF0000) >> 16); new 3966 drivers/net/ethernet/intel/i40e/i40e_ethtool.c static void i40e_print_input_set(struct i40e_vsi *vsi, u64 old, u64 new) new 3973 drivers/net/ethernet/intel/i40e/i40e_ethtool.c new_value = !!(new & I40E_L3_SRC_MASK); new 3980 drivers/net/ethernet/intel/i40e/i40e_ethtool.c new_value = !!(new & I40E_L3_DST_MASK); new 3987 drivers/net/ethernet/intel/i40e/i40e_ethtool.c new_value = !!(new & I40E_L4_SRC_MASK); new 3994 drivers/net/ethernet/intel/i40e/i40e_ethtool.c new_value = !!(new & I40E_L4_DST_MASK); new 4001 drivers/net/ethernet/intel/i40e/i40e_ethtool.c new_value = !!(new & I40E_VERIFY_TAG_MASK); new 4012 drivers/net/ethernet/intel/i40e/i40e_ethtool.c new_value = !!(new & flex_mask); new 4023 drivers/net/ethernet/intel/i40e/i40e_ethtool.c new); new 1233 drivers/net/ethernet/intel/i40e/i40e_main.c struct i40e_new_mac_filter *new; new 1252 drivers/net/ethernet/intel/i40e/i40e_main.c hlist_for_each_entry(new, tmp_add_list, hlist) { new 1253 drivers/net/ethernet/intel/i40e/i40e_main.c if (pvid && new->f->vlan != pvid) new 1254 drivers/net/ethernet/intel/i40e/i40e_main.c new->f->vlan = pvid; new 1255 drivers/net/ethernet/intel/i40e/i40e_main.c else if (vlan_filters && new->f->vlan == I40E_VLAN_ANY) new 1256 drivers/net/ethernet/intel/i40e/i40e_main.c new->f->vlan = 0; new 1257 drivers/net/ethernet/intel/i40e/i40e_main.c else if (!vlan_filters && new->f->vlan == 0) new 1258 drivers/net/ethernet/intel/i40e/i40e_main.c new->f->vlan = I40E_VLAN_ANY; new 1285 drivers/net/ethernet/intel/i40e/i40e_main.c new = kzalloc(sizeof(*new), GFP_ATOMIC); new 1286 drivers/net/ethernet/intel/i40e/i40e_main.c if (!new) new 1289 drivers/net/ethernet/intel/i40e/i40e_main.c new->f = add_head; new 1290 drivers/net/ethernet/intel/i40e/i40e_main.c new->state = add_head->state; new 1293 drivers/net/ethernet/intel/i40e/i40e_main.c hlist_add_head(&new->hlist, tmp_add_list); new 1991 drivers/net/ethernet/intel/i40e/i40e_main.c struct i40e_new_mac_filter *new; new 1994 drivers/net/ethernet/intel/i40e/i40e_main.c hlist_for_each_entry_safe(new, h, from, hlist) { new 1996 drivers/net/ethernet/intel/i40e/i40e_main.c hlist_del(&new->hlist); new 1997 drivers/net/ethernet/intel/i40e/i40e_main.c kfree(new); new 2263 drivers/net/ethernet/intel/i40e/i40e_main.c struct i40e_new_mac_filter *new, *add_head = NULL; new 2320 drivers/net/ethernet/intel/i40e/i40e_main.c new = kzalloc(sizeof(*new), GFP_ATOMIC); new 2321 drivers/net/ethernet/intel/i40e/i40e_main.c if (!new) new 2325 drivers/net/ethernet/intel/i40e/i40e_main.c new->f = f; new 2326 drivers/net/ethernet/intel/i40e/i40e_main.c new->state = f->state; new 2329 drivers/net/ethernet/intel/i40e/i40e_main.c hlist_add_head(&new->hlist, &tmp_add_list); new 2422 drivers/net/ethernet/intel/i40e/i40e_main.c hlist_for_each_entry_safe(new, h, &tmp_add_list, hlist) { new 2426 drivers/net/ethernet/intel/i40e/i40e_main.c if (is_broadcast_ether_addr(new->f->macaddr)) { new 2428 drivers/net/ethernet/intel/i40e/i40e_main.c new->f)) new 2429 drivers/net/ethernet/intel/i40e/i40e_main.c new->state = I40E_FILTER_FAILED; new 2431 drivers/net/ethernet/intel/i40e/i40e_main.c new->state = I40E_FILTER_ACTIVE; new 2437 drivers/net/ethernet/intel/i40e/i40e_main.c add_head = new; new 2440 drivers/net/ethernet/intel/i40e/i40e_main.c new->f->macaddr); new 2441 drivers/net/ethernet/intel/i40e/i40e_main.c if (new->f->vlan == I40E_VLAN_ANY) { new 2446 drivers/net/ethernet/intel/i40e/i40e_main.c cpu_to_le16((u16)(new->f->vlan)); new 2471 drivers/net/ethernet/intel/i40e/i40e_main.c hlist_for_each_entry_safe(new, h, &tmp_add_list, hlist) { new 2473 drivers/net/ethernet/intel/i40e/i40e_main.c if (new->f->state == I40E_FILTER_NEW) new 2474 drivers/net/ethernet/intel/i40e/i40e_main.c new->f->state = new->state; new 2475 drivers/net/ethernet/intel/i40e/i40e_main.c hlist_del(&new->hlist); new 2476 drivers/net/ethernet/intel/i40e/i40e_main.c kfree(new); new 255 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c struct rb_node **new = &(root->rb_node), *parent = NULL; new 258 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c while (*new) { new 259 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c struct res_common *this = rb_entry(*new, struct res_common, new 262 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c parent = *new; new 264 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c new = &((*new)->rb_left); new 266 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c new = &((*new)->rb_right); new 272 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c rb_link_node(&res->node, parent, new); new 894 drivers/net/ethernet/mellanox/mlx5/core/fpga/ipsec.c struct rb_node **new = &root->rb_node, *parent = NULL; new 897 drivers/net/ethernet/mellanox/mlx5/core/fpga/ipsec.c while (*new) { new 899 drivers/net/ethernet/mellanox/mlx5/core/fpga/ipsec.c container_of(*new, struct mlx5_fpga_ipsec_rule, new 902 drivers/net/ethernet/mellanox/mlx5/core/fpga/ipsec.c parent = *new; new 904 drivers/net/ethernet/mellanox/mlx5/core/fpga/ipsec.c new = &((*new)->rb_left); new 906 drivers/net/ethernet/mellanox/mlx5/core/fpga/ipsec.c new = &((*new)->rb_right); new 912 drivers/net/ethernet/mellanox/mlx5/core/fpga/ipsec.c rb_link_node(&rule->node, parent, new); new 79 drivers/net/ethernet/mellanox/mlx5/core/pagealloc.c struct rb_node **new = &root->rb_node; new 85 drivers/net/ethernet/mellanox/mlx5/core/pagealloc.c while (*new) { new 86 drivers/net/ethernet/mellanox/mlx5/core/pagealloc.c parent = *new; new 89 drivers/net/ethernet/mellanox/mlx5/core/pagealloc.c new = &parent->rb_left; new 91 drivers/net/ethernet/mellanox/mlx5/core/pagealloc.c new = &parent->rb_right; new 107 drivers/net/ethernet/mellanox/mlx5/core/pagealloc.c rb_link_node(&nfp->rb_node, parent, new); new 1391 drivers/net/ethernet/mscc/ocelot.c bool new = false; new 1406 drivers/net/ethernet/mscc/ocelot.c new = true; new 1412 drivers/net/ethernet/mscc/ocelot.c if (!new) { new 459 drivers/net/ethernet/netronome/nfp/abm/qdisc.c nfp_abm_stats_calculate(struct nfp_alink_stats *new, new 464 drivers/net/ethernet/netronome/nfp/abm/qdisc.c _bstats_update(bstats, new->tx_bytes - old->tx_bytes, new 465 drivers/net/ethernet/netronome/nfp/abm/qdisc.c new->tx_pkts - old->tx_pkts); new 466 drivers/net/ethernet/netronome/nfp/abm/qdisc.c qstats->qlen += new->backlog_pkts - old->backlog_pkts; new 467 drivers/net/ethernet/netronome/nfp/abm/qdisc.c qstats->backlog += new->backlog_bytes - old->backlog_bytes; new 468 drivers/net/ethernet/netronome/nfp/abm/qdisc.c qstats->overlimits += new->overlimits - old->overlimits; new 469 drivers/net/ethernet/netronome/nfp/abm/qdisc.c qstats->drops += new->drops - old->drops; new 473 drivers/net/ethernet/netronome/nfp/abm/qdisc.c nfp_abm_stats_red_calculate(struct nfp_alink_xstats *new, new 477 drivers/net/ethernet/netronome/nfp/abm/qdisc.c stats->forced_mark += new->ecn_marked - old->ecn_marked; new 478 drivers/net/ethernet/netronome/nfp/abm/qdisc.c stats->pdrop += new->pdrop - old->pdrop; new 959 drivers/net/ethernet/netronome/nfp/nfp_net.h int nfp_net_ring_reconfig(struct nfp_net *nn, struct nfp_net_dp *new, new 3186 drivers/net/ethernet/netronome/nfp/nfp_net_common.c struct nfp_net_dp *new; new 3188 drivers/net/ethernet/netronome/nfp/nfp_net_common.c new = kmalloc(sizeof(*new), GFP_KERNEL); new 3189 drivers/net/ethernet/netronome/nfp/nfp_net_common.c if (!new) new 3192 drivers/net/ethernet/netronome/nfp/nfp_net_common.c *new = nn->dp; new 3195 drivers/net/ethernet/netronome/nfp/nfp_net_common.c new->fl_bufsz = 0; new 3196 drivers/net/ethernet/netronome/nfp/nfp_net_common.c new->tx_rings = NULL; new 3197 drivers/net/ethernet/netronome/nfp/nfp_net_common.c new->rx_rings = NULL; new 3198 drivers/net/ethernet/netronome/nfp/nfp_net_common.c new->num_r_vecs = 0; new 3199 drivers/net/ethernet/netronome/nfp/nfp_net_common.c new->num_stack_tx_rings = 0; new 3201 drivers/net/ethernet/netronome/nfp/nfp_net_common.c return new; new 348 drivers/net/ethernet/pensando/ionic/ionic_lif.c struct ionic_qcq *new; new 368 drivers/net/ethernet/pensando/ionic/ionic_lif.c new = devm_kzalloc(dev, sizeof(*new), GFP_KERNEL); new 369 drivers/net/ethernet/pensando/ionic/ionic_lif.c if (!new) { new 375 drivers/net/ethernet/pensando/ionic/ionic_lif.c new->flags = flags; new 377 drivers/net/ethernet/pensando/ionic/ionic_lif.c new->q.info = devm_kzalloc(dev, sizeof(*new->q.info) * num_descs, new 379 drivers/net/ethernet/pensando/ionic/ionic_lif.c if (!new->q.info) { new 385 drivers/net/ethernet/pensando/ionic/ionic_lif.c new->q.type = type; new 387 drivers/net/ethernet/pensando/ionic/ionic_lif.c err = ionic_q_init(lif, idev, &new->q, index, name, num_descs, new 395 drivers/net/ethernet/pensando/ionic/ionic_lif.c err = ionic_intr_alloc(lif, &new->intr); new 402 drivers/net/ethernet/pensando/ionic/ionic_lif.c err = ionic_bus_get_irq(lif->ionic, new->intr.index); new 408 drivers/net/ethernet/pensando/ionic/ionic_lif.c new->intr.vector = err; new 409 drivers/net/ethernet/pensando/ionic/ionic_lif.c ionic_intr_mask_assert(idev->intr_ctrl, new->intr.index, new 412 drivers/net/ethernet/pensando/ionic/ionic_lif.c new->intr.cpu = new->intr.index % num_online_cpus(); new 413 drivers/net/ethernet/pensando/ionic/ionic_lif.c if (cpu_online(new->intr.cpu)) new 414 drivers/net/ethernet/pensando/ionic/ionic_lif.c cpumask_set_cpu(new->intr.cpu, new 415 drivers/net/ethernet/pensando/ionic/ionic_lif.c &new->intr.affinity_mask); new 417 drivers/net/ethernet/pensando/ionic/ionic_lif.c new->intr.index = INTR_INDEX_NOT_ASSIGNED; new 420 drivers/net/ethernet/pensando/ionic/ionic_lif.c new->cq.info = devm_kzalloc(dev, sizeof(*new->cq.info) * num_descs, new 422 drivers/net/ethernet/pensando/ionic/ionic_lif.c if (!new->cq.info) { new 428 drivers/net/ethernet/pensando/ionic/ionic_lif.c err = ionic_cq_init(lif, &new->cq, &new->intr, num_descs, cq_desc_size); new 434 drivers/net/ethernet/pensando/ionic/ionic_lif.c new->base = dma_alloc_coherent(dev, total_size, &new->base_pa, new 436 drivers/net/ethernet/pensando/ionic/ionic_lif.c if (!new->base) { new 442 drivers/net/ethernet/pensando/ionic/ionic_lif.c new->total_size = total_size; new 444 drivers/net/ethernet/pensando/ionic/ionic_lif.c q_base = new->base; new 445 drivers/net/ethernet/pensando/ionic/ionic_lif.c q_base_pa = new->base_pa; new 454 drivers/net/ethernet/pensando/ionic/ionic_lif.c ionic_q_sg_map(&new->q, sg_base, sg_base_pa); new 457 drivers/net/ethernet/pensando/ionic/ionic_lif.c ionic_q_map(&new->q, q_base, q_base_pa); new 458 drivers/net/ethernet/pensando/ionic/ionic_lif.c ionic_cq_map(&new->cq, cq_base, cq_base_pa); new 459 drivers/net/ethernet/pensando/ionic/ionic_lif.c ionic_cq_bind(&new->cq, &new->q); new 461 drivers/net/ethernet/pensando/ionic/ionic_lif.c *qcq = new; new 466 drivers/net/ethernet/pensando/ionic/ionic_lif.c ionic_intr_free(lif, new->intr.index); new 41 drivers/net/ethernet/pensando/ionic/ionic_txrx.c struct ionic_rxq_desc *new = q->head->desc; new 43 drivers/net/ethernet/pensando/ionic/ionic_txrx.c new->addr = old->addr; new 44 drivers/net/ethernet/pensando/ionic/ionic_txrx.c new->len = old->len; new 2750 drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c unsigned long new; new 2759 drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c if (kstrtoul(buf, 2, &new)) new 2762 drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c if (!netxen_config_bridged_mode(adapter, !!new)) new 2796 drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c unsigned long new; new 2798 drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c if (kstrtoul(buf, 2, &new)) new 2801 drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c if (!!new != !!(adapter->flags & NETXEN_NIC_DIAG_ENABLED)) new 41 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sysfs.c unsigned long new; new 50 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sysfs.c if (kstrtoul(buf, 2, &new)) new 53 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sysfs.c if (!qlcnic_config_bridged_mode(adapter, !!new)) new 78 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sysfs.c unsigned long new; new 80 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sysfs.c if (kstrtoul(buf, 2, &new)) new 83 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sysfs.c if (!!new != !!(adapter->flags & QLCNIC_DIAG_ENABLED)) new 3194 drivers/net/ethernet/sfc/efx.c bool *new) new 3206 drivers/net/ethernet/sfc/efx.c *new = false; new 3211 drivers/net/ethernet/sfc/efx.c *new = true; new 3255 drivers/net/ethernet/sfc/efx.c struct efx_rss_context *ctx, *new; new 3273 drivers/net/ethernet/sfc/efx.c new = kmalloc(sizeof(struct efx_rss_context), GFP_KERNEL); new 3274 drivers/net/ethernet/sfc/efx.c if (!new) new 3276 drivers/net/ethernet/sfc/efx.c new->context_id = EFX_EF10_RSS_CONTEXT_INVALID; new 3277 drivers/net/ethernet/sfc/efx.c new->rx_hash_udp_4tuple = false; new 3280 drivers/net/ethernet/sfc/efx.c new->user_id = id; new 3281 drivers/net/ethernet/sfc/efx.c list_add_tail(&new->list, &ctx->list); new 3282 drivers/net/ethernet/sfc/efx.c return new; new 202 drivers/net/ethernet/sfc/efx.h bool *new); new 898 drivers/net/ethernet/sfc/rx.c bool new; new 954 drivers/net/ethernet/sfc/rx.c rule = efx_rps_hash_add(efx, &req->spec, &new); new 959 drivers/net/ethernet/sfc/rx.c if (new) new 963 drivers/net/ethernet/sfc/rx.c if (!new && rule->rxq_index == rxq_index && new 1362 drivers/net/ethernet/sun/cassini.c cas_page_t *new; new 1367 drivers/net/ethernet/sun/cassini.c new = cas_page_dequeue(cp); new 1368 drivers/net/ethernet/sun/cassini.c if (new) { new 1373 drivers/net/ethernet/sun/cassini.c return new; new 1385 drivers/net/ethernet/sun/cassini.c cas_page_t *new = cas_page_spare(cp, index); new 1386 drivers/net/ethernet/sun/cassini.c if (new) { new 1388 drivers/net/ethernet/sun/cassini.c page0[index] = new; new 2180 drivers/net/ethernet/sun/cassini.c cas_page_t *new; new 2185 drivers/net/ethernet/sun/cassini.c new = cas_page_swap(cp, ring, index); new 2186 drivers/net/ethernet/sun/cassini.c cp->init_rxds[ring][entry].buffer = cpu_to_le64(new->dma_addr); new 2224 drivers/net/ethernet/sun/cassini.c cas_page_t *new = cas_page_dequeue(cp); new 2225 drivers/net/ethernet/sun/cassini.c if (!new) { new 2241 drivers/net/ethernet/sun/cassini.c cpu_to_le64(new->dma_addr); new 2242 drivers/net/ethernet/sun/cassini.c page[entry] = new; new 2139 drivers/net/macsec.c bool new = !!nla_get_u8(tb_rxsc[MACSEC_RXSC_ATTR_ACTIVE]); new 2141 drivers/net/macsec.c if (rx_sc->active != new) new 2142 drivers/net/macsec.c secy->n_rx_sc += new ? 1 : -1; new 2144 drivers/net/macsec.c rx_sc->active = new; new 160 drivers/net/net_failover.c const u64 *new = (const u64 *)_new; new 166 drivers/net/net_failover.c u64 nv = new[i]; new 186 drivers/net/net_failover.c const struct rtnl_link_stats64 *new; new 197 drivers/net/net_failover.c new = dev_get_stats(slave_dev, &temp); new 198 drivers/net/net_failover.c net_failover_fold_stats(stats, new, &nfo_info->primary_stats); new 199 drivers/net/net_failover.c memcpy(&nfo_info->primary_stats, new, sizeof(*new)); new 204 drivers/net/net_failover.c new = dev_get_stats(slave_dev, &temp); new 205 drivers/net/net_failover.c net_failover_fold_stats(stats, new, &nfo_info->standby_stats); new 206 drivers/net/net_failover.c memcpy(&nfo_info->standby_stats, new, sizeof(*new)); new 92 drivers/net/phy/microchip_t1.c int new = 0, rc = 0; new 101 drivers/net/phy/microchip_t1.c new = val | (rc & (mask ^ 0xFFFF)); new 102 drivers/net/phy/microchip_t1.c rc = access_ereg(phydev, PHYACC_ATTR_MODE_WRITE, bank, offset, new); new 498 drivers/net/phy/phy-core.c int new, ret; new 504 drivers/net/phy/phy-core.c new = (ret & ~mask) | set; new 505 drivers/net/phy/phy-core.c if (new == ret) new 508 drivers/net/phy/phy-core.c ret = __phy_write(phydev, regnum, new); new 599 drivers/net/phy/phy-core.c int new, ret; new 605 drivers/net/phy/phy-core.c new = (ret & ~mask) | set; new 606 drivers/net/phy/phy-core.c if (new == ret) new 609 drivers/net/phy/phy-core.c ret = __phy_write_mmd(phydev, devad, regnum, new); new 2049 drivers/net/phy/phy_device.c __ETHTOOL_DECLARE_LINK_MODE_MASK(new); new 2051 drivers/net/phy/phy_device.c linkmode_copy(new, phydev->supported); new 2052 drivers/net/phy/phy_device.c phy_copy_pause_bits(new, phydev->advertising); new 2053 drivers/net/phy/phy_device.c linkmode_copy(phydev->advertising, new); new 293 drivers/net/phy/sfp-bus.c struct sfp_bus *sfp, *new, *found = NULL; new 295 drivers/net/phy/sfp-bus.c new = kzalloc(sizeof(*new), GFP_KERNEL); new 307 drivers/net/phy/sfp-bus.c if (!found && new) { new 308 drivers/net/phy/sfp-bus.c kref_init(&new->kref); new 309 drivers/net/phy/sfp-bus.c new->fwnode = fwnode; new 310 drivers/net/phy/sfp-bus.c list_add(&new->node, &sfp_buses); new 311 drivers/net/phy/sfp-bus.c found = new; new 312 drivers/net/phy/sfp-bus.c new = NULL; new 317 drivers/net/phy/sfp-bus.c kfree(new); new 2280 drivers/net/tun.c struct tun_prog *old, *new = NULL; new 2283 drivers/net/tun.c new = kmalloc(sizeof(*new), GFP_KERNEL); new 2284 drivers/net/tun.c if (!new) new 2286 drivers/net/tun.c new->prog = prog; new 2292 drivers/net/tun.c rcu_assign_pointer(*prog_p, new); new 124 drivers/net/wan/hdlc_fr.c unsigned int new: 1; new 453 drivers/net/wan/hdlc_fr.c pvc->state.new ? " new" : "", new 524 drivers/net/wan/hdlc_fr.c pvc->state.exist = pvc->state.new = 1; new 530 drivers/net/wan/hdlc_fr.c pvc->state.exist && !pvc->state.new) { new 545 drivers/net/wan/hdlc_fr.c if (pvc->state.new) new 580 drivers/net/wan/hdlc_fr.c pvc->state.new = 0; new 589 drivers/net/wan/hdlc_fr.c pvc->state.new = 0; new 750 drivers/net/wan/hdlc_fr.c if (pvc->state.new) { new 751 drivers/net/wan/hdlc_fr.c pvc->state.new = 0; new 792 drivers/net/wan/hdlc_fr.c unsigned int active, new; new 808 drivers/net/wan/hdlc_fr.c new = !! (skb->data[i + 2] & 0x08); new 832 drivers/net/wan/hdlc_fr.c new != pvc->state.new || new 835 drivers/net/wan/hdlc_fr.c pvc->state.new = new; new 851 drivers/net/wan/hdlc_fr.c pvc->state.active = pvc->state.new = 0; new 80 drivers/net/wireless/ath/ath10k/wow.c static void ath10k_wow_convert_8023_to_80211(struct cfg80211_pkt_pattern *new, new 128 drivers/net/wireless/ath/ath10k/wow.c new->pkt_offset = old->pkt_offset + new 131 drivers/net/wireless/ath/ath10k/wow.c new->pkt_offset = old->pkt_offset + new 135 drivers/net/wireless/ath/ath10k/wow.c new->pkt_offset = old->pkt_offset + hdr_len + rfc_len - ETH_HLEN; new 149 drivers/net/wireless/ath/ath10k/wow.c new->pattern_len = hdr_80211_end_offset - new->pkt_offset; new 151 drivers/net/wireless/ath/ath10k/wow.c memcpy((u8 *)new->pattern, new 152 drivers/net/wireless/ath/ath10k/wow.c hdr_80211_pattern + new->pkt_offset, new 153 drivers/net/wireless/ath/ath10k/wow.c new->pattern_len); new 154 drivers/net/wireless/ath/ath10k/wow.c memcpy((u8 *)new->mask, new 155 drivers/net/wireless/ath/ath10k/wow.c hdr_80211_bit_mask + new->pkt_offset, new 156 drivers/net/wireless/ath/ath10k/wow.c new->pattern_len); new 160 drivers/net/wireless/ath/ath10k/wow.c memcpy((u8 *)new->pattern + new->pattern_len, new 163 drivers/net/wireless/ath/ath10k/wow.c memcpy((u8 *)new->mask + new->pattern_len, new 167 drivers/net/wireless/ath/ath10k/wow.c new->pattern_len += total_len - ETH_HLEN; new 420 drivers/net/wireless/ath/ath9k/ar5008_phy.c int tmp, new; new 443 drivers/net/wireless/ath/ath9k/ar5008_phy.c new = tmp | (AR_PHY_TIMING_CTRL4_ENABLE_SPUR_RSSI | new 448 drivers/net/wireless/ath/ath9k/ar5008_phy.c REG_WRITE(ah, AR_PHY_TIMING_CTRL4(0), new); new 450 drivers/net/wireless/ath/ath9k/ar5008_phy.c new = (AR_PHY_SPUR_REG_MASK_RATE_CNTL | new 455 drivers/net/wireless/ath/ath9k/ar5008_phy.c REG_WRITE(ah, AR_PHY_SPUR_REG, new); new 463 drivers/net/wireless/ath/ath9k/ar5008_phy.c new = (AR_PHY_TIMING11_USE_SPUR_IN_AGC | new 466 drivers/net/wireless/ath/ath9k/ar5008_phy.c REG_WRITE(ah, AR_PHY_TIMING11, new); new 51 drivers/net/wireless/ath/ath9k/dynack.c static inline int ath_dynack_ewma(int old, int new) new 54 drivers/net/wireless/ath/ath9k/dynack.c return (new * (EWMA_DIV - EWMA_LEVEL) + new 57 drivers/net/wireless/ath/ath9k/dynack.c return new; new 2199 drivers/net/wireless/ath/ath9k/main.c static u32 fill_chainmask(u32 cap, u32 new) new 2204 drivers/net/wireless/ath/ath9k/main.c for (i = 0; cap && new; i++, cap >>= 1) { new 2208 drivers/net/wireless/ath/ath9k/main.c if (new & BIT(0)) new 2211 drivers/net/wireless/ath/ath9k/main.c new >>= 1; new 2817 drivers/net/wireless/atmel/atmel.c int new = capability & WLAN_CAPABILITY_SHORT_PREAMBLE ? new 2820 drivers/net/wireless/atmel/atmel.c if (priv->preamble != new) { new 2821 drivers/net/wireless/atmel/atmel.c priv->preamble = new; new 2823 drivers/net/wireless/atmel/atmel.c atmel_set_mib8(priv, Local_Mib_Type, LOCAL_MIB_PREAMBLE_TYPE, new); new 4492 drivers/net/wireless/broadcom/b43/phy_n.c struct b43_phy_n_iq_comp new = { }; new 4499 drivers/net/wireless/broadcom/b43/phy_n.c b43_nphy_rx_iq_coeffs(dev, true, &new); new 4501 drivers/net/wireless/broadcom/b43/phy_n.c new = old; new 4554 drivers/net/wireless/broadcom/b43/phy_n.c new.a0 = a & 0x3FF; new 4555 drivers/net/wireless/broadcom/b43/phy_n.c new.b0 = b & 0x3FF; new 4557 drivers/net/wireless/broadcom/b43/phy_n.c new.a0 = b & 0x3FF; new 4558 drivers/net/wireless/broadcom/b43/phy_n.c new.b0 = a & 0x3FF; new 4562 drivers/net/wireless/broadcom/b43/phy_n.c new.a1 = a & 0x3FF; new 4563 drivers/net/wireless/broadcom/b43/phy_n.c new.b1 = b & 0x3FF; new 4565 drivers/net/wireless/broadcom/b43/phy_n.c new.a1 = b & 0x3FF; new 4566 drivers/net/wireless/broadcom/b43/phy_n.c new.b1 = a & 0x3FF; new 4572 drivers/net/wireless/broadcom/b43/phy_n.c new = old; new 4574 drivers/net/wireless/broadcom/b43/phy_n.c b43_nphy_rx_iq_coeffs(dev, true, &new); new 5362 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c struct brcms_c_rateset rs, new; new 5373 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c memcpy(&new, &rs, sizeof(struct brcms_c_rateset)); new 5375 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c (&new, &wlc->bandstate[bandunit]->hw_rateset, true, new 5382 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c memcpy(&new, &rs, sizeof(struct brcms_c_rateset)); new 5383 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c if (brcms_c_rate_hwrs_filter_sort_validate(&new, new 5395 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c memcpy(&wlc->default_bss->rateset, &new, new 5397 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c memcpy(&wlc->bandstate[bandunit]->defrateset, &new, new 817 drivers/net/wireless/realtek/rtw88/pci.c struct sk_buff *skb, *new; new 855 drivers/net/wireless/realtek/rtw88/pci.c new = dev_alloc_skb(new_len); new 856 drivers/net/wireless/realtek/rtw88/pci.c if (WARN_ONCE(!new, "rx routine starvation\n")) new 860 drivers/net/wireless/realtek/rtw88/pci.c skb_put_data(new, skb->data, new_len); new 863 drivers/net/wireless/realtek/rtw88/pci.c rtw_fw_c2h_cmd_rx_irqsafe(rtwdev, pkt_offset, new); new 866 drivers/net/wireless/realtek/rtw88/pci.c skb_pull(new, pkt_offset); new 868 drivers/net/wireless/realtek/rtw88/pci.c rtw_rx_stats(rtwdev, pkt_stat.vif, new); new 869 drivers/net/wireless/realtek/rtw88/pci.c memcpy(new->cb, &rx_status, sizeof(rx_status)); new 870 drivers/net/wireless/realtek/rtw88/pci.c ieee80211_rx_irqsafe(rtwdev->hw, new); new 731 drivers/net/wireless/wl3501_cs.c u8 new = old & (~(WL3501_GCR_ECINT | WL3501_GCR_INT2EC | new 734 drivers/net/wireless/wl3501_cs.c wl3501_outb(new, this->base_addr + WL3501_NIC_GCR); new 748 drivers/net/wireless/wl3501_cs.c u8 new = (old & ~(WL3501_GCR_ECINT | WL3501_GCR_INT2EC)) | new 751 drivers/net/wireless/wl3501_cs.c wl3501_outb(new, this->base_addr + WL3501_NIC_GCR); new 38 drivers/net/xen-netback/hash.c struct xenvif_hash_cache_entry *new, *entry, *oldest; new 42 drivers/net/xen-netback/hash.c new = kmalloc(sizeof(*entry), GFP_ATOMIC); new 43 drivers/net/xen-netback/hash.c if (!new) new 46 drivers/net/xen-netback/hash.c memcpy(new->tag, tag, len); new 47 drivers/net/xen-netback/hash.c new->len = len; new 48 drivers/net/xen-netback/hash.c new->val = val; new 64 drivers/net/xen-netback/hash.c new->seq = atomic_inc_return(&vif->hash.cache.seq); new 65 drivers/net/xen-netback/hash.c list_add_rcu(&new->link, &vif->hash.cache.list); new 77 drivers/net/xen-netback/hash.c kfree(new); new 727 drivers/net/xen-netfront.c int new = xennet_rxidx(queue->rx.req_prod_pvt); new 729 drivers/net/xen-netfront.c BUG_ON(queue->rx_skbs[new]); new 730 drivers/net/xen-netfront.c queue->rx_skbs[new] = skb; new 731 drivers/net/xen-netfront.c queue->grant_rx_ref[new] = ref; new 732 drivers/net/xen-netfront.c RING_GET_REQUEST(&queue->rx, queue->rx.req_prod_pvt)->id = new; new 404 drivers/nfc/s3fwrn5/firmware.c struct s3fwrn5_fw_version *new = (void *) &fw_info->fw.version; new 407 drivers/nfc/s3fwrn5/firmware.c if (new->major > old->major) new 409 drivers/nfc/s3fwrn5/firmware.c if (new->build1 > old->build1) new 411 drivers/nfc/s3fwrn5/firmware.c if (new->build2 > old->build2) new 536 drivers/nvdimm/btt.c int new, ret; new 546 drivers/nvdimm/btt.c new = btt_log_read(arena, i, &log_new, LOG_NEW_ENT); new 547 drivers/nvdimm/btt.c if (new < 0) new 548 drivers/nvdimm/btt.c return new; new 555 drivers/nvdimm/btt.c arena->freelist[i].sub = 1 - new; new 561 drivers/nvdimm/dimm_devs.c resource_size_t map_end, blk_start, new; new 610 drivers/nvdimm/dimm_devs.c new = max(blk_start, min(map_end + 1, res->end + 1)); new 611 drivers/nvdimm/dimm_devs.c if (new != blk_start) { new 612 drivers/nvdimm/dimm_devs.c blk_start = new; new 1959 drivers/nvme/host/core.c u64 new, unsigned flags) new 1969 drivers/nvme/host/core.c return nvme_pr_command(bdev, cdw10, old, new, nvme_cmd_resv_register); new 1985 drivers/nvme/host/core.c static int nvme_pr_preempt(struct block_device *bdev, u64 old, u64 new, new 1989 drivers/nvme/host/core.c return nvme_pr_command(bdev, cdw10, old, new, nvme_cmd_resv_acquire); new 3310 drivers/nvme/host/core.c struct nvme_ns_head *new) new 3317 drivers/nvme/host/core.c if (nvme_ns_ids_valid(&new->ids) && new 3319 drivers/nvme/host/core.c nvme_ns_ids_equal(&new->ids, &h->ids)) new 779 drivers/nvme/host/rdma.c bool new) new 802 drivers/nvme/host/rdma.c if (new) { new 844 drivers/nvme/host/rdma.c if (new) new 847 drivers/nvme/host/rdma.c if (new) new 850 drivers/nvme/host/rdma.c if (new) new 873 drivers/nvme/host/rdma.c static int nvme_rdma_configure_io_queues(struct nvme_rdma_ctrl *ctrl, bool new) new 881 drivers/nvme/host/rdma.c if (new) { new 905 drivers/nvme/host/rdma.c if (new) new 908 drivers/nvme/host/rdma.c if (new) new 983 drivers/nvme/host/rdma.c static int nvme_rdma_setup_ctrl(struct nvme_rdma_ctrl *ctrl, bool new) new 988 drivers/nvme/host/rdma.c ret = nvme_rdma_configure_admin_queue(ctrl, new); new 1020 drivers/nvme/host/rdma.c ret = nvme_rdma_configure_io_queues(ctrl, new); new 1038 drivers/nvme/host/rdma.c nvme_rdma_destroy_io_queues(ctrl, new); new 1041 drivers/nvme/host/rdma.c nvme_rdma_destroy_admin_queue(ctrl, new); new 1664 drivers/nvme/host/tcp.c static int nvme_tcp_configure_io_queues(struct nvme_ctrl *ctrl, bool new) new 1672 drivers/nvme/host/tcp.c if (new) { new 1696 drivers/nvme/host/tcp.c if (new) new 1699 drivers/nvme/host/tcp.c if (new) new 1717 drivers/nvme/host/tcp.c static int nvme_tcp_configure_admin_queue(struct nvme_ctrl *ctrl, bool new) new 1725 drivers/nvme/host/tcp.c if (new) { new 1764 drivers/nvme/host/tcp.c if (new) new 1767 drivers/nvme/host/tcp.c if (new) new 1770 drivers/nvme/host/tcp.c if (new) new 1829 drivers/nvme/host/tcp.c static int nvme_tcp_setup_ctrl(struct nvme_ctrl *ctrl, bool new) new 1834 drivers/nvme/host/tcp.c ret = nvme_tcp_configure_admin_queue(ctrl, new); new 1856 drivers/nvme/host/tcp.c ret = nvme_tcp_configure_io_queues(ctrl, new); new 1873 drivers/nvme/host/tcp.c nvme_tcp_destroy_io_queues(ctrl, new); new 1876 drivers/nvme/host/tcp.c nvme_tcp_destroy_admin_queue(ctrl, new); new 1057 drivers/nvme/target/core.c void nvmet_update_cc(struct nvmet_ctrl *ctrl, u32 new) new 1063 drivers/nvme/target/core.c ctrl->cc = new; new 1065 drivers/nvme/target/core.c if (nvmet_cc_en(new) && !nvmet_cc_en(old)) new 1067 drivers/nvme/target/core.c if (!nvmet_cc_en(new) && nvmet_cc_en(old)) new 1069 drivers/nvme/target/core.c if (nvmet_cc_shn(new) && !nvmet_cc_shn(old)) { new 1073 drivers/nvme/target/core.c if (!nvmet_cc_shn(new) && nvmet_cc_shn(old)) new 394 drivers/nvme/target/nvmet.h void nvmet_update_cc(struct nvmet_ctrl *ctrl, u32 new); new 1585 drivers/of/base.c struct device_node *cur, *new = NULL; new 1651 drivers/of/base.c of_node_put(new); new 1652 drivers/of/base.c new = of_find_node_by_phandle(be32_to_cpup(map)); new 1657 drivers/of/base.c if (!new) new 1660 drivers/of/base.c if (!of_device_is_available(new)) new 1663 drivers/of/base.c ret = of_property_read_u32(new, cells_name, &new_size); new 1703 drivers/of/base.c out_args->np = new; new 1705 drivers/of/base.c cur = new; new 1709 drivers/of/base.c of_node_put(new); new 379 drivers/of/dynamic.c struct property *new; new 381 drivers/of/dynamic.c new = kzalloc(sizeof(*new), allocflags); new 382 drivers/of/dynamic.c if (!new) new 391 drivers/of/dynamic.c new->name = kstrdup(prop->name, allocflags); new 392 drivers/of/dynamic.c new->value = kmemdup(prop->value, prop->length, allocflags); new 393 drivers/of/dynamic.c new->length = prop->length; new 394 drivers/of/dynamic.c if (!new->name || !new->value) new 398 drivers/of/dynamic.c of_property_set_flag(new, OF_DYNAMIC); new 400 drivers/of/dynamic.c return new; new 403 drivers/of/dynamic.c kfree(new->name); new 404 drivers/of/dynamic.c kfree(new->value); new 405 drivers/of/dynamic.c kfree(new); new 686 drivers/of/unittest.c struct property *new; new 688 drivers/of/unittest.c new = __of_prop_dup(&p1, GFP_KERNEL); new 689 drivers/of/unittest.c unittest(new && propcmp(&p1, new), "empty property didn't copy correctly\n"); new 690 drivers/of/unittest.c kfree(new->value); new 691 drivers/of/unittest.c kfree(new->name); new 692 drivers/of/unittest.c kfree(new); new 694 drivers/of/unittest.c new = __of_prop_dup(&p2, GFP_KERNEL); new 695 drivers/of/unittest.c unittest(new && propcmp(&p2, new), "non-empty property didn't copy correctly\n"); new 696 drivers/of/unittest.c kfree(new->value); new 697 drivers/of/unittest.c kfree(new->name); new 698 drivers/of/unittest.c kfree(new); new 498 drivers/oprofile/buffer_sync.c struct task_struct *new; new 536 drivers/oprofile/buffer_sync.c new = (struct task_struct *)val; new 539 drivers/oprofile/buffer_sync.c mm = get_task_mm(new); new 542 drivers/oprofile/buffer_sync.c add_user_ctx_switch(new, cookie); new 571 drivers/parisc/lba_pci.c truncate_pat_collision(struct resource *root, struct resource *new) new 573 drivers/parisc/lba_pci.c unsigned long start = new->start; new 574 drivers/parisc/lba_pci.c unsigned long end = new->end; new 594 drivers/parisc/lba_pci.c new->start = tmp->end + 1; new 604 drivers/parisc/lba_pci.c new->end = tmp->start - 1; new 610 drivers/parisc/lba_pci.c (long)new->start, (long)new->end ); new 466 drivers/pci/controller/pci-aardvark.c int reg, u32 old, u32 new, u32 mask) new 472 drivers/pci/controller/pci-aardvark.c advk_writel(pcie, new, PCIE_CORE_PCIEXP_CAP + reg); new 476 drivers/pci/controller/pci-aardvark.c advk_writel(pcie, new, PCIE_CORE_PCIEXP_CAP + reg); new 477 drivers/pci/controller/pci-aardvark.c if (new & PCI_EXP_LNKCTL_RL) new 485 drivers/pci/controller/pci-aardvark.c if ((new & PCI_EXP_RTCTL_PMEIE) == 0) new 492 drivers/pci/controller/pci-aardvark.c new = (new & PCI_EXP_RTSTA_PME) >> 9; new 493 drivers/pci/controller/pci-aardvark.c advk_writel(pcie, new, PCIE_ISR0_REG); new 474 drivers/pci/controller/pci-mvebu.c int reg, u32 old, u32 new, u32 mask) new 485 drivers/pci/controller/pci-mvebu.c if ((old ^ new) & PCI_COMMAND_IO) new 487 drivers/pci/controller/pci-mvebu.c if ((old ^ new) & PCI_COMMAND_MEMORY) new 523 drivers/pci/controller/pci-mvebu.c int reg, u32 old, u32 new, u32 mask) new 533 drivers/pci/controller/pci-mvebu.c new &= ~(PCI_EXP_DEVCTL_URRE | PCI_EXP_DEVCTL_FERE | new 536 drivers/pci/controller/pci-mvebu.c mvebu_writel(port, new, PCIE_CAP_PCIEXP + PCI_EXP_DEVCTL); new 546 drivers/pci/controller/pci-mvebu.c new &= ~PCI_EXP_LNKCTL_CLKREQ_EN; new 548 drivers/pci/controller/pci-mvebu.c mvebu_writel(port, new, PCIE_CAP_PCIEXP + PCI_EXP_LNKCTL); new 552 drivers/pci/controller/pci-mvebu.c mvebu_writel(port, new, PCIE_RC_RTSTA); new 985 drivers/pci/hotplug/ibmphp_hpc.c static int process_changeinlatch(u8 old, u8 new, struct controller *ctrl) new 992 drivers/pci/hotplug/ibmphp_hpc.c debug("%s - Entry old[%x], new[%x]\n", __func__, old, new); new 997 drivers/pci/hotplug/ibmphp_hpc.c if ((mask & old) != (mask & new)) { new 423 drivers/pci/hotplug/pnv_php.c u16 new, mask; new 432 drivers/pci/hotplug/pnv_php.c new = PCI_EXP_SLTCTL_ATTN_IND_ON; new 434 drivers/pci/hotplug/pnv_php.c new = PCI_EXP_SLTCTL_ATTN_IND_OFF; new 436 drivers/pci/hotplug/pnv_php.c pcie_capability_clear_and_set_word(bridge, PCI_EXP_SLTCTL, mask, new); new 796 drivers/pci/iov.c u32 new; new 829 drivers/pci/iov.c new = region.start; new 830 drivers/pci/iov.c new |= res->flags & ~PCI_BASE_ADDRESS_MEM_MASK; new 833 drivers/pci/iov.c pci_write_config_dword(dev, reg, new); new 835 drivers/pci/iov.c new = region.start >> 16 >> 16; new 836 drivers/pci/iov.c pci_write_config_dword(dev, reg + 4, new); new 387 drivers/pci/pci-bridge-emul.c int mask, ret, old, new, shift; new 389 drivers/pci/pci-bridge-emul.c u32 old, u32 new, u32 mask); new 426 drivers/pci/pci-bridge-emul.c new = old & (~mask | ~behavior[reg / 4].rw); new 429 drivers/pci/pci-bridge-emul.c new |= (value << shift) & (behavior[reg / 4].rw & mask); new 432 drivers/pci/pci-bridge-emul.c new &= ~((value << shift) & (behavior[reg / 4].w1c & mask)); new 434 drivers/pci/pci-bridge-emul.c cfgspace[reg / 4] = new; new 437 drivers/pci/pci-bridge-emul.c write_op(bridge, reg, old, new, mask); new 100 drivers/pci/pci-bridge-emul.h u32 old, u32 new, u32 mask); new 107 drivers/pci/pci-bridge-emul.h u32 old, u32 new, u32 mask); new 4318 drivers/pci/pci.c u16 pci_command, new; new 4323 drivers/pci/pci.c new = pci_command & ~PCI_COMMAND_INTX_DISABLE; new 4325 drivers/pci/pci.c new = pci_command | PCI_COMMAND_INTX_DISABLE; new 4327 drivers/pci/pci.c if (new != pci_command) { new 4330 drivers/pci/pci.c pci_write_config_word(pdev, PCI_COMMAND, new); new 354 drivers/pci/pci.h pci_channel_state_t new) new 359 drivers/pci/pci.h switch (new) { new 387 drivers/pci/pci.h dev->error_state = new; new 22 drivers/pci/pcie/err.c enum pci_ers_result new) new 24 drivers/pci/pcie/err.c if (new == PCI_ERS_RESULT_NO_AER_DRIVER) new 27 drivers/pci/pcie/err.c if (new == PCI_ERS_RESULT_NONE) new 33 drivers/pci/pcie/err.c orig = new; new 36 drivers/pci/pcie/err.c if (new == PCI_ERS_RESULT_NEED_RESET) new 104 drivers/pci/pcie/portdrv.h int pcie_port_service_register(struct pcie_port_service_driver *new); new 105 drivers/pci/pcie/portdrv.h void pcie_port_service_unregister(struct pcie_port_service_driver *new); new 584 drivers/pci/pcie/portdrv_core.c int pcie_port_service_register(struct pcie_port_service_driver *new) new 589 drivers/pci/pcie/portdrv_core.c new->driver.name = new->name; new 590 drivers/pci/pcie/portdrv_core.c new->driver.bus = &pcie_port_bus_type; new 591 drivers/pci/pcie/portdrv_core.c new->driver.probe = pcie_port_probe_service; new 592 drivers/pci/pcie/portdrv_core.c new->driver.remove = pcie_port_remove_service; new 593 drivers/pci/pcie/portdrv_core.c new->driver.shutdown = pcie_port_shutdown_service; new 595 drivers/pci/pcie/portdrv_core.c return driver_register(&new->driver); new 1678 drivers/pci/probe.c u16 orig, toggle, new; new 1683 drivers/pci/probe.c pci_read_config_word(dev, PCI_COMMAND, &new); new 1692 drivers/pci/probe.c if (new != toggle) new 30 drivers/pci/setup-res.c u32 new, check, mask; new 57 drivers/pci/setup-res.c new = region.start; new 61 drivers/pci/setup-res.c new |= res->flags & ~PCI_BASE_ADDRESS_IO_MASK; new 66 drivers/pci/setup-res.c new |= res->flags & ~PCI_BASE_ADDRESS_MEM_MASK; new 82 drivers/pci/setup-res.c new |= PCI_ROM_ADDRESS_ENABLE; new 98 drivers/pci/setup-res.c pci_write_config_dword(dev, reg, new); new 101 drivers/pci/setup-res.c if ((new ^ check) & mask) { new 103 drivers/pci/setup-res.c resno, new, check); new 107 drivers/pci/setup-res.c new = region.start >> 16 >> 16; new 108 drivers/pci/setup-res.c pci_write_config_dword(dev, reg + 4, new); new 110 drivers/pci/setup-res.c if (check != new) { new 112 drivers/pci/setup-res.c resno, new, check); new 460 drivers/pcmcia/ds.c char *new; new 468 drivers/pcmcia/ds.c new = kstrdup(tmp, GFP_KERNEL); new 469 drivers/pcmcia/ds.c if (!new) new 473 drivers/pcmcia/ds.c p_dev->prod_id[i] = new; new 193 drivers/pcmcia/pxa2xx_base.c if (freqs->new > freqs->old) { new 196 drivers/pcmcia/pxa2xx_base.c freqs->new / 1000, (freqs->new / 100) % 10, new 203 drivers/pcmcia/pxa2xx_base.c if (freqs->new < freqs->old) { new 206 drivers/pcmcia/pxa2xx_base.c freqs->new / 1000, (freqs->new / 100) % 10, new 120 drivers/pcmcia/sa11xx_base.c if (freqs->new > freqs->old) new 121 drivers/pcmcia/sa11xx_base.c sa1100_pcmcia_set_mecr(skt, freqs->new); new 125 drivers/pcmcia/sa11xx_base.c if (freqs->new < freqs->old) new 126 drivers/pcmcia/sa11xx_base.c sa1100_pcmcia_set_mecr(skt, freqs->new); new 296 drivers/pcmcia/ti113x.h u8 new, reg = exca_readb(socket, I365_INTCTL); new 298 drivers/pcmcia/ti113x.h new = reg & ~I365_INTR_ENA; new 300 drivers/pcmcia/ti113x.h new |= I365_INTR_ENA; new 301 drivers/pcmcia/ti113x.h if (new != reg) new 302 drivers/pcmcia/ti113x.h exca_writeb(socket, I365_INTCTL, new); new 308 drivers/pcmcia/ti113x.h u8 new, reg = exca_readb(socket, I365_INTCTL); new 310 drivers/pcmcia/ti113x.h new = reg & ~I365_INTR_ENA; new 311 drivers/pcmcia/ti113x.h if (new != reg) new 312 drivers/pcmcia/ti113x.h exca_writeb(socket, I365_INTCTL, new); new 225 drivers/perf/arm_smmuv3_pmu.c u64 new; new 235 drivers/perf/arm_smmuv3_pmu.c new = smmu_pmu_counter_get_value(smmu_pmu, idx); new 243 drivers/perf/arm_smmuv3_pmu.c new = smmu_pmu->counter_mask >> 1; new 244 drivers/perf/arm_smmuv3_pmu.c smmu_pmu_counter_set_value(smmu_pmu, idx, new); new 247 drivers/perf/arm_smmuv3_pmu.c local64_set(&hwc->prev_count, new); new 262 drivers/perf/arm_smmuv3_pmu.c struct perf_event *new) new 264 drivers/perf/arm_smmuv3_pmu.c if (get_filter_enable(new) != get_filter_enable(curr)) new 267 drivers/perf/arm_smmuv3_pmu.c if (!get_filter_enable(new)) new 270 drivers/perf/arm_smmuv3_pmu.c return get_filter_span(new) == get_filter_span(curr) && new 271 drivers/perf/arm_smmuv3_pmu.c get_filter_stream_id(new) == get_filter_stream_id(curr); new 324 drivers/perf/arm_smmuv3_pmu.c struct perf_event *new) new 326 drivers/perf/arm_smmuv3_pmu.c if (new->pmu != curr->pmu) new 329 drivers/perf/arm_smmuv3_pmu.c if (to_smmu_pmu(new->pmu)->global_filter && new 330 drivers/perf/arm_smmuv3_pmu.c !smmu_pmu_check_global_filter(curr, new)) new 365 drivers/perf/qcom_l2_pmu.c u64 new; new 373 drivers/perf/qcom_l2_pmu.c new = L2_CYCLE_COUNTER_RELOAD; new 375 drivers/perf/qcom_l2_pmu.c new = L2_COUNTER_RELOAD; new 377 drivers/perf/qcom_l2_pmu.c local64_set(&hwc->prev_count, new); new 378 drivers/perf/qcom_l2_pmu.c cluster_pmu_counter_set_value(idx, new); new 244 drivers/perf/qcom_l3_pmu.c u64 prev, new; new 252 drivers/perf/qcom_l3_pmu.c new = ((u64)hi << 32) | lo; new 253 drivers/perf/qcom_l3_pmu.c } while (local64_cmpxchg(&event->hw.prev_count, prev, new) != prev); new 255 drivers/perf/qcom_l3_pmu.c local64_add(new - prev, &event->count); new 320 drivers/perf/qcom_l3_pmu.c u32 prev, new; new 324 drivers/perf/qcom_l3_pmu.c new = readl_relaxed(l3pmu->regs + L3_HML3_PM_EVCNTR(idx)); new 325 drivers/perf/qcom_l3_pmu.c } while (local64_cmpxchg(&event->hw.prev_count, prev, new) != prev); new 327 drivers/perf/qcom_l3_pmu.c local64_add(new - prev, &event->count); new 317 drivers/perf/thunderx2_pmu.c s64 prev, delta, new = 0; new 327 drivers/perf/thunderx2_pmu.c new = reg_readl(hwc->event_base); new 328 drivers/perf/thunderx2_pmu.c prev = local64_xchg(&hwc->prev_count, new); new 331 drivers/perf/thunderx2_pmu.c delta = (u32)(((1UL << 32) - prev) + new); new 1830 drivers/platform/x86/asus-wmi.c int new = old; new 1833 drivers/platform/x86/asus-wmi.c new = code - NOTIFY_BRNUP_MIN + 1; new 1835 drivers/platform/x86/asus-wmi.c new = code - NOTIFY_BRNDOWN_MIN; new 1837 drivers/platform/x86/asus-wmi.c bd->props.brightness = new; new 335 drivers/rpmsg/rpmsg_core.c char *new, *old; \ new 337 drivers/rpmsg/rpmsg_core.c new = kstrndup(buf, sz, GFP_KERNEL); \ new 338 drivers/rpmsg/rpmsg_core.c if (!new) \ new 340 drivers/rpmsg/rpmsg_core.c new[strcspn(new, "\n")] = '\0'; \ new 344 drivers/rpmsg/rpmsg_core.c if (strlen(new)) { \ new 345 drivers/rpmsg/rpmsg_core.c rpdev->member = new; \ new 347 drivers/rpmsg/rpmsg_core.c kfree(new); \ new 412 drivers/s390/block/dasd_devmap.c struct dasd_devmap *devmap, *new, *tmp; new 415 drivers/s390/block/dasd_devmap.c new = kzalloc(sizeof(struct dasd_devmap), GFP_KERNEL); new 416 drivers/s390/block/dasd_devmap.c if (!new) new 428 drivers/s390/block/dasd_devmap.c new->devindex = dasd_max_devindex++; new 429 drivers/s390/block/dasd_devmap.c strlcpy(new->bus_id, bus_id, DASD_BUS_ID_SIZE); new 430 drivers/s390/block/dasd_devmap.c new->features = features; new 431 drivers/s390/block/dasd_devmap.c new->device = NULL; new 432 drivers/s390/block/dasd_devmap.c list_add(&new->list, &dasd_hashlists[hash]); new 433 drivers/s390/block/dasd_devmap.c devmap = new; new 434 drivers/s390/block/dasd_devmap.c new = NULL; new 437 drivers/s390/block/dasd_devmap.c kfree(new); new 100 drivers/s390/char/tape_char.c struct idal_buffer *new; new 113 drivers/s390/char/tape_char.c new = idal_buffer_alloc(block_size, 0); new 114 drivers/s390/char/tape_char.c if (IS_ERR(new)) new 120 drivers/s390/char/tape_char.c device->char_data.idal_buf = new; new 214 drivers/s390/char/tape_std.c struct idal_buffer *new; new 238 drivers/s390/char/tape_std.c new = idal_buffer_alloc(count, 0); new 239 drivers/s390/char/tape_std.c if (IS_ERR(new)) new 243 drivers/s390/char/tape_std.c device->char_data.idal_buf = new; new 1782 drivers/s390/char/tty3270.c int new; new 1789 drivers/s390/char/tty3270.c new = L_ECHO(tty) ? TF_INPUT: TF_INPUTN; new 1790 drivers/s390/char/tty3270.c if (new != tp->inattr) { new 1791 drivers/s390/char/tty3270.c tp->inattr = new; new 457 drivers/s390/cio/qdio_main.c int new; new 482 drivers/s390/cio/qdio_main.c new = add_buf(start, count - 1); new 485 drivers/s390/cio/qdio_main.c set_buf_state(q, new, SLSB_P_INPUT_ACK); new 489 drivers/s390/cio/qdio_main.c set_buf_state(q, new, SLSB_P_INPUT_ACK); new 492 drivers/s390/cio/qdio_main.c q->u.in.ack_start = new; new 589 drivers/s390/net/qeth_l3_main.c struct qeth_ipato_entry *new) new 599 drivers/s390/net/qeth_l3_main.c if (ipatoe->proto != new->proto) new 601 drivers/s390/net/qeth_l3_main.c if (!memcmp(ipatoe->addr, new->addr, new 603 drivers/s390/net/qeth_l3_main.c (ipatoe->mask_bits == new->mask_bits)) { new 610 drivers/s390/net/qeth_l3_main.c list_add_tail(&new->entry, &card->ipato.entries); new 392 drivers/sbus/char/bbc_envctrl.c int new; new 394 drivers/sbus/char/bbc_envctrl.c new = maybe_new_ambient_fan_speed(fp); new 395 drivers/sbus/char/bbc_envctrl.c new |= maybe_new_cpu_fan_speed(fp); new 397 drivers/sbus/char/bbc_envctrl.c if (new) new 160 drivers/scsi/fcoe/fcoe_ctlr.c static int fcoe_sysfs_fcf_add(struct fcoe_fcf *new) new 162 drivers/scsi/fcoe/fcoe_ctlr.c struct fcoe_ctlr *fip = new->fip; new 168 drivers/scsi/fcoe/fcoe_ctlr.c new->fabric_name, new->fcf_mac); new 174 drivers/scsi/fcoe/fcoe_ctlr.c temp->fabric_name = new->fabric_name; new 175 drivers/scsi/fcoe/fcoe_ctlr.c temp->switch_name = new->switch_name; new 176 drivers/scsi/fcoe/fcoe_ctlr.c temp->fc_map = new->fc_map; new 177 drivers/scsi/fcoe/fcoe_ctlr.c temp->vfid = new->vfid; new 178 drivers/scsi/fcoe/fcoe_ctlr.c memcpy(temp->mac, new->fcf_mac, ETH_ALEN); new 179 drivers/scsi/fcoe/fcoe_ctlr.c temp->priority = new->pri; new 180 drivers/scsi/fcoe/fcoe_ctlr.c temp->fka_period = new->fka_period; new 213 drivers/scsi/fcoe/fcoe_ctlr.c fcf_dev->priv = new; new 214 drivers/scsi/fcoe/fcoe_ctlr.c new->fcf_dev = fcf_dev; new 218 drivers/scsi/fcoe/fcoe_ctlr.c list_add(&new->list, &fip->fcfs); new 233 drivers/scsi/fcoe/fcoe_ctlr.c static void fcoe_sysfs_fcf_del(struct fcoe_fcf *new) new 235 drivers/scsi/fcoe/fcoe_ctlr.c struct fcoe_ctlr *fip = new->fip; new 239 drivers/scsi/fcoe/fcoe_ctlr.c list_del(&new->list); new 254 drivers/scsi/fcoe/fcoe_ctlr.c fcf_dev = fcoe_fcf_to_fcf_dev(new); new 256 drivers/scsi/fcoe/fcoe_ctlr.c new->fcf_dev = NULL; new 258 drivers/scsi/fcoe/fcoe_ctlr.c kfree(new); new 1021 drivers/scsi/fcoe/fcoe_ctlr.c struct fcoe_fcf new; new 1028 drivers/scsi/fcoe/fcoe_ctlr.c if (fcoe_ctlr_parse_adv(fip, skb, &new)) new 1034 drivers/scsi/fcoe/fcoe_ctlr.c if (fcf->switch_name == new.switch_name && new 1035 drivers/scsi/fcoe/fcoe_ctlr.c fcf->fabric_name == new.fabric_name && new 1036 drivers/scsi/fcoe/fcoe_ctlr.c fcf->fc_map == new.fc_map && new 1037 drivers/scsi/fcoe/fcoe_ctlr.c ether_addr_equal(fcf->fcf_mac, new.fcf_mac)) { new 1050 drivers/scsi/fcoe/fcoe_ctlr.c memcpy(fcf, &new, sizeof(new)); new 1056 drivers/scsi/fcoe/fcoe_ctlr.c new.fabric_name, new.fcf_mac); new 1067 drivers/scsi/fcoe/fcoe_ctlr.c fcf->fd_flags = new.fd_flags; new 1069 drivers/scsi/fcoe/fcoe_ctlr.c fcf->flags = new.flags; new 1073 drivers/scsi/fcoe/fcoe_ctlr.c fip->ctlr_ka_time += new.fka_period; new 1077 drivers/scsi/fcoe/fcoe_ctlr.c fcf->fka_period = new.fka_period; new 1078 drivers/scsi/fcoe/fcoe_ctlr.c memcpy(fcf->fcf_mac, new.fcf_mac, ETH_ALEN); new 2489 drivers/scsi/fcoe/fcoe_ctlr.c static void fcoe_ctlr_vn_add(struct fcoe_ctlr *fip, struct fcoe_rport *new) new 2497 drivers/scsi/fcoe/fcoe_ctlr.c port_id = new->rdata.ids.port_id; new 2515 drivers/scsi/fcoe/fcoe_ctlr.c ids->port_name != new->rdata.ids.port_name) || new 2517 drivers/scsi/fcoe/fcoe_ctlr.c ids->node_name != new->rdata.ids.node_name)) { new 2523 drivers/scsi/fcoe/fcoe_ctlr.c ids->port_name = new->rdata.ids.port_name; new 2524 drivers/scsi/fcoe/fcoe_ctlr.c ids->node_name = new->rdata.ids.node_name; new 2531 drivers/scsi/fcoe/fcoe_ctlr.c frport->fcoe_len = new->fcoe_len; new 2532 drivers/scsi/fcoe/fcoe_ctlr.c frport->flags = new->flags; new 2533 drivers/scsi/fcoe/fcoe_ctlr.c frport->login_count = new->login_count; new 2534 drivers/scsi/fcoe/fcoe_ctlr.c memcpy(frport->enode_mac, new->enode_mac, ETH_ALEN); new 2535 drivers/scsi/fcoe/fcoe_ctlr.c memcpy(frport->vn_mac, new->vn_mac, ETH_ALEN); new 2572 drivers/scsi/fcoe/fcoe_ctlr.c struct fcoe_rport *new) new 2574 drivers/scsi/fcoe/fcoe_ctlr.c if (new->flags & FIP_FL_REC_OR_P2P) { new 2583 drivers/scsi/fcoe/fcoe_ctlr.c if (new->rdata.ids.port_id == fip->port_id) { new 2592 drivers/scsi/fcoe/fcoe_ctlr.c if (new->rdata.ids.port_id == fip->port_id) { new 2593 drivers/scsi/fcoe/fcoe_ctlr.c if (new->rdata.ids.port_name > fip->lp->wwpn) { new 2605 drivers/scsi/fcoe/fcoe_ctlr.c new->rdata.ids.port_id); new 2606 drivers/scsi/fcoe/fcoe_ctlr.c fcoe_ctlr_vn_send(fip, FIP_SC_VN_CLAIM_REP, new->enode_mac, new 2607 drivers/scsi/fcoe/fcoe_ctlr.c min((u32)new->fcoe_len, new 2609 drivers/scsi/fcoe/fcoe_ctlr.c fcoe_ctlr_vn_add(fip, new); new 2614 drivers/scsi/fcoe/fcoe_ctlr.c new->rdata.ids.port_id); new 2627 drivers/scsi/fcoe/fcoe_ctlr.c struct fcoe_rport *new) new 2630 drivers/scsi/fcoe/fcoe_ctlr.c new->rdata.ids.port_id, fcoe_ctlr_state(fip->state)); new 2632 drivers/scsi/fcoe/fcoe_ctlr.c fcoe_ctlr_vn_add(fip, new); new 2643 drivers/scsi/fcoe/fcoe_ctlr.c struct fcoe_rport *new) new 2649 drivers/scsi/fcoe/fcoe_ctlr.c if (new->flags & FIP_FL_REC_OR_P2P) { new 2654 drivers/scsi/fcoe/fcoe_ctlr.c rdata = fc_rport_lookup(lport, new->rdata.ids.port_id); new 2656 drivers/scsi/fcoe/fcoe_ctlr.c if (rdata->ids.node_name == new->rdata.ids.node_name && new 2657 drivers/scsi/fcoe/fcoe_ctlr.c rdata->ids.port_name == new->rdata.ids.port_name) { new 2682 drivers/scsi/fcoe/fcoe_ctlr.c new->rdata.ids.port_id); new 771 drivers/scsi/fcoe/fcoe_sysfs.c static int fcoe_fcf_device_match(struct fcoe_fcf_device *new, new 774 drivers/scsi/fcoe/fcoe_sysfs.c if (new->switch_name == old->switch_name && new 775 drivers/scsi/fcoe/fcoe_sysfs.c new->fabric_name == old->fabric_name && new 776 drivers/scsi/fcoe/fcoe_sysfs.c new->fc_map == old->fc_map && new 777 drivers/scsi/fcoe/fcoe_sysfs.c ether_addr_equal(new->mac, old->mac)) new 342 drivers/scsi/fnic/fnic.h void fnic_update_mac(struct fc_lport *, u8 *new); new 343 drivers/scsi/fnic/fnic.h void fnic_update_mac_locked(struct fnic *, u8 *new); new 730 drivers/scsi/fnic/fnic_fcs.c void fnic_update_mac_locked(struct fnic *fnic, u8 *new) new 735 drivers/scsi/fnic/fnic_fcs.c if (is_zero_ether_addr(new)) new 736 drivers/scsi/fnic/fnic_fcs.c new = ctl; new 737 drivers/scsi/fnic/fnic_fcs.c if (ether_addr_equal(data, new)) new 739 drivers/scsi/fnic/fnic_fcs.c FNIC_FCS_DBG(KERN_DEBUG, fnic->lport->host, "update_mac %pM\n", new); new 742 drivers/scsi/fnic/fnic_fcs.c memcpy(data, new, ETH_ALEN); new 743 drivers/scsi/fnic/fnic_fcs.c if (!ether_addr_equal(new, ctl)) new 744 drivers/scsi/fnic/fnic_fcs.c vnic_dev_add_addr(fnic->vdev, new); new 752 drivers/scsi/fnic/fnic_fcs.c void fnic_update_mac(struct fc_lport *lport, u8 *new) new 757 drivers/scsi/fnic/fnic_fcs.c fnic_update_mac_locked(fnic, new); new 241 drivers/scsi/libfc/fc_rport.c enum fc_rport_state new) new 245 drivers/scsi/libfc/fc_rport.c if (rdata->rp_state != new) new 247 drivers/scsi/libfc/fc_rport.c rdata->rp_state = new; new 1979 drivers/scsi/libsas/sas_expander.c static bool dev_type_flutter(enum sas_device_type new, enum sas_device_type old) new 1981 drivers/scsi/libsas/sas_expander.c if (old == new) new 1987 drivers/scsi/libsas/sas_expander.c if ((old == SAS_SATA_PENDING && new == SAS_END_DEVICE) || new 1988 drivers/scsi/libsas/sas_expander.c (old == SAS_END_DEVICE && new == SAS_SATA_PENDING)) new 357 drivers/scsi/myrb.c struct myrb_error_entry *new, *old; new 366 drivers/scsi/myrb.c new = table + err_table_offset; new 368 drivers/scsi/myrb.c if (new->parity_err == old->parity_err && new 369 drivers/scsi/myrb.c new->soft_err == old->soft_err && new 370 drivers/scsi/myrb.c new->hard_err == old->hard_err && new 371 drivers/scsi/myrb.c new->misc_err == old->misc_err) new 375 drivers/scsi/myrb.c new->parity_err, new->soft_err, new 376 drivers/scsi/myrb.c new->hard_err, new->misc_err); new 401 drivers/scsi/myrb.c struct myrb_ldev_info *new = cb->ldev_info_buf + ldev_num; new 407 drivers/scsi/myrb.c if (new->state == MYRB_DEVICE_OFFLINE) new 411 drivers/scsi/myrb.c ldev_num, myrb_devstate_name(new->state)); new 417 drivers/scsi/myrb.c if (new->state != old->state) new 420 drivers/scsi/myrb.c ldev_num, myrb_devstate_name(new->state)); new 421 drivers/scsi/myrb.c if (new->wb_enabled != old->wb_enabled) new 424 drivers/scsi/myrb.c (new->wb_enabled ? "BACK" : "THRU")); new 425 drivers/scsi/myrb.c memcpy(old, new, sizeof(*new)); new 673 drivers/scsi/myrb.c struct myrb_enquiry old, *new; new 682 drivers/scsi/myrb.c new = cb->enquiry; new 683 drivers/scsi/myrb.c if (new->ldev_count > old.ldev_count) { new 686 drivers/scsi/myrb.c while (++ldev_num < new->ldev_count) new 691 drivers/scsi/myrb.c if (new->ldev_count < old.ldev_count) { new 692 drivers/scsi/myrb.c int ldev_num = new->ldev_count - 1; new 699 drivers/scsi/myrb.c if (new->status.deferred != old.status.deferred) new 702 drivers/scsi/myrb.c (new->status.deferred ? "TRUE" : "FALSE")); new 703 drivers/scsi/myrb.c if (new->ev_seq != old.ev_seq) { new 704 drivers/scsi/myrb.c cb->new_ev_seq = new->ev_seq; new 709 drivers/scsi/myrb.c old.ev_seq, new->ev_seq); new 711 drivers/scsi/myrb.c if ((new->ldev_critical > 0 && new 712 drivers/scsi/myrb.c new->ldev_critical != old.ldev_critical) || new 713 drivers/scsi/myrb.c (new->ldev_offline > 0 && new 714 drivers/scsi/myrb.c new->ldev_offline != old.ldev_offline) || new 715 drivers/scsi/myrb.c (new->ldev_count != old.ldev_count)) { new 718 drivers/scsi/myrb.c new->ldev_critical, new 719 drivers/scsi/myrb.c new->ldev_offline, new 720 drivers/scsi/myrb.c new->ldev_count); new 723 drivers/scsi/myrb.c if (new->pdev_dead > 0 || new 724 drivers/scsi/myrb.c new->pdev_dead != old.pdev_dead || new 730 drivers/scsi/myrb.c if (new->rbld == MYRB_STDBY_RBLD_IN_PROGRESS || new 731 drivers/scsi/myrb.c new->rbld == MYRB_BG_RBLD_IN_PROGRESS || new 735 drivers/scsi/myrb.c cb->rbld_first = (new->ldev_critical < old.ldev_critical); new 738 drivers/scsi/myrb.c switch (new->rbld) { new 770 drivers/scsi/myrb.c else if (new->rbld == MYRB_BG_CHECK_IN_PROGRESS) new 258 drivers/scsi/myrs.c struct myrs_ldev_info *new = ldev_info; new 260 drivers/scsi/myrs.c unsigned long ldev_size = new->cfg_devsize; new 262 drivers/scsi/myrs.c if (new->dev_state != old->dev_state) { new 265 drivers/scsi/myrs.c name = myrs_devstate_name(new->dev_state); new 270 drivers/scsi/myrs.c if ((new->soft_errs != old->soft_errs) || new 271 drivers/scsi/myrs.c (new->cmds_failed != old->cmds_failed) || new 272 drivers/scsi/myrs.c (new->deferred_write_errs != old->deferred_write_errs)) new 275 drivers/scsi/myrs.c ldev_num, new->soft_errs, new 276 drivers/scsi/myrs.c new->cmds_failed, new 277 drivers/scsi/myrs.c new->deferred_write_errs); new 278 drivers/scsi/myrs.c if (new->bg_init_active) new 281 drivers/scsi/myrs.c new->bg_init_lba, ldev_size); new 282 drivers/scsi/myrs.c else if (new->fg_init_active) new 285 drivers/scsi/myrs.c new->fg_init_lba, ldev_size); new 286 drivers/scsi/myrs.c else if (new->migration_active) new 289 drivers/scsi/myrs.c new->migration_lba, ldev_size); new 290 drivers/scsi/myrs.c else if (new->patrol_active) new 293 drivers/scsi/myrs.c new->patrol_lba, ldev_size); new 294 drivers/scsi/myrs.c if (old->bg_init_active && !new->bg_init_active) new 298 drivers/scsi/myrs.c (new->ldev_control.ldev_init_done ? new 3514 drivers/scsi/ncr53c8xx.c ncrcmd opcode, new, old, tmp1, tmp2; new 3615 drivers/scsi/ncr53c8xx.c new = (old & ~RELOC_MASK) + np->paddr; new 3618 drivers/scsi/ncr53c8xx.c new = (old & ~RELOC_MASK) + np->p_script; new 3621 drivers/scsi/ncr53c8xx.c new = (old & ~RELOC_MASK) + np->p_scripth; new 3624 drivers/scsi/ncr53c8xx.c new = (old & ~RELOC_MASK) + np->p_ncb; new 3633 drivers/scsi/ncr53c8xx.c new = vtophys(script_kvars[old & new 3640 drivers/scsi/ncr53c8xx.c new = old; new 3649 drivers/scsi/ncr53c8xx.c *dst++ = cpu_to_scr(new); new 350 drivers/scsi/sym53c8xx_2/sym_fw.c u32 opcode, new, old, tmp1, tmp2; new 505 drivers/scsi/sym53c8xx_2/sym_fw.c new = (old & ~RELOC_MASK) + np->mmio_ba; new 508 drivers/scsi/sym53c8xx_2/sym_fw.c new = (old & ~RELOC_MASK) + np->scripta_ba; new 511 drivers/scsi/sym53c8xx_2/sym_fw.c new = (old & ~RELOC_MASK) + np->scriptb_ba; new 514 drivers/scsi/sym53c8xx_2/sym_fw.c new = (old & ~RELOC_MASK) + np->hcb_ba; new 523 drivers/scsi/sym53c8xx_2/sym_fw.c new = old; new 528 drivers/scsi/sym53c8xx_2/sym_fw.c new = 0; new 534 drivers/scsi/sym53c8xx_2/sym_fw.c *cur++ = cpu_to_scr(new); new 1316 drivers/scsi/sym53c8xx_2/sym_hipd.c goto new; new 1320 drivers/scsi/sym53c8xx_2/sym_hipd.c goto new; new 1325 drivers/scsi/sym53c8xx_2/sym_hipd.c new: new 52 drivers/scsi/sym53c8xx_2/sym_misc.h static inline void __sym_que_add(struct sym_quehead * new, new 56 drivers/scsi/sym53c8xx_2/sym_misc.h flink->blink = new; new 57 drivers/scsi/sym53c8xx_2/sym_misc.h new->flink = flink; new 58 drivers/scsi/sym53c8xx_2/sym_misc.h new->blink = blink; new 59 drivers/scsi/sym53c8xx_2/sym_misc.h blink->flink = new; new 113 drivers/scsi/sym53c8xx_2/sym_misc.h #define sym_insque(new, pos) __sym_que_add(new, pos, (pos)->flink) new 117 drivers/scsi/sym53c8xx_2/sym_misc.h #define sym_insque_head(new, head) __sym_que_add(new, head, (head)->flink) new 130 drivers/scsi/sym53c8xx_2/sym_misc.h #define sym_insque_tail(new, head) __sym_que_add(new, (head)->blink, head) new 470 drivers/soc/bcm/brcmstb/pm/pm-arm.c : [new] "r" (sp) new 202 drivers/soc/ti/knav_qmss_acc.c u32 old, new; new 216 drivers/soc/ti/knav_qmss_acc.c new = old | BIT(queue); new 218 drivers/soc/ti/knav_qmss_acc.c new = old & ~BIT(queue); new 219 drivers/soc/ti/knav_qmss_acc.c acc->open_mask = new; new 223 drivers/soc/ti/knav_qmss_acc.c old, new, acc->name); new 225 drivers/soc/ti/knav_qmss_acc.c if (likely(new == old)) new 228 drivers/soc/ti/knav_qmss_acc.c if (new && !old) { new 244 drivers/soc/ti/knav_qmss_acc.c if (old && !new) { new 246 drivers/staging/comedi/drivers/amplc_pci224.c #define COMBINE(old, new, mask) (((old) & ~(mask)) | ((new) & (mask))) new 432 drivers/staging/comedi/drivers/amplc_pci230.c #define COMBINE(old, new, mask) (((old) & ~(mask)) | ((new) & (mask))) new 125 drivers/staging/fwserial/fwserial.h static inline void peer_set_state(struct fwtty_peer *peer, int new) new 127 drivers/staging/fwserial/fwserial.h peer->state = new; new 667 drivers/staging/greybus/uart.c struct async_icount new; new 675 drivers/staging/greybus/uart.c new = gb_tty->iocount; new 676 drivers/staging/greybus/uart.c gb_tty->oldcount = new; new 679 drivers/staging/greybus/uart.c if ((arg & TIOCM_DSR) && (old.dsr != new.dsr)) new 681 drivers/staging/greybus/uart.c if ((arg & TIOCM_CD) && (old.dcd != new.dcd)) new 683 drivers/staging/greybus/uart.c if ((arg & TIOCM_RI) && (old.rng != new.rng)) new 475 drivers/staging/media/hantro/hantro_h264.c DECLARE_BITMAP(new, ARRAY_SIZE(dec_param->dpb)) = { 0, }; new 510 drivers/staging/media/hantro/hantro_h264.c set_bit(i, new); new 514 drivers/staging/media/hantro/hantro_h264.c for_each_set_bit(i, new, ARRAY_SIZE(dec_param->dpb)) { new 147 drivers/staging/unisys/visorhba/visorhba_main.c char cmdtype, void *new) new 166 drivers/staging/unisys/visorhba/visorhba_main.c if (new) new 167 drivers/staging/unisys/visorhba/visorhba_main.c entry->sent = new; new 206 drivers/staging/uwb/est.c void *new = kmalloc_array(2, actual_size, GFP_ATOMIC); new 207 drivers/staging/uwb/est.c if (new == NULL) new 209 drivers/staging/uwb/est.c memcpy(new, uwb_est, actual_size); new 210 drivers/staging/uwb/est.c memset(new + actual_size, 0, actual_size); new 212 drivers/staging/uwb/est.c uwb_est = new; new 35 drivers/staging/uwb/lc-dev.c int uwb_notifs_register(struct uwb_rc *rc, struct uwb_notifs_handler *new) new 39 drivers/staging/uwb/lc-dev.c list_add(&new->list_node, &rc->notifs_chain.list); new 295 drivers/target/target_core_device.c target_luns_data_has_changed(struct se_node_acl *nacl, struct se_dev_entry *new, new 302 drivers/target/target_core_device.c if (skip_new && tmp == new) new 318 drivers/target/target_core_device.c struct se_dev_entry *orig, *new; new 320 drivers/target/target_core_device.c new = kzalloc(sizeof(*new), GFP_KERNEL); new 321 drivers/target/target_core_device.c if (!new) { new 326 drivers/target/target_core_device.c spin_lock_init(&new->ua_lock); new 327 drivers/target/target_core_device.c INIT_LIST_HEAD(&new->ua_list); new 328 drivers/target/target_core_device.c INIT_LIST_HEAD(&new->lun_link); new 330 drivers/target/target_core_device.c new->mapped_lun = mapped_lun; new 331 drivers/target/target_core_device.c kref_init(&new->pr_kref); new 332 drivers/target/target_core_device.c init_completion(&new->pr_comp); new 334 drivers/target/target_core_device.c new->lun_access_ro = lun_access_ro; new 335 drivers/target/target_core_device.c new->creation_time = get_jiffies_64(); new 336 drivers/target/target_core_device.c new->attach_count++; new 349 drivers/target/target_core_device.c kfree(new); new 358 drivers/target/target_core_device.c kfree(new); new 362 drivers/target/target_core_device.c rcu_assign_pointer(new->se_lun, lun); new 363 drivers/target/target_core_device.c rcu_assign_pointer(new->se_lun_acl, lun_acl); new 365 drivers/target/target_core_device.c hlist_add_head_rcu(&new->link, &nacl->lun_entry_hlist); new 370 drivers/target/target_core_device.c list_add_tail(&new->lun_link, &lun->lun_deve_list); new 376 drivers/target/target_core_device.c target_luns_data_has_changed(nacl, new, true); new 381 drivers/target/target_core_device.c rcu_assign_pointer(new->se_lun, lun); new 382 drivers/target/target_core_device.c rcu_assign_pointer(new->se_lun_acl, lun_acl); new 383 drivers/target/target_core_device.c hlist_add_head_rcu(&new->link, &nacl->lun_entry_hlist); new 387 drivers/target/target_core_device.c list_add_tail(&new->lun_link, &lun->lun_deve_list); new 390 drivers/target/target_core_device.c target_luns_data_has_changed(nacl, new, true); new 203 drivers/thermal/rcar_thermal.c u32 ctemp, old, new; new 224 drivers/thermal/rcar_thermal.c new = rcar_thermal_read(priv, THSSR) & CTEMP; new 225 drivers/thermal/rcar_thermal.c if (new == old) { new 226 drivers/thermal/rcar_thermal.c ctemp = new; new 229 drivers/thermal/rcar_thermal.c old = new; new 62 drivers/thunderbolt/nhi.c u32 old, new; new 96 drivers/thunderbolt/nhi.c new = old | mask; new 98 drivers/thunderbolt/nhi.c new = old & ~mask; new 102 drivers/thunderbolt/nhi.c active ? "enabling" : "disabling", reg, bit, old, new); new 104 drivers/thunderbolt/nhi.c if (new == old) new 109 drivers/thunderbolt/nhi.c iowrite32(new, ring->nhi->iobase + reg); new 5645 drivers/tty/serial/8250/8250_pci.c struct serial_private *new; new 5650 drivers/tty/serial/8250/8250_pci.c new = pciserial_init_ports(dev, priv->board); new 5651 drivers/tty/serial/8250/8250_pci.c if (!IS_ERR(new)) { new 5652 drivers/tty/serial/8250/8250_pci.c pci_set_drvdata(dev, new); new 228 drivers/tty/serial/ar933x_uart.c struct ktermios *new, new 238 drivers/tty/serial/ar933x_uart.c new->c_cflag &= ~CSIZE; new 239 drivers/tty/serial/ar933x_uart.c new->c_cflag |= CS8; new 242 drivers/tty/serial/ar933x_uart.c new->c_cflag &= ~CSTOPB; new 245 drivers/tty/serial/ar933x_uart.c if (new->c_cflag & PARENB) { new 246 drivers/tty/serial/ar933x_uart.c if (!(new->c_cflag & PARODD)) new 255 drivers/tty/serial/ar933x_uart.c new->c_cflag &= ~CMSPAR; new 257 drivers/tty/serial/ar933x_uart.c baud = uart_get_baud_rate(port, new, old, up->min_baud, up->max_baud); new 271 drivers/tty/serial/ar933x_uart.c uart_update_timeout(port, new->c_cflag, baud); new 276 drivers/tty/serial/ar933x_uart.c if ((new->c_cflag & CREAD) == 0) new 300 drivers/tty/serial/ar933x_uart.c if (tty_termios_baud_rate(new)) new 301 drivers/tty/serial/ar933x_uart.c tty_termios_encode_baud_rate(new, baud, baud); new 359 drivers/tty/serial/arc_uart.c arc_serial_set_termios(struct uart_port *port, struct ktermios *new, new 373 drivers/tty/serial/arc_uart.c baud = uart_get_baud_rate(port, new, old, 0, 460800); new 392 drivers/tty/serial/arc_uart.c new->c_cflag &= ~(CMSPAR|CRTSCTS|CSIZE); new 393 drivers/tty/serial/arc_uart.c new->c_cflag |= CS8; new 396 drivers/tty/serial/arc_uart.c tty_termios_copy_hw(new, old); new 399 drivers/tty/serial/arc_uart.c if (tty_termios_baud_rate(new)) new 400 drivers/tty/serial/arc_uart.c tty_termios_encode_baud_rate(new, baud, baud); new 402 drivers/tty/serial/arc_uart.c uart_update_timeout(port, new->c_cflag, baud); new 502 drivers/tty/serial/bcm63xx_uart.c struct ktermios *new, new 523 drivers/tty/serial/bcm63xx_uart.c switch (new->c_cflag & CSIZE) { new 539 drivers/tty/serial/bcm63xx_uart.c if (new->c_cflag & CSTOPB) new 545 drivers/tty/serial/bcm63xx_uart.c if (new->c_cflag & PARENB) new 548 drivers/tty/serial/bcm63xx_uart.c if (new->c_cflag & PARODD) new 553 drivers/tty/serial/bcm63xx_uart.c baud = uart_get_baud_rate(port, new, old, 0, port->uartclk / 16); new 561 drivers/tty/serial/bcm63xx_uart.c if (UART_ENABLE_MS(port, new->c_cflag)) new 568 drivers/tty/serial/bcm63xx_uart.c if (new->c_iflag & INPCK) { new 572 drivers/tty/serial/bcm63xx_uart.c if (new->c_iflag & (IGNBRK | BRKINT)) new 576 drivers/tty/serial/bcm63xx_uart.c if (new->c_iflag & IGNPAR) new 578 drivers/tty/serial/bcm63xx_uart.c if (new->c_iflag & IGNBRK) new 580 drivers/tty/serial/bcm63xx_uart.c if (!(new->c_cflag & CREAD)) new 583 drivers/tty/serial/bcm63xx_uart.c uart_update_timeout(port, new->c_cflag, baud); new 348 drivers/tty/serial/efm32-uart.c struct ktermios *new, struct ktermios *old) new 357 drivers/tty/serial/efm32-uart.c new->c_cflag &= ~(CRTSCTS | CMSPAR); new 359 drivers/tty/serial/efm32-uart.c baud = uart_get_baud_rate(port, new, old, new 363 drivers/tty/serial/efm32-uart.c switch (new->c_cflag & CSIZE) { new 378 drivers/tty/serial/efm32-uart.c if (new->c_cflag & CSTOPB) new 384 drivers/tty/serial/efm32-uart.c if (new->c_cflag & PARENB) { new 385 drivers/tty/serial/efm32-uart.c if (new->c_cflag & PARODD) new 404 drivers/tty/serial/efm32-uart.c if (new->c_iflag & INPCK) new 407 drivers/tty/serial/efm32-uart.c if (new->c_iflag & (IGNBRK | BRKINT | PARMRK)) new 411 drivers/tty/serial/efm32-uart.c if (new->c_iflag & IGNPAR) new 414 drivers/tty/serial/efm32-uart.c if (new->c_iflag & IGNBRK) new 417 drivers/tty/serial/efm32-uart.c uart_update_timeout(port, new->c_cflag, baud); new 411 drivers/tty/serial/lantiq.c struct ktermios *new, struct ktermios *old) new 421 drivers/tty/serial/lantiq.c cflag = new->c_cflag; new 422 drivers/tty/serial/lantiq.c iflag = new->c_iflag; new 432 drivers/tty/serial/lantiq.c new->c_cflag &= ~ CSIZE; new 433 drivers/tty/serial/lantiq.c new->c_cflag |= CS8; new 479 drivers/tty/serial/lantiq.c baud = uart_get_baud_rate(port, new, old, 0, port->uartclk / 16); new 504 drivers/tty/serial/lantiq.c if (tty_termios_baud_rate(new)) new 505 drivers/tty/serial/lantiq.c tty_termios_encode_baud_rate(new, baud, baud); new 105 drivers/tty/serial/mpc52xx_uart.c struct ktermios *new, new 291 drivers/tty/serial/mpc52xx_uart.c struct ktermios *new, new 298 drivers/tty/serial/mpc52xx_uart.c baud = uart_get_baud_rate(port, new, old, new 309 drivers/tty/serial/mpc52xx_uart.c struct ktermios *new, new 318 drivers/tty/serial/mpc52xx_uart.c baud = uart_get_baud_rate(port, new, old, new 537 drivers/tty/serial/mpc52xx_uart.c struct ktermios *new, new 557 drivers/tty/serial/mpc52xx_uart.c baud = uart_get_baud_rate(port, new, old, new 887 drivers/tty/serial/mpc52xx_uart.c struct ktermios *new, new 898 drivers/tty/serial/mpc52xx_uart.c baud = uart_get_baud_rate(port, new, old, new 1174 drivers/tty/serial/mpc52xx_uart.c mpc52xx_uart_set_termios(struct uart_port *port, struct ktermios *new, new 1185 drivers/tty/serial/mpc52xx_uart.c switch (new->c_cflag & CSIZE) { new 1196 drivers/tty/serial/mpc52xx_uart.c if (new->c_cflag & PARENB) { new 1197 drivers/tty/serial/mpc52xx_uart.c if (new->c_cflag & CMSPAR) new 1201 drivers/tty/serial/mpc52xx_uart.c mr1 |= (new->c_cflag & PARODD) ? new 1209 drivers/tty/serial/mpc52xx_uart.c if (new->c_cflag & CSTOPB) new 1212 drivers/tty/serial/mpc52xx_uart.c mr2 |= ((new->c_cflag & CSIZE) == CS5) ? new 1216 drivers/tty/serial/mpc52xx_uart.c if (new->c_cflag & CRTSCTS) { new 1245 drivers/tty/serial/mpc52xx_uart.c baud = psc_ops->set_baudrate(port, new, old); new 1248 drivers/tty/serial/mpc52xx_uart.c uart_update_timeout(port, new->c_cflag, baud); new 1250 drivers/tty/serial/mpc52xx_uart.c if (UART_ENABLE_MS(port, new->c_cflag)) new 528 drivers/tty/serial/pic32_uart.c struct ktermios *new, new 542 drivers/tty/serial/pic32_uart.c if (new->c_cflag & CSTOPB) new 550 drivers/tty/serial/pic32_uart.c if (new->c_cflag & PARENB) { new 551 drivers/tty/serial/pic32_uart.c if (new->c_cflag & PARODD) { new 568 drivers/tty/serial/pic32_uart.c if ((new->c_cflag & CRTSCTS) && sport->hw_flow_ctrl) { new 587 drivers/tty/serial/pic32_uart.c new->c_cflag |= CS8; new 590 drivers/tty/serial/pic32_uart.c new->c_cflag &= ~CMSPAR; new 593 drivers/tty/serial/pic32_uart.c baud = uart_get_baud_rate(port, new, old, 0, port->uartclk / 16); new 596 drivers/tty/serial/pic32_uart.c uart_update_timeout(port, new->c_cflag, baud); new 598 drivers/tty/serial/pic32_uart.c if (tty_termios_baud_rate(new)) new 599 drivers/tty/serial/pic32_uart.c tty_termios_encode_baud_rate(new, baud, baud); new 375 drivers/tty/serial/rp2.c struct ktermios *new, new 382 drivers/tty/serial/rp2.c baud = uart_get_baud_rate(port, new, old, 0, port->uartclk / 16); new 385 drivers/tty/serial/rp2.c if (tty_termios_baud_rate(new)) new 386 drivers/tty/serial/rp2.c tty_termios_encode_baud_rate(new, baud, baud); new 391 drivers/tty/serial/rp2.c port->ignore_status_mask = (new->c_cflag & CREAD) ? 0 : RP2_DUMMY_READ; new 393 drivers/tty/serial/rp2.c __rp2_uart_set_termios(up, new->c_cflag, new->c_iflag, baud_div); new 394 drivers/tty/serial/rp2.c uart_update_timeout(port, new->c_cflag, baud); new 580 drivers/tty/serial/sh-sci.c u16 new, scr = serial_port_in(port, SCSCR); new 582 drivers/tty/serial/sh-sci.c new = scr | SCSCR_TDRQE; new 584 drivers/tty/serial/sh-sci.c new = scr & ~SCSCR_TDRQE; new 585 drivers/tty/serial/sh-sci.c if (new != scr) new 586 drivers/tty/serial/sh-sci.c serial_port_out(port, SCSCR, new); new 128 drivers/tty/serial/tegra-tcu.c struct ktermios *new, new 502 drivers/tty/serial/vr41xx_siu.c static void siu_set_termios(struct uart_port *port, struct ktermios *new, new 510 drivers/tty/serial/vr41xx_siu.c c_cflag = new->c_cflag; new 535 drivers/tty/serial/vr41xx_siu.c baud = uart_get_baud_rate(port, new, old, 0, port->uartclk/16); new 544 drivers/tty/serial/vr41xx_siu.c c_iflag = new->c_iflag; new 273 drivers/tty/tty_ioctl.c void tty_termios_copy_hw(struct ktermios *new, struct ktermios *old) new 277 drivers/tty/tty_ioctl.c new->c_cflag &= HUPCL | CREAD | CLOCAL; new 278 drivers/tty/tty_ioctl.c new->c_cflag |= old->c_cflag & ~(HUPCL | CREAD | CLOCAL); new 279 drivers/tty/tty_ioctl.c new->c_ispeed = old->c_ispeed; new 280 drivers/tty/tty_ioctl.c new->c_ospeed = old->c_ospeed; new 825 drivers/tty/vt/vt.c unsigned short new; new 829 drivers/tty/vt/vt.c new = old ^ vc->vc_complement_mask; new 830 drivers/tty/vt/vt.c scr_writew(new, p); new 834 drivers/tty/vt/vt.c vc->vc_sw->con_putc(vc, new, oldy, oldx); new 2930 drivers/tty/vt/vt.c int vt_kmsg_redirect(int new) new 2934 drivers/tty/vt/vt.c if (new != -1) new 2935 drivers/tty/vt/vt.c return xchg(&kmsg_con, new); new 111 drivers/tty/vt/vt_ioctl.c void vt_event_post(unsigned int event, unsigned int old, unsigned int new) new 129 drivers/tty/vt/vt_ioctl.c ve->event.newev = new + 1; new 771 drivers/usb/atm/usbatm.c struct usbatm_vcc_data *new = NULL; new 805 drivers/usb/atm/usbatm.c new = kzalloc(sizeof(struct usbatm_vcc_data), GFP_KERNEL); new 806 drivers/usb/atm/usbatm.c if (!new) { new 811 drivers/usb/atm/usbatm.c new->vcc = vcc; new 812 drivers/usb/atm/usbatm.c new->vpi = vpi; new 813 drivers/usb/atm/usbatm.c new->vci = vci; new 815 drivers/usb/atm/usbatm.c new->sarb = alloc_skb(usbatm_pdu_length(vcc->qos.rxtp.max_sdu), GFP_KERNEL); new 816 drivers/usb/atm/usbatm.c if (!new->sarb) { new 822 drivers/usb/atm/usbatm.c vcc->dev_data = new; new 825 drivers/usb/atm/usbatm.c instance->cached_vcc = new; new 828 drivers/usb/atm/usbatm.c list_add(&new->list, &instance->vcc_list); new 837 drivers/usb/atm/usbatm.c atm_dbg(instance, "%s: allocated vcc data 0x%p\n", __func__, new); new 842 drivers/usb/atm/usbatm.c kfree(new); new 967 drivers/usb/class/cdc-acm.c struct async_icount old, new; new 972 drivers/usb/class/cdc-acm.c new = acm->iocount; new 973 drivers/usb/class/cdc-acm.c acm->oldcount = new; new 977 drivers/usb/class/cdc-acm.c old.dsr != new.dsr) new 980 drivers/usb/class/cdc-acm.c old.dcd != new.dcd) new 983 drivers/usb/class/cdc-acm.c old.rng != new.rng) new 2061 drivers/usb/gadget/composite.c static void update_unchanged_dev_desc(struct usb_device_descriptor *new, new 2075 drivers/usb/gadget/composite.c idVendor = new->idVendor; new 2076 drivers/usb/gadget/composite.c idProduct = new->idProduct; new 2077 drivers/usb/gadget/composite.c bcdDevice = new->bcdDevice; new 2078 drivers/usb/gadget/composite.c iSerialNumber = new->iSerialNumber; new 2079 drivers/usb/gadget/composite.c iManufacturer = new->iManufacturer; new 2080 drivers/usb/gadget/composite.c iProduct = new->iProduct; new 2082 drivers/usb/gadget/composite.c *new = *old; new 2084 drivers/usb/gadget/composite.c new->idVendor = idVendor; new 2086 drivers/usb/gadget/composite.c new->idProduct = idProduct; new 2088 drivers/usb/gadget/composite.c new->bcdDevice = bcdDevice; new 2090 drivers/usb/gadget/composite.c new->bcdDevice = cpu_to_le16(get_default_bcdDevice()); new 2092 drivers/usb/gadget/composite.c new->iSerialNumber = iSerialNumber; new 2094 drivers/usb/gadget/composite.c new->iManufacturer = iManufacturer; new 2096 drivers/usb/gadget/composite.c new->iProduct = iProduct; new 131 drivers/usb/usbip/vhci_hcd.c u32 new = new_status & bit; new 134 drivers/usb/usbip/vhci_hcd.c if (!prev && new) new 136 drivers/usb/usbip/vhci_hcd.c else if (prev && !new) new 141 drivers/usb/usbip/vhci_hcd.c if (prev || new) { new 51 drivers/usb/usbip/vudc.h unsigned new:1; new 120 drivers/usb/usbip/vudc_rx.c urb_p->new = 1; new 359 drivers/usb/usbip/vudc_transfer.c if (ep == &udc->ep[0] && urb_p->new) { new 361 drivers/usb/usbip/vudc_transfer.c urb_p->new = 0; new 153 drivers/vfio/vfio_iommu_type1.c static void vfio_link_dma(struct vfio_iommu *iommu, struct vfio_dma *new) new 162 drivers/vfio/vfio_iommu_type1.c if (new->iova + new->size <= dma->iova) new 168 drivers/vfio/vfio_iommu_type1.c rb_link_node(&new->node, parent, link); new 169 drivers/vfio/vfio_iommu_type1.c rb_insert_color(&new->node, &iommu->dma_list); new 199 drivers/vfio/vfio_iommu_type1.c struct vfio_pfn *new) new 209 drivers/vfio/vfio_iommu_type1.c if (new->iova < vpfn->iova) new 215 drivers/vfio/vfio_iommu_type1.c rb_link_node(&new->node, parent, link); new 216 drivers/vfio/vfio_iommu_type1.c rb_insert_color(&new->node, &dma->pfn_list); new 1419 drivers/vfio/vfio_iommu_type1.c struct device **old = data, *new; new 1421 drivers/vfio/vfio_iommu_type1.c new = vfio_mdev_get_iommu_device(dev); new 1422 drivers/vfio/vfio_iommu_type1.c if (!new || (*old && *old != new)) new 1425 drivers/vfio/vfio_iommu_type1.c *old = new; new 2374 drivers/vhost/vhost.c u16 old, new; new 2391 drivers/vhost/vhost.c new = (vq->last_used_idx += count); new 2396 drivers/vhost/vhost.c if (unlikely((u16)(new - vq->signalled_used) < (u16)(new - old))) new 2440 drivers/vhost/vhost.c __u16 old, new; new 2462 drivers/vhost/vhost.c new = vq->signalled_used = vq->last_used_idx; new 2472 drivers/vhost/vhost.c return vring_need_event(vhost16_to_cpu(vq, event), new, old); new 185 drivers/vhost/vringh.c struct kvec *new; new 193 drivers/vhost/vringh.c new = krealloc(iov->iov, new_num * sizeof(struct iovec), gfp); new 195 drivers/vhost/vringh.c new = kmalloc_array(new_num, sizeof(struct iovec), gfp); new 196 drivers/vhost/vringh.c if (new) { new 197 drivers/vhost/vringh.c memcpy(new, iov->iov, new 202 drivers/vhost/vringh.c if (!new) new 204 drivers/vhost/vringh.c iov->iov = new; new 2051 drivers/video/fbdev/atafb.c register unsigned char old, new; new 2064 drivers/video/fbdev/atafb.c new = st_mfp.tim_dt_b; new 2067 drivers/video/fbdev/atafb.c old = new; new 2068 drivers/video/fbdev/atafb.c new = st_mfp.tim_dt_b; new 2069 drivers/video/fbdev/atafb.c } while (old != new); new 592 drivers/virtio/virtio_ring.c u16 new, old; new 601 drivers/virtio/virtio_ring.c new = vq->split.avail_idx_shadow; new 610 drivers/virtio/virtio_ring.c new, old); new 1235 drivers/virtio/virtio_ring.c u16 new, old, off_wrap, flags, wrap_counter, event_idx; new 1254 drivers/virtio/virtio_ring.c new = vq->packed.next_avail_idx; new 1275 drivers/virtio/virtio_ring.c needs_kick = vring_need_event(event_idx, new, old); new 145 drivers/watchdog/machzwd.c static inline void zf_set_status(unsigned char new) new 147 drivers/watchdog/machzwd.c zf_writeb(STATUS, new); new 158 drivers/watchdog/machzwd.c static inline void zf_set_control(unsigned short new) new 160 drivers/watchdog/machzwd.c zf_writew(CONTROL, new); new 169 drivers/watchdog/machzwd.c static inline void zf_set_timer(unsigned short new, unsigned char n) new 173 drivers/watchdog/machzwd.c zf_writew(COUNTER_1, new); new 176 drivers/watchdog/machzwd.c zf_writeb(COUNTER_2, new > 0xff ? 0xff : new); new 234 drivers/xen/events/events_fifo.c event_word_t new, old, w; new 240 drivers/xen/events/events_fifo.c new = old & ~(1 << EVTCHN_FIFO_MASKED); new 241 drivers/xen/events/events_fifo.c w = sync_cmpxchg(word, old, new); new 260 drivers/xen/events/events_fifo.c event_word_t new, old, w; new 266 drivers/xen/events/events_fifo.c new = (w & ~((1 << EVTCHN_FIFO_LINKED) new 268 drivers/xen/events/events_fifo.c } while ((w = sync_cmpxchg(word, old, new)) != old); new 109 drivers/xen/evtchn.c struct rb_node **new = &(u->evtchns.rb_node), *parent = NULL; new 113 drivers/xen/evtchn.c while (*new) { new 116 drivers/xen/evtchn.c this = rb_entry(*new, struct user_evtchn, node); new 118 drivers/xen/evtchn.c parent = *new; new 120 drivers/xen/evtchn.c new = &((*new)->rb_left); new 122 drivers/xen/evtchn.c new = &((*new)->rb_right); new 128 drivers/xen/evtchn.c rb_link_node(&evtchn->node, parent, new); new 99 drivers/xen/xen-acpi-memhotplug.c struct acpi_memory_info *info, *new; new 116 drivers/xen/xen-acpi-memhotplug.c new = kzalloc(sizeof(struct acpi_memory_info), GFP_KERNEL); new 117 drivers/xen/xen-acpi-memhotplug.c if (!new) new 120 drivers/xen/xen-acpi-memhotplug.c INIT_LIST_HEAD(&new->list); new 121 drivers/xen/xen-acpi-memhotplug.c new->caching = address64.info.mem.caching; new 122 drivers/xen/xen-acpi-memhotplug.c new->write_protect = address64.info.mem.write_protect; new 123 drivers/xen/xen-acpi-memhotplug.c new->start_addr = address64.address.minimum; new 124 drivers/xen/xen-acpi-memhotplug.c new->length = address64.address.address_length; new 125 drivers/xen/xen-acpi-memhotplug.c list_add_tail(&new->list, &mem_device->res_list); new 484 drivers/xen/xen-pciback/pci_stub.c static void pcistub_device_id_add_list(struct pcistub_device_id *new, new 502 drivers/xen/xen-pciback/pci_stub.c new->domain = domain; new 503 drivers/xen/xen-pciback/pci_stub.c new->bus = bus; new 504 drivers/xen/xen-pciback/pci_stub.c new->devfn = devfn; new 505 drivers/xen/xen-pciback/pci_stub.c list_add_tail(&new->slot_list, &pcistub_device_ids); new 511 drivers/xen/xen-pciback/pci_stub.c kfree(new); new 896 drivers/xen/xen-scsiback.c struct v2p_entry *new; new 947 drivers/xen/xen-scsiback.c new = kmalloc(sizeof(struct v2p_entry), GFP_KERNEL); new 948 drivers/xen/xen-scsiback.c if (new == NULL) { new 963 drivers/xen/xen-scsiback.c kref_init(&new->kref); new 964 drivers/xen/xen-scsiback.c new->v = *v; new 965 drivers/xen/xen-scsiback.c new->tpg = tpg; new 966 drivers/xen/xen-scsiback.c new->lun = unpacked_lun; new 967 drivers/xen/xen-scsiback.c list_add_tail(&new->l, &info->v2p_entry_lists); new 977 drivers/xen/xen-scsiback.c kfree(new); new 143 fs/9p/v9fs.h struct super_block *sb, int new); new 149 fs/9p/v9fs.h struct super_block *sb, int new); new 488 fs/9p/vfs_inode.c int new) new 498 fs/9p/vfs_inode.c if (new) new 532 fs/9p/vfs_inode.c struct super_block *sb, int new) new 541 fs/9p/vfs_inode.c inode = v9fs_qid_iget(sb, &st->qid, st, new); new 101 fs/9p/vfs_inode_dotl.c int new) new 109 fs/9p/vfs_inode_dotl.c if (new) new 147 fs/9p/vfs_inode_dotl.c struct super_block *sb, int new) new 156 fs/9p/vfs_inode_dotl.c inode = v9fs_qid_iget_dotl(sb, &st->qid, fid, st, new); new 163 fs/adfs/dir_f.c if (dir->dirhead.startmasseq != dir->dirtail.new.endmasseq || new 164 fs/adfs/dir_f.c memcmp(&dir->dirhead.startname, &dir->dirtail.new.endname, 4)) new 171 fs/adfs/dir_f.c if (adfs_dir_checkbyte(dir) != dir->dirtail.new.dircheckbyte) new 340 fs/adfs/dir_f.c dir->parent_id = adfs_readval(dir->dirtail.new.dirparent, 3); new 400 fs/adfs/dir_f.c if (dir->dirhead.startmasseq != dir->dirtail.new.endmasseq || new 401 fs/adfs/dir_f.c memcmp(&dir->dirhead.startname, &dir->dirtail.new.endname, 4)) new 408 fs/adfs/dir_f.c if (adfs_dir_checkbyte(dir) != dir->dirtail.new.dircheckbyte) new 59 fs/adfs/dir_f.h } new; new 321 fs/affs/bitmap.c u32 old, new; new 325 fs/affs/bitmap.c new = old & mask; new 327 fs/affs/bitmap.c ((__be32 *)bh->b_data)[offset] = cpu_to_be32(new); new 30 fs/afs/callback.c struct afs_cb_interest *new; new 37 fs/afs/callback.c new = kzalloc(sizeof(struct afs_cb_interest), GFP_KERNEL); new 38 fs/afs/callback.c if (!new) { new 48 fs/afs/callback.c refcount_set(&new->usage, 1); new 49 fs/afs/callback.c new->sb = vnode->vfs_inode.i_sb; new 50 fs/afs/callback.c new->vid = vnode->volume->vid; new 51 fs/afs/callback.c new->server = afs_get_server(server, afs_server_trace_get_new_cbi); new 52 fs/afs/callback.c INIT_HLIST_NODE(&new->cb_vlink); new 75 fs/afs/callback.c new->vol_interest = vi; new 76 fs/afs/callback.c hlist_add_head(&new->cb_vlink, &vi->cb_interests); new 80 fs/afs/callback.c return new; new 93 fs/afs/callback.c struct afs_cb_interest *cbi, *vcbi, *new, *old; new 141 fs/afs/callback.c new = afs_create_interest(server, vnode); new 142 fs/afs/callback.c if (!new) new 147 fs/afs/callback.c entry->cb_interest = afs_get_cb_interest(new); new 148 fs/afs/callback.c cbi = new; new 149 fs/afs/callback.c new = NULL; new 154 fs/afs/callback.c afs_put_cb_interest(afs_v2net(vnode), new); new 19 fs/afs/dir_silly.c struct dentry *old, struct dentry *new, new 27 fs/afs/dir_silly.c _enter("%pd,%pd", old, new); new 40 fs/afs/dir_silly.c dvnode, new->d_name.name, new 63 fs/afs/dir_silly.c afs_edit_dir_add(dvnode, &new->d_name, new 17 fs/afs/flock.c static void afs_fl_copy_lock(struct file_lock *new, struct file_lock *fl); new 859 fs/afs/flock.c static void afs_fl_copy_lock(struct file_lock *new, struct file_lock *fl) new 865 fs/afs/flock.c new->fl_u.afs.debug_id = atomic_inc_return(&afs_file_lock_debug_id); new 868 fs/afs/flock.c trace_afs_flock_op(vnode, new, afs_flock_op_copy_lock); new 869 fs/afs/flock.c list_add(&new->fl_u.afs.link, &fl->fl_u.afs.link); new 146 fs/afs/security.c struct afs_permits *permits, *xpermits, *replacement, *zap, *new = NULL; new 223 fs/afs/security.c new = kzalloc(sizeof(struct afs_permits) + new 225 fs/afs/security.c if (!new) new 228 fs/afs/security.c refcount_set(&new->usage, 1); new 229 fs/afs/security.c new->nr_permits = size; new 234 fs/afs/security.c new->permits[j].key = key; new 235 fs/afs/security.c new->permits[j].access = caller_access; new 238 fs/afs/security.c new->permits[j].key = permits->permits[i].key; new 239 fs/afs/security.c new->permits[j].access = permits->permits[i].access; new 245 fs/afs/security.c new->permits[j].key = key; new 246 fs/afs/security.c new->permits[j].access = caller_access; new 249 fs/afs/security.c afs_hash_permits(new); new 254 fs/afs/security.c hash_for_each_possible(afs_permits_cache, xpermits, hash_node, new->h) { new 255 fs/afs/security.c if (xpermits->h != new->h || new 257 fs/afs/security.c xpermits->nr_permits != new->nr_permits || new 258 fs/afs/security.c memcmp(xpermits->permits, new->permits, new 259 fs/afs/security.c new->nr_permits * sizeof(struct afs_permit)) != 0) new 270 fs/afs/security.c for (i = 0; i < new->nr_permits; i++) new 271 fs/afs/security.c key_get(new->permits[i].key); new 272 fs/afs/security.c hash_add_rcu(afs_permits_cache, &new->hash_node, new->h); new 273 fs/afs/security.c replacement = new; new 274 fs/afs/security.c new = NULL; new 279 fs/afs/security.c kfree(new); new 101 fs/afs/server_list.c bool afs_annotate_server_list(struct afs_server_list *new, new 107 fs/afs/server_list.c if (old->nr_servers != new->nr_servers) new 111 fs/afs/server_list.c if (old->servers[i].server != new->servers[i].server) new 119 fs/afs/server_list.c for (j = 0; j < new->nr_servers; j++) { new 120 fs/afs/server_list.c if (new->servers[j].server == cur) { new 121 fs/afs/server_list.c new->preferred = j; new 131 fs/afs/server_list.c while (i < old->nr_servers && j < new->nr_servers) { new 132 fs/afs/server_list.c if (new->servers[j].server == old->servers[i].server) { new 135 fs/afs/server_list.c new->servers[j].cb_interest = cbi; new 143 fs/afs/server_list.c if (new->servers[j].server < old->servers[i].server) { new 222 fs/afs/volume.c struct afs_server_list *new, *old, *discard; new 249 fs/afs/volume.c new = afs_alloc_server_list(volume->cell, key, new 251 fs/afs/volume.c if (IS_ERR(new)) { new 252 fs/afs/volume.c ret = PTR_ERR(new); new 258 fs/afs/volume.c discard = new; new 260 fs/afs/volume.c if (afs_annotate_server_list(new, old)) { new 261 fs/afs/volume.c new->seq = volume->servers_seq + 1; new 262 fs/afs/volume.c volume->servers = new; new 377 fs/aio.c static int aio_migratepage(struct address_space *mapping, struct page *new, new 425 fs/aio.c get_page(new); new 427 fs/aio.c rc = migrate_page_move_mapping(mapping, new, old, 1); new 429 fs/aio.c put_page(new); new 438 fs/aio.c migrate_page_copy(new, old); new 440 fs/aio.c ctx->ring_pages[idx] = new; new 296 fs/autofs/root.c struct dentry *new; new 298 fs/autofs/root.c new = d_lookup(parent, &dentry->d_name); new 299 fs/autofs/root.c if (!new) new 301 fs/autofs/root.c ino = autofs_dentry_ino(new); new 304 fs/autofs/root.c path->dentry = new; new 313 fs/autofs/waitq.c struct dentry *new = NULL; new 329 fs/autofs/waitq.c new = d_lookup(parent, &dentry->d_name); new 330 fs/autofs/waitq.c if (new) new 331 fs/autofs/waitq.c dentry = new; new 339 fs/autofs/waitq.c if (new) new 340 fs/autofs/waitq.c dput(new); new 148 fs/bfs/dir.c struct dentry *new) new 155 fs/bfs/dir.c err = bfs_add_entry(dir, &new->d_name, inode->i_ino); new 164 fs/bfs/dir.c d_instantiate(new, inode); new 36 fs/bfs/file.c struct buffer_head *bh, *new; new 41 fs/bfs/file.c new = sb_getblk(sb, to); new 42 fs/bfs/file.c memcpy(new->b_data, bh->b_data, bh->b_size); new 43 fs/bfs/file.c mark_buffer_dirty(new); new 45 fs/bfs/file.c brelse(new); new 402 fs/btrfs/ctree.c struct rb_node **new; new 411 fs/btrfs/ctree.c new = &tm_root->rb_node; new 412 fs/btrfs/ctree.c while (*new) { new 413 fs/btrfs/ctree.c cur = rb_entry(*new, struct tree_mod_elem, node); new 414 fs/btrfs/ctree.c parent = *new; new 416 fs/btrfs/ctree.c new = &((*new)->rb_left); new 418 fs/btrfs/ctree.c new = &((*new)->rb_right); new 420 fs/btrfs/ctree.c new = &((*new)->rb_left); new 422 fs/btrfs/ctree.c new = &((*new)->rb_right); new 427 fs/btrfs/ctree.c rb_link_node(&tm->node, parent, new); new 2853 fs/btrfs/ctree.h void btrfs_merge_delalloc_extent(struct inode *inode, struct extent_state *new, new 2871 fs/btrfs/ctree.h struct btrfs_root *root, int *new, new 43 fs/btrfs/extent_io.c void btrfs_leak_debug_add(struct list_head *new, struct list_head *head) new 48 fs/btrfs/extent_io.c list_add(new, head); new 106 fs/btrfs/extent_io.c #define btrfs_leak_debug_add(new, head) do {} while (0) new 2881 fs/btrfs/extent_io.c struct bio *new; new 2884 fs/btrfs/extent_io.c new = bio_clone_fast(bio, GFP_NOFS, &btrfs_bioset); new 2885 fs/btrfs/extent_io.c btrfs_bio = btrfs_io_bio(new); new 2888 fs/btrfs/extent_io.c return new; new 4964 fs/btrfs/extent_io.c struct extent_buffer *new; new 4967 fs/btrfs/extent_io.c new = __alloc_extent_buffer(src->fs_info, src->start, src->len); new 4968 fs/btrfs/extent_io.c if (new == NULL) new 4974 fs/btrfs/extent_io.c btrfs_release_extent_buffer(new); new 4977 fs/btrfs/extent_io.c attach_extent_buffer_page(new, p); new 4980 fs/btrfs/extent_io.c new->pages[i] = p; new 4984 fs/btrfs/extent_io.c set_bit(EXTENT_BUFFER_UPTODATE, &new->bflags); new 4985 fs/btrfs/extent_io.c set_bit(EXTENT_BUFFER_UNMAPPED, &new->bflags); new 4987 fs/btrfs/extent_io.c return new; new 497 fs/btrfs/extent_map.c struct extent_map *new, new 504 fs/btrfs/extent_map.c rb_replace_node_cached(&cur->rb_node, &new->rb_node, &tree->map); new 507 fs/btrfs/extent_map.c setup_extent_mapping(tree, new, modified); new 92 fs/btrfs/extent_map.h struct extent_map *new, new 1776 fs/btrfs/inode.c void btrfs_merge_delalloc_extent(struct inode *inode, struct extent_state *new, new 1786 fs/btrfs/inode.c if (new->start > other->start) new 1787 fs/btrfs/inode.c new_size = new->end - other->start + 1; new 1789 fs/btrfs/inode.c new_size = other->end - new->start + 1; new 1819 fs/btrfs/inode.c old_size = new->end - new->start + 1; new 2459 fs/btrfs/inode.c struct new_sa_defrag_extent *new; new 2544 fs/btrfs/inode.c struct new_sa_defrag_extent *new = old->new; new 2545 fs/btrfs/inode.c struct btrfs_path *path = new->path; new 2550 fs/btrfs/inode.c struct inode *inode = new->inode; new 2651 fs/btrfs/inode.c backref_insert(&new->root, backref); new 2660 fs/btrfs/inode.c struct new_sa_defrag_extent *new) new 2662 fs/btrfs/inode.c struct btrfs_fs_info *fs_info = btrfs_sb(new->inode->i_sb); new 2666 fs/btrfs/inode.c new->path = path; new 2668 fs/btrfs/inode.c list_for_each_entry_safe(old, tmp, &new->head, list) { new 2683 fs/btrfs/inode.c if (list_empty(&new->head)) new 2691 fs/btrfs/inode.c struct new_sa_defrag_extent *new) new 2693 fs/btrfs/inode.c if (btrfs_file_extent_disk_bytenr(leaf, fi) != new->bytenr) new 2699 fs/btrfs/inode.c if (btrfs_file_extent_compression(leaf, fi) != new->compress_type) new 2725 fs/btrfs/inode.c struct new_sa_defrag_extent *new = old->new; new 2726 fs/btrfs/inode.c struct btrfs_fs_info *fs_info = btrfs_sb(new->inode->i_sb); new 2851 fs/btrfs/inode.c relink_is_mergable(leaf, fi, new)) { new 2876 fs/btrfs/inode.c btrfs_set_file_extent_disk_bytenr(leaf, item, new->bytenr); new 2877 fs/btrfs/inode.c btrfs_set_file_extent_disk_num_bytes(leaf, item, new->disk_len); new 2878 fs/btrfs/inode.c btrfs_set_file_extent_offset(leaf, item, start - new->file_pos); new 2880 fs/btrfs/inode.c btrfs_set_file_extent_ram_bytes(leaf, item, new->len); new 2883 fs/btrfs/inode.c btrfs_set_file_extent_compression(leaf, item, new->compress_type); new 2891 fs/btrfs/inode.c btrfs_init_generic_ref(&ref, BTRFS_ADD_DELAYED_REF, new->bytenr, new 2892 fs/btrfs/inode.c new->disk_len, 0); new 2894 fs/btrfs/inode.c new->file_pos); /* start - extent_offset */ new 2913 fs/btrfs/inode.c static void free_sa_defrag_extent(struct new_sa_defrag_extent *new) new 2917 fs/btrfs/inode.c if (!new) new 2920 fs/btrfs/inode.c list_for_each_entry_safe(old, tmp, &new->head, list) { new 2923 fs/btrfs/inode.c kfree(new); new 2926 fs/btrfs/inode.c static void relink_file_extents(struct new_sa_defrag_extent *new) new 2928 fs/btrfs/inode.c struct btrfs_fs_info *fs_info = btrfs_sb(new->inode->i_sb); new 2939 fs/btrfs/inode.c if (!record_extent_backrefs(path, new)) { new 2946 fs/btrfs/inode.c node = rb_first(&new->root); new 2949 fs/btrfs/inode.c rb_erase(node, &new->root); new 2968 fs/btrfs/inode.c free_sa_defrag_extent(new); new 2983 fs/btrfs/inode.c struct new_sa_defrag_extent *new; new 2986 fs/btrfs/inode.c new = kmalloc(sizeof(*new), GFP_NOFS); new 2987 fs/btrfs/inode.c if (!new) new 2990 fs/btrfs/inode.c new->inode = inode; new 2991 fs/btrfs/inode.c new->file_pos = ordered->file_offset; new 2992 fs/btrfs/inode.c new->len = ordered->len; new 2993 fs/btrfs/inode.c new->bytenr = ordered->start; new 2994 fs/btrfs/inode.c new->disk_len = ordered->disk_len; new 2995 fs/btrfs/inode.c new->compress_type = ordered->compress_type; new 2996 fs/btrfs/inode.c new->root = RB_ROOT; new 2997 fs/btrfs/inode.c INIT_LIST_HEAD(&new->head); new 3005 fs/btrfs/inode.c key.offset = new->file_pos; new 3042 fs/btrfs/inode.c if (key.offset >= new->file_pos + new->len) new 3048 fs/btrfs/inode.c if (key.offset + num_bytes < new->file_pos) new 3061 fs/btrfs/inode.c offset = max(new->file_pos, key.offset); new 3062 fs/btrfs/inode.c end = min(new->file_pos + new->len, key.offset + num_bytes); new 3068 fs/btrfs/inode.c old->new = new; new 3070 fs/btrfs/inode.c list_add_tail(&old->list, &new->head); new 3079 fs/btrfs/inode.c return new; new 3084 fs/btrfs/inode.c free_sa_defrag_extent(new); new 3115 fs/btrfs/inode.c struct new_sa_defrag_extent *new = NULL; new 3189 fs/btrfs/inode.c new = record_old_file_extents(inode, ordered_extent); new 3315 fs/btrfs/inode.c if (new) { new 3317 fs/btrfs/inode.c free_sa_defrag_extent(new); new 3320 fs/btrfs/inode.c relink_file_extents(new); new 5773 fs/btrfs/inode.c struct rb_node *new = &BTRFS_I(inode)->rb_node; new 5792 fs/btrfs/inode.c rb_replace_node(parent, new, &root->inode_tree); new 5798 fs/btrfs/inode.c rb_link_node(new, parent, p); new 5799 fs/btrfs/inode.c rb_insert_color(new, &root->inode_tree); new 5864 fs/btrfs/inode.c struct btrfs_root *root, int *new, new 5880 fs/btrfs/inode.c if (new) new 5881 fs/btrfs/inode.c *new = 1; new 5899 fs/btrfs/inode.c struct btrfs_root *root, int *new) new 5901 fs/btrfs/inode.c return btrfs_iget_path(s, location, root, new, NULL); new 2800 fs/btrfs/send.c struct recorded_ref *new; new 2802 fs/btrfs/send.c new = kmalloc(sizeof(*ref), GFP_KERNEL); new 2803 fs/btrfs/send.c if (!new) new 2806 fs/btrfs/send.c new->dir = ref->dir; new 2807 fs/btrfs/send.c new->dir_gen = ref->dir_gen; new 2808 fs/btrfs/send.c new->full_path = NULL; new 2809 fs/btrfs/send.c INIT_LIST_HEAD(&new->list); new 2810 fs/btrfs/send.c list_add_tail(&new->list, list); new 5888 fs/btrfs/volumes.c struct btrfs_bio_stripe *new = new 5893 fs/btrfs/volumes.c new->physical = old->physical; new 5894 fs/btrfs/volumes.c new->length = old->length; new 5895 fs/btrfs/volumes.c new->dev = dev_replace->tgtdev; new 45 fs/cachefiles/internal.h uint8_t new; /* T if object new */ new 543 fs/cachefiles/namei.c object->new = !inode; new 636 fs/cachefiles/namei.c if (!object->new) { new 672 fs/cachefiles/namei.c if (object->new) { new 704 fs/cachefiles/namei.c object->new = 0; new 18 fs/cachefiles/security.c struct cred *new; new 23 fs/cachefiles/security.c new = prepare_kernel_cred(current); new 24 fs/cachefiles/security.c if (!new) { new 30 fs/cachefiles/security.c ret = set_security_override_from_ctx(new, cache->secctx); new 32 fs/cachefiles/security.c put_cred(new); new 39 fs/cachefiles/security.c cache->cache_cred = new; new 79 fs/cachefiles/security.c struct cred *new; new 86 fs/cachefiles/security.c new = prepare_creds(); new 87 fs/cachefiles/security.c if (!new) new 94 fs/cachefiles/security.c ret = set_create_files_as(new, d_backing_inode(root)); new 96 fs/cachefiles/security.c abort_creds(new); new 103 fs/cachefiles/security.c cache->cache_cred = new; new 464 fs/ceph/caps.c struct ceph_cap *new) new 473 fs/ceph/caps.c if (new->mds < cap->mds) new 475 fs/ceph/caps.c else if (new->mds > cap->mds) new 481 fs/ceph/caps.c rb_link_node(&new->ci_node, parent, p); new 482 fs/ceph/caps.c rb_insert_color(&new->ci_node, &ci->i_caps); new 267 fs/ceph/quota.c bool ceph_quota_is_same_realm(struct inode *old, struct inode *new) new 282 fs/ceph/quota.c new_realm = get_quota_realm(mdsc, new, false); new 84 fs/ceph/snap.c struct ceph_snap_realm *new) new 93 fs/ceph/snap.c if (new->ino < r->ino) new 95 fs/ceph/snap.c else if (new->ino > r->ino) new 101 fs/ceph/snap.c rb_link_node(&new->node, parent, p); new 102 fs/ceph/snap.c rb_insert_color(&new->node, root); new 1048 fs/ceph/super.c struct ceph_fs_client *new = data; new 1049 fs/ceph/super.c struct ceph_mount_options *fsopt = new->mount_options; new 1050 fs/ceph/super.c struct ceph_options *opt = new->client->options; new 1173 fs/ceph/super.h extern bool ceph_quota_is_same_realm(struct inode *old, struct inode *new); new 443 fs/ceph/xattr.c int new = 0; new 488 fs/ceph/xattr.c new = 1; new 520 fs/ceph/xattr.c if (new) { new 377 fs/char_dev.c struct cdev *new = NULL; new 389 fs/char_dev.c new = container_of(kobj, struct cdev, kobj); new 395 fs/char_dev.c inode->i_cdev = p = new; new 397 fs/char_dev.c new = NULL; new 403 fs/char_dev.c cdev_put(new); new 3617 fs/cifs/connect.c struct cifs_sb_info *new = mnt_data->cifs_sb; new 3619 fs/cifs/connect.c unsigned int newflags = new->mnt_cifs_flags & CIFS_MOUNT_MASK; new 3634 fs/cifs/connect.c if (new->wsize && new->wsize < old->wsize) new 3637 fs/cifs/connect.c if (new->rsize && new->rsize < old->rsize) new 3640 fs/cifs/connect.c if (!uid_eq(old->mnt_uid, new->mnt_uid) || !gid_eq(old->mnt_gid, new->mnt_gid)) new 3643 fs/cifs/connect.c if (old->mnt_file_mode != new->mnt_file_mode || new 3644 fs/cifs/connect.c old->mnt_dir_mode != new->mnt_dir_mode) new 3647 fs/cifs/connect.c if (strcmp(old->local_nls->charset, new->local_nls->charset)) new 3650 fs/cifs/connect.c if (old->actimeo != new->actimeo) new 3660 fs/cifs/connect.c struct cifs_sb_info *new = mnt_data->cifs_sb; new 3663 fs/cifs/connect.c bool new_set = (new->mnt_cifs_flags & CIFS_MOUNT_USE_PREFIX_PATH) && new 3664 fs/cifs/connect.c new->prepath; new 3666 fs/cifs/connect.c if (old_set && new_set && !strcmp(new->prepath, old->prepath)) new 5357 fs/cifs/connect.c struct rb_node **new = &(root->rb_node), *parent = NULL; new 5360 fs/cifs/connect.c while (*new) { new 5361 fs/cifs/connect.c tlink = rb_entry(*new, struct tcon_link, tl_rbnode); new 5362 fs/cifs/connect.c parent = *new; new 5365 fs/cifs/connect.c new = &((*new)->rb_left); new 5367 fs/cifs/connect.c new = &((*new)->rb_right); new 5370 fs/cifs/connect.c rb_link_node(&new_tlink->tl_rbnode, parent, new); new 1082 fs/cifs/dfs_cache.c static int dup_vol(struct smb_vol *vol, struct smb_vol *new) new 1084 fs/cifs/dfs_cache.c memcpy(new, vol, sizeof(*new)); new 1087 fs/cifs/dfs_cache.c new->username = kstrndup(vol->username, strlen(vol->username), new 1089 fs/cifs/dfs_cache.c if (!new->username) new 1093 fs/cifs/dfs_cache.c new->password = kstrndup(vol->password, strlen(vol->password), new 1095 fs/cifs/dfs_cache.c if (!new->password) new 1100 fs/cifs/dfs_cache.c new->UNC = kstrndup(vol->UNC, strlen(vol->UNC), GFP_KERNEL); new 1101 fs/cifs/dfs_cache.c if (!new->UNC) new 1105 fs/cifs/dfs_cache.c new->domainname = kstrndup(vol->domainname, new 1107 fs/cifs/dfs_cache.c if (!new->domainname) new 1111 fs/cifs/dfs_cache.c new->iocharset = kstrndup(vol->iocharset, new 1113 fs/cifs/dfs_cache.c if (!new->iocharset) new 1118 fs/cifs/dfs_cache.c new->prepath = kstrndup(vol->prepath, strlen(vol->prepath), new 1120 fs/cifs/dfs_cache.c if (!new->prepath) new 1127 fs/cifs/dfs_cache.c kfree(new->iocharset); new 1129 fs/cifs/dfs_cache.c kfree(new->domainname); new 1131 fs/cifs/dfs_cache.c kfree(new->UNC); new 1133 fs/cifs/dfs_cache.c kzfree(new->password); new 1135 fs/cifs/dfs_cache.c kfree(new->username); new 1136 fs/cifs/dfs_cache.c kfree(new); new 166 fs/compat_ioctl.c void __user *new = compat_alloc_user_space(sizeof(sg_io_hdr_t) + new 168 fs/compat_ioctl.c if (new > top) new 171 fs/compat_ioctl.c sgio = new; new 218 fs/configfs/dir.c const unsigned char *new) new 225 fs/configfs/dir.c if (strcmp(existing, new)) new 549 fs/coredump.c static int umh_pipe_setup(struct subprocess_info *info, struct cred *new) new 2519 fs/dcache.c struct dentry *new = d_alloc(parent, name); new 2523 fs/dcache.c if (unlikely(!new)) new 2542 fs/dcache.c dput(new); new 2605 fs/dcache.c dput(new); new 2610 fs/dcache.c new->d_flags |= DCACHE_PAR_LOOKUP; new 2611 fs/dcache.c new->d_wait = wq; new 2612 fs/dcache.c hlist_bl_add_head_rcu(&new->d_u.d_in_lookup_hash, b); new 2614 fs/dcache.c return new; new 3011 fs/dcache.c struct dentry *new = __d_find_any_alias(inode); new 3012 fs/dcache.c if (unlikely(new)) { new 3016 fs/dcache.c if (unlikely(d_ancestor(new, dentry))) { new 3018 fs/dcache.c dput(new); new 3019 fs/dcache.c new = ERR_PTR(-ELOOP); new 3026 fs/dcache.c } else if (!IS_ROOT(new)) { new 3027 fs/dcache.c struct dentry *old_parent = dget(new->d_parent); new 3028 fs/dcache.c int err = __d_unalias(inode, dentry, new); new 3031 fs/dcache.c dput(new); new 3032 fs/dcache.c new = ERR_PTR(err); new 3036 fs/dcache.c __d_move(new, dentry, false); new 3040 fs/dcache.c return new; new 243 fs/dlm/config.c int new; new 500 fs/dlm/config.c nd->new = 1; /* set to 0 once it's been read by dlm_nodeid_list() */ new 806 fs/dlm/config.c node->new = nd->new; new 810 fs/dlm/config.c nd->new = 0; new 18 fs/dlm/config.h int new; new 1298 fs/dlm/lock.c static void lkb_add_ordered(struct list_head *new, struct list_head *head, new 1307 fs/dlm/lock.c __list_add(new, lkb->lkb_statequeue.prev, &lkb->lkb_statequeue); new 290 fs/dlm/member.c static void add_ordered_member(struct dlm_ls *ls, struct dlm_member *new) new 294 fs/dlm/member.c struct list_head *newlist = &new->list; new 299 fs/dlm/member.c if (new->nodeid < memb->nodeid) new 542 fs/dlm/member.c if (node && !node->new) new 595 fs/dlm/member.c int new; new 618 fs/dlm/member.c new = test_and_clear_bit(LSFL_RUNNING, &ls->ls_flags); new 637 fs/dlm/member.c if (new) { new 1134 fs/eventpoll.c static inline bool list_add_tail_lockless(struct list_head *new, new 1145 fs/eventpoll.c if (cmpxchg(&new->next, new, head) != new) new 1156 fs/eventpoll.c prev = xchg(&head->prev, new); new 1163 fs/eventpoll.c prev->next = new; new 1164 fs/eventpoll.c new->prev = prev; new 1932 fs/exec.c void set_binfmt(struct linux_binfmt *new) new 1939 fs/exec.c mm->binfmt = new; new 1940 fs/exec.c if (new) new 1941 fs/exec.c __module_get(new->module); new 625 fs/ext2/inode.c u32 *bno, bool *new, bool *boundary, new 762 fs/ext2/inode.c *new = true; new 786 fs/ext2/inode.c bool new = false, boundary = false; new 790 fs/ext2/inode.c ret = ext2_get_blocks(inode, iblock, max_blocks, &bno, &new, &boundary, new 797 fs/ext2/inode.c if (new) new 813 fs/ext2/inode.c bool new = false, boundary = false; new 818 fs/ext2/inode.c &bno, &new, &boundary, flags & IOMAP_WRITE); new 838 fs/ext2/inode.c if (new) new 1262 fs/ext4/namei.c struct dx_entry *old = frame->at, *new = old + 1; new 1267 fs/ext4/namei.c memmove(new + 1, new, (char *)(entries + count) - (char *)(new)); new 1268 fs/ext4/namei.c dx_set_hash(new, hash); new 1269 fs/ext4/namei.c dx_set_block(new, block); new 3651 fs/ext4/namei.c struct ext4_renament new = { new 3662 fs/ext4/namei.c if (new.inode && new.inode->i_nlink == 0) { new 3663 fs/ext4/namei.c EXT4_ERROR_INODE(new.inode, new 3676 fs/ext4/namei.c retval = dquot_initialize(new.dir); new 3682 fs/ext4/namei.c if (new.inode) { new 3683 fs/ext4/namei.c retval = dquot_initialize(new.inode); new 3701 fs/ext4/namei.c new.bh = ext4_find_entry(new.dir, &new.dentry->d_name, new 3702 fs/ext4/namei.c &new.de, &new.inlined); new 3703 fs/ext4/namei.c if (IS_ERR(new.bh)) { new 3704 fs/ext4/namei.c retval = PTR_ERR(new.bh); new 3705 fs/ext4/namei.c new.bh = NULL; new 3708 fs/ext4/namei.c if (new.bh) { new 3709 fs/ext4/namei.c if (!new.inode) { new 3710 fs/ext4/namei.c brelse(new.bh); new 3711 fs/ext4/namei.c new.bh = NULL; new 3714 fs/ext4/namei.c if (new.inode && !test_opt(new.dir->i_sb, NO_AUTO_DA_ALLOC)) new 3735 fs/ext4/namei.c if (IS_DIRSYNC(old.dir) || IS_DIRSYNC(new.dir)) new 3739 fs/ext4/namei.c if (new.inode) { new 3741 fs/ext4/namei.c if (!ext4_empty_dir(new.inode)) new 3745 fs/ext4/namei.c if (new.dir != old.dir && EXT4_DIR_LINK_MAX(new.dir)) new 3759 fs/ext4/namei.c force_reread = (new.dir->i_ino == old.dir->i_ino && new 3760 fs/ext4/namei.c ext4_test_inode_flag(new.dir, EXT4_INODE_INLINE_DATA)); new 3774 fs/ext4/namei.c if (!new.bh) { new 3775 fs/ext4/namei.c retval = ext4_add_entry(handle, new.dentry, old.inode); new 3779 fs/ext4/namei.c retval = ext4_setent(handle, &new, new 3785 fs/ext4/namei.c force_reread = !ext4_test_inode_flag(new.dir, new 3802 fs/ext4/namei.c if (new.inode) { new 3803 fs/ext4/namei.c ext4_dec_count(handle, new.inode); new 3804 fs/ext4/namei.c new.inode->i_ctime = current_time(new.inode); new 3809 fs/ext4/namei.c retval = ext4_rename_dir_finish(handle, &old, new.dir->i_ino); new 3814 fs/ext4/namei.c if (new.inode) { new 3818 fs/ext4/namei.c clear_nlink(new.inode); new 3820 fs/ext4/namei.c ext4_inc_count(handle, new.dir); new 3821 fs/ext4/namei.c ext4_update_dx_flag(new.dir); new 3822 fs/ext4/namei.c ext4_mark_inode_dirty(handle, new.dir); new 3826 fs/ext4/namei.c if (new.inode) { new 3827 fs/ext4/namei.c ext4_mark_inode_dirty(handle, new.inode); new 3828 fs/ext4/namei.c if (!new.inode->i_nlink) new 3829 fs/ext4/namei.c ext4_orphan_add(handle, new.inode); new 3836 fs/ext4/namei.c brelse(new.bh); new 3857 fs/ext4/namei.c struct ext4_renament new = { new 3877 fs/ext4/namei.c retval = dquot_initialize(new.dir); new 3895 fs/ext4/namei.c new.bh = ext4_find_entry(new.dir, &new.dentry->d_name, new 3896 fs/ext4/namei.c &new.de, &new.inlined); new 3897 fs/ext4/namei.c if (IS_ERR(new.bh)) { new 3898 fs/ext4/namei.c retval = PTR_ERR(new.bh); new 3899 fs/ext4/namei.c new.bh = NULL; new 3904 fs/ext4/namei.c if (!new.bh || le32_to_cpu(new.de->inode) != new.inode->i_ino) new 3916 fs/ext4/namei.c if (IS_DIRSYNC(old.dir) || IS_DIRSYNC(new.dir)) new 3925 fs/ext4/namei.c if (S_ISDIR(new.inode->i_mode)) { new 3926 fs/ext4/namei.c new.is_dir = true; new 3927 fs/ext4/namei.c retval = ext4_rename_dir_prepare(handle, &new); new 3936 fs/ext4/namei.c if (old.dir != new.dir && old.is_dir != new.is_dir) { new 3938 fs/ext4/namei.c new.dir_nlink_delta = -old.dir_nlink_delta; new 3941 fs/ext4/namei.c (new.dir_nlink_delta > 0 && EXT4_DIR_LINK_MAX(new.dir))) new 3945 fs/ext4/namei.c new_file_type = new.de->file_type; new 3946 fs/ext4/namei.c retval = ext4_setent(handle, &new, old.inode->i_ino, old.de->file_type); new 3950 fs/ext4/namei.c retval = ext4_setent(handle, &old, new.inode->i_ino, new_file_type); new 3960 fs/ext4/namei.c new.inode->i_ctime = ctime; new 3962 fs/ext4/namei.c ext4_mark_inode_dirty(handle, new.inode); new 3965 fs/ext4/namei.c retval = ext4_rename_dir_finish(handle, &old, new.dir->i_ino); new 3969 fs/ext4/namei.c if (new.dir_bh) { new 3970 fs/ext4/namei.c retval = ext4_rename_dir_finish(handle, &new, old.dir->i_ino); new 3975 fs/ext4/namei.c ext4_update_dir_count(handle, &new); new 3980 fs/ext4/namei.c brelse(new.dir_bh); new 3982 fs/ext4/namei.c brelse(new.bh); new 432 fs/f2fs/namei.c struct dentry *new; new 501 fs/f2fs/namei.c new = d_splice_alias(inode, dentry); new 502 fs/f2fs/namei.c err = PTR_ERR_OR_ZERO(new); new 504 fs/f2fs/namei.c return new; new 146 fs/f2fs/node.c struct nat_entry *new; new 149 fs/f2fs/node.c new = f2fs_kmem_cache_alloc(nat_entry_slab, GFP_F2FS_ZERO); new 151 fs/f2fs/node.c new = kmem_cache_alloc(nat_entry_slab, GFP_F2FS_ZERO); new 152 fs/f2fs/node.c if (new) { new 153 fs/f2fs/node.c nat_set_nid(new, nid); new 154 fs/f2fs/node.c nat_reset_flag(new); new 156 fs/f2fs/node.c return new; new 404 fs/f2fs/node.c struct nat_entry *new, *e; new 406 fs/f2fs/node.c new = __alloc_nat_entry(nid, false); new 407 fs/f2fs/node.c if (!new) new 413 fs/f2fs/node.c e = __init_nat_entry(nm_i, new, ne, false); new 420 fs/f2fs/node.c if (e != new) new 421 fs/f2fs/node.c __free_nat_entry(new); new 429 fs/f2fs/node.c struct nat_entry *new = __alloc_nat_entry(ni->nid, true); new 434 fs/f2fs/node.c e = __init_nat_entry(nm_i, new, NULL, true); new 447 fs/f2fs/node.c if (e != new) new 448 fs/f2fs/node.c __free_nat_entry(new); new 188 fs/f2fs/segment.c struct inmem_pages *new; new 194 fs/f2fs/segment.c new = f2fs_kmem_cache_alloc(inmem_entry_slab, GFP_NOFS); new 197 fs/f2fs/segment.c new->page = page; new 198 fs/f2fs/segment.c INIT_LIST_HEAD(&new->list); new 203 fs/f2fs/segment.c list_add_tail(&new->list, &F2FS_I(inode)->inmem_pages); new 3497 fs/f2fs/segment.c struct page *new; new 3524 fs/f2fs/segment.c new = f2fs_get_meta_page(sbi, blk_addr); new 3525 fs/f2fs/segment.c if (IS_ERR(new)) new 3526 fs/f2fs/segment.c return PTR_ERR(new); new 3527 fs/f2fs/segment.c sum = (struct f2fs_summary_block *)page_address(new); new 3561 fs/f2fs/segment.c f2fs_put_page(new, 1); new 118 fs/fat/cache.c struct fat_cache_id *new) new 124 fs/fat/cache.c if (p->fcluster == new->fcluster) { new 125 fs/fat/cache.c BUG_ON(p->dcluster != new->dcluster); new 126 fs/fat/cache.c if (new->nr_contig > p->nr_contig) new 127 fs/fat/cache.c p->nr_contig = new->nr_contig; new 134 fs/fat/cache.c static void fat_cache_add(struct inode *inode, struct fat_cache_id *new) new 138 fs/fat/cache.c if (new->fcluster == -1) /* dummy cache */ new 142 fs/fat/cache.c if (new->id != FAT_CACHE_VALID && new 143 fs/fat/cache.c new->id != MSDOS_I(inode)->cache_valid_id) new 146 fs/fat/cache.c cache = fat_cache_merge(inode, new); new 161 fs/fat/cache.c cache = fat_cache_merge(inode, new); new 172 fs/fat/cache.c cache->fcluster = new->fcluster; new 173 fs/fat/cache.c cache->dcluster = new->dcluster; new 174 fs/fat/cache.c cache->nr_contig = new->nr_contig; new 388 fs/fat/fat.h int new, int wait); new 154 fs/fat/fatent.c static void fat12_ent_put(struct fat_entry *fatent, int new) new 158 fs/fat/fatent.c if (new == FAT_ENT_EOF) new 159 fs/fat/fatent.c new = EOF_FAT12; new 163 fs/fat/fatent.c *ent12_p[0] = (new << 4) | (*ent12_p[0] & 0x0f); new 164 fs/fat/fatent.c *ent12_p[1] = new >> 4; new 166 fs/fat/fatent.c *ent12_p[0] = new & 0xff; new 167 fs/fat/fatent.c *ent12_p[1] = (*ent12_p[1] & 0xf0) | (new >> 8); new 176 fs/fat/fatent.c static void fat16_ent_put(struct fat_entry *fatent, int new) new 178 fs/fat/fatent.c if (new == FAT_ENT_EOF) new 179 fs/fat/fatent.c new = EOF_FAT16; new 181 fs/fat/fatent.c *fatent->u.ent16_p = cpu_to_le16(new); new 185 fs/fat/fatent.c static void fat32_ent_put(struct fat_entry *fatent, int new) new 187 fs/fat/fatent.c WARN_ON(new & 0xf0000000); new 188 fs/fat/fatent.c new |= le32_to_cpu(*fatent->u.ent32_p) & ~0x0fffffff; new 189 fs/fat/fatent.c *fatent->u.ent32_p = cpu_to_le32(new); new 409 fs/fat/fatent.c int new, int wait) new 415 fs/fat/fatent.c ops->ent_put(fatent, new); new 903 fs/fcntl.c void fasync_free(struct fasync_struct *new) new 905 fs/fcntl.c kmem_cache_free(fasync_cache, new); new 915 fs/fcntl.c struct fasync_struct *fasync_insert_entry(int fd, struct file *filp, struct fasync_struct **fapp, struct fasync_struct *new) new 931 fs/fcntl.c rwlock_init(&new->fa_lock); new 932 fs/fcntl.c new->magic = FASYNC_MAGIC; new 933 fs/fcntl.c new->fa_file = filp; new 934 fs/fcntl.c new->fa_fd = fd; new 935 fs/fcntl.c new->fa_next = *fapp; new 936 fs/fcntl.c rcu_assign_pointer(*fapp, new); new 951 fs/fcntl.c struct fasync_struct *new; new 953 fs/fcntl.c new = fasync_alloc(); new 954 fs/fcntl.c if (!new) new 964 fs/fcntl.c if (fasync_insert_entry(fd, filp, fapp, new)) { new 965 fs/fcntl.c fasync_free(new); new 50 fs/fs_struct.c static inline int replace_path(struct path *p, const struct path *old, const struct path *new) new 54 fs/fs_struct.c *p = *new; new 2200 fs/fuse/dev.c static int fuse_device_clone(struct fuse_conn *fc, struct file *new) new 2204 fs/fuse/dev.c if (new->private_data) new 2211 fs/fuse/dev.c new->private_data = fud; new 1332 fs/gfs2/bmap.c int gfs2_extent_map(struct inode *inode, u64 lblock, int *new, u64 *dblock, unsigned *extlen) new 1336 fs/gfs2/bmap.c int create = *new; new 1340 fs/gfs2/bmap.c BUG_ON(!new); new 1347 fs/gfs2/bmap.c *new = 1; new 1349 fs/gfs2/bmap.c *new = 0; new 53 fs/gfs2/bmap.h extern int gfs2_extent_map(struct inode *inode, u64 lblock, int *new, new 165 fs/gfs2/dir.c int new = 0; new 195 fs/gfs2/dir.c new = 1; new 196 fs/gfs2/dir.c error = gfs2_extent_map(&ip->i_inode, lblock, &new, new 205 fs/gfs2/dir.c if (amount == sdp->sd_jbsize || new) new 292 fs/gfs2/dir.c int new; new 299 fs/gfs2/dir.c new = 0; new 300 fs/gfs2/dir.c error = gfs2_extent_map(&ip->i_inode, lblock, &new, new 1009 fs/gfs2/dir.c struct gfs2_dirent *dent = NULL, *prev = NULL, *next = NULL, *new; new 1098 fs/gfs2/dir.c new = gfs2_dirent_split_alloc(inode, nbh, &str, ptr); new 1099 fs/gfs2/dir.c if (IS_ERR(new)) { new 1100 fs/gfs2/dir.c error = PTR_ERR(new); new 1104 fs/gfs2/dir.c new->de_inum = dent->de_inum; /* No endian worries */ new 1105 fs/gfs2/dir.c new->de_type = dent->de_type; /* No endian worries */ new 745 fs/gfs2/glock.c struct gfs2_glock *new) new 758 fs/gfs2/glock.c if (new) { new 760 fs/gfs2/glock.c &new->gl_node, ht_parms); new 884 fs/gfs2/log.c static void gfs2_merge_trans(struct gfs2_trans *old, struct gfs2_trans *new) new 888 fs/gfs2/log.c old->tr_num_buf_new += new->tr_num_buf_new; new 889 fs/gfs2/log.c old->tr_num_databuf_new += new->tr_num_databuf_new; new 890 fs/gfs2/log.c old->tr_num_buf_rm += new->tr_num_buf_rm; new 891 fs/gfs2/log.c old->tr_num_databuf_rm += new->tr_num_databuf_rm; new 892 fs/gfs2/log.c old->tr_num_revoke += new->tr_num_revoke; new 894 fs/gfs2/log.c list_splice_tail_init(&new->tr_databuf, &old->tr_databuf); new 895 fs/gfs2/log.c list_splice_tail_init(&new->tr_buf, &old->tr_buf); new 476 fs/gfs2/lops.c struct bio *new; new 478 fs/gfs2/lops.c new = bio_alloc(GFP_NOIO, nr_iovecs); new 479 fs/gfs2/lops.c bio_copy_dev(new, prev); new 480 fs/gfs2/lops.c new->bi_iter.bi_sector = bio_end_sector(prev); new 481 fs/gfs2/lops.c new->bi_opf = prev->bi_opf; new 482 fs/gfs2/lops.c new->bi_write_hint = prev->bi_write_hint; new 483 fs/gfs2/lops.c bio_chain(new, prev); new 485 fs/gfs2/lops.c return new; new 1368 fs/gfs2/quota.c int new = 0; new 1369 fs/gfs2/quota.c error = gfs2_extent_map(&ip->i_inode, x, &new, &dblock, &extlen); new 37 fs/gfs2/recovery.c int new = 0; new 42 fs/gfs2/recovery.c error = gfs2_extent_map(&ip->i_inode, blk, &new, &dblock, &extlen); new 812 fs/gfs2/xattr.c struct gfs2_ea_header *new = (struct gfs2_ea_header *)((char *)ea + new 820 fs/gfs2/xattr.c new->ea_rec_len = cpu_to_be32(new_size); new 821 fs/gfs2/xattr.c new->ea_flags = last; new 823 fs/gfs2/xattr.c return new; new 1147 fs/inode.c struct inode *new = alloc_inode(sb); new 1149 fs/inode.c if (new) { new 1150 fs/inode.c new->i_state = 0; new 1151 fs/inode.c inode = inode_insert5(new, hashval, test, set, data); new 1152 fs/inode.c if (unlikely(inode != new)) new 1153 fs/inode.c destroy_inode(new); new 17 fs/isofs/dir.c int isofs_name_translate(struct iso_directory_record *de, char *new, struct inode *inode) new 44 fs/isofs/dir.c new[i] = c; new 26 fs/jffs2/nodelist.c void jffs2_add_fd_to_list(struct jffs2_sb_info *c, struct jffs2_full_dirent *new, struct jffs2_full_dirent **list) new 30 fs/jffs2/nodelist.c dbg_dentlist("add dirent \"%s\", ino #%u\n", new->name, new->ino); new 32 fs/jffs2/nodelist.c while ((*prev) && (*prev)->nhash <= new->nhash) { new 33 fs/jffs2/nodelist.c if ((*prev)->nhash == new->nhash && !strcmp((*prev)->name, new->name)) { new 35 fs/jffs2/nodelist.c if (new->version < (*prev)->version) { new 38 fs/jffs2/nodelist.c jffs2_mark_node_obsolete(c, new->raw); new 39 fs/jffs2/nodelist.c jffs2_free_full_dirent(new); new 43 fs/jffs2/nodelist.c new->next = (*prev)->next; new 49 fs/jffs2/nodelist.c *prev = new; new 55 fs/jffs2/nodelist.c new->next = *prev; new 56 fs/jffs2/nodelist.c *prev = new; new 436 fs/jffs2/nodelist.c void jffs2_add_ino_cache (struct jffs2_sb_info *c, struct jffs2_inode_cache *new) new 441 fs/jffs2/nodelist.c if (!new->ino) new 442 fs/jffs2/nodelist.c new->ino = ++c->highest_ino; new 444 fs/jffs2/nodelist.c dbg_inocache("add %p (ino #%u)\n", new, new->ino); new 446 fs/jffs2/nodelist.c prev = &c->inocache_list[new->ino % c->inocache_hashsize]; new 448 fs/jffs2/nodelist.c while ((*prev) && (*prev)->ino < new->ino) { new 451 fs/jffs2/nodelist.c new->next = *prev; new 452 fs/jffs2/nodelist.c *prev = new; new 364 fs/jffs2/nodelist.h void jffs2_add_fd_to_list(struct jffs2_sb_info *c, struct jffs2_full_dirent *new, struct jffs2_full_dirent **list); new 367 fs/jffs2/nodelist.h void jffs2_add_ino_cache (struct jffs2_sb_info *c, struct jffs2_inode_cache *new); new 505 fs/jffs2/nodemgmt.c struct jffs2_raw_node_ref *new; new 530 fs/jffs2/nodemgmt.c new = jffs2_link_node_ref(c, jeb, ofs, len, ic); new 552 fs/jffs2/nodemgmt.c return new; new 87 fs/jffs2/wbuf.c struct jffs2_inodirty *new; new 95 fs/jffs2/wbuf.c new = kmalloc(sizeof(*new), GFP_KERNEL); new 96 fs/jffs2/wbuf.c if (!new) { new 102 fs/jffs2/wbuf.c new->ino = ino; new 103 fs/jffs2/wbuf.c new->next = c->wbuf_inodes; new 104 fs/jffs2/wbuf.c c->wbuf_inodes = new; new 577 fs/jfs/jfs_metapage.c unsigned long new) new 613 fs/jfs/jfs_metapage.c if (new && (PSIZE == PAGE_SIZE)) { new 642 fs/jfs/jfs_metapage.c if (!new) { new 671 fs/jfs/jfs_metapage.c if (new) { new 52 fs/jfs/jfs_metapage.h int absolute, unsigned long new); new 415 fs/kernfs/file.c struct mempolicy *new) new 429 fs/kernfs/file.c ret = of->vm_ops->set_policy(vma, new); new 99 fs/lockd/clntproc.c struct nlm_lockowner *res, *new = NULL; new 105 fs/lockd/clntproc.c new = kmalloc(sizeof(*new), GFP_KERNEL); new 108 fs/lockd/clntproc.c if (res == NULL && new != NULL) { new 109 fs/lockd/clntproc.c res = new; new 110 fs/lockd/clntproc.c refcount_set(&new->count, 1); new 111 fs/lockd/clntproc.c new->owner = owner; new 112 fs/lockd/clntproc.c new->pid = __nlm_alloc_pid(host); new 113 fs/lockd/clntproc.c new->host = nlm_get_host(host); new 114 fs/lockd/clntproc.c list_add(&new->list, &host->h_lockowners); new 115 fs/lockd/clntproc.c new = NULL; new 119 fs/lockd/clntproc.c kfree(new); new 457 fs/lockd/clntproc.c static void nlmclnt_locks_copy_lock(struct file_lock *new, struct file_lock *fl) new 460 fs/lockd/clntproc.c new->fl_u.nfs_fl.state = fl->fl_u.nfs_fl.state; new 461 fs/lockd/clntproc.c new->fl_u.nfs_fl.owner = nlmclnt_get_lockowner(fl->fl_u.nfs_fl.owner); new 462 fs/lockd/clntproc.c list_add_tail(&new->fl_u.nfs_fl.list, &fl->fl_u.nfs_fl.owner->host->h_granted); new 277 fs/lockd/mon.c struct nsm_handle *new; new 279 fs/lockd/mon.c new = kzalloc(sizeof(*new) + hostname_len + 1, GFP_KERNEL); new 280 fs/lockd/mon.c if (unlikely(new == NULL)) new 283 fs/lockd/mon.c refcount_set(&new->sm_count, 1); new 284 fs/lockd/mon.c new->sm_name = (char *)(new + 1); new 285 fs/lockd/mon.c memcpy(nsm_addr(new), sap, salen); new 286 fs/lockd/mon.c new->sm_addrlen = salen; new 287 fs/lockd/mon.c nsm_init_private(new); new 289 fs/lockd/mon.c if (rpc_ntop(nsm_addr(new), new->sm_addrbuf, new 290 fs/lockd/mon.c sizeof(new->sm_addrbuf)) == 0) new 291 fs/lockd/mon.c (void)snprintf(new->sm_addrbuf, sizeof(new->sm_addrbuf), new 293 fs/lockd/mon.c memcpy(new->sm_name, hostname, hostname_len); new 294 fs/lockd/mon.c new->sm_name[hostname_len] = '\0'; new 296 fs/lockd/mon.c return new; new 319 fs/lockd/mon.c struct nsm_handle *cached, *new = NULL; new 343 fs/lockd/mon.c kfree(new); new 351 fs/lockd/mon.c if (new != NULL) { new 352 fs/lockd/mon.c list_add(&new->sm_link, &ln->nsm_handles); new 355 fs/lockd/mon.c new->sm_name, new->sm_addrbuf); new 356 fs/lockd/mon.c return new; new 361 fs/lockd/mon.c new = nsm_create_handle(sap, salen, hostname, hostname_len); new 362 fs/lockd/mon.c if (unlikely(new == NULL)) new 365 fs/lockd/svclock.c struct nlm_lockowner *res, *new = NULL; new 372 fs/lockd/svclock.c new = kmalloc(sizeof(*res), GFP_KERNEL); new 375 fs/lockd/svclock.c if (res == NULL && new != NULL) { new 376 fs/lockd/svclock.c res = new; new 378 fs/lockd/svclock.c refcount_set(&new->count, 1); new 379 fs/lockd/svclock.c new->pid = pid; new 380 fs/lockd/svclock.c new->host = nlm_get_host(host); new 381 fs/lockd/svclock.c list_add(&new->list, &host->h_lockowners); new 382 fs/lockd/svclock.c new = NULL; new 387 fs/lockd/svclock.c kfree(new); new 398 fs/lockd/svclock.c static void nlmsvc_locks_copy_lock(struct file_lock *new, struct file_lock *fl) new 401 fs/lockd/svclock.c new->fl_owner = nlmsvc_get_lockowner(nlm_lo); new 409 fs/locks.c void locks_copy_conflock(struct file_lock *new, struct file_lock *fl) new 411 fs/locks.c new->fl_owner = fl->fl_owner; new 412 fs/locks.c new->fl_pid = fl->fl_pid; new 413 fs/locks.c new->fl_file = NULL; new 414 fs/locks.c new->fl_flags = fl->fl_flags; new 415 fs/locks.c new->fl_type = fl->fl_type; new 416 fs/locks.c new->fl_start = fl->fl_start; new 417 fs/locks.c new->fl_end = fl->fl_end; new 418 fs/locks.c new->fl_lmops = fl->fl_lmops; new 419 fs/locks.c new->fl_ops = NULL; new 428 fs/locks.c void locks_copy_lock(struct file_lock *new, struct file_lock *fl) new 431 fs/locks.c WARN_ON_ONCE(new->fl_ops); new 433 fs/locks.c locks_copy_conflock(new, fl); new 435 fs/locks.c new->fl_file = fl->fl_file; new 436 fs/locks.c new->fl_ops = fl->fl_ops; new 440 fs/locks.c fl->fl_ops->fl_copy_lock(new, fl); new 445 fs/locks.c static void locks_move_blocks(struct file_lock *new, struct file_lock *fl) new 457 fs/locks.c list_splice_init(&fl->fl_blocked_requests, &new->fl_blocked_requests); new 458 fs/locks.c list_for_each_entry(f, &new->fl_blocked_requests, fl_blocked_member) new 459 fs/locks.c f->fl_blocker = new; new 2111 fs/locks.c struct fasync_struct *new; new 2118 fs/locks.c new = fasync_alloc(); new 2119 fs/locks.c if (!new) { new 2123 fs/locks.c new->fa_fd = fd; new 2125 fs/locks.c error = vfs_setlease(filp, arg, &fl, (void **)&new); new 2128 fs/locks.c if (new) new 2129 fs/locks.c fasync_free(new); new 702 fs/namespace.c struct mountpoint *mp, *new = NULL; new 717 fs/namespace.c if (!new) new 718 fs/namespace.c new = kmalloc(sizeof(struct mountpoint), GFP_KERNEL); new 719 fs/namespace.c if (!new) new 737 fs/namespace.c new->m_dentry = dget(dentry); new 738 fs/namespace.c new->m_count = 1; new 739 fs/namespace.c hlist_add_head(&new->m_hash, mp_hash(dentry)); new 740 fs/namespace.c INIT_HLIST_HEAD(&new->m_list); new 743 fs/namespace.c mp = new; new 744 fs/namespace.c new = NULL; new 746 fs/namespace.c kfree(new); new 3219 fs/namespace.c struct mount *new; new 3240 fs/namespace.c new = copy_tree(old, old->mnt.mnt_root, copy_flags); new 3241 fs/namespace.c if (IS_ERR(new)) { new 3244 fs/namespace.c return ERR_CAST(new); new 3248 fs/namespace.c lock_mnt_tree(new); new 3251 fs/namespace.c new_ns->root = new; new 3252 fs/namespace.c list_add_tail(&new_ns->list, &new->mnt_list); new 3260 fs/namespace.c q = new; new 3275 fs/namespace.c q = next_mnt(q, new); new 3619 fs/namespace.c struct path new, old, root; new 3628 fs/namespace.c LOOKUP_FOLLOW | LOOKUP_DIRECTORY, &new); new 3637 fs/namespace.c error = security_sb_pivotroot(&old, &new); new 3648 fs/namespace.c new_mnt = real_mount(new.mnt); new 3662 fs/namespace.c if (d_unlinked(new.dentry)) new 3672 fs/namespace.c if (new.mnt->mnt_root != new.dentry) new 3677 fs/namespace.c if (!is_path_reachable(old_mnt, old.dentry, &new)) new 3680 fs/namespace.c if (!is_path_reachable(new_mnt, new.dentry, &root)) new 3699 fs/namespace.c chroot_fs_refs(&root, &new); new 3710 fs/namespace.c path_put(&new); new 181 fs/nfs/blocklayout/blocklayout.h struct pnfs_block_extent *new); new 137 fs/nfs/blocklayout/extent_tree.c struct pnfs_block_extent *new, bool merge_ok) new 146 fs/nfs/blocklayout/extent_tree.c if (new->be_f_offset < be->be_f_offset) { new 147 fs/nfs/blocklayout/extent_tree.c if (merge_ok && ext_can_merge(new, be)) { new 148 fs/nfs/blocklayout/extent_tree.c be->be_f_offset = new->be_f_offset; new 150 fs/nfs/blocklayout/extent_tree.c be->be_v_offset = new->be_v_offset; new 151 fs/nfs/blocklayout/extent_tree.c be->be_length += new->be_length; new 156 fs/nfs/blocklayout/extent_tree.c } else if (new->be_f_offset >= ext_f_end(be)) { new 157 fs/nfs/blocklayout/extent_tree.c if (merge_ok && ext_can_merge(be, new)) { new 158 fs/nfs/blocklayout/extent_tree.c be->be_length += new->be_length; new 168 fs/nfs/blocklayout/extent_tree.c rb_link_node(&new->be_node, parent, p); new 169 fs/nfs/blocklayout/extent_tree.c rb_insert_color(&new->be_node, root); new 172 fs/nfs/blocklayout/extent_tree.c nfs4_put_deviceid_node(new->be_device); new 173 fs/nfs/blocklayout/extent_tree.c kfree(new); new 201 fs/nfs/blocklayout/extent_tree.c struct pnfs_block_extent *new; new 203 fs/nfs/blocklayout/extent_tree.c new = kzalloc(sizeof(*new), GFP_ATOMIC); new 204 fs/nfs/blocklayout/extent_tree.c if (!new) new 209 fs/nfs/blocklayout/extent_tree.c new->be_f_offset = end; new 211 fs/nfs/blocklayout/extent_tree.c new->be_v_offset = new 214 fs/nfs/blocklayout/extent_tree.c new->be_length = len2; new 215 fs/nfs/blocklayout/extent_tree.c new->be_state = be->be_state; new 216 fs/nfs/blocklayout/extent_tree.c new->be_tag = be->be_tag; new 217 fs/nfs/blocklayout/extent_tree.c new->be_device = nfs4_get_deviceid(be->be_device); new 219 fs/nfs/blocklayout/extent_tree.c __ext_tree_insert(root, new, true); new 255 fs/nfs/blocklayout/extent_tree.c ext_tree_insert(struct pnfs_block_layout *bl, struct pnfs_block_extent *new) new 261 fs/nfs/blocklayout/extent_tree.c switch (new->be_state) { new 277 fs/nfs/blocklayout/extent_tree.c be = __ext_tree_search(root, new->be_f_offset); new 278 fs/nfs/blocklayout/extent_tree.c if (!be || be->be_f_offset >= ext_f_end(new)) { new 279 fs/nfs/blocklayout/extent_tree.c __ext_tree_insert(root, new, true); new 280 fs/nfs/blocklayout/extent_tree.c } else if (new->be_f_offset >= be->be_f_offset) { new 281 fs/nfs/blocklayout/extent_tree.c if (ext_f_end(new) <= ext_f_end(be)) { new 282 fs/nfs/blocklayout/extent_tree.c nfs4_put_deviceid_node(new->be_device); new 283 fs/nfs/blocklayout/extent_tree.c kfree(new); new 285 fs/nfs/blocklayout/extent_tree.c sector_t new_len = ext_f_end(new) - ext_f_end(be); new 286 fs/nfs/blocklayout/extent_tree.c sector_t diff = new->be_length - new_len; new 288 fs/nfs/blocklayout/extent_tree.c new->be_f_offset += diff; new 289 fs/nfs/blocklayout/extent_tree.c new->be_v_offset += diff; new 290 fs/nfs/blocklayout/extent_tree.c new->be_length = new_len; new 293 fs/nfs/blocklayout/extent_tree.c } else if (ext_f_end(new) <= ext_f_end(be)) { new 294 fs/nfs/blocklayout/extent_tree.c new->be_length = be->be_f_offset - new->be_f_offset; new 295 fs/nfs/blocklayout/extent_tree.c __ext_tree_insert(root, new, true); new 298 fs/nfs/blocklayout/extent_tree.c sector_t new_len = ext_f_end(new) - ext_f_end(be); new 299 fs/nfs/blocklayout/extent_tree.c sector_t diff = new->be_length - new_len; new 301 fs/nfs/blocklayout/extent_tree.c split = kmemdup(new, sizeof(*new), GFP_ATOMIC); new 308 fs/nfs/blocklayout/extent_tree.c split->be_device = nfs4_get_deviceid(new->be_device); new 311 fs/nfs/blocklayout/extent_tree.c new->be_f_offset += diff; new 312 fs/nfs/blocklayout/extent_tree.c new->be_v_offset += diff; new 313 fs/nfs/blocklayout/extent_tree.c new->be_length = new_len; new 383 fs/nfs/blocklayout/extent_tree.c struct pnfs_block_extent *new; new 386 fs/nfs/blocklayout/extent_tree.c new = kzalloc(sizeof(*new), GFP_ATOMIC); new 387 fs/nfs/blocklayout/extent_tree.c if (!new) new 392 fs/nfs/blocklayout/extent_tree.c new->be_f_offset = split; new 394 fs/nfs/blocklayout/extent_tree.c new->be_v_offset = be->be_v_offset + be->be_length; new 395 fs/nfs/blocklayout/extent_tree.c new->be_length = orig_len - be->be_length; new 396 fs/nfs/blocklayout/extent_tree.c new->be_state = be->be_state; new 397 fs/nfs/blocklayout/extent_tree.c new->be_tag = be->be_tag; new 398 fs/nfs/blocklayout/extent_tree.c new->be_device = nfs4_get_deviceid(be->be_device); new 400 fs/nfs/blocklayout/extent_tree.c __ext_tree_insert(root, new, false); new 216 fs/nfs/callback_proc.c const nfs4_stateid *new) new 225 fs/nfs/callback_proc.c if (!nfs4_stateid_match_other(&lo->plh_stateid, new)) new 228 fs/nfs/callback_proc.c newseq = be32_to_cpu(new->seqid); new 396 fs/nfs/client.c struct nfs_client *clp, *new = NULL; new 412 fs/nfs/client.c if (new) new 413 fs/nfs/client.c new->rpc_ops->free_client(new); new 418 fs/nfs/client.c if (new) { new 419 fs/nfs/client.c list_add_tail(&new->cl_share_link, new 422 fs/nfs/client.c new->cl_flags = cl_init->init_flags; new 423 fs/nfs/client.c return rpc_ops->init_client(new, cl_init); new 428 fs/nfs/client.c new = rpc_ops->alloc_client(cl_init); new 429 fs/nfs/client.c } while (!IS_ERR(new)); new 431 fs/nfs/client.c return new; new 76 fs/nfs/dns_resolve.c struct nfs_dns_ent *new; new 79 fs/nfs/dns_resolve.c new = container_of(cnew, struct nfs_dns_ent, h); new 82 fs/nfs/dns_resolve.c memcpy(&new->addr, &key->addr, key->addrlen); new 83 fs/nfs/dns_resolve.c new->addrlen = key->addrlen; new 89 fs/nfs/dns_resolve.c struct nfs_dns_ent *new; new 92 fs/nfs/dns_resolve.c new = container_of(cnew, struct nfs_dns_ent, h); new 95 fs/nfs/dns_resolve.c kfree(new->hostname); new 96 fs/nfs/dns_resolve.c new->hostname = kstrndup(key->hostname, key->namelen, GFP_KERNEL); new 97 fs/nfs/dns_resolve.c if (new->hostname) { new 98 fs/nfs/dns_resolve.c new->namelen = key->namelen; new 101 fs/nfs/dns_resolve.c new->namelen = 0; new 102 fs/nfs/dns_resolve.c new->addrlen = 0; new 217 fs/nfs/dns_resolve.c struct nfs_dns_ent *new, new 223 fs/nfs/dns_resolve.c &new->h, &key->h, new 309 fs/nfs/flexfilelayout/flexfilelayout.c ff_lseg_merge(struct pnfs_layout_segment *new, new 316 fs/nfs/flexfilelayout/flexfilelayout.c if (new->pls_range.iomode != old->pls_range.iomode) new 320 fs/nfs/flexfilelayout/flexfilelayout.c if (old_end < new->pls_range.offset) new 322 fs/nfs/flexfilelayout/flexfilelayout.c new_end = pnfs_calc_offset_end(new->pls_range.offset, new 323 fs/nfs/flexfilelayout/flexfilelayout.c new->pls_range.length); new 330 fs/nfs/flexfilelayout/flexfilelayout.c if (new->pls_range.offset < old->pls_range.offset) new 331 fs/nfs/flexfilelayout/flexfilelayout.c new->pls_range.offset = old->pls_range.offset; new 332 fs/nfs/flexfilelayout/flexfilelayout.c new->pls_range.length = pnfs_calc_offset_length(new->pls_range.offset, new 335 fs/nfs/flexfilelayout/flexfilelayout.c set_bit(NFS_LSEG_ROC, &new->pls_flags); new 877 fs/nfs/inode.c struct nfs_lock_context *res, *new = NULL; new 884 fs/nfs/inode.c new = kmalloc(sizeof(*new), GFP_KERNEL); new 885 fs/nfs/inode.c if (new == NULL) new 887 fs/nfs/inode.c nfs_init_lock_context(new); new 891 fs/nfs/inode.c new->open_context = get_nfs_open_context(ctx); new 892 fs/nfs/inode.c if (new->open_context) { new 893 fs/nfs/inode.c list_add_tail_rcu(&new->list, new 895 fs/nfs/inode.c res = new; new 896 fs/nfs/inode.c new = NULL; new 901 fs/nfs/inode.c kfree(new); new 722 fs/nfs/nfs2xdr.c const struct qstr *new = args->new_name; new 725 fs/nfs/nfs2xdr.c encode_diropargs(xdr, args->new_dir, new->name, new->len); new 1192 fs/nfs/nfs3xdr.c const struct qstr *new = args->new_name; new 1195 fs/nfs/nfs3xdr.c encode_diropargs3(xdr, args->new_dir, new->name, new->len); new 78 fs/nfs/nfs4client.c struct nfs4_ds_server *new) new 88 fs/nfs/nfs4client.c if (new) new 89 fs/nfs/nfs4client.c list_add_rcu(&new->list, &ds_clp->cl_ds_clients); new 90 fs/nfs/nfs4client.c dss = new; new 134 fs/nfs/nfs4client.c struct nfs4_ds_server *dss, *new; new 140 fs/nfs/nfs4client.c new = nfs4_alloc_ds_server(ds_clp, flavor); new 141 fs/nfs/nfs4client.c if (IS_ERR(new)) new 142 fs/nfs/nfs4client.c return ERR_CAST(new); new 143 fs/nfs/nfs4client.c dss = nfs4_add_ds_client(ds_clp, flavor, new); new 144 fs/nfs/nfs4client.c if (dss != new) new 145 fs/nfs/nfs4client.c nfs4_free_ds_server(new); new 492 fs/nfs/nfs4client.c static int nfs4_match_client(struct nfs_client *pos, struct nfs_client *new, new 497 fs/nfs/nfs4client.c if (pos->rpc_ops != new->rpc_ops) new 500 fs/nfs/nfs4client.c if (pos->cl_minorversion != new->cl_minorversion) new 524 fs/nfs/nfs4client.c if (pos->cl_clientid != new->cl_clientid) new 530 fs/nfs/nfs4client.c if (!nfs4_match_client_owner_id(pos, new)) new 549 fs/nfs/nfs4client.c int nfs40_walk_client_list(struct nfs_client *new, new 553 fs/nfs/nfs4client.c struct nfs_net *nn = net_generic(new->cl_net, nfs_net_id); new 556 fs/nfs/nfs4client.c .clientid = new->cl_clientid, new 557 fs/nfs/nfs4client.c .confirm = new->cl_confirm, new 564 fs/nfs/nfs4client.c if (pos == new) new 567 fs/nfs/nfs4client.c status = nfs4_match_client(pos, new, &prev, nn); new 579 fs/nfs/nfs4client.c if ((new != pos) && nfs4_same_verifier(&pos->cl_confirm, new 580 fs/nfs/nfs4client.c &new->cl_confirm)) new 599 fs/nfs/nfs4client.c nfs4_swap_callback_idents(pos, new); new 600 fs/nfs/nfs4client.c pos->cl_confirm = new->cl_confirm; new 715 fs/nfs/nfs4client.c int nfs41_walk_client_list(struct nfs_client *new, new 719 fs/nfs/nfs4client.c struct nfs_net *nn = net_generic(new->cl_net, nfs_net_id); new 726 fs/nfs/nfs4client.c if (pos == new) new 729 fs/nfs/nfs4client.c status = nfs4_match_client(pos, new, &prev, nn); new 741 fs/nfs/nfs4client.c new->cl_serverowner)) new 175 fs/nfs/nfs4namespace.c struct rpc_clnt *new; new 179 fs/nfs/nfs4namespace.c new = rpc_clone_client_set_auth(clnt, pflavor); new 180 fs/nfs/nfs4namespace.c if (IS_ERR(new)) new 187 fs/nfs/nfs4namespace.c cred = rpcauth_lookupcred(new->cl_auth, 0); new 189 fs/nfs/nfs4namespace.c rpc_shutdown_client(new); new 193 fs/nfs/nfs4namespace.c return new; new 217 fs/nfs/nfs4namespace.c struct rpc_clnt *new; new 228 fs/nfs/nfs4namespace.c new = ERR_PTR(err); new 232 fs/nfs/nfs4namespace.c new = nfs_find_best_sec(clnt, NFS_SERVER(inode), flavors); new 236 fs/nfs/nfs4namespace.c return new; new 428 fs/nfs/nfs4state.c nfs4_insert_state_owner_locked(struct nfs4_state_owner *new) new 430 fs/nfs/nfs4state.c struct nfs_server *server = new->so_server; new 439 fs/nfs/nfs4state.c cmp = cred_fscmp(new->so_cred, sp->so_cred); new 452 fs/nfs/nfs4state.c rb_link_node(&new->so_server_node, parent, p); new 453 fs/nfs/nfs4state.c rb_insert_color(&new->so_server_node, &server->state_owners); new 454 fs/nfs/nfs4state.c return new; new 577 fs/nfs/nfs4state.c struct nfs4_state_owner *sp, *new; new 584 fs/nfs/nfs4state.c new = nfs4_alloc_state_owner(server, cred, gfp_flags); new 585 fs/nfs/nfs4state.c if (new == NULL) new 588 fs/nfs/nfs4state.c sp = nfs4_insert_state_owner_locked(new); new 590 fs/nfs/nfs4state.c if (sp != new) new 591 fs/nfs/nfs4state.c nfs4_free_state_owner(new); new 720 fs/nfs/nfs4state.c struct nfs4_state *state, *new; new 728 fs/nfs/nfs4state.c new = nfs4_alloc_open_state(); new 732 fs/nfs/nfs4state.c if (state == NULL && new != NULL) { new 733 fs/nfs/nfs4state.c state = new; new 747 fs/nfs/nfs4state.c if (new) new 748 fs/nfs/nfs4state.c nfs4_free_open_state(new); new 901 fs/nfs/nfs4state.c struct nfs4_lock_state *lsp, *new = NULL; new 908 fs/nfs/nfs4state.c if (new != NULL) { new 909 fs/nfs/nfs4state.c list_add(&new->ls_locks, &state->lock_states); new 911 fs/nfs/nfs4state.c lsp = new; new 912 fs/nfs/nfs4state.c new = NULL; new 916 fs/nfs/nfs4state.c new = nfs4_alloc_lock_state(state, owner); new 917 fs/nfs/nfs4state.c if (new == NULL) new 921 fs/nfs/nfs4state.c if (new != NULL) new 922 fs/nfs/nfs4state.c nfs4_free_lock_state(state->owner->so_server, new); new 1073 fs/nfs/nfs4state.c struct nfs_seqid *new; new 1075 fs/nfs/nfs4state.c new = kmalloc(sizeof(*new), gfp_mask); new 1076 fs/nfs/nfs4state.c if (new == NULL) new 1078 fs/nfs/nfs4state.c new->sequence = counter; new 1079 fs/nfs/nfs4state.c INIT_LIST_HEAD(&new->list); new 1080 fs/nfs/nfs4state.c new->task = NULL; new 1081 fs/nfs/nfs4state.c return new; new 175 fs/nfs/nfs4super.c struct nfs_referral_count *p, *new; new 178 fs/nfs/nfs4super.c new = kmalloc(sizeof(*new), GFP_KERNEL); new 179 fs/nfs/nfs4super.c if (!new) new 181 fs/nfs/nfs4super.c new->task = current; new 182 fs/nfs/nfs4super.c new->referral_count = 1; new 193 fs/nfs/nfs4super.c list_add(&new->list, &nfs_referral_count_list); new 194 fs/nfs/nfs4super.c new = NULL; new 197 fs/nfs/nfs4super.c kfree(new); new 67 fs/nfs/pagelist.c unsigned int new = pos - hdr->io_start; new 69 fs/nfs/pagelist.c if (hdr->good_bytes > new) { new 70 fs/nfs/pagelist.c hdr->good_bytes = new; new 908 fs/nfs/pnfs.c pnfs_set_layout_stateid(struct pnfs_layout_hdr *lo, const nfs4_stateid *new, new 914 fs/nfs/pnfs.c newseq = be32_to_cpu(new->seqid); new 917 fs/nfs/pnfs.c nfs4_stateid_copy(&lo->plh_stateid, new); new 924 fs/nfs/pnfs.c nfs4_stateid_copy(&lo->plh_stateid, new); new 932 fs/nfs/pnfs.c new_barrier = be32_to_cpu(new->seqid); new 1673 fs/nfs/pnfs.c struct pnfs_layout_hdr *new = NULL; new 1680 fs/nfs/pnfs.c new = alloc_init_layout_hdr(ino, ctx, gfp_flags); new 1684 fs/nfs/pnfs.c nfsi->layout = new; new 1685 fs/nfs/pnfs.c return new; new 1686 fs/nfs/pnfs.c } else if (new != NULL) new 1687 fs/nfs/pnfs.c pnfs_free_layout_hdr(new); new 269 fs/nfs/pnfs.h const nfs4_stateid *new, new 191 fs/nfs/pnfs_dev.c struct nfs4_deviceid_node *d, *new; new 197 fs/nfs/pnfs_dev.c new = nfs4_get_device_info(server, id, cred, gfp_mask); new 198 fs/nfs/pnfs_dev.c if (!new) new 199 fs/nfs/pnfs_dev.c return new; new 205 fs/nfs/pnfs_dev.c server->pnfs_curr_ld->free_deviceid_node(new); new 208 fs/nfs/pnfs_dev.c hlist_add_head_rcu(&new->node, &nfs4_deviceid_cache[hash]); new 209 fs/nfs/pnfs_dev.c atomic_inc(&new->ref); new 212 fs/nfs/pnfs_dev.c return new; new 120 fs/nfs/read.c struct nfs_page *new; new 128 fs/nfs/read.c new = nfs_create_request(ctx, page, 0, len); new 129 fs/nfs/read.c if (IS_ERR(new)) { new 131 fs/nfs/read.c return PTR_ERR(new); new 138 fs/nfs/read.c if (!nfs_pageio_add_request(&pgio, new)) { new 139 fs/nfs/read.c nfs_list_remove_request(new); new 140 fs/nfs/read.c nfs_readpage_release(new, pgio.pg_error); new 293 fs/nfs/read.c unsigned int new = pos - hdr->io_start; new 295 fs/nfs/read.c if (hdr->good_bytes > new) { new 296 fs/nfs/read.c hdr->good_bytes = new; new 376 fs/nfs/read.c struct nfs_page *new; new 384 fs/nfs/read.c new = nfs_create_request(desc->ctx, page, 0, len); new 385 fs/nfs/read.c if (IS_ERR(new)) new 390 fs/nfs/read.c if (!nfs_pageio_add_request(desc->pgio, new)) { new 391 fs/nfs/read.c nfs_list_remove_request(new); new 393 fs/nfs/read.c nfs_readpage_release(new, error); new 398 fs/nfs/read.c error = PTR_ERR(new); new 2530 fs/nfs/super.c const struct nfs_server *new) new 2537 fs/nfs/super.c if (new->client && new->client->cl_cred) new 2538 fs/nfs/super.c newns = new->client->cl_cred->user_ns; new 25 fs/nfsd/auth.c struct cred *new; new 33 fs/nfsd/auth.c new = prepare_creds(); new 34 fs/nfsd/auth.c if (!new) new 37 fs/nfsd/auth.c new->fsuid = rqstp->rq_cred.cr_uid; new 38 fs/nfsd/auth.c new->fsgid = rqstp->rq_cred.cr_gid; new 43 fs/nfsd/auth.c new->fsuid = exp->ex_anon_uid; new 44 fs/nfsd/auth.c new->fsgid = exp->ex_anon_gid; new 49 fs/nfsd/auth.c if (uid_eq(new->fsuid, GLOBAL_ROOT_UID)) new 50 fs/nfsd/auth.c new->fsuid = exp->ex_anon_uid; new 51 fs/nfsd/auth.c if (gid_eq(new->fsgid, GLOBAL_ROOT_GID)) new 52 fs/nfsd/auth.c new->fsgid = exp->ex_anon_gid; new 71 fs/nfsd/auth.c if (uid_eq(new->fsuid, INVALID_UID)) new 72 fs/nfsd/auth.c new->fsuid = exp->ex_anon_uid; new 73 fs/nfsd/auth.c if (gid_eq(new->fsgid, INVALID_GID)) new 74 fs/nfsd/auth.c new->fsgid = exp->ex_anon_gid; new 76 fs/nfsd/auth.c set_groups(new, gi); new 79 fs/nfsd/auth.c if (!uid_eq(new->fsuid, GLOBAL_ROOT_UID)) new 80 fs/nfsd/auth.c new->cap_effective = cap_drop_nfsd_set(new->cap_effective); new 82 fs/nfsd/auth.c new->cap_effective = cap_raise_nfsd_set(new->cap_effective, new 83 fs/nfsd/auth.c new->cap_permitted); new 85 fs/nfsd/auth.c put_cred(override_creds(new)); new 86 fs/nfsd/auth.c put_cred(new); new 91 fs/nfsd/auth.c abort_creds(new); new 68 fs/nfsd/export.c static struct svc_expkey *svc_expkey_update(struct cache_detail *cd, struct svc_expkey *new, new 195 fs/nfsd/export.c struct svc_expkey *new = container_of(b, struct svc_expkey, h); new 197 fs/nfsd/export.c if (orig->ek_fsidtype != new->ek_fsidtype || new 198 fs/nfsd/export.c orig->ek_client != new->ek_client || new 199 fs/nfsd/export.c memcmp(orig->ek_fsid, new->ek_fsid, key_len(orig->ek_fsidtype)) != 0) new 207 fs/nfsd/export.c struct svc_expkey *new = container_of(cnew, struct svc_expkey, h); new 211 fs/nfsd/export.c new->ek_client = item->ek_client; new 212 fs/nfsd/export.c new->ek_fsidtype = item->ek_fsidtype; new 214 fs/nfsd/export.c memcpy(new->ek_fsid, item->ek_fsid, sizeof(new->ek_fsid)); new 220 fs/nfsd/export.c struct svc_expkey *new = container_of(cnew, struct svc_expkey, h); new 223 fs/nfsd/export.c new->ek_path = item->ek_path; new 289 fs/nfsd/export.c svc_expkey_update(struct cache_detail *cd, struct svc_expkey *new, new 293 fs/nfsd/export.c int hash = svc_expkey_hash(new); new 295 fs/nfsd/export.c ch = sunrpc_cache_update(cd, &new->h, &old->h, hash); new 352 fs/nfsd/export.c static struct svc_export *svc_export_update(struct svc_export *new, new 707 fs/nfsd/export.c struct svc_export *new = container_of(b, struct svc_export, h); new 708 fs/nfsd/export.c return orig->ex_client == new->ex_client && new 709 fs/nfsd/export.c path_equal(&orig->ex_path, &new->ex_path); new 714 fs/nfsd/export.c struct svc_export *new = container_of(cnew, struct svc_export, h); new 718 fs/nfsd/export.c new->ex_client = item->ex_client; new 719 fs/nfsd/export.c new->ex_path = item->ex_path; new 721 fs/nfsd/export.c new->ex_fslocs.locations = NULL; new 722 fs/nfsd/export.c new->ex_fslocs.locations_count = 0; new 723 fs/nfsd/export.c new->ex_fslocs.migrated = 0; new 724 fs/nfsd/export.c new->ex_layout_types = 0; new 725 fs/nfsd/export.c new->ex_uuid = NULL; new 726 fs/nfsd/export.c new->cd = item->cd; new 731 fs/nfsd/export.c struct svc_export *new = container_of(cnew, struct svc_export, h); new 735 fs/nfsd/export.c new->ex_flags = item->ex_flags; new 736 fs/nfsd/export.c new->ex_anon_uid = item->ex_anon_uid; new 737 fs/nfsd/export.c new->ex_anon_gid = item->ex_anon_gid; new 738 fs/nfsd/export.c new->ex_fsid = item->ex_fsid; new 739 fs/nfsd/export.c new->ex_devid_map = item->ex_devid_map; new 741 fs/nfsd/export.c new->ex_uuid = item->ex_uuid; new 743 fs/nfsd/export.c new->ex_fslocs.locations = item->ex_fslocs.locations; new 745 fs/nfsd/export.c new->ex_fslocs.locations_count = item->ex_fslocs.locations_count; new 747 fs/nfsd/export.c new->ex_fslocs.migrated = item->ex_fslocs.migrated; new 749 fs/nfsd/export.c new->ex_layout_types = item->ex_layout_types; new 750 fs/nfsd/export.c new->ex_nflavors = item->ex_nflavors; new 752 fs/nfsd/export.c new->ex_flavors[i] = item->ex_flavors[i]; new 804 fs/nfsd/export.c svc_export_update(struct svc_export *new, struct svc_export *old) new 809 fs/nfsd/export.c ch = sunrpc_cache_update(old->cd, &new->h, &old->h, hash); new 124 fs/nfsd/filecache.c struct nfsd_file_mark *nfm = NULL, *new; new 143 fs/nfsd/filecache.c new = kmem_cache_alloc(nfsd_file_mark_slab, GFP_KERNEL); new 144 fs/nfsd/filecache.c if (!new) new 146 fs/nfsd/filecache.c fsnotify_init_mark(&new->nfm_mark, nfsd_file_fsnotify_group); new 147 fs/nfsd/filecache.c new->nfm_mark.mask = FS_ATTRIB|FS_DELETE_SELF; new 148 fs/nfsd/filecache.c atomic_set(&new->nfm_ref, 1); new 150 fs/nfsd/filecache.c err = fsnotify_add_inode_mark(&new->nfm_mark, inode, 0); new 162 fs/nfsd/filecache.c nfm = new; new 164 fs/nfsd/filecache.c fsnotify_put_mark(&new->nfm_mark); new 791 fs/nfsd/filecache.c struct nfsd_file *nf, *new; new 811 fs/nfsd/filecache.c new = nfsd_file_alloc(inode, may_flags, hashval, net); new 812 fs/nfsd/filecache.c if (!new) { new 823 fs/nfsd/filecache.c nfsd_file_slab_free(&new->nf_rcu); new 867 fs/nfsd/filecache.c nf = new; new 79 fs/nfsd/nfs4idmap.c struct ent *new = container_of(cnew, struct ent, h); new 82 fs/nfsd/nfs4idmap.c new->id = itm->id; new 83 fs/nfsd/nfs4idmap.c new->type = itm->type; new 85 fs/nfsd/nfs4idmap.c strlcpy(new->name, itm->name, sizeof(new->name)); new 86 fs/nfsd/nfs4idmap.c strlcpy(new->authname, itm->authname, sizeof(new->authname)); new 277 fs/nfsd/nfs4idmap.c idtoname_update(struct cache_detail *cd, struct ent *new, struct ent *old) new 279 fs/nfsd/nfs4idmap.c struct cache_head *ch = sunrpc_cache_update(cd, &new->h, &old->h, new 280 fs/nfsd/nfs4idmap.c idtoname_hash(new)); new 435 fs/nfsd/nfs4idmap.c nametoid_update(struct cache_detail *cd, struct ent *new, struct ent *old) new 437 fs/nfsd/nfs4idmap.c struct cache_head *ch = sunrpc_cache_update(cd, &new->h, &old->h, new 438 fs/nfsd/nfs4idmap.c nametoid_hash(new)); new 368 fs/nfsd/nfs4layouts.c layouts_try_merge(struct nfsd4_layout_seg *lo, struct nfsd4_layout_seg *new) new 370 fs/nfsd/nfs4layouts.c if (lo->iomode != new->iomode) new 372 fs/nfsd/nfs4layouts.c if (layout_end(new) < lo->offset) new 374 fs/nfsd/nfs4layouts.c if (layout_end(lo) < new->offset) new 377 fs/nfsd/nfs4layouts.c lo->offset = min(lo->offset, new->offset); new 378 fs/nfsd/nfs4layouts.c layout_update_len(lo, max(layout_end(lo), layout_end(new))); new 406 fs/nfsd/nfs4layouts.c struct nfs4_layout *lp, *new = NULL; new 421 fs/nfsd/nfs4layouts.c new = kmem_cache_alloc(nfs4_layout_cache, GFP_KERNEL); new 422 fs/nfsd/nfs4layouts.c if (!new) new 424 fs/nfsd/nfs4layouts.c memcpy(&new->lo_seg, seg, sizeof(lp->lo_seg)); new 425 fs/nfsd/nfs4layouts.c new->lo_state = ls; new 438 fs/nfsd/nfs4layouts.c list_add_tail(&new->lo_perstate, &ls->ls_layouts); new 439 fs/nfsd/nfs4layouts.c new = NULL; new 445 fs/nfsd/nfs4layouts.c if (new) new 446 fs/nfsd/nfs4layouts.c kmem_cache_free(nfs4_layout_cache, new); new 75 fs/nfsd/nfs4recover.c struct cred *new; new 77 fs/nfsd/nfs4recover.c new = prepare_creds(); new 78 fs/nfsd/nfs4recover.c if (!new) new 81 fs/nfsd/nfs4recover.c new->fsuid = GLOBAL_ROOT_UID; new 82 fs/nfsd/nfs4recover.c new->fsgid = GLOBAL_ROOT_GID; new 83 fs/nfsd/nfs4recover.c *original_creds = override_creds(new); new 84 fs/nfsd/nfs4recover.c put_cred(new); new 1062 fs/nfsd/nfs4recover.c struct cld_upcall *new, *tmp; new 1065 fs/nfsd/nfs4recover.c new = kzalloc(sizeof(*new), GFP_KERNEL); new 1066 fs/nfsd/nfs4recover.c if (!new) new 1067 fs/nfsd/nfs4recover.c return new; new 1079 fs/nfsd/nfs4recover.c init_completion(&new->cu_done); new 1080 fs/nfsd/nfs4recover.c new->cu_u.cu_msg.cm_vers = nn->client_tracking_ops->version; new 1081 fs/nfsd/nfs4recover.c put_unaligned(cn->cn_xid++, &new->cu_u.cu_msg.cm_xid); new 1082 fs/nfsd/nfs4recover.c new->cu_net = cn; new 1083 fs/nfsd/nfs4recover.c list_add(&new->cu_list, &cn->cn_list); new 1086 fs/nfsd/nfs4recover.c dprintk("%s: allocated xid %u\n", __func__, new->cu_u.cu_msg.cm_xid); new 1088 fs/nfsd/nfs4recover.c return new; new 676 fs/nfsd/nfs4state.c find_or_hash_clnt_odstate(struct nfs4_file *fp, struct nfs4_clnt_odstate *new) new 681 fs/nfsd/nfs4state.c if (!new) new 684 fs/nfsd/nfs4state.c cl = new->co_client; new 693 fs/nfsd/nfs4state.c co = new; new 695 fs/nfsd/nfs4state.c hash_clnt_odstate_locked(new); new 812 fs/nfsd/nfs4state.c int new; /* index into 'set' */ new 828 fs/nfsd/nfs4state.c memset(bd->set[bd->new], 0, new 830 fs/nfsd/nfs4state.c bd->new = 1-bd->new; new 857 fs/nfsd/nfs4state.c __set_bit(hash&255, bd->set[bd->new]); new 858 fs/nfsd/nfs4state.c __set_bit((hash>>8)&255, bd->set[bd->new]); new 859 fs/nfsd/nfs4state.c __set_bit((hash>>16)&255, bd->set[bd->new]); new 1626 fs/nfsd/nfs4state.c struct nfsd4_session *new; new 1633 fs/nfsd/nfs4state.c new = kzalloc(sizeof(*new) + mem, GFP_KERNEL); new 1634 fs/nfsd/nfs4state.c if (!new) new 1638 fs/nfsd/nfs4state.c new->se_slots[i] = kzalloc(slotsize, GFP_KERNEL); new 1639 fs/nfsd/nfs4state.c if (!new->se_slots[i]) new 1643 fs/nfsd/nfs4state.c memcpy(&new->se_fchannel, fattrs, sizeof(struct nfsd4_channel_attrs)); new 1644 fs/nfsd/nfs4state.c memcpy(&new->se_bchannel, battrs, sizeof(struct nfsd4_channel_attrs)); new 1646 fs/nfsd/nfs4state.c return new; new 1649 fs/nfsd/nfs4state.c kfree(new->se_slots[i]); new 1650 fs/nfsd/nfs4state.c kfree(new); new 1764 fs/nfsd/nfs4state.c static void init_session(struct svc_rqst *rqstp, struct nfsd4_session *new, struct nfs4_client *clp, struct nfsd4_create_session *cses) new 1769 fs/nfsd/nfs4state.c new->se_client = clp; new 1770 fs/nfsd/nfs4state.c gen_sessionid(new); new 1772 fs/nfsd/nfs4state.c INIT_LIST_HEAD(&new->se_conns); new 1774 fs/nfsd/nfs4state.c new->se_cb_seq_nr = 1; new 1775 fs/nfsd/nfs4state.c new->se_flags = cses->flags; new 1776 fs/nfsd/nfs4state.c new->se_cb_prog = cses->callback_prog; new 1777 fs/nfsd/nfs4state.c new->se_cb_sec = cses->cb_sec; new 1778 fs/nfsd/nfs4state.c atomic_set(&new->se_ref, 0); new 1779 fs/nfsd/nfs4state.c idx = hash_sessionid(&new->se_sessionid); new 1780 fs/nfsd/nfs4state.c list_add(&new->se_hash, &nn->sessionid_hashtbl[idx]); new 1782 fs/nfsd/nfs4state.c list_add(&new->se_perclnt, &clp->cl_sessions); new 2636 fs/nfsd/nfs4state.c struct rb_node **new = &(root->rb_node), *parent = NULL; new 2639 fs/nfsd/nfs4state.c while (*new) { new 2640 fs/nfsd/nfs4state.c clp = rb_entry(*new, struct nfs4_client, cl_namenode); new 2641 fs/nfsd/nfs4state.c parent = *new; new 2644 fs/nfsd/nfs4state.c new = &((*new)->rb_left); new 2646 fs/nfsd/nfs4state.c new = &((*new)->rb_right); new 2649 fs/nfsd/nfs4state.c rb_link_node(&new_clp->cl_namenode, parent, new); new 2903 fs/nfsd/nfs4state.c nfsd4_set_ex_flags(struct nfs4_client *new, struct nfsd4_exchange_id *clid) new 2906 fs/nfsd/nfs4state.c new->cl_exchange_flags |= EXCHGID4_FLAG_USE_PNFS_MDS; new 2908 fs/nfsd/nfs4state.c new->cl_exchange_flags |= EXCHGID4_FLAG_USE_NON_PNFS; new 2912 fs/nfsd/nfs4state.c new->cl_exchange_flags |= EXCHGID4_FLAG_SUPP_MOVED_REFER; new 2915 fs/nfsd/nfs4state.c clid->flags = new->cl_exchange_flags; new 2961 fs/nfsd/nfs4state.c struct nfs4_client *conf, *new; new 2979 fs/nfsd/nfs4state.c new = create_client(exid->clname, rqstp, &verf); new 2980 fs/nfsd/nfs4state.c if (new == NULL) new 2982 fs/nfsd/nfs4state.c status = copy_impl_id(new, exid); new 3013 fs/nfsd/nfs4state.c if (!new->cl_cred.cr_principal && new 3014 fs/nfsd/nfs4state.c !new->cl_cred.cr_raw_principal) { new 3018 fs/nfsd/nfs4state.c new->cl_mach_cred = true; new 3089 fs/nfsd/nfs4state.c new->cl_minorversion = cstate->minorversion; new 3090 fs/nfsd/nfs4state.c new->cl_spo_must_allow.u.words[0] = exid->spo_must_allow[0]; new 3091 fs/nfsd/nfs4state.c new->cl_spo_must_allow.u.words[1] = exid->spo_must_allow[1]; new 3093 fs/nfsd/nfs4state.c add_to_unconfirmed(new); new 3094 fs/nfsd/nfs4state.c swap(new, conf); new 3109 fs/nfsd/nfs4state.c if (new) new 3110 fs/nfsd/nfs4state.c expire_client(new); new 3262 fs/nfsd/nfs4state.c struct nfsd4_session *new; new 3280 fs/nfsd/nfs4state.c new = alloc_session(&cr_ses->fore_channel, &cr_ses->back_channel); new 3281 fs/nfsd/nfs4state.c if (!new) new 3339 fs/nfsd/nfs4state.c init_session(rqstp, new, conf, cr_ses); new 3340 fs/nfsd/nfs4state.c nfsd4_get_session_locked(new); new 3342 fs/nfsd/nfs4state.c memcpy(cr_ses->sessionid.data, new->se_sessionid.data, new 3351 fs/nfsd/nfs4state.c nfsd4_init_conn(rqstp, conn, new); new 3352 fs/nfsd/nfs4state.c nfsd4_put_session(new); new 3362 fs/nfsd/nfs4state.c __free_session(new); new 3502 fs/nfsd/nfs4state.c static __be32 nfsd4_sequence_check_conn(struct nfsd4_conn *new, struct nfsd4_session *ses) new 3510 fs/nfsd/nfs4state.c c = __nfsd4_find_conn(new->cn_xprt, ses); new 3516 fs/nfsd/nfs4state.c __nfsd4_hash_conn(new, ses); new 3518 fs/nfsd/nfs4state.c ret = nfsd4_register_conn(new); new 3521 fs/nfsd/nfs4state.c nfsd4_conn_lost(&new->cn_xpt_user); new 3525 fs/nfsd/nfs4state.c free_conn(new); new 3808 fs/nfsd/nfs4state.c struct nfs4_client *conf, *new; new 3813 fs/nfsd/nfs4state.c new = create_client(clname, rqstp, &clverifier); new 3814 fs/nfsd/nfs4state.c if (new == NULL) new 3838 fs/nfsd/nfs4state.c copy_clid(new, conf); new 3839 fs/nfsd/nfs4state.c gen_confirm(new, nn); new 3842 fs/nfsd/nfs4state.c new->cl_minorversion = 0; new 3843 fs/nfsd/nfs4state.c gen_callback(new, setclid, rqstp); new 3844 fs/nfsd/nfs4state.c add_to_unconfirmed(new); new 3845 fs/nfsd/nfs4state.c setclid->se_clientid.cl_boot = new->cl_clientid.cl_boot; new 3846 fs/nfsd/nfs4state.c setclid->se_clientid.cl_id = new->cl_clientid.cl_id; new 3847 fs/nfsd/nfs4state.c memcpy(setclid->se_confirm.data, new->cl_confirm.data, sizeof(setclid->se_confirm.data)); new 3848 fs/nfsd/nfs4state.c new = NULL; new 3852 fs/nfsd/nfs4state.c if (new) new 3853 fs/nfsd/nfs4state.c free_client(new); new 4322 fs/nfsd/nfs4state.c find_or_add_file(struct nfs4_file *new, struct knfsd_fh *fh) new 4336 fs/nfsd/nfs4state.c nfsd4_init_file(fh, hashval, new); new 4337 fs/nfsd/nfs4state.c fp = new; new 6316 fs/nfsd/nfs4state.c bool *new) new 6323 fs/nfsd/nfs4state.c *new = false; new 6338 fs/nfsd/nfs4state.c *new = true; new 6368 fs/nfsd/nfs4state.c struct nfs4_ol_stateid **plst, bool *new) new 6393 fs/nfsd/nfs4state.c lst = find_or_create_lock_stateid(lo, fi, inode, ost, new); new 6426 fs/nfsd/nfs4state.c bool new = false; new 6470 fs/nfsd/nfs4state.c &lock_stp, &new); new 6610 fs/nfsd/nfs4state.c if (status && new) new 2169 fs/nfsd/nfs4xdr.c struct dentry **new; new 2170 fs/nfsd/nfs4xdr.c new = krealloc(components, new 2171 fs/nfsd/nfs4xdr.c sizeof(*new) * (ncomponents + 16), new 2173 fs/nfsd/nfs4xdr.c if (!new) new 2175 fs/nfsd/nfs4xdr.c components = new; new 228 fs/nfsd/nfsfh.c struct cred *new = prepare_creds(); new 229 fs/nfsd/nfsfh.c if (!new) { new 233 fs/nfsd/nfsfh.c new->cap_effective = new 234 fs/nfsd/nfsfh.c cap_raise_nfsd_set(new->cap_effective, new 235 fs/nfsd/nfsfh.c new->cap_permitted); new 236 fs/nfsd/nfsfh.c put_cred(override_creds(new)); new 237 fs/nfsd/nfsfh.c put_cred(new); new 690 fs/nfsd/nfssvc.c int new = nthreads[i] * NFSD_MAXSERVS / tot; new 691 fs/nfsd/nfssvc.c tot -= (nthreads[i] - new); new 692 fs/nfsd/nfssvc.c nthreads[i] = new; new 170 fs/nfsd/xdr4.h } new; new 185 fs/nfsd/xdr4.h #define lk_new_open_seqid v.new.open_seqid new 186 fs/nfsd/xdr4.h #define lk_new_open_stateid v.new.open_stateid new 187 fs/nfsd/xdr4.h #define lk_new_lock_seqid v.new.lock_seqid new 188 fs/nfsd/xdr4.h #define lk_new_clientid v.new.clientid new 189 fs/nfsd/xdr4.h #define lk_new_owner v.new.owner new 745 fs/nilfs2/the_nilfs.c struct nilfs_root *root, *new; new 752 fs/nilfs2/the_nilfs.c new = kzalloc(sizeof(*root), GFP_KERNEL); new 753 fs/nilfs2/the_nilfs.c if (!new) new 772 fs/nilfs2/the_nilfs.c kfree(new); new 777 fs/nilfs2/the_nilfs.c new->cno = cno; new 778 fs/nilfs2/the_nilfs.c new->ifile = NULL; new 779 fs/nilfs2/the_nilfs.c new->nilfs = nilfs; new 780 fs/nilfs2/the_nilfs.c refcount_set(&new->count, 1); new 781 fs/nilfs2/the_nilfs.c atomic64_set(&new->inodes_count, 0); new 782 fs/nilfs2/the_nilfs.c atomic64_set(&new->blocks_count, 0); new 784 fs/nilfs2/the_nilfs.c rb_link_node(&new->rb_node, parent, p); new 785 fs/nilfs2/the_nilfs.c rb_insert_color(&new->rb_node, &nilfs->ns_cptree); new 789 fs/nilfs2/the_nilfs.c err = nilfs_sysfs_create_snapshot_group(new); new 791 fs/nilfs2/the_nilfs.c kfree(new); new 792 fs/nilfs2/the_nilfs.c new = NULL; new 795 fs/nilfs2/the_nilfs.c return new; new 23 fs/notify/fanotify/fanotify.c struct fanotify_event *old, *new; new 27 fs/notify/fanotify/fanotify.c new = FANOTIFY_E(new_fsn); new 29 fs/notify/fanotify/fanotify.c if (old_fsn->objectid != new_fsn->objectid || old->pid != new->pid || new 30 fs/notify/fanotify/fanotify.c old->fh_type != new->fh_type || old->fh_len != new->fh_len) new 34 fs/notify/fanotify/fanotify.c return old->path.mnt == new->path.mnt && new 35 fs/notify/fanotify/fanotify.c old->path.dentry == new->path.dentry; new 45 fs/notify/fanotify/fanotify.c return (old->mask & FS_ISDIR) == (new->mask & FS_ISDIR) && new 46 fs/notify/fanotify/fanotify.c fanotify_fid_equal(&old->fid, &new->fid, old->fh_len); new 57 fs/notify/fanotify/fanotify.c struct fanotify_event *new; new 60 fs/notify/fanotify/fanotify.c new = FANOTIFY_E(event); new 67 fs/notify/fanotify/fanotify.c if (fanotify_is_perm_event(new->mask)) new 72 fs/notify/fanotify/fanotify.c FANOTIFY_E(test_event)->mask |= new->mask; new 35 fs/notify/inotify/inotify_fsnotify.c struct inotify_event_info *old, *new; new 38 fs/notify/inotify/inotify_fsnotify.c new = INOTIFY_E(new_fsn); new 41 fs/notify/inotify/inotify_fsnotify.c if ((old->mask == new->mask) && new 43 fs/notify/inotify/inotify_fsnotify.c (old->name_len == new->name_len) && new 44 fs/notify/inotify/inotify_fsnotify.c (!old->name_len || !strcmp(old->name, new->name))) new 606 fs/ocfs2/aops.c unsigned int to, int new) new 637 fs/ocfs2/aops.c if (new) new 668 fs/ocfs2/aops.c if (ret == 0 || !new) new 957 fs/ocfs2/aops.c int new) new 970 fs/ocfs2/aops.c new = new | ((i_size_read(inode) <= page_offset(page)) && new 977 fs/ocfs2/aops.c if (new) new 980 fs/ocfs2/aops.c new); new 983 fs/ocfs2/aops.c map_from, map_to, new); new 991 fs/ocfs2/aops.c if (new) { new 1001 fs/ocfs2/aops.c BUG_ON(!new); new 1007 fs/ocfs2/aops.c cluster_start, cluster_end, new); new 1024 fs/ocfs2/aops.c if (new && !PageUptodate(page)) new 1040 fs/ocfs2/aops.c unsigned user_len, int new, new 1057 fs/ocfs2/aops.c if (new) { new 1128 fs/ocfs2/aops.c u32 *phys, unsigned int new, new 1142 fs/ocfs2/aops.c if (new) { new 1223 fs/ocfs2/aops.c if (ret && new) new 1339 fs/ocfs2/aops.c struct ocfs2_unwritten_extent *ue = NULL, *new = NULL; new 1362 fs/ocfs2/aops.c if (new == NULL) { new 1364 fs/ocfs2/aops.c new = kmalloc(sizeof(struct ocfs2_unwritten_extent), new 1366 fs/ocfs2/aops.c if (new == NULL) { new 1373 fs/ocfs2/aops.c new->ue_cpos = desc->c_cpos; new 1374 fs/ocfs2/aops.c new->ue_phys = desc->c_phys; new 1376 fs/ocfs2/aops.c list_add_tail(&new->ue_ip_node, &oi->ip_unwritten_list); new 1377 fs/ocfs2/aops.c list_add_tail(&new->ue_node, &wc->w_unwritten_list); new 1379 fs/ocfs2/aops.c new = NULL; new 1383 fs/ocfs2/aops.c kfree(new); new 20 fs/ocfs2/aops.h unsigned int to, int new); new 1069 fs/ocfs2/ocfs2_trace.h unsigned int old, unsigned int new, unsigned int len, new 1071 fs/ocfs2/ocfs2_trace.h TP_ARGS(ino, cpos, old, new, len, ext_flags), new 1076 fs/ocfs2/ocfs2_trace.h __field(unsigned int, new) new 1084 fs/ocfs2/ocfs2_trace.h __entry->new = new; new 1089 fs/ocfs2/ocfs2_trace.h __entry->ino, __entry->cpos, __entry->old, __entry->new, new 224 fs/ocfs2/quota_global.c int err = 0, new = 0, ja_type; new 250 fs/ocfs2/quota_global.c new = 1; new 259 fs/ocfs2/quota_global.c !new) { new 273 fs/ocfs2/quota_global.c if (new) new 215 fs/ocfs2/refcounttree.c struct ocfs2_refcount_tree *new) new 217 fs/ocfs2/refcounttree.c u64 rf_blkno = new->rf_blkno; new 240 fs/ocfs2/refcounttree.c rb_link_node(&new->rf_node, parent, p); new 241 fs/ocfs2/refcounttree.c rb_insert_color(&new->rf_node, &osb->osb_rf_lock_tree); new 289 fs/ocfs2/refcounttree.c static inline void ocfs2_init_refcount_tree_ci(struct ocfs2_refcount_tree *new, new 292 fs/ocfs2/refcounttree.c ocfs2_metadata_cache_init(&new->rf_ci, &ocfs2_refcount_caching_ops); new 293 fs/ocfs2/refcounttree.c mutex_init(&new->rf_io_mutex); new 294 fs/ocfs2/refcounttree.c new->rf_sb = sb; new 295 fs/ocfs2/refcounttree.c spin_lock_init(&new->rf_lock); new 299 fs/ocfs2/refcounttree.c struct ocfs2_refcount_tree *new, new 302 fs/ocfs2/refcounttree.c init_rwsem(&new->rf_sem); new 303 fs/ocfs2/refcounttree.c ocfs2_refcount_lock_res_init(&new->rf_lockres, osb, new 310 fs/ocfs2/refcounttree.c struct ocfs2_refcount_tree *new; new 312 fs/ocfs2/refcounttree.c new = kzalloc(sizeof(struct ocfs2_refcount_tree), GFP_NOFS); new 313 fs/ocfs2/refcounttree.c if (!new) new 316 fs/ocfs2/refcounttree.c new->rf_blkno = rf_blkno; new 317 fs/ocfs2/refcounttree.c kref_init(&new->rf_getcnt); new 318 fs/ocfs2/refcounttree.c ocfs2_init_refcount_tree_ci(new, osb->sb); new 320 fs/ocfs2/refcounttree.c return new; new 327 fs/ocfs2/refcounttree.c struct ocfs2_refcount_tree *tree, *new = NULL; new 342 fs/ocfs2/refcounttree.c new = ocfs2_allocate_refcount_tree(osb, rf_blkno); new 343 fs/ocfs2/refcounttree.c if (!new) { new 356 fs/ocfs2/refcounttree.c ret = ocfs2_read_refcount_block(&new->rf_ci, rf_blkno, &ref_root_bh); new 359 fs/ocfs2/refcounttree.c ocfs2_metadata_cache_exit(&new->rf_ci); new 360 fs/ocfs2/refcounttree.c kfree(new); new 365 fs/ocfs2/refcounttree.c new->rf_generation = le32_to_cpu(ref_rb->rf_generation); new 366 fs/ocfs2/refcounttree.c ocfs2_init_refcount_tree_lock(osb, new, rf_blkno, new 367 fs/ocfs2/refcounttree.c new->rf_generation); new 368 fs/ocfs2/refcounttree.c ocfs2_metadata_cache_purge(&new->rf_ci); new 375 fs/ocfs2/refcounttree.c ocfs2_insert_refcount_tree(osb, new); new 377 fs/ocfs2/refcounttree.c tree = new; new 378 fs/ocfs2/refcounttree.c new = NULL; new 387 fs/ocfs2/refcounttree.c if (new) new 388 fs/ocfs2/refcounttree.c ocfs2_free_refcount_tree(new); new 3114 fs/ocfs2/refcounttree.c u32 new, u32 len, new 3122 fs/ocfs2/refcounttree.c cpos, old, new, len, ext_flags); new 3127 fs/ocfs2/refcounttree.c cpos, old, new, len); new 3135 fs/ocfs2/refcounttree.c cpos, new, len, ext_flags, new 307 fs/ocfs2/reservations.c struct ocfs2_alloc_reservation *new) new 316 fs/ocfs2/reservations.c trace_ocfs2_resv_insert(new->r_start, new->r_len); new 323 fs/ocfs2/reservations.c if (new->r_start < tmp->r_start) { new 330 fs/ocfs2/reservations.c BUG_ON(ocfs2_resv_end(new) >= tmp->r_start); new 331 fs/ocfs2/reservations.c } else if (new->r_start > ocfs2_resv_end(tmp)) { new 340 fs/ocfs2/reservations.c rb_link_node(&new->r_node, parent, p); new 341 fs/ocfs2/reservations.c rb_insert_color(&new->r_node, root); new 342 fs/ocfs2/reservations.c new->r_flags |= OCFS2_RESV_FLAG_INUSE; new 344 fs/ocfs2/reservations.c ocfs2_resv_mark_lru(resmap, new); new 436 fs/ocfs2/super.c struct inode *new = NULL; new 440 fs/ocfs2/super.c new = ocfs2_iget(osb, osb->root_blkno, OCFS2_FI_FLAG_SYSFILE, 0); new 441 fs/ocfs2/super.c if (IS_ERR(new)) { new 442 fs/ocfs2/super.c status = PTR_ERR(new); new 446 fs/ocfs2/super.c osb->root_inode = new; new 448 fs/ocfs2/super.c new = ocfs2_iget(osb, osb->system_dir_blkno, OCFS2_FI_FLAG_SYSFILE, 0); new 449 fs/ocfs2/super.c if (IS_ERR(new)) { new 450 fs/ocfs2/super.c status = PTR_ERR(new); new 454 fs/ocfs2/super.c osb->sys_root_inode = new; new 460 fs/ocfs2/super.c new = ocfs2_get_system_file_inode(osb, i, osb->slot_num); new 461 fs/ocfs2/super.c if (!new) { new 470 fs/ocfs2/super.c iput(new); new 481 fs/ocfs2/super.c struct inode *new = NULL; new 490 fs/ocfs2/super.c new = ocfs2_get_system_file_inode(osb, i, osb->slot_num); new 491 fs/ocfs2/super.c if (!new) { new 499 fs/ocfs2/super.c iput(new); new 306 fs/ocfs2/uptodate.c struct ocfs2_meta_cache_item *new) new 308 fs/ocfs2/uptodate.c sector_t block = new->c_block; new 334 fs/ocfs2/uptodate.c rb_link_node(&new->c_node, parent, p); new 335 fs/ocfs2/uptodate.c rb_insert_color(&new->c_node, &ci->ci_cache.ci_tree); new 391 fs/ocfs2/uptodate.c struct ocfs2_meta_cache_item *new = NULL; new 399 fs/ocfs2/uptodate.c new = kmem_cache_alloc(ocfs2_uptodate_cachep, GFP_NOFS); new 400 fs/ocfs2/uptodate.c if (!new) { new 404 fs/ocfs2/uptodate.c new->c_block = block; new 433 fs/ocfs2/uptodate.c __ocfs2_insert_cache_tree(ci, new); new 436 fs/ocfs2/uptodate.c new = NULL; new 438 fs/ocfs2/uptodate.c if (new) new 439 fs/ocfs2/uptodate.c kmem_cache_free(ocfs2_uptodate_cachep, new); new 364 fs/ocfs2/xattr.c u64 xb_blkno, int new) new 379 fs/ocfs2/xattr.c if (new) new 134 fs/orangefs/dir.c struct orangefs_dir_part *part, *new; new 147 fs/orangefs/dir.c new = (void *)op->downcall.trailer_buf; new 148 fs/orangefs/dir.c new->next = NULL; new 149 fs/orangefs/dir.c new->len = op->downcall.trailer_size - new 152 fs/orangefs/dir.c od->part = new; new 154 fs/orangefs/dir.c part->next = new; new 605 fs/orangefs/orangefs-debugfs.c char *new; new 614 fs/orangefs/orangefs-debugfs.c new = kzalloc(DEBUG_HELP_STRING_SIZE, GFP_KERNEL); new 615 fs/orangefs/orangefs-debugfs.c if (!new) { new 630 fs/orangefs/orangefs-debugfs.c strlcat(new, client_title, string_size); new 641 fs/orangefs/orangefs-debugfs.c kfree(new); new 646 fs/orangefs/orangefs-debugfs.c strlcat(new, "\t", string_size); new 647 fs/orangefs/orangefs-debugfs.c strlcat(new, cdm_array[i].keyword, string_size); new 648 fs/orangefs/orangefs-debugfs.c strlcat(new, "\n", string_size); new 652 fs/orangefs/orangefs-debugfs.c strlcat(new, "\n", string_size); new 653 fs/orangefs/orangefs-debugfs.c strlcat(new, kernel_title, string_size); new 656 fs/orangefs/orangefs-debugfs.c strlcat(new, "\t", string_size); new 657 fs/orangefs/orangefs-debugfs.c strlcat(new, s_kmod_keyword_mask_map[i].keyword, string_size); new 658 fs/orangefs/orangefs-debugfs.c result_size = strlcat(new, "\n", string_size); new 663 fs/orangefs/orangefs-debugfs.c kfree(new); new 668 fs/orangefs/orangefs-debugfs.c debug_help_string = new; new 672 fs/orangefs/orangefs-debugfs.c strlcat(debug_help_string, new, string_size); new 39 fs/overlayfs/copy_up.c int ovl_copy_xattr(struct dentry *old, struct dentry *new) new 47 fs/overlayfs/copy_up.c !(new->d_inode->i_opflags & IOP_XATTR)) new 90 fs/overlayfs/copy_up.c void *new; new 92 fs/overlayfs/copy_up.c new = krealloc(value, size, GFP_KERNEL); new 93 fs/overlayfs/copy_up.c if (!new) { new 97 fs/overlayfs/copy_up.c value = new; new 109 fs/overlayfs/copy_up.c error = vfs_setxattr(new, name, value, size, 0); new 119 fs/overlayfs/copy_up.c static int ovl_copy_up_data(struct path *old, struct path *new, loff_t len) new 135 fs/overlayfs/copy_up.c new_file = ovl_path_open(new, O_LARGEFILE | O_WRONLY); new 497 fs/overlayfs/copy_up.c struct cred *new; new 504 fs/overlayfs/copy_up.c cc->old = cc->new = NULL; new 505 fs/overlayfs/copy_up.c err = security_inode_copy_up(dentry, &cc->new); new 509 fs/overlayfs/copy_up.c if (cc->new) new 510 fs/overlayfs/copy_up.c cc->old = override_creds(cc->new); new 517 fs/overlayfs/copy_up.c if (cc->new) { new 519 fs/overlayfs/copy_up.c put_cred(cc->new); new 664 fs/overlayfs/dir.c struct dentry *new) new 677 fs/overlayfs/dir.c err = ovl_copy_up(new->d_parent); new 694 fs/overlayfs/dir.c err = ovl_create_or_link(new, inode, new 1021 fs/overlayfs/dir.c struct inode *newdir, struct dentry *new, new 1036 fs/overlayfs/dir.c bool new_is_dir = d_is_dir(new); new 1052 fs/overlayfs/dir.c if (!overwrite && !ovl_can_move(new)) new 1055 fs/overlayfs/dir.c if (overwrite && new_is_dir && !ovl_pure_upper(new)) { new 1056 fs/overlayfs/dir.c err = ovl_check_empty_dir(new, &list); new 1063 fs/overlayfs/dir.c if (!ovl_dentry_is_whiteout(new)) { new 1070 fs/overlayfs/dir.c } else if (is_dir && ovl_dentry_is_whiteout(new)) { new 1084 fs/overlayfs/dir.c err = ovl_copy_up(new->d_parent); new 1088 fs/overlayfs/dir.c err = ovl_copy_up(new); new 1091 fs/overlayfs/dir.c } else if (d_inode(new)) { new 1092 fs/overlayfs/dir.c err = ovl_nlink_start(new); new 1102 fs/overlayfs/dir.c opaquedir = ovl_clear_empty(new, &list); new 1111 fs/overlayfs/dir.c new_upperdir = ovl_dentry_upper(new->d_parent); new 1121 fs/overlayfs/dir.c err = ovl_set_impure(new->d_parent, new_upperdir); new 1125 fs/overlayfs/dir.c if (!overwrite && ovl_type_origin(new)) { new 1144 fs/overlayfs/dir.c newdentry = lookup_one_len(new->d_name.name, new_upperdir, new 1145 fs/overlayfs/dir.c new->d_name.len); new 1151 fs/overlayfs/dir.c new_opaque = ovl_dentry_is_opaque(new); new 1154 fs/overlayfs/dir.c if (d_inode(new) && ovl_dentry_upper(new)) { new 1159 fs/overlayfs/dir.c if (!ovl_matches_upper(new, newdentry)) new 1179 fs/overlayfs/dir.c else if (is_dir && !old_opaque && ovl_type_merge(new->d_parent)) new 1184 fs/overlayfs/dir.c if (!overwrite && ovl_type_merge_or_lower(new)) new 1185 fs/overlayfs/dir.c err = ovl_set_redirect(new, samedir); new 1188 fs/overlayfs/dir.c err = ovl_set_opaque_xerr(new, newdentry, -EXDEV); new 1200 fs/overlayfs/dir.c if (overwrite && d_inode(new)) { new 1202 fs/overlayfs/dir.c clear_nlink(d_inode(new)); new 1204 fs/overlayfs/dir.c drop_nlink(d_inode(new)); new 1208 fs/overlayfs/dir.c (!overwrite && ovl_type_origin(new))); new 1209 fs/overlayfs/dir.c ovl_dir_modified(new->d_parent, ovl_type_origin(old) || new 1210 fs/overlayfs/dir.c (d_inode(new) && ovl_type_origin(new))); new 1214 fs/overlayfs/dir.c if (d_inode(new) && ovl_dentry_upper(new)) new 1215 fs/overlayfs/dir.c ovl_copyattr(d_inode(newdentry), d_inode(new)); new 1226 fs/overlayfs/dir.c ovl_nlink_end(new); new 425 fs/overlayfs/overlayfs.h int ovl_copy_xattr(struct dentry *old, struct dentry *new); new 627 fs/pipe.c unsigned long old, unsigned long new) new 629 fs/pipe.c return atomic_long_add_return(new - old, &user->pipe_bufs); new 94 fs/proc/generic.c struct rb_node **new = &root->rb_node, *parent = NULL; new 97 fs/proc/generic.c while (*new) { new 98 fs/proc/generic.c struct proc_dir_entry *this = rb_entry(*new, new 103 fs/proc/generic.c parent = *new; new 105 fs/proc/generic.c new = &(*new)->rb_left; new 107 fs/proc/generic.c new = &(*new)->rb_right; new 113 fs/proc/generic.c rb_link_node(&de->subdir_node, parent, new); new 81 fs/proc/kcore.c void __init kclist_add(struct kcore_list *new, void *addr, size_t size, new 84 fs/proc/kcore.c new->addr = (unsigned long)addr; new 85 fs/proc/kcore.c new->size = size; new 86 fs/proc/kcore.c new->type = type; new 88 fs/proc/kcore.c list_add_tail(&new->list, &kclist_head); new 968 fs/proc/proc_sysctl.c struct ctl_dir *new; new 972 fs/proc/proc_sysctl.c new = kzalloc(sizeof(*new) + sizeof(struct ctl_node) + new 975 fs/proc/proc_sysctl.c if (!new) new 978 fs/proc/proc_sysctl.c node = (struct ctl_node *)(new + 1); new 985 fs/proc/proc_sysctl.c init_header(&new->header, set->dir.header.root, set, node, table); new 987 fs/proc/proc_sysctl.c return new; new 1006 fs/proc/proc_sysctl.c struct ctl_dir *subdir, *new = NULL; new 1017 fs/proc/proc_sysctl.c new = new_dir(set, name, namelen); new 1020 fs/proc/proc_sysctl.c if (!new) new 1031 fs/proc/proc_sysctl.c err = insert_header(dir, &new->header); new 1035 fs/proc/proc_sysctl.c subdir = new; new 1046 fs/proc/proc_sysctl.c if (new) new 1047 fs/proc/proc_sysctl.c drop_sysctl_table(&new->header); new 1452 fs/proc/proc_sysctl.c struct ctl_table *new; new 1459 fs/proc/proc_sysctl.c for (new = files, entry = table; entry->procname; entry++) { new 1462 fs/proc/proc_sysctl.c *new = *entry; new 1463 fs/proc/proc_sysctl.c new++; new 1088 fs/proc/vmcore.c struct vmcore *new; new 1108 fs/proc/vmcore.c new = get_new_element(); new 1109 fs/proc/vmcore.c if (!new) new 1111 fs/proc/vmcore.c new->paddr = start; new 1112 fs/proc/vmcore.c new->size = size; new 1113 fs/proc/vmcore.c list_add_tail(&new->list, vc_list); new 1131 fs/proc/vmcore.c struct vmcore *new; new 1151 fs/proc/vmcore.c new = get_new_element(); new 1152 fs/proc/vmcore.c if (!new) new 1154 fs/proc/vmcore.c new->paddr = start; new 1155 fs/proc/vmcore.c new->size = size; new 1156 fs/proc/vmcore.c list_add_tail(&new->list, vc_list); new 56 fs/pstore/ram_core.c int new; new 63 fs/pstore/ram_core.c new = old + a; new 64 fs/pstore/ram_core.c while (unlikely(new >= prz->buffer_size)) new 65 fs/pstore/ram_core.c new -= prz->buffer_size; new 66 fs/pstore/ram_core.c atomic_set(&prz->buffer->start, new); new 78 fs/pstore/ram_core.c size_t new; new 88 fs/pstore/ram_core.c new = old + a; new 89 fs/pstore/ram_core.c if (new > prz->buffer_size) new 90 fs/pstore/ram_core.c new = prz->buffer_size; new 91 fs/pstore/ram_core.c atomic_set(&prz->buffer->size, new); new 184 fs/reiserfs/file.c int new; new 205 fs/reiserfs/file.c new = buffer_new(bh); new 223 fs/reiserfs/file.c (new || page->index >= i_size_index)) { new 227 fs/reiserfs/item_ops.c static void start_new_sequence(__u32 * start, int *len, __u32 new) new 229 fs/reiserfs/item_ops.c *start = new; new 233 fs/reiserfs/item_ops.c static int sequence_finished(__u32 start, int *len, __u32 new) new 238 fs/reiserfs/item_ops.c if (start == 0 && new == 0) { new 242 fs/reiserfs/item_ops.c if (start != 0 && (start + *len) == new) { new 84 fs/signalfd.c struct signalfd_siginfo new; new 91 fs/signalfd.c memset(&new, 0, sizeof(new)); new 97 fs/signalfd.c new.ssi_signo = kinfo->si_signo; new 98 fs/signalfd.c new.ssi_errno = kinfo->si_errno; new 99 fs/signalfd.c new.ssi_code = kinfo->si_code; new 102 fs/signalfd.c new.ssi_pid = kinfo->si_pid; new 103 fs/signalfd.c new.ssi_uid = kinfo->si_uid; new 106 fs/signalfd.c new.ssi_tid = kinfo->si_tid; new 107 fs/signalfd.c new.ssi_overrun = kinfo->si_overrun; new 108 fs/signalfd.c new.ssi_ptr = (long) kinfo->si_ptr; new 109 fs/signalfd.c new.ssi_int = kinfo->si_int; new 112 fs/signalfd.c new.ssi_band = kinfo->si_band; new 113 fs/signalfd.c new.ssi_fd = kinfo->si_fd; new 125 fs/signalfd.c new.ssi_addr = (long) kinfo->si_addr; new 127 fs/signalfd.c new.ssi_trapno = kinfo->si_trapno; new 131 fs/signalfd.c new.ssi_addr = (long) kinfo->si_addr; new 133 fs/signalfd.c new.ssi_trapno = kinfo->si_trapno; new 135 fs/signalfd.c new.ssi_addr_lsb = (short) kinfo->si_addr_lsb; new 138 fs/signalfd.c new.ssi_pid = kinfo->si_pid; new 139 fs/signalfd.c new.ssi_uid = kinfo->si_uid; new 140 fs/signalfd.c new.ssi_status = kinfo->si_status; new 141 fs/signalfd.c new.ssi_utime = kinfo->si_utime; new 142 fs/signalfd.c new.ssi_stime = kinfo->si_stime; new 148 fs/signalfd.c new.ssi_pid = kinfo->si_pid; new 149 fs/signalfd.c new.ssi_uid = kinfo->si_uid; new 150 fs/signalfd.c new.ssi_ptr = (long) kinfo->si_ptr; new 151 fs/signalfd.c new.ssi_int = kinfo->si_int; new 154 fs/signalfd.c new.ssi_call_addr = (long) kinfo->si_call_addr; new 155 fs/signalfd.c new.ssi_syscall = kinfo->si_syscall; new 156 fs/signalfd.c new.ssi_arch = kinfo->si_arch; new 160 fs/signalfd.c if (copy_to_user(uinfo, &new, sizeof(struct signalfd_siginfo))) new 167 fs/sysfs/symlink.c const char *old, const char *new, const void *new_ns) new 192 fs/sysfs/symlink.c result = kernfs_rename_ns(kn, parent, new, new_ns); new 436 fs/timerfd.c const struct itimerspec64 *new, new 444 fs/timerfd.c !itimerspec64_valid(new)) new 500 fs/timerfd.c ret = timerfd_setup(ctx, flags, new); new 543 fs/timerfd.c struct itimerspec64 new, old; new 546 fs/timerfd.c if (get_itimerspec64(&new, utmr)) new 548 fs/timerfd.c ret = do_timerfd_settime(ufd, flags, &new, &old); new 571 fs/timerfd.c struct itimerspec64 new, old; new 574 fs/timerfd.c if (get_old_itimerspec32(&new, utmr)) new 576 fs/timerfd.c ret = do_timerfd_settime(ufd, flags, &new, &old); new 148 fs/ubifs/misc.h dev->new = cpu_to_le32(new_encode_dev(rdev)); new 149 fs/ubifs/misc.h return sizeof(dev->new); new 55 fs/ubifs/orphan.c orphan->new = 1; new 121 fs/ubifs/orphan.c if (o->new) { new 238 fs/ubifs/orphan.c ubifs_assert(c, orphan->new); new 240 fs/ubifs/orphan.c orphan->new = 0; new 446 fs/ubifs/orphan.c if (orphan->new) new 512 fs/ubifs/orphan.c ubifs_assert(c, !orphan->new); new 204 fs/ubifs/super.c if (ui->data_len == sizeof(dev->new)) new 205 fs/ubifs/super.c rdev = new_decode_dev(le32_to_cpu(dev->new)); new 476 fs/ubifs/ubifs-media.h __le32 new; new 932 fs/ubifs/ubifs.h unsigned new:1; new 421 fs/udf/inode.c int err, new; new 433 fs/udf/inode.c new = 0; new 443 fs/udf/inode.c phys = inode_getblk(inode, block, &err, &new); new 447 fs/udf/inode.c if (new) new 689 fs/udf/inode.c int *err, int *new) new 707 fs/udf/inode.c *new = 0; new 899 fs/udf/inode.c *new = 1; new 257 fs/ufs/inode.c int *new, struct page *locked_page) new 297 fs/ufs/inode.c if (new) new 298 fs/ufs/inode.c *new = 1; new 336 fs/ufs/inode.c int *new, struct page *locked_page) new 375 fs/ufs/inode.c if (new) new 376 fs/ufs/inode.c *new = 1; new 400 fs/ufs/inode.c int err = 0, new = 0; new 443 fs/ufs/inode.c &err, &new, bh_result->b_page); new 452 fs/ufs/inode.c fragment, &err, &new, bh_result->b_page); new 458 fs/ufs/inode.c if (new) new 83 fs/userfaultfd.c struct userfaultfd_ctx *new; new 616 fs/userfaultfd.c struct userfaultfd_ctx *new; new 618 fs/userfaultfd.c new = (struct userfaultfd_ctx *) new 621 fs/userfaultfd.c release_new_ctx = new; new 685 fs/userfaultfd.c ctx = fctx->new; new 712 fs/userfaultfd.c fctx->new = ctx; new 728 fs/userfaultfd.c ewq.msg.arg.reserved.reserved1 = (unsigned long)fctx->new; new 1020 fs/userfaultfd.c struct userfaultfd_ctx *new, new 1025 fs/userfaultfd.c fd = anon_inode_getfd("[userfaultfd]", &userfaultfd_fops, new, new 1026 fs/userfaultfd.c O_RDWR | (new->flags & UFFD_SHARED_FCNTL_FLAGS)); new 1002 fs/xfs/libxfs/xfs_alloc.c xfs_agblock_t new; new 1048 fs/xfs/libxfs/xfs_alloc.c *slena, &new); new 2203 fs/xfs/libxfs/xfs_alloc.c struct xfs_extent_free_item *new; /* new element */ new 2208 fs/xfs/libxfs/xfs_alloc.c new = kmem_zone_alloc(xfs_bmap_free_item_zone, 0); new 2209 fs/xfs/libxfs/xfs_alloc.c new->xefi_startblock = XFS_AGB_TO_FSB(mp, agno, agbno); new 2210 fs/xfs/libxfs/xfs_alloc.c new->xefi_blockcount = 1; new 2211 fs/xfs/libxfs/xfs_alloc.c new->xefi_oinfo = *oinfo; new 2215 fs/xfs/libxfs/xfs_alloc.c xfs_defer_add(tp, XFS_DEFER_OPS_TYPE_AGFL_FREE, &new->xefi_list); new 58 fs/xfs/libxfs/xfs_alloc_btree.c union xfs_btree_ptr *new, new 78 fs/xfs/libxfs/xfs_alloc_btree.c new->s = cpu_to_be32(bno); new 537 fs/xfs/libxfs/xfs_bmap.c struct xfs_extent_free_item *new; /* new element */ new 556 fs/xfs/libxfs/xfs_bmap.c new = kmem_zone_alloc(xfs_bmap_free_item_zone, 0); new 557 fs/xfs/libxfs/xfs_bmap.c new->xefi_startblock = bno; new 558 fs/xfs/libxfs/xfs_bmap.c new->xefi_blockcount = (xfs_extlen_t)len; new 560 fs/xfs/libxfs/xfs_bmap.c new->xefi_oinfo = *oinfo; new 562 fs/xfs/libxfs/xfs_bmap.c new->xefi_oinfo = XFS_RMAP_OINFO_SKIP_UPDATE; new 563 fs/xfs/libxfs/xfs_bmap.c new->xefi_skip_discard = skip_discard; new 567 fs/xfs/libxfs/xfs_bmap.c xfs_defer_add(tp, XFS_DEFER_OPS_TYPE_FREE, &new->xefi_list); new 1172 fs/xfs/libxfs/xfs_bmap.c struct xfs_bmbt_irec new; new 1256 fs/xfs/libxfs/xfs_bmap.c xfs_bmbt_disk_get_all(frp, &new); new 1257 fs/xfs/libxfs/xfs_bmap.c fa = xfs_bmap_validate_extent(ip, whichfork, &new); new 1265 fs/xfs/libxfs/xfs_bmap.c xfs_iext_insert(ip, &icur, &new, state); new 1533 fs/xfs/libxfs/xfs_bmap.c struct xfs_bmbt_irec *new = &bma->got; new 1556 fs/xfs/libxfs/xfs_bmap.c ASSERT(!isnullstartblock(new->br_startblock)); new 1570 fs/xfs/libxfs/xfs_bmap.c new_endoff = new->br_startoff + new->br_blockcount; new 1572 fs/xfs/libxfs/xfs_bmap.c ASSERT(PREV.br_startoff <= new->br_startoff); new 1582 fs/xfs/libxfs/xfs_bmap.c if (PREV.br_startoff == new->br_startoff) new 1598 fs/xfs/libxfs/xfs_bmap.c LEFT.br_startoff + LEFT.br_blockcount == new->br_startoff && new 1599 fs/xfs/libxfs/xfs_bmap.c LEFT.br_startblock + LEFT.br_blockcount == new->br_startblock && new 1600 fs/xfs/libxfs/xfs_bmap.c LEFT.br_state == new->br_state && new 1601 fs/xfs/libxfs/xfs_bmap.c LEFT.br_blockcount + new->br_blockcount <= MAXEXTLEN) new 1617 fs/xfs/libxfs/xfs_bmap.c new->br_startblock + new->br_blockcount == RIGHT.br_startblock && new 1618 fs/xfs/libxfs/xfs_bmap.c new->br_state == RIGHT.br_state && new 1619 fs/xfs/libxfs/xfs_bmap.c new->br_blockcount + RIGHT.br_blockcount <= MAXEXTLEN && new 1624 fs/xfs/libxfs/xfs_bmap.c LEFT.br_blockcount + new->br_blockcount + RIGHT.br_blockcount new 1703 fs/xfs/libxfs/xfs_bmap.c PREV.br_startblock = new->br_startblock; new 1705 fs/xfs/libxfs/xfs_bmap.c PREV.br_state = new->br_state; new 1732 fs/xfs/libxfs/xfs_bmap.c PREV.br_startblock = new->br_startblock; new 1733 fs/xfs/libxfs/xfs_bmap.c PREV.br_state = new->br_state; new 1741 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_lookup_eq(bma->cur, new, &i); new 1758 fs/xfs/libxfs/xfs_bmap.c temp = PREV.br_blockcount - new->br_blockcount; new 1762 fs/xfs/libxfs/xfs_bmap.c LEFT.br_blockcount += new->br_blockcount; new 1765 fs/xfs/libxfs/xfs_bmap.c PREV.br_startoff += new->br_blockcount; new 1791 fs/xfs/libxfs/xfs_bmap.c xfs_iext_update_extent(bma->ip, state, &bma->icur, new); new 1797 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_lookup_eq(bma->cur, new, &i); new 1815 fs/xfs/libxfs/xfs_bmap.c temp = PREV.br_blockcount - new->br_blockcount; new 1834 fs/xfs/libxfs/xfs_bmap.c RIGHT.br_startoff = new->br_startoff; new 1835 fs/xfs/libxfs/xfs_bmap.c RIGHT.br_startblock = new->br_startblock; new 1836 fs/xfs/libxfs/xfs_bmap.c RIGHT.br_blockcount += new->br_blockcount; new 1851 fs/xfs/libxfs/xfs_bmap.c temp = PREV.br_blockcount - new->br_blockcount; new 1868 fs/xfs/libxfs/xfs_bmap.c xfs_iext_update_extent(bma->ip, state, &bma->icur, new); new 1874 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_lookup_eq(bma->cur, new, &i); new 1892 fs/xfs/libxfs/xfs_bmap.c temp = PREV.br_blockcount - new->br_blockcount; new 1927 fs/xfs/libxfs/xfs_bmap.c LEFT = *new; new 1939 fs/xfs/libxfs/xfs_bmap.c PREV.br_blockcount = new->br_startoff - PREV.br_startoff; new 1954 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_lookup_eq(bma->cur, new, &i); new 1991 fs/xfs/libxfs/xfs_bmap.c xfs_rmap_map_extent(bma->tp, bma->ip, whichfork, new); new 2041 fs/xfs/libxfs/xfs_bmap.c xfs_bmbt_irec_t *new, /* new data to add to file extents */ new 2061 fs/xfs/libxfs/xfs_bmap.c ASSERT(!isnullstartblock(new->br_startblock)); new 2074 fs/xfs/libxfs/xfs_bmap.c ASSERT(new->br_state != PREV.br_state); new 2075 fs/xfs/libxfs/xfs_bmap.c new_endoff = new->br_startoff + new->br_blockcount; new 2076 fs/xfs/libxfs/xfs_bmap.c ASSERT(PREV.br_startoff <= new->br_startoff); new 2083 fs/xfs/libxfs/xfs_bmap.c if (PREV.br_startoff == new->br_startoff) new 2099 fs/xfs/libxfs/xfs_bmap.c LEFT.br_startoff + LEFT.br_blockcount == new->br_startoff && new 2100 fs/xfs/libxfs/xfs_bmap.c LEFT.br_startblock + LEFT.br_blockcount == new->br_startblock && new 2101 fs/xfs/libxfs/xfs_bmap.c LEFT.br_state == new->br_state && new 2102 fs/xfs/libxfs/xfs_bmap.c LEFT.br_blockcount + new->br_blockcount <= MAXEXTLEN) new 2118 fs/xfs/libxfs/xfs_bmap.c new->br_startblock + new->br_blockcount == RIGHT.br_startblock && new 2119 fs/xfs/libxfs/xfs_bmap.c new->br_state == RIGHT.br_state && new 2120 fs/xfs/libxfs/xfs_bmap.c new->br_blockcount + RIGHT.br_blockcount <= MAXEXTLEN && new 2125 fs/xfs/libxfs/xfs_bmap.c LEFT.br_blockcount + new->br_blockcount + RIGHT.br_blockcount new 2212 fs/xfs/libxfs/xfs_bmap.c PREV.br_state = new->br_state; new 2247 fs/xfs/libxfs/xfs_bmap.c PREV.br_state = new->br_state; new 2254 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_lookup_eq(cur, new, &i); new 2269 fs/xfs/libxfs/xfs_bmap.c LEFT.br_blockcount += new->br_blockcount; new 2272 fs/xfs/libxfs/xfs_bmap.c PREV.br_startoff += new->br_blockcount; new 2273 fs/xfs/libxfs/xfs_bmap.c PREV.br_startblock += new->br_blockcount; new 2274 fs/xfs/libxfs/xfs_bmap.c PREV.br_blockcount -= new->br_blockcount; new 2306 fs/xfs/libxfs/xfs_bmap.c PREV.br_startoff += new->br_blockcount; new 2307 fs/xfs/libxfs/xfs_bmap.c PREV.br_startblock += new->br_blockcount; new 2308 fs/xfs/libxfs/xfs_bmap.c PREV.br_blockcount -= new->br_blockcount; new 2311 fs/xfs/libxfs/xfs_bmap.c xfs_iext_insert(ip, icur, new, state); new 2325 fs/xfs/libxfs/xfs_bmap.c cur->bc_rec.b = *new; new 2338 fs/xfs/libxfs/xfs_bmap.c PREV.br_blockcount -= new->br_blockcount; new 2340 fs/xfs/libxfs/xfs_bmap.c RIGHT.br_startoff = new->br_startoff; new 2341 fs/xfs/libxfs/xfs_bmap.c RIGHT.br_startblock = new->br_startblock; new 2342 fs/xfs/libxfs/xfs_bmap.c RIGHT.br_blockcount += new->br_blockcount; new 2374 fs/xfs/libxfs/xfs_bmap.c PREV.br_blockcount -= new->br_blockcount; new 2378 fs/xfs/libxfs/xfs_bmap.c xfs_iext_insert(ip, icur, new, state); new 2393 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_lookup_eq(cur, new, &i); new 2410 fs/xfs/libxfs/xfs_bmap.c PREV.br_blockcount = new->br_startoff - PREV.br_startoff; new 2412 fs/xfs/libxfs/xfs_bmap.c r[0] = *new; new 2416 fs/xfs/libxfs/xfs_bmap.c r[1].br_startblock = new->br_startblock + new->br_blockcount; new 2448 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_lookup_eq(cur, new, &i); new 2473 fs/xfs/libxfs/xfs_bmap.c xfs_rmap_convert_extent(mp, tp, ip, whichfork, new); new 2510 fs/xfs/libxfs/xfs_bmap.c xfs_bmbt_irec_t *new) /* new data to add to file extents */ new 2521 fs/xfs/libxfs/xfs_bmap.c ASSERT(isnullstartblock(new->br_startblock)); new 2547 fs/xfs/libxfs/xfs_bmap.c left.br_startoff + left.br_blockcount == new->br_startoff && new 2548 fs/xfs/libxfs/xfs_bmap.c left.br_blockcount + new->br_blockcount <= MAXEXTLEN) new 2552 fs/xfs/libxfs/xfs_bmap.c new->br_startoff + new->br_blockcount == right.br_startoff && new 2553 fs/xfs/libxfs/xfs_bmap.c new->br_blockcount + right.br_blockcount <= MAXEXTLEN && new 2555 fs/xfs/libxfs/xfs_bmap.c (left.br_blockcount + new->br_blockcount + new 2569 fs/xfs/libxfs/xfs_bmap.c temp = left.br_blockcount + new->br_blockcount + new 2573 fs/xfs/libxfs/xfs_bmap.c startblockval(new->br_startblock) + new 2591 fs/xfs/libxfs/xfs_bmap.c temp = left.br_blockcount + new->br_blockcount; new 2594 fs/xfs/libxfs/xfs_bmap.c startblockval(new->br_startblock); new 2610 fs/xfs/libxfs/xfs_bmap.c temp = new->br_blockcount + right.br_blockcount; new 2611 fs/xfs/libxfs/xfs_bmap.c oldlen = startblockval(new->br_startblock) + new 2615 fs/xfs/libxfs/xfs_bmap.c right.br_startoff = new->br_startoff; new 2628 fs/xfs/libxfs/xfs_bmap.c xfs_iext_insert(ip, icur, new, state); new 2652 fs/xfs/libxfs/xfs_bmap.c struct xfs_bmbt_irec *new, new 2667 fs/xfs/libxfs/xfs_bmap.c ASSERT(!isnullstartblock(new->br_startblock)); new 2696 fs/xfs/libxfs/xfs_bmap.c left.br_startoff + left.br_blockcount == new->br_startoff && new 2697 fs/xfs/libxfs/xfs_bmap.c left.br_startblock + left.br_blockcount == new->br_startblock && new 2698 fs/xfs/libxfs/xfs_bmap.c left.br_state == new->br_state && new 2699 fs/xfs/libxfs/xfs_bmap.c left.br_blockcount + new->br_blockcount <= MAXEXTLEN) new 2703 fs/xfs/libxfs/xfs_bmap.c new->br_startoff + new->br_blockcount == right.br_startoff && new 2704 fs/xfs/libxfs/xfs_bmap.c new->br_startblock + new->br_blockcount == right.br_startblock && new 2705 fs/xfs/libxfs/xfs_bmap.c new->br_state == right.br_state && new 2706 fs/xfs/libxfs/xfs_bmap.c new->br_blockcount + right.br_blockcount <= MAXEXTLEN && new 2708 fs/xfs/libxfs/xfs_bmap.c left.br_blockcount + new->br_blockcount + new 2723 fs/xfs/libxfs/xfs_bmap.c left.br_blockcount += new->br_blockcount + right.br_blockcount; new 2760 fs/xfs/libxfs/xfs_bmap.c left.br_blockcount += new->br_blockcount; new 2787 fs/xfs/libxfs/xfs_bmap.c right.br_startoff = new->br_startoff; new 2788 fs/xfs/libxfs/xfs_bmap.c right.br_startblock = new->br_startblock; new 2789 fs/xfs/libxfs/xfs_bmap.c right.br_blockcount += new->br_blockcount; new 2812 fs/xfs/libxfs/xfs_bmap.c xfs_iext_insert(ip, icur, new, state); new 2819 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_lookup_eq(cur, new, &i); new 2833 fs/xfs/libxfs/xfs_bmap.c xfs_rmap_map_extent(tp, ip, whichfork, new); new 4735 fs/xfs/libxfs/xfs_bmap.c struct xfs_bmbt_irec new; new 4821 fs/xfs/libxfs/xfs_bmap.c new.br_blockcount = got_endoff - del_endoff; new 4822 fs/xfs/libxfs/xfs_bmap.c new_indlen = xfs_bmap_worst_indlen(ip, new.br_blockcount); new 4830 fs/xfs/libxfs/xfs_bmap.c new.br_startoff = del_endoff; new 4831 fs/xfs/libxfs/xfs_bmap.c new.br_state = got->br_state; new 4832 fs/xfs/libxfs/xfs_bmap.c new.br_startblock = nullstartblock((int)new_indlen); new 4836 fs/xfs/libxfs/xfs_bmap.c xfs_iext_insert(ip, icur, &new, state); new 4863 fs/xfs/libxfs/xfs_bmap.c struct xfs_bmbt_irec new; new 4912 fs/xfs/libxfs/xfs_bmap.c new.br_startoff = del_endoff; new 4913 fs/xfs/libxfs/xfs_bmap.c new.br_blockcount = got_endoff - del_endoff; new 4914 fs/xfs/libxfs/xfs_bmap.c new.br_state = got->br_state; new 4915 fs/xfs/libxfs/xfs_bmap.c new.br_startblock = del->br_startblock + del->br_blockcount; new 4919 fs/xfs/libxfs/xfs_bmap.c xfs_iext_insert(ip, icur, &new, state); new 4951 fs/xfs/libxfs/xfs_bmap.c xfs_bmbt_irec_t new; /* new record to be inserted */ new 5081 fs/xfs/libxfs/xfs_bmap.c new.br_startoff = del_endoff; new 5082 fs/xfs/libxfs/xfs_bmap.c new.br_blockcount = got_endoff - del_endoff; new 5083 fs/xfs/libxfs/xfs_bmap.c new.br_state = got.br_state; new 5084 fs/xfs/libxfs/xfs_bmap.c new.br_startblock = del_endblock; new 5094 fs/xfs/libxfs/xfs_bmap.c cur->bc_rec.b = new; new 5134 fs/xfs/libxfs/xfs_bmap.c xfs_iext_insert(ip, icur, &new, state); new 5583 fs/xfs/libxfs/xfs_bmap.c struct xfs_bmbt_irec new; new 5594 fs/xfs/libxfs/xfs_bmap.c new = *left; new 5595 fs/xfs/libxfs/xfs_bmap.c new.br_blockcount = blockcount; new 5626 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_update(cur, &new); new 5639 fs/xfs/libxfs/xfs_bmap.c &new); new 5643 fs/xfs/libxfs/xfs_bmap.c memcpy(&new, got, sizeof(new)); new 5644 fs/xfs/libxfs/xfs_bmap.c new.br_startoff = left->br_startoff + left->br_blockcount; new 5645 fs/xfs/libxfs/xfs_bmap.c xfs_rmap_map_extent(tp, ip, whichfork, &new); new 5923 fs/xfs/libxfs/xfs_bmap.c struct xfs_bmbt_irec new; /* split extent */ new 5960 fs/xfs/libxfs/xfs_bmap.c new.br_startoff = split_fsb; new 5961 fs/xfs/libxfs/xfs_bmap.c new.br_startblock = got.br_startblock + gotblkcnt; new 5962 fs/xfs/libxfs/xfs_bmap.c new.br_blockcount = got.br_blockcount - gotblkcnt; new 5963 fs/xfs/libxfs/xfs_bmap.c new.br_state = got.br_state; new 5988 fs/xfs/libxfs/xfs_bmap.c xfs_iext_insert(ip, &icur, &new, 0); new 5993 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_lookup_eq(cur, &new, &i); new 236 fs/xfs/libxfs/xfs_bmap.h struct xfs_bmbt_irec *new, int *logflagsp); new 166 fs/xfs/libxfs/xfs_bmap_btree.c struct xfs_btree_cur *new; new 168 fs/xfs/libxfs/xfs_bmap_btree.c new = xfs_bmbt_init_cursor(cur->bc_mp, cur->bc_tp, new 175 fs/xfs/libxfs/xfs_bmap_btree.c new->bc_private.b.flags = cur->bc_private.b.flags; new 177 fs/xfs/libxfs/xfs_bmap_btree.c return new; new 198 fs/xfs/libxfs/xfs_bmap_btree.c union xfs_btree_ptr *new, new 268 fs/xfs/libxfs/xfs_bmap_btree.c new->l = cpu_to_be64(args.fsbno); new 403 fs/xfs/libxfs/xfs_btree.c xfs_btree_cur_t *new; /* new cursor value */ new 412 fs/xfs/libxfs/xfs_btree.c new = cur->bc_ops->dup_cursor(cur); new 417 fs/xfs/libxfs/xfs_btree.c new->bc_rec = cur->bc_rec; new 422 fs/xfs/libxfs/xfs_btree.c for (i = 0; i < new->bc_nlevels; i++) { new 423 fs/xfs/libxfs/xfs_btree.c new->bc_ptrs[i] = cur->bc_ptrs[i]; new 424 fs/xfs/libxfs/xfs_btree.c new->bc_ra[i] = cur->bc_ra[i]; new 432 fs/xfs/libxfs/xfs_btree.c xfs_btree_del_cursor(new, error); new 437 fs/xfs/libxfs/xfs_btree.c new->bc_bufs[i] = bp; new 439 fs/xfs/libxfs/xfs_btree.c *ncur = new; new 3029 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_block *new; /* new (root) btree block */ new 3050 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_get_buf_block(cur, &lptr, &new, &nbp); new 3108 fs/xfs/libxfs/xfs_btree.c xfs_btree_key_addr(cur, 1, new)); new 3110 fs/xfs/libxfs/xfs_btree.c xfs_btree_key_addr(cur, 2, new)); new 3118 fs/xfs/libxfs/xfs_btree.c xfs_btree_key_addr(cur, 1, new)); new 3120 fs/xfs/libxfs/xfs_btree.c xfs_btree_key_addr(cur, 2, new)); new 3126 fs/xfs/libxfs/xfs_btree.c xfs_btree_ptr_addr(cur, 1, new), &lptr, 1); new 3128 fs/xfs/libxfs/xfs_btree.c xfs_btree_ptr_addr(cur, 2, new), &rptr, 1); new 485 fs/xfs/libxfs/xfs_dir2_data.c xfs_dir2_data_free_t new; /* new bestfree entry */ new 492 fs/xfs/libxfs/xfs_dir2_data.c new.length = dup->length; new 493 fs/xfs/libxfs/xfs_dir2_data.c new.offset = cpu_to_be16((char *)dup - (char *)hdr); new 498 fs/xfs/libxfs/xfs_dir2_data.c if (be16_to_cpu(new.length) > be16_to_cpu(dfp[0].length)) { new 501 fs/xfs/libxfs/xfs_dir2_data.c dfp[0] = new; new 505 fs/xfs/libxfs/xfs_dir2_data.c if (be16_to_cpu(new.length) > be16_to_cpu(dfp[1].length)) { new 507 fs/xfs/libxfs/xfs_dir2_data.c dfp[1] = new; new 511 fs/xfs/libxfs/xfs_dir2_data.c if (be16_to_cpu(new.length) > be16_to_cpu(dfp[2].length)) { new 512 fs/xfs/libxfs/xfs_dir2_data.c dfp[2] = new; new 74 fs/xfs/libxfs/xfs_ialloc_btree.c union xfs_btree_ptr *new, new 103 fs/xfs/libxfs/xfs_ialloc_btree.c new->s = cpu_to_be32(XFS_FSB_TO_AGBNO(args.mp, args.fsbno)); new 112 fs/xfs/libxfs/xfs_ialloc_btree.c union xfs_btree_ptr *new, new 115 fs/xfs/libxfs/xfs_ialloc_btree.c return __xfs_inobt_alloc_block(cur, start, new, stat, XFS_AG_RESV_NONE); new 122 fs/xfs/libxfs/xfs_ialloc_btree.c union xfs_btree_ptr *new, new 126 fs/xfs/libxfs/xfs_ialloc_btree.c return xfs_inobt_alloc_block(cur, start, new, stat); new 127 fs/xfs/libxfs/xfs_ialloc_btree.c return __xfs_inobt_alloc_block(cur, start, new, stat, new 457 fs/xfs/libxfs/xfs_iext_tree.c struct xfs_iext_node *new = kmem_zalloc(NODE_SIZE, KM_NOFS); new 464 fs/xfs/libxfs/xfs_iext_tree.c *nodep = new; new 472 fs/xfs/libxfs/xfs_iext_tree.c new->keys[i] = node->keys[nr_keep + i]; new 473 fs/xfs/libxfs/xfs_iext_tree.c new->ptrs[i] = node->ptrs[nr_keep + i]; new 480 fs/xfs/libxfs/xfs_iext_tree.c *nodep = new; new 488 fs/xfs/libxfs/xfs_iext_tree.c new->keys[i] = XFS_IEXT_KEY_INVALID; new 489 fs/xfs/libxfs/xfs_iext_tree.c return new; new 499 fs/xfs/libxfs/xfs_iext_tree.c struct xfs_iext_node *node, *new; new 506 fs/xfs/libxfs/xfs_iext_tree.c new = NULL; new 515 fs/xfs/libxfs/xfs_iext_tree.c new = xfs_iext_split_node(&node, &pos, &nr_entries); new 521 fs/xfs/libxfs/xfs_iext_tree.c if (node != new && pos == 0 && nr_entries > 0) new 531 fs/xfs/libxfs/xfs_iext_tree.c if (new) { new 532 fs/xfs/libxfs/xfs_iext_tree.c offset = new->keys[0]; new 533 fs/xfs/libxfs/xfs_iext_tree.c ptr = new; new 545 fs/xfs/libxfs/xfs_iext_tree.c struct xfs_iext_leaf *new = kmem_zalloc(NODE_SIZE, KM_NOFS); new 552 fs/xfs/libxfs/xfs_iext_tree.c cur->leaf = new; new 559 fs/xfs/libxfs/xfs_iext_tree.c new->recs[i] = leaf->recs[nr_keep + i]; new 564 fs/xfs/libxfs/xfs_iext_tree.c cur->leaf = new; new 572 fs/xfs/libxfs/xfs_iext_tree.c leaf->next->prev = new; new 573 fs/xfs/libxfs/xfs_iext_tree.c new->next = leaf->next; new 574 fs/xfs/libxfs/xfs_iext_tree.c new->prev = leaf; new 575 fs/xfs/libxfs/xfs_iext_tree.c leaf->next = new; new 576 fs/xfs/libxfs/xfs_iext_tree.c return new; new 600 fs/xfs/libxfs/xfs_iext_tree.c void *new; new 606 fs/xfs/libxfs/xfs_iext_tree.c new = kmem_realloc(ifp->if_u1.if_root, new_size, KM_NOFS); new 607 fs/xfs/libxfs/xfs_iext_tree.c memset(new + ifp->if_bytes, 0, new_size - ifp->if_bytes); new 608 fs/xfs/libxfs/xfs_iext_tree.c ifp->if_u1.if_root = new; new 609 fs/xfs/libxfs/xfs_iext_tree.c cur->leaf = new; new 633 fs/xfs/libxfs/xfs_iext_tree.c struct xfs_iext_leaf *new = NULL; new 649 fs/xfs/libxfs/xfs_iext_tree.c new = xfs_iext_split_leaf(cur, &nr_entries); new 655 fs/xfs/libxfs/xfs_iext_tree.c if (cur->leaf != new && cur->pos == 0 && nr_entries > 0) { new 667 fs/xfs/libxfs/xfs_iext_tree.c if (new) new 668 fs/xfs/libxfs/xfs_iext_tree.c xfs_iext_insert_node(ifp, xfs_iext_leaf_key(new, 0), new, 2); new 982 fs/xfs/libxfs/xfs_iext_tree.c struct xfs_bmbt_irec *new) new 992 fs/xfs/libxfs/xfs_iext_tree.c if (new->br_startoff != old.br_startoff) { new 994 fs/xfs/libxfs/xfs_iext_tree.c new->br_startoff, 1, cur->leaf); new 999 fs/xfs/libxfs/xfs_iext_tree.c xfs_iext_set(cur_rec(cur), new); new 210 fs/xfs/libxfs/xfs_inode_fork.c struct xfs_bmbt_irec new; new 236 fs/xfs/libxfs/xfs_inode_fork.c xfs_bmbt_disk_get_all(dp, &new); new 237 fs/xfs/libxfs/xfs_inode_fork.c fa = xfs_bmap_validate_extent(ip, whichfork, &new); new 245 fs/xfs/libxfs/xfs_inode_fork.c xfs_iext_insert(ip, &icur, &new, state); new 57 fs/xfs/libxfs/xfs_refcount_btree.c union xfs_btree_ptr *new, new 87 fs/xfs/libxfs/xfs_refcount_btree.c new->s = cpu_to_be32(args.agbno); new 83 fs/xfs/libxfs/xfs_rmap_btree.c union xfs_btree_ptr *new, new 108 fs/xfs/libxfs/xfs_rmap_btree.c new->s = cpu_to_be32(bno); new 770 fs/xfs/libxfs/xfs_rtbitmap.c xfs_rtblock_t *new, /* out: first block not matching */ new 829 fs/xfs/libxfs/xfs_rtbitmap.c *new = start + i; new 875 fs/xfs/libxfs/xfs_rtbitmap.c *new = start + i; new 920 fs/xfs/libxfs/xfs_rtbitmap.c *new = start + i; new 930 fs/xfs/libxfs/xfs_rtbitmap.c *new = start + i; new 946 fs/xfs/libxfs/xfs_rtbitmap.c xfs_rtblock_t new; /* dummy for xfs_rtcheck_range */ new 950 fs/xfs/libxfs/xfs_rtbitmap.c error = xfs_rtcheck_range(mp, tp, bno, len, 0, &new, &stat); new 736 fs/xfs/xfs_aops.c struct bio *new; new 738 fs/xfs/xfs_aops.c new = bio_alloc(GFP_NOFS, BIO_MAX_PAGES); new 739 fs/xfs/xfs_aops.c bio_copy_dev(new, prev);/* also copies over blkcg information */ new 740 fs/xfs/xfs_aops.c new->bi_iter.bi_sector = bio_end_sector(prev); new 741 fs/xfs/xfs_aops.c new->bi_opf = prev->bi_opf; new 742 fs/xfs/xfs_aops.c new->bi_write_hint = prev->bi_write_hint; new 744 fs/xfs/xfs_aops.c bio_chain(prev, new); new 747 fs/xfs/xfs_aops.c return new; new 30 fs/xfs/xfs_extent_busy.c struct xfs_extent_busy *new; new 36 fs/xfs/xfs_extent_busy.c new = kmem_zalloc(sizeof(struct xfs_extent_busy), 0); new 37 fs/xfs/xfs_extent_busy.c new->agno = agno; new 38 fs/xfs/xfs_extent_busy.c new->bno = bno; new 39 fs/xfs/xfs_extent_busy.c new->length = len; new 40 fs/xfs/xfs_extent_busy.c INIT_LIST_HEAD(&new->list); new 41 fs/xfs/xfs_extent_busy.c new->flags = flags; new 46 fs/xfs/xfs_extent_busy.c pag = xfs_perag_get(tp->t_mountp, new->agno); new 53 fs/xfs/xfs_extent_busy.c if (new->bno < busyp->bno) { new 55 fs/xfs/xfs_extent_busy.c ASSERT(new->bno + new->length <= busyp->bno); new 56 fs/xfs/xfs_extent_busy.c } else if (new->bno > busyp->bno) { new 64 fs/xfs/xfs_extent_busy.c rb_link_node(&new->rb_node, parent, rbp); new 65 fs/xfs/xfs_extent_busy.c rb_insert_color(&new->rb_node, &pag->pagb_tree); new 67 fs/xfs/xfs_extent_busy.c list_add(&new->list, &tp->t_busy); new 36 fs/xfs/xfs_fsops.c xfs_rfsblock_t new; new 53 fs/xfs/xfs_fsops.c new = nb; /* use new as a temporary here */ new 54 fs/xfs/xfs_fsops.c nb_mod = do_div(new, mp->m_sb.sb_agblocks); new 55 fs/xfs/xfs_fsops.c nagcount = new + (nb_mod != 0); new 62 fs/xfs/xfs_fsops.c new = nb - mp->m_sb.sb_dblocks; new 92 fs/xfs/xfs_fsops.c id.agno--, new -= id.agsize) { new 113 fs/xfs/xfs_fsops.c if (new) { new 114 fs/xfs/xfs_fsops.c error = xfs_ag_extend_space(mp, tp, &id, new); new 146 fs/xfs/xfs_fsops.c if (new) { new 123 fs/xfs/xfs_log.c int64_t new, old; new 137 fs/xfs/xfs_log.c new = xlog_assign_grant_head_val(cycle, space); new 138 fs/xfs/xfs_log.c head_val = atomic64_cmpxchg(head, old, new); new 149 fs/xfs/xfs_log.c int64_t new, old; new 166 fs/xfs/xfs_log.c new = xlog_assign_grant_head_val(cycle, space); new 167 fs/xfs/xfs_log.c head_val = atomic64_cmpxchg(head, old, new); new 140 fs/xfs/xfs_quota.h #define xfs_qm_vop_chown(tp, ip, old, new) (NULL) new 107 fs/xfs/xfs_rtalloc.h xfs_rtblock_t *new, int *stat); new 651 include/asm-generic/atomic-instrumented.h atomic_cmpxchg(atomic_t *v, int old, int new) new 654 include/asm-generic/atomic-instrumented.h return arch_atomic_cmpxchg(v, old, new); new 661 include/asm-generic/atomic-instrumented.h atomic_cmpxchg_acquire(atomic_t *v, int old, int new) new 664 include/asm-generic/atomic-instrumented.h return arch_atomic_cmpxchg_acquire(v, old, new); new 671 include/asm-generic/atomic-instrumented.h atomic_cmpxchg_release(atomic_t *v, int old, int new) new 674 include/asm-generic/atomic-instrumented.h return arch_atomic_cmpxchg_release(v, old, new); new 681 include/asm-generic/atomic-instrumented.h atomic_cmpxchg_relaxed(atomic_t *v, int old, int new) new 684 include/asm-generic/atomic-instrumented.h return arch_atomic_cmpxchg_relaxed(v, old, new); new 691 include/asm-generic/atomic-instrumented.h atomic_try_cmpxchg(atomic_t *v, int *old, int new) new 695 include/asm-generic/atomic-instrumented.h return arch_atomic_try_cmpxchg(v, old, new); new 702 include/asm-generic/atomic-instrumented.h atomic_try_cmpxchg_acquire(atomic_t *v, int *old, int new) new 706 include/asm-generic/atomic-instrumented.h return arch_atomic_try_cmpxchg_acquire(v, old, new); new 713 include/asm-generic/atomic-instrumented.h atomic_try_cmpxchg_release(atomic_t *v, int *old, int new) new 717 include/asm-generic/atomic-instrumented.h return arch_atomic_try_cmpxchg_release(v, old, new); new 724 include/asm-generic/atomic-instrumented.h atomic_try_cmpxchg_relaxed(atomic_t *v, int *old, int new) new 728 include/asm-generic/atomic-instrumented.h return arch_atomic_try_cmpxchg_relaxed(v, old, new); new 1461 include/asm-generic/atomic-instrumented.h atomic64_cmpxchg(atomic64_t *v, s64 old, s64 new) new 1464 include/asm-generic/atomic-instrumented.h return arch_atomic64_cmpxchg(v, old, new); new 1471 include/asm-generic/atomic-instrumented.h atomic64_cmpxchg_acquire(atomic64_t *v, s64 old, s64 new) new 1474 include/asm-generic/atomic-instrumented.h return arch_atomic64_cmpxchg_acquire(v, old, new); new 1481 include/asm-generic/atomic-instrumented.h atomic64_cmpxchg_release(atomic64_t *v, s64 old, s64 new) new 1484 include/asm-generic/atomic-instrumented.h return arch_atomic64_cmpxchg_release(v, old, new); new 1491 include/asm-generic/atomic-instrumented.h atomic64_cmpxchg_relaxed(atomic64_t *v, s64 old, s64 new) new 1494 include/asm-generic/atomic-instrumented.h return arch_atomic64_cmpxchg_relaxed(v, old, new); new 1501 include/asm-generic/atomic-instrumented.h atomic64_try_cmpxchg(atomic64_t *v, s64 *old, s64 new) new 1505 include/asm-generic/atomic-instrumented.h return arch_atomic64_try_cmpxchg(v, old, new); new 1512 include/asm-generic/atomic-instrumented.h atomic64_try_cmpxchg_acquire(atomic64_t *v, s64 *old, s64 new) new 1516 include/asm-generic/atomic-instrumented.h return arch_atomic64_try_cmpxchg_acquire(v, old, new); new 1523 include/asm-generic/atomic-instrumented.h atomic64_try_cmpxchg_release(atomic64_t *v, s64 *old, s64 new) new 1527 include/asm-generic/atomic-instrumented.h return arch_atomic64_try_cmpxchg_release(v, old, new); new 1534 include/asm-generic/atomic-instrumented.h atomic64_try_cmpxchg_relaxed(atomic64_t *v, s64 *old, s64 new) new 1538 include/asm-generic/atomic-instrumented.h return arch_atomic64_try_cmpxchg_relaxed(v, old, new); new 410 include/asm-generic/atomic-long.h atomic_long_cmpxchg(atomic_long_t *v, long old, long new) new 412 include/asm-generic/atomic-long.h return atomic64_cmpxchg(v, old, new); new 416 include/asm-generic/atomic-long.h atomic_long_cmpxchg_acquire(atomic_long_t *v, long old, long new) new 418 include/asm-generic/atomic-long.h return atomic64_cmpxchg_acquire(v, old, new); new 422 include/asm-generic/atomic-long.h atomic_long_cmpxchg_release(atomic_long_t *v, long old, long new) new 424 include/asm-generic/atomic-long.h return atomic64_cmpxchg_release(v, old, new); new 428 include/asm-generic/atomic-long.h atomic_long_cmpxchg_relaxed(atomic_long_t *v, long old, long new) new 430 include/asm-generic/atomic-long.h return atomic64_cmpxchg_relaxed(v, old, new); new 434 include/asm-generic/atomic-long.h atomic_long_try_cmpxchg(atomic_long_t *v, long *old, long new) new 436 include/asm-generic/atomic-long.h return atomic64_try_cmpxchg(v, (s64 *)old, new); new 440 include/asm-generic/atomic-long.h atomic_long_try_cmpxchg_acquire(atomic_long_t *v, long *old, long new) new 442 include/asm-generic/atomic-long.h return atomic64_try_cmpxchg_acquire(v, (s64 *)old, new); new 446 include/asm-generic/atomic-long.h atomic_long_try_cmpxchg_release(atomic_long_t *v, long *old, long new) new 448 include/asm-generic/atomic-long.h return atomic64_try_cmpxchg_release(v, (s64 *)old, new); new 452 include/asm-generic/atomic-long.h atomic_long_try_cmpxchg_relaxed(atomic_long_t *v, long *old, long new) new 454 include/asm-generic/atomic-long.h return atomic64_try_cmpxchg_relaxed(v, (s64 *)old, new); new 904 include/asm-generic/atomic-long.h atomic_long_cmpxchg(atomic_long_t *v, long old, long new) new 906 include/asm-generic/atomic-long.h return atomic_cmpxchg(v, old, new); new 910 include/asm-generic/atomic-long.h atomic_long_cmpxchg_acquire(atomic_long_t *v, long old, long new) new 912 include/asm-generic/atomic-long.h return atomic_cmpxchg_acquire(v, old, new); new 916 include/asm-generic/atomic-long.h atomic_long_cmpxchg_release(atomic_long_t *v, long old, long new) new 918 include/asm-generic/atomic-long.h return atomic_cmpxchg_release(v, old, new); new 922 include/asm-generic/atomic-long.h atomic_long_cmpxchg_relaxed(atomic_long_t *v, long old, long new) new 924 include/asm-generic/atomic-long.h return atomic_cmpxchg_relaxed(v, old, new); new 928 include/asm-generic/atomic-long.h atomic_long_try_cmpxchg(atomic_long_t *v, long *old, long new) new 930 include/asm-generic/atomic-long.h return atomic_try_cmpxchg(v, (int *)old, new); new 934 include/asm-generic/atomic-long.h atomic_long_try_cmpxchg_acquire(atomic_long_t *v, long *old, long new) new 936 include/asm-generic/atomic-long.h return atomic_try_cmpxchg_acquire(v, (int *)old, new); new 940 include/asm-generic/atomic-long.h atomic_long_try_cmpxchg_release(atomic_long_t *v, long *old, long new) new 942 include/asm-generic/atomic-long.h return atomic_try_cmpxchg_release(v, (int *)old, new); new 946 include/asm-generic/atomic-long.h atomic_long_try_cmpxchg_relaxed(atomic_long_t *v, long *old, long new) new 948 include/asm-generic/atomic-long.h return atomic_try_cmpxchg_relaxed(v, (int *)old, new); new 196 include/asm-generic/atomic.h #define atomic_cmpxchg(v, old, new) (cmpxchg(&((v)->counter), (old), (new))) new 52 include/asm-generic/atomic64.h extern s64 atomic64_xchg(atomic64_t *v, s64 new); new 16 include/asm-generic/cmpxchg-local.h unsigned long old, unsigned long new, int size) new 30 include/asm-generic/cmpxchg-local.h *(u8 *)ptr = (u8)new; new 34 include/asm-generic/cmpxchg-local.h *(u16 *)ptr = (u16)new; new 38 include/asm-generic/cmpxchg-local.h *(u32 *)ptr = (u32)new; new 42 include/asm-generic/cmpxchg-local.h *(u64 *)ptr = (u64)new; new 55 include/asm-generic/cmpxchg-local.h u64 old, u64 new) new 63 include/asm-generic/cmpxchg-local.h *(u64 *)ptr = new; new 462 include/drm/drm_mm.h void drm_mm_replace_node(struct drm_mm_node *old, struct drm_mm_node *new); new 540 include/drm/ttm/ttm_bo_driver.h ttm_flag_masked(uint32_t *old, uint32_t new, uint32_t mask) new 542 include/drm/ttm/ttm_bo_driver.h *old ^= (*old ^ new) & mask; new 872 include/linux/atomic-fallback.h atomic_cmpxchg_acquire(atomic_t *v, int old, int new) new 874 include/linux/atomic-fallback.h int ret = atomic_cmpxchg_relaxed(v, old, new); new 883 include/linux/atomic-fallback.h atomic_cmpxchg_release(atomic_t *v, int old, int new) new 886 include/linux/atomic-fallback.h return atomic_cmpxchg_relaxed(v, old, new); new 893 include/linux/atomic-fallback.h atomic_cmpxchg(atomic_t *v, int old, int new) new 897 include/linux/atomic-fallback.h ret = atomic_cmpxchg_relaxed(v, old, new); new 915 include/linux/atomic-fallback.h atomic_try_cmpxchg(atomic_t *v, int *old, int new) new 918 include/linux/atomic-fallback.h r = atomic_cmpxchg(v, o, new); new 928 include/linux/atomic-fallback.h atomic_try_cmpxchg_acquire(atomic_t *v, int *old, int new) new 931 include/linux/atomic-fallback.h r = atomic_cmpxchg_acquire(v, o, new); new 941 include/linux/atomic-fallback.h atomic_try_cmpxchg_release(atomic_t *v, int *old, int new) new 944 include/linux/atomic-fallback.h r = atomic_cmpxchg_release(v, o, new); new 954 include/linux/atomic-fallback.h atomic_try_cmpxchg_relaxed(atomic_t *v, int *old, int new) new 957 include/linux/atomic-fallback.h r = atomic_cmpxchg_relaxed(v, o, new); new 969 include/linux/atomic-fallback.h atomic_try_cmpxchg_acquire(atomic_t *v, int *old, int new) new 971 include/linux/atomic-fallback.h bool ret = atomic_try_cmpxchg_relaxed(v, old, new); new 980 include/linux/atomic-fallback.h atomic_try_cmpxchg_release(atomic_t *v, int *old, int new) new 983 include/linux/atomic-fallback.h return atomic_try_cmpxchg_relaxed(v, old, new); new 990 include/linux/atomic-fallback.h atomic_try_cmpxchg(atomic_t *v, int *old, int new) new 994 include/linux/atomic-fallback.h ret = atomic_try_cmpxchg_relaxed(v, old, new); new 1982 include/linux/atomic-fallback.h atomic64_cmpxchg_acquire(atomic64_t *v, s64 old, s64 new) new 1984 include/linux/atomic-fallback.h s64 ret = atomic64_cmpxchg_relaxed(v, old, new); new 1993 include/linux/atomic-fallback.h atomic64_cmpxchg_release(atomic64_t *v, s64 old, s64 new) new 1996 include/linux/atomic-fallback.h return atomic64_cmpxchg_relaxed(v, old, new); new 2003 include/linux/atomic-fallback.h atomic64_cmpxchg(atomic64_t *v, s64 old, s64 new) new 2007 include/linux/atomic-fallback.h ret = atomic64_cmpxchg_relaxed(v, old, new); new 2025 include/linux/atomic-fallback.h atomic64_try_cmpxchg(atomic64_t *v, s64 *old, s64 new) new 2028 include/linux/atomic-fallback.h r = atomic64_cmpxchg(v, o, new); new 2038 include/linux/atomic-fallback.h atomic64_try_cmpxchg_acquire(atomic64_t *v, s64 *old, s64 new) new 2041 include/linux/atomic-fallback.h r = atomic64_cmpxchg_acquire(v, o, new); new 2051 include/linux/atomic-fallback.h atomic64_try_cmpxchg_release(atomic64_t *v, s64 *old, s64 new) new 2054 include/linux/atomic-fallback.h r = atomic64_cmpxchg_release(v, o, new); new 2064 include/linux/atomic-fallback.h atomic64_try_cmpxchg_relaxed(atomic64_t *v, s64 *old, s64 new) new 2067 include/linux/atomic-fallback.h r = atomic64_cmpxchg_relaxed(v, o, new); new 2079 include/linux/atomic-fallback.h atomic64_try_cmpxchg_acquire(atomic64_t *v, s64 *old, s64 new) new 2081 include/linux/atomic-fallback.h bool ret = atomic64_try_cmpxchg_relaxed(v, old, new); new 2090 include/linux/atomic-fallback.h atomic64_try_cmpxchg_release(atomic64_t *v, s64 *old, s64 new) new 2093 include/linux/atomic-fallback.h return atomic64_try_cmpxchg_relaxed(v, old, new); new 2100 include/linux/atomic-fallback.h atomic64_try_cmpxchg(atomic64_t *v, s64 *old, s64 new) new 2104 include/linux/atomic-fallback.h ret = atomic64_try_cmpxchg_relaxed(v, old, new); new 373 include/linux/audit.h const struct cred *new, new 375 include/linux/audit.h extern void __audit_log_capset(const struct cred *new, const struct cred *old); new 450 include/linux/audit.h const struct cred *new, new 454 include/linux/audit.h return __audit_log_bprm_fcaps(bprm, new, old); new 458 include/linux/audit.h static inline void audit_log_capset(const struct cred *new, new 462 include/linux/audit.h __audit_log_capset(new, old); new 611 include/linux/audit.h const struct cred *new, new 616 include/linux/audit.h static inline void audit_log_capset(const struct cred *new, new 144 include/linux/binfmts.h extern void set_binfmt(struct linux_binfmt *new); new 188 include/linux/bitmap.h const unsigned long *old, const unsigned long *new, unsigned int nbits); new 190 include/linux/bitmap.h const unsigned long *old, const unsigned long *new, int bits); new 127 include/linux/buffer_head.h BUFFER_FNS(New, new) new 168 include/linux/cpufreq.h unsigned int new; new 185 include/linux/devfreq.h unsigned long new; new 35 include/linux/edac.h void edac_set_report_status(int new); new 1226 include/linux/fs.h static inline void locks_copy_conflock(struct file_lock *new, struct file_lock *fl) new 1231 include/linux/fs.h static inline void locks_copy_lock(struct file_lock *new, struct file_lock *fl) new 43 include/linux/hdlcdrv.h unsigned char new; new 45 include/linux/hdlcdrv.h new = buf->shreg & 1; new 48 include/linux/hdlcdrv.h if (new) { new 195 include/linux/hugetlb.h #define move_hugetlb_state(old, new, reason) do {} while (0) new 1228 include/linux/hyperv.h int vmbus_allocate_mmio(struct resource **new, struct hv_device *device_obj, new 181 include/linux/ioport.h extern struct resource *request_resource_conflict(struct resource *root, struct resource *new); new 182 include/linux/ioport.h extern int request_resource(struct resource *root, struct resource *new); new 183 include/linux/ioport.h extern int release_resource(struct resource *new); new 184 include/linux/ioport.h void release_child_resources(struct resource *new); new 188 include/linux/ioport.h extern struct resource *insert_resource_conflict(struct resource *parent, struct resource *new); new 189 include/linux/ioport.h extern int insert_resource(struct resource *parent, struct resource *new); new 190 include/linux/ioport.h extern void insert_resource_expand_to_fit(struct resource *root, struct resource *new); new 193 include/linux/ioport.h extern int allocate_resource(struct resource *root, struct resource *new, new 257 include/linux/ioport.h struct resource *new); new 258 include/linux/ioport.h extern void devm_release_resource(struct device *dev, struct resource *new); new 563 include/linux/irq.h extern int setup_irq(unsigned int irq, struct irqaction *new); new 565 include/linux/irq.h extern int setup_percpu_irq(unsigned int irq, struct irqaction *new); new 200 include/linux/iversion.h u64 cur, old, new; new 220 include/linux/iversion.h new = (cur & ~I_VERSION_QUERIED) + I_VERSION_INCREMENT; new 222 include/linux/iversion.h old = atomic64_cmpxchg(&inode->i_version, cur, new); new 307 include/linux/iversion.h u64 cur, old, new; new 322 include/linux/iversion.h new = cur | I_VERSION_QUERIED; new 323 include/linux/iversion.h old = atomic64_cmpxchg(&inode->i_version, cur, new); new 45 include/linux/kcore.h void kclist_add(struct kcore_list *new, void *addr, size_t size, int type) new 692 include/linux/kvm_host.h const struct kvm_memory_slot *new, new 33 include/linux/list.h extern bool __list_add_valid(struct list_head *new, new 38 include/linux/list.h static inline bool __list_add_valid(struct list_head *new, new 56 include/linux/list.h static inline void __list_add(struct list_head *new, new 60 include/linux/list.h if (!__list_add_valid(new, prev, next)) new 63 include/linux/list.h next->prev = new; new 64 include/linux/list.h new->next = next; new 65 include/linux/list.h new->prev = prev; new 66 include/linux/list.h WRITE_ONCE(prev->next, new); new 77 include/linux/list.h static inline void list_add(struct list_head *new, struct list_head *head) new 79 include/linux/list.h __list_add(new, head, head->next); new 91 include/linux/list.h static inline void list_add_tail(struct list_head *new, struct list_head *head) new 93 include/linux/list.h __list_add(new, head->prev, head); new 152 include/linux/list.h struct list_head *new) new 154 include/linux/list.h new->next = old->next; new 155 include/linux/list.h new->next->prev = new; new 156 include/linux/list.h new->prev = old->prev; new 157 include/linux/list.h new->prev->next = new; new 161 include/linux/list.h struct list_head *new) new 163 include/linux/list.h list_replace(old, new); new 839 include/linux/list.h struct hlist_head *new) new 841 include/linux/list.h new->first = old->first; new 842 include/linux/list.h if (new->first) new 843 include/linux/list.h new->first->pprev = &new->first; new 207 include/linux/llist.h static inline bool llist_add(struct llist_node *new, struct llist_head *head) new 209 include/linux/llist.h return llist_add_batch(new, new, head); new 1473 include/linux/lsm_hooks.h int (*capset)(struct cred *new, const struct cred *old, new 1524 include/linux/lsm_hooks.h struct cred *new); new 1594 include/linux/lsm_hooks.h int (*inode_copy_up)(struct dentry *src, struct cred **new); new 1623 include/linux/lsm_hooks.h int (*cred_prepare)(struct cred *new, const struct cred *old, new 1625 include/linux/lsm_hooks.h void (*cred_transfer)(struct cred *new, const struct cred *old); new 1627 include/linux/lsm_hooks.h int (*kernel_act_as)(struct cred *new, u32 secid); new 1628 include/linux/lsm_hooks.h int (*kernel_create_files_as)(struct cred *new, struct inode *inode); new 1634 include/linux/lsm_hooks.h int (*task_fix_setuid)(struct cred *new, const struct cred *old, new 13 include/linux/mISDNdsp.h void *(*new)(const char *arg); new 901 include/linux/memcontrol.h static inline void mem_cgroup_migrate(struct page *old, struct page *new) new 134 include/linux/mempolicy.h struct mempolicy *new); new 146 include/linux/mempolicy.h extern void mpol_rebind_task(struct task_struct *tsk, const nodemask_t *new); new 147 include/linux/mempolicy.h extern void mpol_rebind_mm(struct mm_struct *mm, nodemask_t *new); new 263 include/linux/mempolicy.h const nodemask_t *new) new 267 include/linux/mempolicy.h static inline void mpol_rebind_mm(struct mm_struct *mm, nodemask_t *new) new 68 include/linux/migrate.h extern int migrate_pages(struct list_head *l, new_page_t new, free_page_t free, new 84 include/linux/migrate.h static inline int migrate_pages(struct list_head *l, new_page_t new, new 505 include/linux/mm.h int (*set_policy)(struct vm_area_struct *vma, struct mempolicy *new); new 1479 include/linux/mm.h extern void truncate_pagecache(struct inode *inode, loff_t new); new 2304 include/linux/mm.h unsigned long new, new 2310 include/linux/mm.h if (((new - start) + (end_data - start_data)) > rlim) new 581 include/linux/mtd/mtd.h extern void register_mtd_user (struct mtd_notifier *new); new 549 include/linux/netdevice.h unsigned long val, new; new 559 include/linux/netdevice.h new = val | NAPIF_STATE_MISSED; new 560 include/linux/netdevice.h } while (cmpxchg(&n->state, val, new) != val); new 434 include/linux/netfilter.h static inline void nf_ct_attach(struct sk_buff *new, struct sk_buff *skb) {} new 37 include/linux/netfilter/nf_conntrack_h323.h void nf_conntrack_h245_expect(struct nf_conn *new, new 39 include/linux/netfilter/nf_conntrack_h323.h void nf_conntrack_q931_expect(struct nf_conn *new, new 345 include/linux/nodemask.h #define node_remap(oldbit, old, new) \ new 346 include/linux/nodemask.h __node_remap((oldbit), &(old), &(new), MAX_NUMNODES) new 353 include/linux/nodemask.h #define nodes_remap(dst, src, old, new) \ new 354 include/linux/nodemask.h __nodes_remap(&(dst), &(src), &(old), &(new), MAX_NUMNODES) new 70 include/linux/nsproxy.h void switch_task_namespaces(struct task_struct *tsk, struct nsproxy *new); new 613 include/linux/pagemap.h int replace_page_cache_page(struct page *old, struct page *new, gfp_t gfp_mask); new 99 include/linux/pid.h extern void transfer_pid(struct task_struct *old, struct task_struct *new, new 64 include/linux/rbtree.h extern void rb_replace_node(struct rb_node *victim, struct rb_node *new, new 66 include/linux/rbtree.h extern void rb_replace_node_rcu(struct rb_node *victim, struct rb_node *new, new 153 include/linux/rbtree.h struct rb_node *new, new 157 include/linux/rbtree.h root->rb_leftmost = new; new 158 include/linux/rbtree.h rb_replace_node(victim, new, &root->rb_root); new 29 include/linux/rbtree_augmented.h void (*copy)(struct rb_node *old, struct rb_node *new); new 30 include/linux/rbtree_augmented.h void (*rotate)(struct rb_node *old, struct rb_node *new); new 34 include/linux/rbtree_augmented.h void (*augment_rotate)(struct rb_node *old, struct rb_node *new)); new 90 include/linux/rbtree_augmented.h RBSTRUCT *new = rb_entry(rb_new, RBSTRUCT, RBFIELD); \ new 91 include/linux/rbtree_augmented.h new->RBAUGMENTED = old->RBAUGMENTED; \ new 97 include/linux/rbtree_augmented.h RBSTRUCT *new = rb_entry(rb_new, RBSTRUCT, RBFIELD); \ new 98 include/linux/rbtree_augmented.h new->RBAUGMENTED = old->RBAUGMENTED; \ new 169 include/linux/rbtree_augmented.h __rb_change_child(struct rb_node *old, struct rb_node *new, new 174 include/linux/rbtree_augmented.h WRITE_ONCE(parent->rb_left, new); new 176 include/linux/rbtree_augmented.h WRITE_ONCE(parent->rb_right, new); new 178 include/linux/rbtree_augmented.h WRITE_ONCE(root->rb_node, new); new 182 include/linux/rbtree_augmented.h __rb_change_child_rcu(struct rb_node *old, struct rb_node *new, new 187 include/linux/rbtree_augmented.h rcu_assign_pointer(parent->rb_left, new); new 189 include/linux/rbtree_augmented.h rcu_assign_pointer(parent->rb_right, new); new 191 include/linux/rbtree_augmented.h rcu_assign_pointer(root->rb_node, new); new 195 include/linux/rbtree_augmented.h void (*augment_rotate)(struct rb_node *old, struct rb_node *new)); new 67 include/linux/rculist.h static inline void __list_add_rcu(struct list_head *new, new 70 include/linux/rculist.h if (!__list_add_valid(new, prev, next)) new 73 include/linux/rculist.h new->next = next; new 74 include/linux/rculist.h new->prev = prev; new 75 include/linux/rculist.h rcu_assign_pointer(list_next_rcu(prev), new); new 76 include/linux/rculist.h next->prev = new; new 95 include/linux/rculist.h static inline void list_add_rcu(struct list_head *new, struct list_head *head) new 97 include/linux/rculist.h __list_add_rcu(new, head, head->next); new 116 include/linux/rculist.h static inline void list_add_tail_rcu(struct list_head *new, new 119 include/linux/rculist.h __list_add_rcu(new, head->prev, head); new 189 include/linux/rculist.h struct list_head *new) new 191 include/linux/rculist.h new->next = old->next; new 192 include/linux/rculist.h new->prev = old->prev; new 193 include/linux/rculist.h rcu_assign_pointer(list_next_rcu(new->prev), new); new 194 include/linux/rculist.h new->next->prev = new; new 487 include/linux/rculist.h struct hlist_node *new) new 491 include/linux/rculist.h new->next = next; new 492 include/linux/rculist.h new->pprev = old->pprev; new 493 include/linux/rculist.h rcu_assign_pointer(*(struct hlist_node __rcu **)new->pprev, new); new 495 include/linux/rculist.h new->next->pprev = &new->next; new 244 include/linux/rmap.h void remove_migration_ptes(struct page *old, struct page *new, bool locked); new 507 include/linux/sbitmap.h int new = sbq_index_inc(old); new 508 include/linux/sbitmap.h atomic_cmpxchg(index, old, new); new 24 include/linux/sched/task_stack.h #define setup_thread_stack(new,old) do { } while(0) new 138 include/linux/security.h extern int cap_capset(struct cred *new, const struct cred *old, new 153 include/linux/security.h extern int cap_task_fix_setuid(struct cred *new, const struct cred *old, int flags); new 265 include/linux/security.h int security_capset(struct cred *new, const struct cred *old, new 313 include/linux/security.h struct cred *new); new 355 include/linux/security.h int security_inode_copy_up(struct dentry *src, struct cred **new); new 379 include/linux/security.h int security_prepare_creds(struct cred *new, const struct cred *old, gfp_t gfp); new 380 include/linux/security.h void security_transfer_creds(struct cred *new, const struct cred *old); new 382 include/linux/security.h int security_kernel_act_as(struct cred *new, u32 secid); new 383 include/linux/security.h int security_kernel_create_files_as(struct cred *new, struct inode *inode); new 389 include/linux/security.h int security_task_fix_setuid(struct cred *new, const struct cred *old, new 527 include/linux/security.h static inline int security_capset(struct cred *new, new 533 include/linux/security.h return cap_capset(new, old, effective, inheritable, permitted); new 708 include/linux/security.h struct cred *new) new 873 include/linux/security.h static inline int security_inode_copy_up(struct dentry *src, struct cred **new) new 976 include/linux/security.h static inline int security_prepare_creds(struct cred *new, new 983 include/linux/security.h static inline void security_transfer_creds(struct cred *new, new 1022 include/linux/security.h static inline int security_task_fix_setuid(struct cred *new, new 1026 include/linux/security.h return cap_task_fix_setuid(new, old, flags); new 33 include/linux/serial_8250.h struct ktermios *new, new 52 include/linux/serial_core.h void (*set_termios)(struct uart_port *, struct ktermios *new, new 112 include/linux/serial_core.h struct ktermios *new, new 1108 include/linux/skbuff.h void skb_copy_header(struct sk_buff *new, const struct sk_buff *old); new 178 include/linux/string.h char *strreplace(char *s, char old, char new); new 91 include/linux/sunrpc/cache.h int (*match)(struct cache_head *orig, struct cache_head *new); new 92 include/linux/sunrpc/cache.h void (*init)(struct cache_head *orig, struct cache_head *new); new 93 include/linux/sunrpc/cache.h void (*update)(struct cache_head *orig, struct cache_head *new); new 174 include/linux/sunrpc/cache.h struct cache_head *new, struct cache_head *old, int hash); new 163 include/linux/sunrpc/svcauth.h extern struct auth_domain *auth_domain_lookup(char *name, struct auth_domain *new); new 10 include/linux/swap_cgroup.h unsigned short old, unsigned short new); new 249 include/linux/swapops.h struct page *new); new 283 include/linux/swapops.h struct page *new) new 633 include/linux/syscalls.h struct old_itimerspec32 __user *new, new 1050 include/linux/syscalls.h asmlinkage long sys_symlink(const char __user *old, const char __user *new); new 539 include/linux/tty.h extern void tty_termios_copy_hw(struct ktermios *new, struct ktermios *old); new 231 include/linux/uio.h const void *dup_iter(struct iov_iter *new, struct iov_iter *old, gfp_t flags); new 29 include/linux/umh.h int (*init)(struct subprocess_info *info, struct cred *new); new 40 include/linux/umh.h int (*init)(struct subprocess_info *info, struct cred *new), new 44 include/linux/umh.h int (*init)(struct subprocess_info *info, struct cred *new), new 53 include/linux/usb/gadget_configfs.h struct struct_in *new; \ new 57 include/linux/usb/gadget_configfs.h new = kzalloc(sizeof(*new), GFP_KERNEL); \ new 58 include/linux/usb/gadget_configfs.h if (!new) \ new 61 include/linux/usb/gadget_configfs.h ret = check_user_usb_string(name, &new->stringtab_dev); \ new 64 include/linux/usb/gadget_configfs.h config_group_init_type_name(&new->group, name, \ new 70 include/linux/usb/gadget_configfs.h if (gs->stringtab_dev.language == new->stringtab_dev.language) \ new 78 include/linux/usb/gadget_configfs.h list_add_tail(&new->list, &gi->string_list); \ new 79 include/linux/usb/gadget_configfs.h return &new->group; \ new 81 include/linux/usb/gadget_configfs.h kfree(new); \ new 115 include/linux/user_namespace.h extern int create_user_ns(struct cred *new); new 146 include/linux/user_namespace.h static inline int create_user_ns(struct cred *new) new 17 include/linux/vt.h extern int vt_kmsg_redirect(int new); new 21 include/linux/vt.h static inline int vt_kmsg_redirect(int new) new 130 include/linux/vt_kern.h void vt_event_post(unsigned int event, unsigned int old, unsigned int new); new 83 include/media/drv-intf/cx2341x.h const struct cx2341x_mpeg_params *new); new 1080 include/media/v4l2-ctrls.h void v4l2_ctrl_replace(struct v4l2_event *old, const struct v4l2_event *new); new 1091 include/media/v4l2-ctrls.h void v4l2_ctrl_merge(const struct v4l2_event *old, struct v4l2_event *new); new 126 include/media/v4l2-dev.h enum v4l2_priority new); new 49 include/media/v4l2-event.h void (*replace)(struct v4l2_event *old, const struct v4l2_event *new); new 50 include/media/v4l2-event.h void (*merge)(const struct v4l2_event *old, struct v4l2_event *new); new 137 include/net/checksum.h static inline void csum_replace2(__sum16 *sum, __be16 old, __be16 new) new 139 include/net/checksum.h *sum = ~csum16_add(csum16_sub(~(*sum), old), new); new 441 include/net/ip_fib.h void fib_nhc_update_mtu(struct fib_nh_common *nhc, u32 new, u32 orig); new 1545 include/net/ip_vs.h static inline __wsum ip_vs_check_diff4(__be32 old, __be32 new, __wsum oldsum) new 1547 include/net/ip_vs.h __be32 diff[2] = { ~old, new }; new 1553 include/net/ip_vs.h static inline __wsum ip_vs_check_diff16(const __be32 *old, const __be32 *new, new 1557 include/net/ip_vs.h new[3], new[2], new[1], new[0] }; new 1563 include/net/ip_vs.h static inline __wsum ip_vs_check_diff2(__be16 old, __be16 new, __wsum oldsum) new 1565 include/net/ip_vs.h __be16 diff[2] = { ~old, new }; new 477 include/net/ndisc.h const u8 *lladdr, u8 new, u32 flags, u8 icmp6_type, new 327 include/net/neighbour.h int neigh_update(struct neighbour *neigh, const u8 *lladdr, u8 new, u32 flags, new 20 include/net/netevent.h struct dst_entry *new; new 30 include/net/netfilter/nf_conntrack_expect.h void (*expectfn)(struct nf_conn *new, new 328 include/net/netfilter/nf_tables.h void *(*new)(struct nft_set *, new 648 include/net/sch_generic.h struct Qdisc *new, struct Qdisc *old, new 662 include/net/sch_generic.h struct Qdisc *new, struct Qdisc *old, new 1152 include/net/sch_generic.h static inline struct Qdisc *qdisc_replace(struct Qdisc *sch, struct Qdisc *new, new 1159 include/net/sch_generic.h *pold = new; new 56 include/net/sctp/checksum.h __wsum new; new 59 include/net/sctp/checksum.h new = ~__skb_checksum(skb, offset, skb->len - offset, ~(__wsum)0, new 63 include/net/sctp/checksum.h return cpu_to_le32((__force __u32)new); new 388 include/net/sctp/structs.h void sctp_stream_update(struct sctp_stream *stream, struct sctp_stream *new); new 2131 include/net/sctp/structs.h struct sctp_association *new); new 1801 include/net/tcp.h static inline void tcp_insert_write_queue_before(struct sk_buff *new, new 1805 include/net/tcp.h __skb_queue_before(&sk->sk_write_queue, skb, new); new 1871 include/net/tcp.h struct sk_buff *new) new 1874 include/net/tcp.h tcp_sk(sk)->highest_sack = new; new 4520 include/rdma/ib_verbs.h const struct rdma_ah_attr *new); new 614 include/sound/soc-dapm.h unsigned char new:1; /* cnew complete */ new 158 include/sound/soc-dpcm.h int stream, struct snd_soc_dapm_widget_list **list, int new); new 92 include/trace/events/filemap.h __field(errseq_t, new) new 105 include/trace/events/filemap.h __entry->new = file->f_wb_err; new 111 include/trace/events/filemap.h __entry->new) new 372 include/trace/events/kvm.h TP_PROTO(bool grow, unsigned int vcpu_id, unsigned int new, new 374 include/trace/events/kvm.h TP_ARGS(grow, vcpu_id, new, old), new 379 include/trace/events/kvm.h __field(unsigned int, new) new 386 include/trace/events/kvm.h __entry->new = new; new 392 include/trace/events/kvm.h __entry->new, new 397 include/trace/events/kvm.h #define trace_kvm_halt_poll_ns_grow(vcpu_id, new, old) \ new 398 include/trace/events/kvm.h trace_kvm_halt_poll_ns(true, vcpu_id, new, old) new 399 include/trace/events/kvm.h #define trace_kvm_halt_poll_ns_shrink(vcpu_id, new, old) \ new 400 include/trace/events/kvm.h trace_kvm_halt_poll_ns(false, vcpu_id, new, old) new 74 include/trace/events/neigh.h TP_PROTO(struct neighbour *n, const u8 *lladdr, u8 new, new 77 include/trace/events/neigh.h TP_ARGS(n, lladdr, new, flags, nlmsg_pid), new 136 include/trace/events/neigh.h __entry->new_state = new; new 761 include/xen/interface/xen.h } new; new 48 init/do_mounts_initrd.c static int init_linuxrc(struct subprocess_info *info, struct cred *new) new 1368 ipc/mqueue.c static int do_mq_getsetattr(int mqdes, struct mq_attr *new, struct mq_attr *old) new 1374 ipc/mqueue.c if (new && (new->mq_flags & (~O_NONBLOCK))) new 1395 ipc/mqueue.c if (new) { new 1396 ipc/mqueue.c audit_mq_getsetattr(mqdes, new); new 1398 ipc/mqueue.c if (new->mq_flags & O_NONBLOCK) new 1418 ipc/mqueue.c struct mq_attr *new = NULL, *old = NULL; new 1421 ipc/mqueue.c new = &mqstat; new 1422 ipc/mqueue.c if (copy_from_user(new, u_mqstat, sizeof(struct mq_attr))) new 1428 ipc/mqueue.c ret = do_mq_getsetattr(mqdes, new, old); new 1511 ipc/mqueue.c struct mq_attr *new = NULL, *old = NULL; new 1514 ipc/mqueue.c new = &mqstat; new 1515 ipc/mqueue.c if (get_compat_mq_attr(new, u_mqstat)) new 1521 ipc/mqueue.c ret = do_mq_getsetattr(mqdes, new, old); new 180 ipc/namespace.c static int ipcns_install(struct nsproxy *nsproxy, struct ns_common *new) new 182 ipc/namespace.c struct ipc_namespace *ns = to_ipc_ns(new); new 1892 ipc/sem.c struct sem_undo *un, *new; new 1923 ipc/sem.c new = kzalloc(sizeof(struct sem_undo) + sizeof(short)*nsems, GFP_KERNEL); new 1924 ipc/sem.c if (!new) { new 1935 ipc/sem.c kfree(new); new 1946 ipc/sem.c kfree(new); new 1950 ipc/sem.c new->semadj = (short *) &new[1]; new 1951 ipc/sem.c new->ulp = ulp; new 1952 ipc/sem.c new->semid = semid; new 1954 ipc/sem.c list_add_rcu(&new->list_proc, &ulp->list_proc); new 1956 ipc/sem.c list_add(&new->list_id, &sma->list_id); new 1957 ipc/sem.c un = new; new 460 ipc/shm.c static int shm_set_policy(struct vm_area_struct *vma, struct mempolicy *new) new 467 ipc/shm.c err = sfd->vm_ops->set_policy(vma, new); new 201 ipc/util.c static inline int ipc_idr_alloc(struct ipc_ids *ids, struct kern_ipc_perm *new) new 245 ipc/util.c new->seq = ids->seq; new 250 ipc/util.c idr_replace(&ids->ipcs_idr, new, idx); new 253 ipc/util.c new->seq = ipcid_to_seqx(next_id); new 254 ipc/util.c idx = idr_alloc(&ids->ipcs_idr, new, ipcid_to_idx(next_id), new 258 ipc/util.c new->id = (new->seq << ipcmni_seq_shift()) + idx; new 277 ipc/util.c int ipc_addid(struct ipc_ids *ids, struct kern_ipc_perm *new, int limit) new 284 ipc/util.c refcount_set(&new->refcount, 1); new 294 ipc/util.c spin_lock_init(&new->lock); new 296 ipc/util.c spin_lock(&new->lock); new 299 ipc/util.c new->cuid = new->uid = euid; new 300 ipc/util.c new->gid = new->cgid = egid; new 302 ipc/util.c new->deleted = false; new 304 ipc/util.c idx = ipc_idr_alloc(ids, new); new 307 ipc/util.c if (idx >= 0 && new->key != IPC_PRIVATE) { new 308 ipc/util.c err = rhashtable_insert_fast(&ids->key_ht, &new->khtnode, new 316 ipc/util.c new->deleted = true; new 317 ipc/util.c spin_unlock(&new->lock); new 380 kernel/audit.c static int audit_log_config_change(char *function_name, u32 new, u32 old, new 389 kernel/audit.c audit_log_format(ab, "op=set %s=%u old=%u ", function_name, new, old); new 399 kernel/audit.c static int audit_do_config_change(char *function_name, u32 *to_change, u32 new) new 411 kernel/audit.c rc = audit_log_config_change(function_name, new, old, allow_changes); new 418 kernel/audit.c *to_change = new; new 1388 kernel/audit.c char *old, *new; new 1401 kernel/audit.c new = audit_unpack_string(&bufp, &msglen, sizes[1]); new 1402 kernel/audit.c if (IS_ERR(new)) { new 1403 kernel/audit.c err = PTR_ERR(new); new 1408 kernel/audit.c err = audit_tag_tree(old, new); new 1415 kernel/audit.c audit_log_untrustedstring(ab, new); new 1419 kernel/audit.c kfree(new); new 272 kernel/audit.h extern int audit_dupe_exe(struct audit_krule *new, struct audit_krule *old); new 284 kernel/audit.h extern int audit_tag_tree(char *old, char *new); new 316 kernel/audit.h #define audit_tag_tree(old, new) -EINVAL new 295 kernel/audit_tree.c static void replace_chunk(struct audit_chunk *new, struct audit_chunk *old) new 300 kernel/audit_tree.c new->key = old->key; new 301 kernel/audit_tree.c list_splice_init(&old->trees, &new->trees); new 302 kernel/audit_tree.c list_for_each_entry(owner, &new->trees, same_root) new 303 kernel/audit_tree.c owner->root = new; new 310 kernel/audit_tree.c new->owners[i].owner = owner; new 311 kernel/audit_tree.c new->owners[i].index = old->owners[j].index - j + i; new 315 kernel/audit_tree.c list_replace_init(&old->owners[j].list, &new->owners[i].list); new 317 kernel/audit_tree.c replace_mark_chunk(old->mark, new); new 324 kernel/audit_tree.c list_replace_rcu(&old->hash, &new->hash); new 353 kernel/audit_tree.c struct audit_chunk *new; new 379 kernel/audit_tree.c new = alloc_chunk(size); new 380 kernel/audit_tree.c if (!new) new 388 kernel/audit_tree.c replace_chunk(new, chunk); new 872 kernel/audit_tree.c int audit_tag_tree(char *old, char *new) new 880 kernel/audit_tree.c err = kern_path(new, 0, &path2); new 205 kernel/audit_watch.c struct audit_watch *new; new 211 kernel/audit_watch.c new = audit_init_watch(path); new 212 kernel/audit_watch.c if (IS_ERR(new)) { new 217 kernel/audit_watch.c new->dev = old->dev; new 218 kernel/audit_watch.c new->ino = old->ino; new 220 kernel/audit_watch.c new->parent = old->parent; new 223 kernel/audit_watch.c return new; new 522 kernel/audit_watch.c int audit_dupe_exe(struct audit_krule *new, struct audit_krule *old) new 531 kernel/audit_watch.c audit_mark = audit_alloc_mark(new, pathname, strlen(pathname)); new 536 kernel/audit_watch.c new->exe = audit_mark; new 815 kernel/auditfilter.c struct audit_krule *new; new 823 kernel/auditfilter.c new = &entry->rule; new 824 kernel/auditfilter.c new->flags = old->flags; new 825 kernel/auditfilter.c new->pflags = old->pflags; new 826 kernel/auditfilter.c new->listnr = old->listnr; new 827 kernel/auditfilter.c new->action = old->action; new 829 kernel/auditfilter.c new->mask[i] = old->mask[i]; new 830 kernel/auditfilter.c new->prio = old->prio; new 831 kernel/auditfilter.c new->buflen = old->buflen; new 832 kernel/auditfilter.c new->inode_f = old->inode_f; new 833 kernel/auditfilter.c new->field_count = old->field_count; new 842 kernel/auditfilter.c new->tree = old->tree; new 843 kernel/auditfilter.c memcpy(new->fields, old->fields, sizeof(struct audit_field) * fcount); new 848 kernel/auditfilter.c switch (new->fields[i].type) { new 859 kernel/auditfilter.c err = audit_dupe_lsm_field(&new->fields[i], new 867 kernel/auditfilter.c new->filterkey = fk; new 870 kernel/auditfilter.c err = audit_dupe_exe(new, old); new 874 kernel/auditfilter.c if (new->exe) new 875 kernel/auditfilter.c audit_remove_mark(new->exe); new 883 kernel/auditfilter.c new->watch = old->watch; new 2442 kernel/auditsc.c const struct cred *new, const struct cred *old) new 2469 kernel/auditsc.c ax->new_pcap.permitted = new->cap_permitted; new 2470 kernel/auditsc.c ax->new_pcap.inheritable = new->cap_inheritable; new 2471 kernel/auditsc.c ax->new_pcap.effective = new->cap_effective; new 2472 kernel/auditsc.c ax->new_pcap.ambient = new->cap_ambient; new 2484 kernel/auditsc.c void __audit_log_capset(const struct cred *new, const struct cred *old) new 2488 kernel/auditsc.c context->capset.cap.effective = new->cap_effective; new 2489 kernel/auditsc.c context->capset.cap.inheritable = new->cap_effective; new 2490 kernel/auditsc.c context->capset.cap.permitted = new->cap_permitted; new 2491 kernel/auditsc.c context->capset.cap.ambient = new->cap_ambient; new 88 kernel/bpf/local_storage.c struct rb_node **new = &(root->rb_node), *parent = NULL; new 90 kernel/bpf/local_storage.c while (*new) { new 93 kernel/bpf/local_storage.c this = container_of(*new, struct bpf_cgroup_storage, node); new 95 kernel/bpf/local_storage.c parent = *new; new 98 kernel/bpf/local_storage.c new = &((*new)->rb_left); new 101 kernel/bpf/local_storage.c new = &((*new)->rb_right); new 108 kernel/bpf/local_storage.c rb_link_node(&storage->node, parent, new); new 132 kernel/bpf/local_storage.c struct bpf_storage_buffer *new; new 154 kernel/bpf/local_storage.c new = kmalloc_node(sizeof(struct bpf_storage_buffer) + new 158 kernel/bpf/local_storage.c if (!new) new 161 kernel/bpf/local_storage.c memcpy(&new->data[0], value, map->value_size); new 162 kernel/bpf/local_storage.c check_and_init_map_lock(map, new->data); new 164 kernel/bpf/local_storage.c new = xchg(&storage->buf, new); new 165 kernel/bpf/local_storage.c kfree_rcu(new, rcu); new 7311 kernel/bpf/verifier.c struct bpf_verifier_state *cur = env->cur_state, *new; new 7463 kernel/bpf/verifier.c new = &new_sl->state; new 7464 kernel/bpf/verifier.c err = copy_verifier_state(new, cur); new 7466 kernel/bpf/verifier.c free_verifier_state(new, false); new 7470 kernel/bpf/verifier.c new->insn_idx = insn_idx; new 7471 kernel/bpf/verifier.c WARN_ONCE(new->branches != 1, new 7472 kernel/bpf/verifier.c "BUG is_state_visited:branches_to_explore=%d insn %d\n", new->branches, insn_idx); new 7474 kernel/bpf/verifier.c cur->parent = new; new 7494 kernel/bpf/verifier.c cur->frame[j]->regs[i].parent = &new->frame[j]->regs[i]; new 7502 kernel/bpf/verifier.c struct bpf_func_state *newframe = new->frame[j]; new 227 kernel/capability.c struct cred *new; new 265 kernel/capability.c new = prepare_creds(); new 266 kernel/capability.c if (!new) new 269 kernel/capability.c ret = security_capset(new, current_cred(), new 274 kernel/capability.c audit_log_capset(new, current_cred()); new 276 kernel/capability.c return commit_creds(new); new 279 kernel/capability.c abort_creds(new); new 3628 kernel/cgroup/cgroup.c struct psi_trigger *new; new 3638 kernel/cgroup/cgroup.c new = psi_trigger_create(&cgrp->psi, buf, nbytes, res); new 3639 kernel/cgroup/cgroup.c if (IS_ERR(new)) { new 3641 kernel/cgroup/cgroup.c return PTR_ERR(new); new 3644 kernel/cgroup/cgroup.c psi_trigger_replace(&of->priv, new); new 148 kernel/cgroup/pids.c int64_t new = atomic64_add_return(num, &p->counter); new 156 kernel/cgroup/pids.c if (new > limit) new 265 kernel/cgroup/rdma.c s64 new; new 284 kernel/cgroup/rdma.c new = rpool->resources[index].usage + 1; new 285 kernel/cgroup/rdma.c if (new > rpool->resources[index].max) { new 289 kernel/cgroup/rdma.c rpool->resources[index].usage = new; new 215 kernel/cred.c struct cred *new; new 217 kernel/cred.c new = kmem_cache_zalloc(cred_jar, GFP_KERNEL); new 218 kernel/cred.c if (!new) new 221 kernel/cred.c atomic_set(&new->usage, 1); new 223 kernel/cred.c new->magic = CRED_MAGIC; new 226 kernel/cred.c if (security_cred_alloc_blank(new, GFP_KERNEL_ACCOUNT) < 0) new 229 kernel/cred.c return new; new 232 kernel/cred.c abort_creds(new); new 254 kernel/cred.c struct cred *new; new 258 kernel/cred.c new = kmem_cache_alloc(cred_jar, GFP_KERNEL); new 259 kernel/cred.c if (!new) new 262 kernel/cred.c kdebug("prepare_creds() alloc %p", new); new 265 kernel/cred.c memcpy(new, old, sizeof(struct cred)); new 267 kernel/cred.c new->non_rcu = 0; new 268 kernel/cred.c atomic_set(&new->usage, 1); new 269 kernel/cred.c set_cred_subscribers(new, 0); new 270 kernel/cred.c get_group_info(new->group_info); new 271 kernel/cred.c get_uid(new->user); new 272 kernel/cred.c get_user_ns(new->user_ns); new 275 kernel/cred.c key_get(new->session_keyring); new 276 kernel/cred.c key_get(new->process_keyring); new 277 kernel/cred.c key_get(new->thread_keyring); new 278 kernel/cred.c key_get(new->request_key_auth); new 282 kernel/cred.c new->security = NULL; new 285 kernel/cred.c if (security_prepare_creds(new, old, GFP_KERNEL_ACCOUNT) < 0) new 287 kernel/cred.c validate_creds(new); new 288 kernel/cred.c return new; new 291 kernel/cred.c abort_creds(new); new 302 kernel/cred.c struct cred *new; new 304 kernel/cred.c new = prepare_creds(); new 305 kernel/cred.c if (!new) new 306 kernel/cred.c return new; new 310 kernel/cred.c key_put(new->thread_keyring); new 311 kernel/cred.c new->thread_keyring = NULL; new 314 kernel/cred.c key_put(new->process_keyring); new 315 kernel/cred.c new->process_keyring = NULL; new 318 kernel/cred.c return new; new 332 kernel/cred.c struct cred *new; new 355 kernel/cred.c new = prepare_creds(); new 356 kernel/cred.c if (!new) new 360 kernel/cred.c ret = create_user_ns(new); new 368 kernel/cred.c if (new->thread_keyring) { new 369 kernel/cred.c key_put(new->thread_keyring); new 370 kernel/cred.c new->thread_keyring = NULL; new 372 kernel/cred.c install_thread_keyring_to_cred(new); new 379 kernel/cred.c key_put(new->process_keyring); new 380 kernel/cred.c new->process_keyring = NULL; new 384 kernel/cred.c atomic_inc(&new->user->processes); new 385 kernel/cred.c p->cred = p->real_cred = get_cred(new); new 386 kernel/cred.c alter_cred_subscribers(new, 2); new 387 kernel/cred.c validate_creds(new); new 391 kernel/cred.c put_cred(new); new 434 kernel/cred.c int commit_creds(struct cred *new) new 439 kernel/cred.c kdebug("commit_creds(%p{%d,%d})", new, new 440 kernel/cred.c atomic_read(&new->usage), new 441 kernel/cred.c read_cred_subscribers(new)); new 447 kernel/cred.c validate_creds(new); new 449 kernel/cred.c BUG_ON(atomic_read(&new->usage) < 1); new 451 kernel/cred.c get_cred(new); /* we will require a ref for the subj creds too */ new 454 kernel/cred.c if (!uid_eq(old->euid, new->euid) || new 455 kernel/cred.c !gid_eq(old->egid, new->egid) || new 456 kernel/cred.c !uid_eq(old->fsuid, new->fsuid) || new 457 kernel/cred.c !gid_eq(old->fsgid, new->fsgid) || new 458 kernel/cred.c !cred_cap_issubset(old, new)) { new 475 kernel/cred.c if (!uid_eq(new->fsuid, old->fsuid)) new 476 kernel/cred.c key_fsuid_changed(new); new 477 kernel/cred.c if (!gid_eq(new->fsgid, old->fsgid)) new 478 kernel/cred.c key_fsgid_changed(new); new 484 kernel/cred.c alter_cred_subscribers(new, 2); new 485 kernel/cred.c if (new->user != old->user) new 486 kernel/cred.c atomic_inc(&new->user->processes); new 487 kernel/cred.c rcu_assign_pointer(task->real_cred, new); new 488 kernel/cred.c rcu_assign_pointer(task->cred, new); new 489 kernel/cred.c if (new->user != old->user) new 494 kernel/cred.c if (!uid_eq(new->uid, old->uid) || new 495 kernel/cred.c !uid_eq(new->euid, old->euid) || new 496 kernel/cred.c !uid_eq(new->suid, old->suid) || new 497 kernel/cred.c !uid_eq(new->fsuid, old->fsuid)) new 500 kernel/cred.c if (!gid_eq(new->gid, old->gid) || new 501 kernel/cred.c !gid_eq(new->egid, old->egid) || new 502 kernel/cred.c !gid_eq(new->sgid, old->sgid) || new 503 kernel/cred.c !gid_eq(new->fsgid, old->fsgid)) new 520 kernel/cred.c void abort_creds(struct cred *new) new 522 kernel/cred.c kdebug("abort_creds(%p{%d,%d})", new, new 523 kernel/cred.c atomic_read(&new->usage), new 524 kernel/cred.c read_cred_subscribers(new)); new 527 kernel/cred.c BUG_ON(read_cred_subscribers(new) != 0); new 529 kernel/cred.c BUG_ON(atomic_read(&new->usage) < 1); new 530 kernel/cred.c put_cred(new); new 541 kernel/cred.c const struct cred *override_creds(const struct cred *new) new 545 kernel/cred.c kdebug("override_creds(%p{%d,%d})", new, new 546 kernel/cred.c atomic_read(&new->usage), new 547 kernel/cred.c read_cred_subscribers(new)); new 550 kernel/cred.c validate_creds(new); new 563 kernel/cred.c get_new_cred((struct cred *)new); new 564 kernel/cred.c alter_cred_subscribers(new, 1); new 565 kernel/cred.c rcu_assign_pointer(current->cred, new); new 684 kernel/cred.c struct cred *new; new 686 kernel/cred.c new = kmem_cache_alloc(cred_jar, GFP_KERNEL); new 687 kernel/cred.c if (!new) new 690 kernel/cred.c kdebug("prepare_kernel_cred() alloc %p", new); new 699 kernel/cred.c *new = *old; new 700 kernel/cred.c new->non_rcu = 0; new 701 kernel/cred.c atomic_set(&new->usage, 1); new 702 kernel/cred.c set_cred_subscribers(new, 0); new 703 kernel/cred.c get_uid(new->user); new 704 kernel/cred.c get_user_ns(new->user_ns); new 705 kernel/cred.c get_group_info(new->group_info); new 708 kernel/cred.c new->session_keyring = NULL; new 709 kernel/cred.c new->process_keyring = NULL; new 710 kernel/cred.c new->thread_keyring = NULL; new 711 kernel/cred.c new->request_key_auth = NULL; new 712 kernel/cred.c new->jit_keyring = KEY_REQKEY_DEFL_THREAD_KEYRING; new 716 kernel/cred.c new->security = NULL; new 718 kernel/cred.c if (security_prepare_creds(new, old, GFP_KERNEL_ACCOUNT) < 0) new 722 kernel/cred.c validate_creds(new); new 723 kernel/cred.c return new; new 726 kernel/cred.c put_cred(new); new 740 kernel/cred.c int set_security_override(struct cred *new, u32 secid) new 742 kernel/cred.c return security_kernel_act_as(new, secid); new 756 kernel/cred.c int set_security_override_from_ctx(struct cred *new, const char *secctx) new 765 kernel/cred.c return set_security_override(new, secid); new 778 kernel/cred.c int set_create_files_as(struct cred *new, struct inode *inode) new 782 kernel/cred.c new->fsuid = inode->i_uid; new 783 kernel/cred.c new->fsgid = inode->i_gid; new 784 kernel/cred.c return security_kernel_create_files_as(new, inode); new 2688 kernel/debug/kdb/kdb_main.c kdbtab_t *new = kmalloc_array(kdb_max_commands - new 2691 kernel/debug/kdb/kdb_main.c sizeof(*new), new 2693 kernel/debug/kdb/kdb_main.c if (!new) { new 2699 kernel/debug/kdb/kdb_main.c memcpy(new, kdb_commands, new 2700 kernel/debug/kdb/kdb_main.c (kdb_max_commands - KDB_BASE_CMD_MAX) * sizeof(*new)); new 2703 kernel/debug/kdb/kdb_main.c memset(new + kdb_max_commands - KDB_BASE_CMD_MAX, 0, new 2704 kernel/debug/kdb/kdb_main.c kdb_command_extend * sizeof(*new)); new 2705 kernel/debug/kdb/kdb_main.c kdb_commands = new; new 7085 kernel/events/core.c int new) new 7099 kernel/events/core.c .type = new ? PERF_RECORD_FORK : PERF_RECORD_EXIT, new 1042 kernel/events/uprobes.c register_for_each_vma(struct uprobe *uprobe, struct uprobe_consumer *new) new 1044 kernel/events/uprobes.c bool is_register = !!new; new 1075 kernel/events/uprobes.c if (consumer_filter(new, new 356 kernel/fork.c struct vm_area_struct *new = kmem_cache_alloc(vm_area_cachep, GFP_KERNEL); new 358 kernel/fork.c if (new) { new 359 kernel/fork.c *new = *orig; new 360 kernel/fork.c INIT_LIST_HEAD(&new->anon_vma_chain); new 362 kernel/fork.c return new; new 536 kernel/futex.c u64 new = atomic64_add_return(1, &i_seq); new 537 kernel/futex.c if (WARN_ON_ONCE(!new)) new 540 kernel/futex.c old = atomic64_cmpxchg_relaxed(&inode->i_sequence, 0, new); new 543 kernel/futex.c return new; new 123 kernel/groups.c void set_groups(struct cred *new, struct group_info *group_info) new 125 kernel/groups.c put_group_info(new->group_info); new 127 kernel/groups.c new->group_info = group_info; new 141 kernel/groups.c struct cred *new; new 143 kernel/groups.c new = prepare_creds(); new 144 kernel/groups.c if (!new) new 147 kernel/groups.c set_groups(new, group_info); new 148 kernel/groups.c return commit_creds(new); new 1135 kernel/irq/manage.c static int irq_setup_forced_threading(struct irqaction *new) new 1139 kernel/irq/manage.c if (new->flags & (IRQF_NO_THREAD | IRQF_PERCPU | IRQF_ONESHOT)) new 1146 kernel/irq/manage.c if (new->handler == irq_default_primary_handler) new 1149 kernel/irq/manage.c new->flags |= IRQF_ONESHOT; new 1156 kernel/irq/manage.c if (new->handler && new->thread_fn) { new 1158 kernel/irq/manage.c new->secondary = kzalloc(sizeof(struct irqaction), GFP_KERNEL); new 1159 kernel/irq/manage.c if (!new->secondary) new 1161 kernel/irq/manage.c new->secondary->handler = irq_forced_secondary_handler; new 1162 kernel/irq/manage.c new->secondary->thread_fn = new->thread_fn; new 1163 kernel/irq/manage.c new->secondary->dev_id = new->dev_id; new 1164 kernel/irq/manage.c new->secondary->irq = new->irq; new 1165 kernel/irq/manage.c new->secondary->name = new->name; new 1168 kernel/irq/manage.c set_bit(IRQTF_FORCED_THREAD, &new->thread_flags); new 1169 kernel/irq/manage.c new->thread_fn = new->handler; new 1170 kernel/irq/manage.c new->handler = irq_default_primary_handler; new 1225 kernel/irq/manage.c setup_irq_thread(struct irqaction *new, unsigned int irq, bool secondary) new 1233 kernel/irq/manage.c t = kthread_create(irq_thread, new, "irq/%d-%s", irq, new 1234 kernel/irq/manage.c new->name); new 1236 kernel/irq/manage.c t = kthread_create(irq_thread, new, "irq/%d-s-%s", irq, new 1237 kernel/irq/manage.c new->name); new 1251 kernel/irq/manage.c new->thread = get_task_struct(t); new 1261 kernel/irq/manage.c set_bit(IRQTF_AFFINITY, &new->thread_flags); new 1280 kernel/irq/manage.c __setup_irq(unsigned int irq, struct irq_desc *desc, struct irqaction *new) new 1294 kernel/irq/manage.c new->irq = irq; new 1300 kernel/irq/manage.c if (!(new->flags & IRQF_TRIGGER_MASK)) new 1301 kernel/irq/manage.c new->flags |= irqd_get_trigger_type(&desc->irq_data); new 1309 kernel/irq/manage.c if (!new->thread_fn) { new 1318 kernel/irq/manage.c new->handler = irq_nested_primary_handler; new 1321 kernel/irq/manage.c ret = irq_setup_forced_threading(new); new 1332 kernel/irq/manage.c if (new->thread_fn && !nested) { new 1333 kernel/irq/manage.c ret = setup_irq_thread(new, irq, false); new 1336 kernel/irq/manage.c if (new->secondary) { new 1337 kernel/irq/manage.c ret = setup_irq_thread(new->secondary, irq, true); new 1353 kernel/irq/manage.c new->flags &= ~IRQF_ONESHOT; new 1376 kernel/irq/manage.c new->name, irq, desc->irq_data.chip->name); new 1403 kernel/irq/manage.c new->name, irq, desc->irq_data.chip->name); new 1415 kernel/irq/manage.c oldtype = new->flags & IRQF_TRIGGER_MASK; new 1419 kernel/irq/manage.c if (!((old->flags & new->flags) & IRQF_SHARED) || new 1420 kernel/irq/manage.c (oldtype != (new->flags & IRQF_TRIGGER_MASK)) || new 1421 kernel/irq/manage.c ((old->flags ^ new->flags) & IRQF_ONESHOT)) new 1426 kernel/irq/manage.c (new->flags & IRQF_PERCPU)) new 1448 kernel/irq/manage.c if (new->flags & IRQF_ONESHOT) { new 1477 kernel/irq/manage.c new->thread_mask = 1UL << ffz(thread_mask); new 1479 kernel/irq/manage.c } else if (new->handler == irq_default_primary_handler && new 1506 kernel/irq/manage.c if (new->flags & IRQF_TRIGGER_MASK) { new 1508 kernel/irq/manage.c new->flags & IRQF_TRIGGER_MASK); new 1533 kernel/irq/manage.c if (new->flags & IRQF_PERCPU) { new 1538 kernel/irq/manage.c if (new->flags & IRQF_ONESHOT) new 1542 kernel/irq/manage.c if (new->flags & IRQF_NOBALANCING) { new 1556 kernel/irq/manage.c WARN_ON_ONCE(new->flags & IRQF_SHARED); new 1561 kernel/irq/manage.c } else if (new->flags & IRQF_TRIGGER_MASK) { new 1562 kernel/irq/manage.c unsigned int nmsk = new->flags & IRQF_TRIGGER_MASK; new 1571 kernel/irq/manage.c *old_ptr = new; new 1573 kernel/irq/manage.c irq_pm_install_action(desc, new); new 1592 kernel/irq/manage.c irq_setup_timings(desc, new); new 1598 kernel/irq/manage.c if (new->thread) new 1599 kernel/irq/manage.c wake_up_process(new->thread); new 1600 kernel/irq/manage.c if (new->secondary) new 1601 kernel/irq/manage.c wake_up_process(new->secondary->thread); new 1604 kernel/irq/manage.c new->dir = NULL; new 1605 kernel/irq/manage.c register_handler_proc(irq, new); new 1609 kernel/irq/manage.c if (!(new->flags & IRQF_PROBE_SHARED)) { new 1611 kernel/irq/manage.c irq, new->flags, new->name, old->flags, old->name); new 1628 kernel/irq/manage.c if (new->thread) { new 1629 kernel/irq/manage.c struct task_struct *t = new->thread; new 1631 kernel/irq/manage.c new->thread = NULL; new 1635 kernel/irq/manage.c if (new->secondary && new->secondary->thread) { new 1636 kernel/irq/manage.c struct task_struct *t = new->secondary->thread; new 1638 kernel/irq/manage.c new->secondary->thread = NULL; new 229 kernel/locking/mutex.c unsigned long old, new; new 236 kernel/locking/mutex.c new = (owner & MUTEX_FLAG_WAITERS); new 237 kernel/locking/mutex.c new |= (unsigned long)task; new 239 kernel/locking/mutex.c new |= MUTEX_FLAG_PICKUP; new 241 kernel/locking/mutex.c old = atomic_long_cmpxchg_release(&lock->owner, owner, new); new 221 kernel/locking/qspinlock.c u32 old, new, val = atomic_read(&lock->val); new 224 kernel/locking/qspinlock.c new = (val & _Q_LOCKED_PENDING_MASK) | tail; new 230 kernel/locking/qspinlock.c old = atomic_cmpxchg_relaxed(&lock->val, val, new); new 136 kernel/locking/qspinlock_paravirt.h int old, new; new 145 kernel/locking/qspinlock_paravirt.h new = (val & ~_Q_PENDING_MASK) | _Q_LOCKED_VAL; new 146 kernel/locking/qspinlock_paravirt.h val = atomic_cmpxchg_acquire(&lock->val, old, new); new 566 kernel/locking/rwsem.c long count, new; new 577 kernel/locking/rwsem.c new = count; new 583 kernel/locking/rwsem.c new |= RWSEM_FLAG_HANDOFF; new 585 kernel/locking/rwsem.c new |= RWSEM_WRITER_LOCKED; new 586 kernel/locking/rwsem.c new &= ~RWSEM_FLAG_HANDOFF; new 589 kernel/locking/rwsem.c new &= ~RWSEM_FLAG_WAITERS; new 591 kernel/locking/rwsem.c } while (!atomic_long_try_cmpxchg_acquire(&sem->count, &count, new)); new 597 kernel/locking/rwsem.c if (new & RWSEM_FLAG_HANDOFF) new 720 kernel/locking/rwsem.c struct task_struct *new, *owner; new 737 kernel/locking/rwsem.c new = rwsem_owner_flags(sem, &new_flags); new 738 kernel/locking/rwsem.c if ((new != owner) || (new_flags != flags)) { new 739 kernel/locking/rwsem.c state = rwsem_owner_state(new, new_flags, nonspinnable); new 213 kernel/nsproxy.c void switch_task_namespaces(struct task_struct *p, struct nsproxy *new) new 221 kernel/nsproxy.c p->nsproxy = new; new 290 kernel/pid.c struct pid *new) new 299 kernel/pid.c *pid_ptr = new; new 321 kernel/pid.c void transfer_pid(struct task_struct *old, struct task_struct *new, new 325 kernel/pid.c new->thread_pid = old->thread_pid; new 326 kernel/pid.c hlist_replace_rcu(&old->pid_links[type], &new->pid_links[type]); new 386 kernel/pid_namespace.c struct pid_namespace *ancestor, *new = to_pid_ns(ns); new 388 kernel/pid_namespace.c if (!ns_capable(new->user_ns, CAP_SYS_ADMIN) || new 400 kernel/pid_namespace.c if (new->level < active->level) new 403 kernel/pid_namespace.c ancestor = new; new 410 kernel/pid_namespace.c nsproxy->pid_ns_for_children = get_pid_ns(new); new 132 kernel/power/swap.c struct rb_node **new = &(swsusp_extents.rb_node); new 137 kernel/power/swap.c while (*new) { new 138 kernel/power/swap.c ext = rb_entry(*new, struct swsusp_extent, node); new 139 kernel/power/swap.c parent = *new; new 146 kernel/power/swap.c new = &((*new)->rb_left); new 153 kernel/power/swap.c new = &((*new)->rb_right); new 166 kernel/power/swap.c rb_link_node(&ext->node, parent, new); new 123 kernel/rcu/rcu.h static inline bool rcu_seq_completed_gp(unsigned long old, unsigned long new) new 125 kernel/rcu/rcu.h return ULONG_CMP_LT(old, new & ~RCU_SEQ_STATE_MASK); new 131 kernel/rcu/rcu.h static inline bool rcu_seq_new_gp(unsigned long old, unsigned long new) new 134 kernel/rcu/rcu.h new); new 141 kernel/rcu/rcu.h static inline unsigned long rcu_seq_diff(unsigned long new, unsigned long old) new 145 kernel/rcu/rcu.h if (old == new) new 151 kernel/rcu/rcu.h rnd_diff = (new & ~RCU_SEQ_STATE_MASK) - new 153 kernel/rcu/rcu.h ((new & RCU_SEQ_STATE_MASK) || (old & RCU_SEQ_STATE_MASK)); new 135 kernel/rcu/rcuperf.c unsigned long (*gp_diff)(unsigned long new, unsigned long old); new 298 kernel/rcu/rcuperf.c static unsigned long rcuperf_seq_diff(unsigned long new, unsigned long old) new 301 kernel/rcu/rcuperf.c return new - old; new 302 kernel/rcu/rcuperf.c return cur_ops->gp_diff(new, old); new 298 kernel/rcu/rcutorture.c unsigned long (*gp_diff)(unsigned long new, unsigned long old); new 726 kernel/rcu/rcutorture.c static unsigned long rcutorture_seq_diff(unsigned long new, unsigned long old) new 729 kernel/rcu/rcutorture.c return new - old; new 730 kernel/rcu/rcutorture.c return cur_ops->gp_diff(new, old); new 344 kernel/rcu/tree.c int new; new 351 kernel/rcu/tree.c new = old | RCU_DYNTICK_CTRL_MASK; new 352 kernel/rcu/tree.c } while (atomic_cmpxchg(&rdp->dynticks, old, new) != old); new 748 kernel/relay.c void *old, *new; new 780 kernel/relay.c new = buf->start + new_subbuf * buf->chan->subbuf_size; new 782 kernel/relay.c if (!buf->chan->cb->subbuf_start(buf, new, old, buf->prev_padding)) { new 786 kernel/relay.c buf->data = new; new 185 kernel/resource.c static struct resource * __request_resource(struct resource *root, struct resource *new) new 187 kernel/resource.c resource_size_t start = new->start; new 188 kernel/resource.c resource_size_t end = new->end; new 201 kernel/resource.c new->sibling = tmp; new 202 kernel/resource.c *p = new; new 203 kernel/resource.c new->parent = root; new 279 kernel/resource.c struct resource *request_resource_conflict(struct resource *root, struct resource *new) new 284 kernel/resource.c conflict = __request_resource(root, new); new 296 kernel/resource.c int request_resource(struct resource *root, struct resource *new) new 300 kernel/resource.c conflict = request_resource_conflict(root, new); new 593 kernel/resource.c struct resource *new, new 598 kernel/resource.c struct resource tmp = *new, avail, alloc; new 624 kernel/resource.c avail.flags = new->flags & ~IORESOURCE_UNSET; new 632 kernel/resource.c new->start = alloc.start; new 633 kernel/resource.c new->end = alloc.end; new 651 kernel/resource.c static int find_resource(struct resource *root, struct resource *new, new 655 kernel/resource.c return __find_resource(root, NULL, new, size, constraint); new 673 kernel/resource.c struct resource new = *old; new 678 kernel/resource.c if ((err = __find_resource(root, old, &new, newsize, constraint))) new 681 kernel/resource.c if (resource_contains(&new, old)) { new 682 kernel/resource.c old->start = new.start; new 683 kernel/resource.c old->end = new.end; new 692 kernel/resource.c if (resource_contains(old, &new)) { new 693 kernel/resource.c old->start = new.start; new 694 kernel/resource.c old->end = new.end; new 697 kernel/resource.c *old = new; new 719 kernel/resource.c int allocate_resource(struct resource *root, struct resource *new, new 740 kernel/resource.c if ( new->parent ) { new 743 kernel/resource.c return reallocate_resource(root, new, size, &constraint); new 747 kernel/resource.c err = find_resource(root, new, size, &constraint); new 748 kernel/resource.c if (err >= 0 && __request_resource(root, new)) new 781 kernel/resource.c static struct resource * __insert_resource(struct resource *parent, struct resource *new) new 786 kernel/resource.c first = __request_resource(parent, new); new 792 kernel/resource.c if (WARN_ON(first == new)) /* duplicated insertion */ new 795 kernel/resource.c if ((first->start > new->start) || (first->end < new->end)) new 797 kernel/resource.c if ((first->start == new->start) && (first->end == new->end)) new 803 kernel/resource.c if (next->start < new->start || next->end > new->end) new 807 kernel/resource.c if (next->sibling->start > new->end) new 811 kernel/resource.c new->parent = parent; new 812 kernel/resource.c new->sibling = next->sibling; new 813 kernel/resource.c new->child = first; new 817 kernel/resource.c next->parent = new; new 820 kernel/resource.c parent->child = new; new 825 kernel/resource.c next->sibling = new; new 846 kernel/resource.c struct resource *insert_resource_conflict(struct resource *parent, struct resource *new) new 851 kernel/resource.c conflict = __insert_resource(parent, new); new 866 kernel/resource.c int insert_resource(struct resource *parent, struct resource *new) new 870 kernel/resource.c conflict = insert_resource_conflict(parent, new); new 883 kernel/resource.c void insert_resource_expand_to_fit(struct resource *root, struct resource *new) new 885 kernel/resource.c if (new->parent) new 892 kernel/resource.c conflict = __insert_resource(root, new); new 899 kernel/resource.c if (conflict->start < new->start) new 900 kernel/resource.c new->start = conflict->start; new 901 kernel/resource.c if (conflict->end > new->end) new 902 kernel/resource.c new->end = conflict->end; new 904 kernel/resource.c printk("Expanded resource %s due to conflict with %s\n", new->name, conflict->name); new 1380 kernel/resource.c struct resource *new) new 1388 kernel/resource.c *ptr = new; new 1390 kernel/resource.c conflict = request_resource_conflict(root, new); new 1393 kernel/resource.c new, conflict->name, conflict); new 1417 kernel/resource.c void devm_release_resource(struct device *dev, struct resource *new) new 1420 kernel/resource.c new)); new 4946 kernel/sched/fair.c u64 new, old = ktime_to_ns(cfs_b->period); new 4953 kernel/sched/fair.c new = old * 2; new 4954 kernel/sched/fair.c if (new < max_cfs_quota_period) { new 4955 kernel/sched/fair.c cfs_b->period = ns_to_ktime(new); new 4961 kernel/sched/fair.c div_u64(new, NSEC_PER_USEC), new 1149 kernel/sched/psi.c void psi_trigger_replace(void **trigger_ptr, struct psi_trigger *new) new 1156 kernel/sched/psi.c rcu_assign_pointer(*trigger_ptr, new); new 1197 kernel/sched/psi.c struct psi_trigger *new; new 1211 kernel/sched/psi.c new = psi_trigger_create(&psi_system, buf, nbytes, res); new 1212 kernel/sched/psi.c if (IS_ERR(new)) new 1213 kernel/sched/psi.c return PTR_ERR(new); new 1218 kernel/sched/psi.c psi_trigger_replace(&seq->private, new); new 2180 kernel/sched/topology.c struct sched_domain_attr *new, int idx_new) new 2185 kernel/sched/topology.c if (!new && !cur) new 2191 kernel/sched/topology.c new ? (new + idx_new) : &tmp, new 1731 kernel/seccomp.c const char *new = names; new 1741 kernel/seccomp.c new = "?"; new 1743 kernel/seccomp.c new = "(none)"; new 1746 kernel/seccomp.c new = "?"; new 1755 kernel/seccomp.c return audit_seccomp_actions_logged(new, old, !ret); new 3252 kernel/signal.c struct compat_siginfo new; new 3253 kernel/signal.c memset(&new, 0, sizeof(new)); new 3255 kernel/signal.c new.si_signo = from->si_signo; new 3256 kernel/signal.c new.si_errno = from->si_errno; new 3257 kernel/signal.c new.si_code = from->si_code; new 3260 kernel/signal.c new.si_pid = from->si_pid; new 3261 kernel/signal.c new.si_uid = from->si_uid; new 3264 kernel/signal.c new.si_tid = from->si_tid; new 3265 kernel/signal.c new.si_overrun = from->si_overrun; new 3266 kernel/signal.c new.si_int = from->si_int; new 3269 kernel/signal.c new.si_band = from->si_band; new 3270 kernel/signal.c new.si_fd = from->si_fd; new 3273 kernel/signal.c new.si_addr = ptr_to_compat(from->si_addr); new 3275 kernel/signal.c new.si_trapno = from->si_trapno; new 3279 kernel/signal.c new.si_addr = ptr_to_compat(from->si_addr); new 3281 kernel/signal.c new.si_trapno = from->si_trapno; new 3283 kernel/signal.c new.si_addr_lsb = from->si_addr_lsb; new 3286 kernel/signal.c new.si_addr = ptr_to_compat(from->si_addr); new 3288 kernel/signal.c new.si_trapno = from->si_trapno; new 3290 kernel/signal.c new.si_lower = ptr_to_compat(from->si_lower); new 3291 kernel/signal.c new.si_upper = ptr_to_compat(from->si_upper); new 3294 kernel/signal.c new.si_addr = ptr_to_compat(from->si_addr); new 3296 kernel/signal.c new.si_trapno = from->si_trapno; new 3298 kernel/signal.c new.si_pkey = from->si_pkey; new 3301 kernel/signal.c new.si_pid = from->si_pid; new 3302 kernel/signal.c new.si_uid = from->si_uid; new 3303 kernel/signal.c new.si_status = from->si_status; new 3306 kernel/signal.c new._sifields._sigchld_x32._utime = from->si_utime; new 3307 kernel/signal.c new._sifields._sigchld_x32._stime = from->si_stime; new 3311 kernel/signal.c new.si_utime = from->si_utime; new 3312 kernel/signal.c new.si_stime = from->si_stime; new 3316 kernel/signal.c new.si_pid = from->si_pid; new 3317 kernel/signal.c new.si_uid = from->si_uid; new 3318 kernel/signal.c new.si_int = from->si_int; new 3321 kernel/signal.c new.si_call_addr = ptr_to_compat(from->si_call_addr); new 3322 kernel/signal.c new.si_syscall = from->si_syscall; new 3323 kernel/signal.c new.si_arch = from->si_arch; new 3327 kernel/signal.c if (copy_to_user(to, &new, sizeof(struct compat_siginfo))) new 4048 kernel/signal.c stack_t new, old; new 4050 kernel/signal.c if (uss && copy_from_user(&new, uss, sizeof(stack_t))) new 4052 kernel/signal.c err = do_sigaltstack(uss ? &new : NULL, uoss ? &old : NULL, new 4062 kernel/signal.c stack_t new; new 4063 kernel/signal.c if (copy_from_user(&new, uss, sizeof(stack_t))) new 4065 kernel/signal.c (void)do_sigaltstack(&new, NULL, current_user_stack_pointer(), new 354 kernel/sys.c struct cred *new; new 366 kernel/sys.c new = prepare_creds(); new 367 kernel/sys.c if (!new) new 376 kernel/sys.c new->gid = krgid; new 385 kernel/sys.c new->egid = kegid; new 392 kernel/sys.c new->sgid = new->egid; new 393 kernel/sys.c new->fsgid = new->egid; new 395 kernel/sys.c return commit_creds(new); new 398 kernel/sys.c abort_creds(new); new 416 kernel/sys.c struct cred *new; new 424 kernel/sys.c new = prepare_creds(); new 425 kernel/sys.c if (!new) new 431 kernel/sys.c new->gid = new->egid = new->sgid = new->fsgid = kgid; new 433 kernel/sys.c new->egid = new->fsgid = kgid; new 437 kernel/sys.c return commit_creds(new); new 440 kernel/sys.c abort_creds(new); new 452 kernel/sys.c static int set_user(struct cred *new) new 456 kernel/sys.c new_user = alloc_uid(new->uid); new 473 kernel/sys.c free_uid(new->user); new 474 kernel/sys.c new->user = new_user; new 497 kernel/sys.c struct cred *new; new 509 kernel/sys.c new = prepare_creds(); new 510 kernel/sys.c if (!new) new 516 kernel/sys.c new->uid = kruid; new 524 kernel/sys.c new->euid = keuid; new 532 kernel/sys.c if (!uid_eq(new->uid, old->uid)) { new 533 kernel/sys.c retval = set_user(new); new 539 kernel/sys.c new->suid = new->euid; new 540 kernel/sys.c new->fsuid = new->euid; new 542 kernel/sys.c retval = security_task_fix_setuid(new, old, LSM_SETID_RE); new 546 kernel/sys.c return commit_creds(new); new 549 kernel/sys.c abort_creds(new); new 573 kernel/sys.c struct cred *new; new 581 kernel/sys.c new = prepare_creds(); new 582 kernel/sys.c if (!new) new 588 kernel/sys.c new->suid = new->uid = kuid; new 590 kernel/sys.c retval = set_user(new); new 594 kernel/sys.c } else if (!uid_eq(kuid, old->uid) && !uid_eq(kuid, new->suid)) { new 598 kernel/sys.c new->fsuid = new->euid = kuid; new 600 kernel/sys.c retval = security_task_fix_setuid(new, old, LSM_SETID_ID); new 604 kernel/sys.c return commit_creds(new); new 607 kernel/sys.c abort_creds(new); new 625 kernel/sys.c struct cred *new; new 642 kernel/sys.c new = prepare_creds(); new 643 kernel/sys.c if (!new) new 662 kernel/sys.c new->uid = kruid; new 664 kernel/sys.c retval = set_user(new); new 670 kernel/sys.c new->euid = keuid; new 672 kernel/sys.c new->suid = ksuid; new 673 kernel/sys.c new->fsuid = new->euid; new 675 kernel/sys.c retval = security_task_fix_setuid(new, old, LSM_SETID_RES); new 679 kernel/sys.c return commit_creds(new); new 682 kernel/sys.c abort_creds(new); new 717 kernel/sys.c struct cred *new; new 732 kernel/sys.c new = prepare_creds(); new 733 kernel/sys.c if (!new) new 751 kernel/sys.c new->gid = krgid; new 753 kernel/sys.c new->egid = kegid; new 755 kernel/sys.c new->sgid = ksgid; new 756 kernel/sys.c new->fsgid = new->egid; new 758 kernel/sys.c return commit_creds(new); new 761 kernel/sys.c abort_creds(new); new 800 kernel/sys.c struct cred *new; new 811 kernel/sys.c new = prepare_creds(); new 812 kernel/sys.c if (!new) new 819 kernel/sys.c new->fsuid = kuid; new 820 kernel/sys.c if (security_task_fix_setuid(new, old, LSM_SETID_FS) == 0) new 825 kernel/sys.c abort_creds(new); new 829 kernel/sys.c commit_creds(new); new 844 kernel/sys.c struct cred *new; new 855 kernel/sys.c new = prepare_creds(); new 856 kernel/sys.c if (!new) new 863 kernel/sys.c new->fsgid = kgid; new 868 kernel/sys.c abort_creds(new); new 872 kernel/sys.c commit_creds(new); new 1611 kernel/sys.c struct rlimit old, new; new 1622 kernel/sys.c rlim64_to_rlim(&new64, &new); new 1640 kernel/sys.c ret = do_prlimit(tsk, resource, new_rlim ? &new : NULL, new 570 kernel/time/clockevents.c struct clock_event_device *new) new 583 kernel/time/clockevents.c if (new) { new 584 kernel/time/clockevents.c BUG_ON(!clockevent_state_detached(new)); new 585 kernel/time/clockevents.c clockevents_shutdown(new); new 562 kernel/time/posix-cpu-timers.c struct itimerspec64 *new, struct itimerspec64 *old) new 579 kernel/time/posix-cpu-timers.c new_expires = ktime_to_ns(timespec64_to_ktime(new->it_value)); new 673 kernel/time/posix-cpu-timers.c timer->it_interval = timespec64_to_ktime(new->it_interval); new 940 kernel/time/posix-timers.c struct old_itimerspec32 __user *, new, new 947 kernel/time/posix-timers.c if (!new) new 949 kernel/time/posix-timers.c if (get_old_itimerspec32(&new_spec, new)) new 53 kernel/time/tick-internal.h struct clock_event_device *new); new 1363 kernel/time/timekeeping.c struct clocksource *new, *old; new 1366 kernel/time/timekeeping.c new = (struct clocksource *) data; new 1376 kernel/time/timekeeping.c if (try_module_get(new->owner)) { new 1377 kernel/time/timekeeping.c if (!new->enable || new->enable(new) == 0) { new 1379 kernel/time/timekeeping.c tk_setup_internals(tk, new); new 1384 kernel/time/timekeeping.c module_put(new->owner); new 1045 kernel/trace/ring_buffer.c struct buffer_page *new) new 1054 kernel/trace/ring_buffer.c ret = cmpxchg(ptr, val, (unsigned long)&new->list); new 5992 kernel/trace/trace_events_hist.c goto new; new 6013 kernel/trace/trace_events_hist.c new: new 79 kernel/trace/trace_stat.c struct rb_node **new = &(root->rb_node), *parent = NULL; new 91 kernel/trace/trace_stat.c while (*new) { new 95 kernel/trace/trace_stat.c this = container_of(*new, struct stat_node, node); new 98 kernel/trace/trace_stat.c parent = *new; new 100 kernel/trace/trace_stat.c new = &((*new)->rb_left); new 102 kernel/trace/trace_stat.c new = &((*new)->rb_right); new 105 kernel/trace/trace_stat.c rb_link_node(&data->node, parent, new); new 472 kernel/trace/trace_uprobe.c static int validate_ref_ctr_offset(struct trace_uprobe *new) new 476 kernel/trace/trace_uprobe.c struct inode *new_inode = d_real_inode(new->path.dentry); new 480 kernel/trace/trace_uprobe.c new->offset == tmp->offset && new 481 kernel/trace/trace_uprobe.c new->ref_ctr_offset != tmp->ref_ctr_offset) { new 132 kernel/tracepoint.c struct tracepoint_func *old, *new; new 153 kernel/tracepoint.c new = allocate_probes(nr_probes + 2); new 154 kernel/tracepoint.c if (new == NULL) new 159 kernel/tracepoint.c memcpy(new, old, nr_probes * sizeof(struct tracepoint_func)); new 162 kernel/tracepoint.c memcpy(new, old, pos * sizeof(struct tracepoint_func)); new 164 kernel/tracepoint.c memcpy(new + pos + 1, old + pos, new 169 kernel/tracepoint.c new[pos] = *tp_func; new 170 kernel/tracepoint.c new[nr_probes + 1].func = NULL; new 171 kernel/tracepoint.c *funcs = new; new 180 kernel/tracepoint.c struct tracepoint_func *old, *new; new 210 kernel/tracepoint.c new = allocate_probes(nr_probes - nr_del + 1); new 211 kernel/tracepoint.c if (new == NULL) new 216 kernel/tracepoint.c new[j++] = old[i]; new 217 kernel/tracepoint.c new[nr_probes - nr_del].func = NULL; new 218 kernel/tracepoint.c *funcs = new; new 128 kernel/ucount.c struct ucounts *ucounts, *new; new 135 kernel/ucount.c new = kzalloc(sizeof(*new), GFP_KERNEL); new 136 kernel/ucount.c if (!new) new 139 kernel/ucount.c new->ns = ns; new 140 kernel/ucount.c new->uid = uid; new 141 kernel/ucount.c new->count = 0; new 146 kernel/ucount.c kfree(new); new 148 kernel/ucount.c hlist_add_head(&new->node, hashent); new 149 kernel/ucount.c ucounts = new; new 71 kernel/umh.c struct cred *new; new 85 kernel/umh.c new = prepare_kernel_cred(current); new 86 kernel/umh.c if (!new) new 90 kernel/umh.c new->cap_bset = cap_intersect(usermodehelper_bset, new->cap_bset); new 91 kernel/umh.c new->cap_inheritable = cap_intersect(usermodehelper_inheritable, new 92 kernel/umh.c new->cap_inheritable); new 96 kernel/umh.c retval = sub_info->init(sub_info, new); new 98 kernel/umh.c abort_creds(new); new 103 kernel/umh.c commit_creds(new); new 381 kernel/umh.c int (*init)(struct subprocess_info *info, struct cred *new), new 409 kernel/umh.c int (*init)(struct subprocess_info *info, struct cred *new), new 435 kernel/umh.c static int umh_pipe_setup(struct subprocess_info *info, struct cred *new) new 178 kernel/user.c struct user_struct *up, *new; new 185 kernel/user.c new = kmem_cache_zalloc(uid_cachep, GFP_KERNEL); new 186 kernel/user.c if (!new) new 189 kernel/user.c new->uid = uid; new 190 kernel/user.c refcount_set(&new->__count, 1); new 191 kernel/user.c ratelimit_state_init(&new->ratelimit, HZ, 100); new 192 kernel/user.c ratelimit_set_flags(&new->ratelimit, RATELIMIT_MSG_ON_RELEASE); new 201 kernel/user.c kmem_cache_free(uid_cachep, new); new 203 kernel/user.c uid_hash_insert(new, hashent); new 204 kernel/user.c up = new; new 69 kernel/user_namespace.c int create_user_ns(struct cred *new) new 71 kernel/user_namespace.c struct user_namespace *ns, *parent_ns = new->user_ns; new 72 kernel/user_namespace.c kuid_t owner = new->euid; new 73 kernel/user_namespace.c kgid_t group = new->egid; new 139 kernel/user_namespace.c set_cred_user_ns(new, ns); new 143 kernel/utsname.c static int utsns_install(struct nsproxy *nsproxy, struct ns_common *new) new 145 kernel/utsname.c struct uts_namespace *ns = to_uts_ns(new); new 163 lib/atomic64.c s64 atomic64_xchg(atomic64_t *v, s64 new) new 171 lib/atomic64.c v->counter = new; new 79 lib/atomic64_test.c #define XCHG_FAMILY_TEST(bit, init, new) \ new 81 lib/atomic64_test.c FAMILY_TEST(TEST_ARGS, bit, xchg, init, init, new, new); \ new 84 lib/atomic64_test.c #define CMPXCHG_FAMILY_TEST(bit, init, new, wrong) \ new 87 lib/atomic64_test.c init, init, new, init, new); \ new 89 lib/atomic64_test.c init, init, init, wrong, new); \ new 792 lib/bitmap.c const unsigned long *old, const unsigned long *new, new 801 lib/bitmap.c w = bitmap_weight(new, nbits); new 808 lib/bitmap.c set_bit(bitmap_ord_to_pos(new, n % w, nbits), dst); new 839 lib/bitmap.c const unsigned long *new, int bits) new 841 lib/bitmap.c int w = bitmap_weight(new, bits); new 846 lib/bitmap.c return bitmap_ord_to_pos(new, n % w, bits); new 471 lib/btree.c unsigned long *new; new 473 lib/btree.c new = btree_node_alloc(head, gfp); new 474 lib/btree.c if (!new) new 478 lib/btree.c new, level + 1, gfp); new 480 lib/btree.c mempool_free(new, head->mempool); new 484 lib/btree.c setkey(geo, new, i, bkey(geo, node, i)); new 485 lib/btree.c setval(geo, new, i, bval(geo, node, i)); new 26 lib/crc-t10dif.c struct crypto_shash *new, *old; new 40 lib/crc-t10dif.c new = crypto_alloc_shash("crct10dif", 0, 0); new 41 lib/crc-t10dif.c if (IS_ERR(new)) { new 45 lib/crc-t10dif.c rcu_assign_pointer(crct10dif_tfm, new); new 166 lib/debugobjects.c struct debug_obj *new[ODEBUG_BATCH_SIZE]; new 170 lib/debugobjects.c new[cnt] = kmem_cache_zalloc(obj_cache, gfp); new 171 lib/debugobjects.c if (!new[cnt]) new 179 lib/debugobjects.c hlist_add_head(&new[--cnt]->node, &obj_pool); new 1299 lib/debugobjects.c struct debug_obj *obj, *new; new 1327 lib/debugobjects.c new = hlist_entry(obj_pool.first, typeof(*obj), node); new 1328 lib/debugobjects.c hlist_del(&new->node); new 1330 lib/debugobjects.c *new = *obj; new 1331 lib/debugobjects.c hlist_add_head(&new->node, &db->list); new 78 lib/errseq.c errseq_t new; new 81 lib/errseq.c new = (old & ~(MAX_ERRNO|ERRSEQ_SEEN)) | -err; new 85 lib/errseq.c new += ERRSEQ_CTR_INC; new 88 lib/errseq.c if (new == old) { new 89 lib/errseq.c cur = new; new 94 lib/errseq.c cur = cmpxchg(eseq, old, new); new 100 lib/errseq.c if (likely(cur == old || cur == new)) new 177 lib/errseq.c errseq_t old, new; new 198 lib/errseq.c new = old | ERRSEQ_SEEN; new 199 lib/errseq.c if (new != old) new 200 lib/errseq.c cmpxchg(eseq, old, new); new 201 lib/errseq.c *since = new; new 202 lib/errseq.c err = -(new & MAX_ERRNO); new 1596 lib/iov_iter.c const void *dup_iter(struct iov_iter *new, struct iov_iter *old, gfp_t flags) new 1598 lib/iov_iter.c *new = *old; new 1599 lib/iov_iter.c if (unlikely(iov_iter_is_pipe(new))) { new 1603 lib/iov_iter.c if (unlikely(iov_iter_is_discard(new))) new 1605 lib/iov_iter.c if (iov_iter_is_bvec(new)) new 1606 lib/iov_iter.c return new->bvec = kmemdup(new->bvec, new 1607 lib/iov_iter.c new->nr_segs * sizeof(struct bio_vec), new 1611 lib/iov_iter.c return new->iov = kmemdup(new->iov, new 1612 lib/iov_iter.c new->nr_segs * sizeof(struct iovec), new 20 lib/list_debug.c bool __list_add_valid(struct list_head *new, struct list_head *prev, new 29 lib/list_debug.c CHECK_DATA_CORRUPTION(new == prev || new == next, new 31 lib/list_debug.c new, prev, next)) new 17 lib/lockref.c struct lockref new = old, prev = old; \ new 21 lib/lockref.c new.lock_count); \ new 47 lib/lockref.c new.count++; new 68 lib/lockref.c new.count++; new 96 lib/lockref.c new.count--; new 123 lib/lockref.c new.count++; new 149 lib/lockref.c new.count--; new 153 lib/lockref.c return new.count; new 167 lib/lockref.c new.count--; new 204 lib/lockref.c new.count++; new 66 lib/lru_cache.c unsigned long old, new, val; new 69 lib/lru_cache.c new = old | LC_LOCKED; new 70 lib/lru_cache.c val = cmpxchg(&lc->flags, old, new); new 112 lib/math/prime_numbers.c struct primes *new; new 128 lib/math/prime_numbers.c new = kmalloc(sizeof(*new) + bitmap_size(sz), new 130 lib/math/prime_numbers.c if (!new) new 136 lib/math/prime_numbers.c kfree(new); new 144 lib/math/prime_numbers.c bitmap_fill(new->primes, sz); new 145 lib/math/prime_numbers.c bitmap_copy(new->primes, p->primes, p->sz); new 146 lib/math/prime_numbers.c for (y = 2UL; y < sz; y = find_next_bit(new->primes, sz, y + 1)) new 147 lib/math/prime_numbers.c new->last = clear_multiples(y, new->primes, p->sz, sz); new 148 lib/math/prime_numbers.c new->sz = sz; new 150 lib/math/prime_numbers.c BUG_ON(new->last <= x); new 152 lib/math/prime_numbers.c rcu_assign_pointer(primes, new); new 12 lib/memcat_p.c void **p = a, **new; new 23 lib/memcat_p.c new = kmalloc_array(nr, sizeof(void *), GFP_KERNEL); new 24 lib/memcat_p.c if (!new) new 29 lib/memcat_p.c new[nr] = *p; new 31 lib/memcat_p.c return new; new 75 lib/rbtree.c __rb_rotate_set_parents(struct rb_node *old, struct rb_node *new, new 79 lib/rbtree.c new->__rb_parent_color = old->__rb_parent_color; new 80 lib/rbtree.c rb_set_parent_color(old, new, color); new 81 lib/rbtree.c __rb_change_child(old, new, parent, root); new 86 lib/rbtree.c void (*augment_rotate)(struct rb_node *old, struct rb_node *new)) new 228 lib/rbtree.c void (*augment_rotate)(struct rb_node *old, struct rb_node *new)) new 411 lib/rbtree.c void (*augment_rotate)(struct rb_node *old, struct rb_node *new)) new 425 lib/rbtree.c static inline void dummy_copy(struct rb_node *old, struct rb_node *new) {} new 426 lib/rbtree.c static inline void dummy_rotate(struct rb_node *old, struct rb_node *new) {} new 457 lib/rbtree.c void (*augment_rotate)(struct rb_node *old, struct rb_node *new)) new 553 lib/rbtree.c void rb_replace_node(struct rb_node *victim, struct rb_node *new, new 559 lib/rbtree.c *new = *victim; new 563 lib/rbtree.c rb_set_parent(victim->rb_left, new); new 565 lib/rbtree.c rb_set_parent(victim->rb_right, new); new 566 lib/rbtree.c __rb_change_child(victim, new, parent, root); new 570 lib/rbtree.c void rb_replace_node_rcu(struct rb_node *victim, struct rb_node *new, new 576 lib/rbtree.c *new = *victim; new 580 lib/rbtree.c rb_set_parent(victim->rb_left, new); new 582 lib/rbtree.c rb_set_parent(victim->rb_right, new); new 588 lib/rbtree.c __rb_change_child_rcu(victim, new, parent, root); new 34 lib/rbtree_test.c struct rb_node **new = &root->rb_root.rb_node, *parent = NULL; new 37 lib/rbtree_test.c while (*new) { new 38 lib/rbtree_test.c parent = *new; new 40 lib/rbtree_test.c new = &parent->rb_left; new 42 lib/rbtree_test.c new = &parent->rb_right; new 45 lib/rbtree_test.c rb_link_node(&node->rb, parent, new); new 51 lib/rbtree_test.c struct rb_node **new = &root->rb_root.rb_node, *parent = NULL; new 55 lib/rbtree_test.c while (*new) { new 56 lib/rbtree_test.c parent = *new; new 58 lib/rbtree_test.c new = &parent->rb_left; new 60 lib/rbtree_test.c new = &parent->rb_right; new 65 lib/rbtree_test.c rb_link_node(&node->rb, parent, new); new 88 lib/rbtree_test.c struct rb_node **new = &root->rb_root.rb_node, *rb_parent = NULL; new 93 lib/rbtree_test.c while (*new) { new 94 lib/rbtree_test.c rb_parent = *new; new 99 lib/rbtree_test.c new = &parent->rb.rb_left; new 101 lib/rbtree_test.c new = &parent->rb.rb_right; new 105 lib/rbtree_test.c rb_link_node(&node->rb, rb_parent, new); new 112 lib/rbtree_test.c struct rb_node **new = &root->rb_root.rb_node, *rb_parent = NULL; new 118 lib/rbtree_test.c while (*new) { new 119 lib/rbtree_test.c rb_parent = *new; new 124 lib/rbtree_test.c new = &parent->rb.rb_left; new 126 lib/rbtree_test.c new = &parent->rb.rb_right; new 132 lib/rbtree_test.c rb_link_node(&node->rb, rb_parent, new); new 66 lib/refcount.c unsigned int new, val = atomic_read(&r->refs); new 75 lib/refcount.c new = val + i; new 76 lib/refcount.c if (new < val) new 77 lib/refcount.c new = UINT_MAX; new 79 lib/refcount.c } while (!atomic_try_cmpxchg_relaxed(&r->refs, &val, new)); new 81 lib/refcount.c WARN_ONCE(new == UINT_MAX, "refcount_t: saturated; leaking memory.\n"); new 123 lib/refcount.c unsigned int new, val = atomic_read(&r->refs); new 126 lib/refcount.c new = val + 1; new 131 lib/refcount.c if (unlikely(!new)) new 134 lib/refcount.c } while (!atomic_try_cmpxchg_relaxed(&r->refs, &val, new)); new 136 lib/refcount.c WARN_ONCE(new == UINT_MAX, "refcount_t: saturated; leaking memory.\n"); new 182 lib/refcount.c unsigned int new, val = atomic_read(&r->refs); new 188 lib/refcount.c new = val - i; new 189 lib/refcount.c if (new > val) { new 190 lib/refcount.c WARN_ONCE(new > val, "refcount_t: underflow; use-after-free.\n"); new 194 lib/refcount.c } while (!atomic_try_cmpxchg_release(&r->refs, &val, new)); new 196 lib/refcount.c if (!new) { new 277 lib/refcount.c unsigned int new, val = atomic_read(&r->refs); new 286 lib/refcount.c new = val - 1; new 287 lib/refcount.c if (new > val) { new 288 lib/refcount.c WARN_ONCE(new > val, "refcount_t: underflow; use-after-free.\n"); new 292 lib/refcount.c } while (!atomic_try_cmpxchg_release(&r->refs, &val, new)); new 276 lib/stackdepot.c struct stack_record *new = new 279 lib/stackdepot.c if (new) { new 280 lib/stackdepot.c new->next = *bucket; new 285 lib/stackdepot.c smp_store_release(bucket, new); new 286 lib/stackdepot.c found = new; new 1084 lib/string.c char *strreplace(char *s, char old, char new) new 1088 lib/string.c *s = new; new 339 lib/test_firmware.c long new; new 341 lib/test_firmware.c ret = kstrtol(buf, 10, &new); new 345 lib/test_firmware.c if (new > U8_MAX) new 349 lib/test_firmware.c *(u8 *)cfg = new; new 880 lib/test_kmod.c unsigned long new; new 883 lib/test_kmod.c ret = kstrtoul(buf, 10, &new); new 887 lib/test_kmod.c if (new > UINT_MAX) new 893 lib/test_kmod.c *(unsigned int *)config = new; new 918 lib/test_kmod.c unsigned long new; new 920 lib/test_kmod.c ret = kstrtoul(buf, 10, &new); new 924 lib/test_kmod.c if (new < min || new > max) new 928 lib/test_kmod.c *config = new; new 940 lib/test_kmod.c long new; new 942 lib/test_kmod.c ret = kstrtol(buf, 10, &new); new 946 lib/test_kmod.c if (new < INT_MIN || new > INT_MAX) new 950 lib/test_kmod.c *config = new; new 811 mm/filemap.c int replace_page_cache_page(struct page *old, struct page *new, gfp_t gfp_mask) new 820 mm/filemap.c VM_BUG_ON_PAGE(!PageLocked(new), new); new 821 mm/filemap.c VM_BUG_ON_PAGE(new->mapping, new); new 823 mm/filemap.c get_page(new); new 824 mm/filemap.c new->mapping = mapping; new 825 mm/filemap.c new->index = offset; new 828 mm/filemap.c xas_store(&xas, new); new 833 mm/filemap.c __dec_node_page_state(new, NR_FILE_PAGES); new 834 mm/filemap.c if (!PageHuge(new)) new 835 mm/filemap.c __inc_node_page_state(new, NR_FILE_PAGES); new 837 mm/filemap.c __dec_node_page_state(new, NR_SHMEM); new 838 mm/filemap.c if (PageSwapBacked(new)) new 839 mm/filemap.c __inc_node_page_state(new, NR_SHMEM); new 841 mm/filemap.c mem_cgroup_migrate(old, new); new 3046 mm/huge_memory.c void remove_migration_pmd(struct page_vma_mapped_walk *pvmw, struct page *new) new 3059 mm/huge_memory.c get_page(new); new 3060 mm/huge_memory.c pmde = pmd_mkold(mk_huge_pmd(new, vma->vm_page_prot)); new 3067 mm/huge_memory.c if (PageAnon(new)) new 3068 mm/huge_memory.c page_add_anon_rmap(new, vma, mmun_start, true); new 3070 mm/huge_memory.c page_add_file_rmap(new, true); new 3072 mm/huge_memory.c if ((vma->vm_flags & VM_LOCKED) && !PageDoubleMap(new)) new 3073 mm/huge_memory.c mlock_vma_page(new); new 389 mm/internal.h static inline void mlock_migrate_page(struct page *new, struct page *old) { } new 1555 mm/ksm.c struct rb_node **new; new 1570 mm/ksm.c new = &root->rb_node; new 1573 mm/ksm.c while (*new) { new 1578 mm/ksm.c stable_node = rb_entry(*new, struct stable_node, node); new 1634 mm/ksm.c parent = *new; new 1636 mm/ksm.c new = &parent->rb_left; new 1638 mm/ksm.c new = &parent->rb_right; new 1703 mm/ksm.c rb_link_node(&page_node->node, parent, new); new 1808 mm/ksm.c struct rb_node **new; new 1818 mm/ksm.c new = &root->rb_node; new 1820 mm/ksm.c while (*new) { new 1825 mm/ksm.c stable_node = rb_entry(*new, struct stable_node, node); new 1869 mm/ksm.c parent = *new; new 1871 mm/ksm.c new = &parent->rb_left; new 1873 mm/ksm.c new = &parent->rb_right; new 1890 mm/ksm.c rb_link_node(&stable_node_dup->node, parent, new); new 1927 mm/ksm.c struct rb_node **new; new 1934 mm/ksm.c new = &root->rb_node; new 1936 mm/ksm.c while (*new) { new 1942 mm/ksm.c tree_rmap_item = rb_entry(*new, struct rmap_item, node); new 1957 mm/ksm.c parent = *new; new 1960 mm/ksm.c new = &parent->rb_left; new 1963 mm/ksm.c new = &parent->rb_right; new 1982 mm/ksm.c rb_link_node(&rmap_item->node, parent, new); new 404 mm/list_lru.c struct list_lru_memcg *old, *new; new 410 mm/list_lru.c new = kvmalloc(sizeof(*new) + new_size * sizeof(void *), GFP_KERNEL); new 411 mm/list_lru.c if (!new) new 414 mm/list_lru.c if (__memcg_init_list_lru_node(new, old_size, new_size)) { new 415 mm/list_lru.c kvfree(new); new 419 mm/list_lru.c memcpy(&new->lru, &old->lru, old_size * sizeof(void *)); new 429 mm/list_lru.c rcu_assign_pointer(nlru->memcg_lrus, new); new 334 mm/memcontrol.c struct memcg_shrinker_map *new, *old; new 346 mm/memcontrol.c new = kvmalloc(sizeof(*new) + size, GFP_KERNEL); new 347 mm/memcontrol.c if (!new) new 351 mm/memcontrol.c memset(new->map, (int)0xff, old_size); new 352 mm/memcontrol.c memset((void *)new->map + old_size, 0, size - old_size); new 354 mm/memcontrol.c rcu_assign_pointer(memcg->nodeinfo[nid]->shrinker_map, new); new 4123 mm/memcontrol.c struct mem_cgroup_threshold_ary *new; new 4150 mm/memcontrol.c new = kmalloc(struct_size(new, entries, size), GFP_KERNEL); new 4151 mm/memcontrol.c if (!new) { new 4155 mm/memcontrol.c new->size = size; new 4159 mm/memcontrol.c memcpy(new->entries, thresholds->primary->entries, (size - 1) * new 4164 mm/memcontrol.c new->entries[size - 1].eventfd = eventfd; new 4165 mm/memcontrol.c new->entries[size - 1].threshold = threshold; new 4168 mm/memcontrol.c sort(new->entries, size, sizeof(struct mem_cgroup_threshold), new 4172 mm/memcontrol.c new->current_threshold = -1; new 4174 mm/memcontrol.c if (new->entries[i].threshold <= usage) { new 4180 mm/memcontrol.c ++new->current_threshold; new 4189 mm/memcontrol.c rcu_assign_pointer(thresholds->primary, new); new 4216 mm/memcontrol.c struct mem_cgroup_threshold_ary *new; new 4246 mm/memcontrol.c new = thresholds->spare; new 4254 mm/memcontrol.c kfree(new); new 4255 mm/memcontrol.c new = NULL; new 4259 mm/memcontrol.c new->size = size; new 4262 mm/memcontrol.c new->current_threshold = -1; new 4267 mm/memcontrol.c new->entries[j] = thresholds->primary->entries[i]; new 4268 mm/memcontrol.c if (new->entries[j].threshold <= usage) { new 4274 mm/memcontrol.c ++new->current_threshold; new 4283 mm/memcontrol.c rcu_assign_pointer(thresholds->primary, new); new 4289 mm/memcontrol.c if (!new) { new 408 mm/memory.c pgtable_t new = pte_alloc_one(mm); new 409 mm/memory.c if (!new) new 430 mm/memory.c pmd_populate(mm, pmd, new); new 431 mm/memory.c new = NULL; new 434 mm/memory.c if (new) new 435 mm/memory.c pte_free(mm, new); new 441 mm/memory.c pte_t *new = pte_alloc_one_kernel(&init_mm); new 442 mm/memory.c if (!new) new 449 mm/memory.c pmd_populate_kernel(&init_mm, pmd, new); new 450 mm/memory.c new = NULL; new 453 mm/memory.c if (new) new 454 mm/memory.c pte_free_kernel(&init_mm, new); new 4051 mm/memory.c p4d_t *new = p4d_alloc_one(mm, address); new 4052 mm/memory.c if (!new) new 4059 mm/memory.c p4d_free(mm, new); new 4061 mm/memory.c pgd_populate(mm, pgd, new); new 4074 mm/memory.c pud_t *new = pud_alloc_one(mm, address); new 4075 mm/memory.c if (!new) new 4084 mm/memory.c p4d_populate(mm, p4d, new); new 4086 mm/memory.c pud_free(mm, new); new 4090 mm/memory.c pgd_populate(mm, p4d, new); new 4092 mm/memory.c pud_free(mm, new); new 4107 mm/memory.c pmd_t *new = pmd_alloc_one(mm, address); new 4108 mm/memory.c if (!new) new 4117 mm/memory.c pud_populate(mm, pud, new); new 4119 mm/memory.c pmd_free(mm, new); new 4123 mm/memory.c pgd_populate(mm, pud, new); new 4125 mm/memory.c pmd_free(mm, new); new 367 mm/mempolicy.c void mpol_rebind_task(struct task_struct *tsk, const nodemask_t *new) new 369 mm/mempolicy.c mpol_rebind_policy(tsk->mempolicy, new); new 378 mm/mempolicy.c void mpol_rebind_mm(struct mm_struct *mm, nodemask_t *new) new 384 mm/mempolicy.c mpol_rebind_policy(vma->vm_policy, new); new 703 mm/mempolicy.c struct mempolicy *new; new 710 mm/mempolicy.c new = mpol_dup(pol); new 711 mm/mempolicy.c if (IS_ERR(new)) new 712 mm/mempolicy.c return PTR_ERR(new); new 715 mm/mempolicy.c err = vma->vm_ops->set_policy(vma, new); new 721 mm/mempolicy.c vma->vm_policy = new; /* protected by mmap_sem */ new 726 mm/mempolicy.c mpol_put(new); new 795 mm/mempolicy.c struct mempolicy *new, *old; new 802 mm/mempolicy.c new = mpol_new(mode, flags, nodes); new 803 mm/mempolicy.c if (IS_ERR(new)) { new 804 mm/mempolicy.c ret = PTR_ERR(new); new 809 mm/mempolicy.c ret = mpol_set_nodemask(new, nodes, scratch); new 812 mm/mempolicy.c mpol_put(new); new 816 mm/mempolicy.c current->mempolicy = new; new 817 mm/mempolicy.c if (new && new->mode == MPOL_INTERLEAVE) new 1222 mm/mempolicy.c struct mempolicy *new; new 1247 mm/mempolicy.c new = mpol_new(mode, mode_flags, nmask); new 1248 mm/mempolicy.c if (IS_ERR(new)) new 1249 mm/mempolicy.c return PTR_ERR(new); new 1252 mm/mempolicy.c new->flags |= MPOL_F_MOF; new 1258 mm/mempolicy.c if (!new) new 1276 mm/mempolicy.c err = mpol_set_nodemask(new, nmask, scratch); new 1295 mm/mempolicy.c err = mbind_range(mm, start, end, new); new 1318 mm/mempolicy.c mpol_put(new); new 1468 mm/mempolicy.c nodemask_t *new; new 1475 mm/mempolicy.c new = &scratch->mask2; new 1481 mm/mempolicy.c err = get_nodes(new, new_nodes, maxnode); new 1510 mm/mempolicy.c if (!nodes_subset(*new, task_nodes) && !capable(CAP_SYS_NICE)) { new 1516 mm/mempolicy.c nodes_and(*new, *new, task_nodes); new 1517 mm/mempolicy.c if (nodes_empty(*new)) new 1532 mm/mempolicy.c err = do_migrate_pages(mm, old, new, new 1673 mm/mempolicy.c unsigned long __user *new = NULL; new 1685 mm/mempolicy.c new = old + size / sizeof(unsigned long); new 1692 mm/mempolicy.c if (new == NULL) new 1693 mm/mempolicy.c new = compat_alloc_user_space(size); new 1694 mm/mempolicy.c if (copy_to_user(new, nodes_addr(tmp_mask), size)) new 1697 mm/mempolicy.c return kernel_migrate_pages(pid, nr_bits + 1, old, new); new 2224 mm/mempolicy.c struct mempolicy *new = kmem_cache_alloc(policy_cache, GFP_KERNEL); new 2226 mm/mempolicy.c if (!new) new 2232 mm/mempolicy.c *new = *old; new 2235 mm/mempolicy.c *new = *old; new 2239 mm/mempolicy.c mpol_rebind_policy(new, &mems); new 2241 mm/mempolicy.c atomic_set(&new->refcnt, 1); new 2242 mm/mempolicy.c return new; new 2321 mm/mempolicy.c static void sp_insert(struct shared_policy *sp, struct sp_node *new) new 2330 mm/mempolicy.c if (new->start < nd->start) new 2332 mm/mempolicy.c else if (new->end > nd->end) new 2337 mm/mempolicy.c rb_link_node(&new->nd, parent, p); new 2338 mm/mempolicy.c rb_insert_color(&new->nd, &sp->root); new 2339 mm/mempolicy.c pr_debug("inserting %lx-%lx: %d\n", new->start, new->end, new 2340 mm/mempolicy.c new->policy ? new->policy->mode : 0); new 2506 mm/mempolicy.c unsigned long end, struct sp_node *new) new 2545 mm/mempolicy.c if (new) new 2546 mm/mempolicy.c sp_insert(sp, new); new 2589 mm/mempolicy.c struct mempolicy *new; new 2595 mm/mempolicy.c new = mpol_new(mpol->mode, mpol->flags, &mpol->w.user_nodemask); new 2596 mm/mempolicy.c if (IS_ERR(new)) new 2600 mm/mempolicy.c ret = mpol_set_nodemask(new, &mpol->w.user_nodemask, scratch); new 2608 mm/mempolicy.c mpol_set_shared_policy(sp, &pvma, new); /* adds ref */ new 2611 mm/mempolicy.c mpol_put(new); /* drop initial ref */ new 2623 mm/mempolicy.c struct sp_node *new = NULL; new 2633 mm/mempolicy.c new = sp_alloc(vma->vm_pgoff, vma->vm_pgoff + sz, npol); new 2634 mm/mempolicy.c if (!new) new 2637 mm/mempolicy.c err = shared_policy_replace(info, vma->vm_pgoff, vma->vm_pgoff+sz, new); new 2638 mm/mempolicy.c if (err && new) new 2639 mm/mempolicy.c sp_free(new); new 2798 mm/mempolicy.c struct mempolicy *new = NULL; new 2883 mm/mempolicy.c new = mpol_new(mode, mode_flags, &nodes); new 2884 mm/mempolicy.c if (IS_ERR(new)) new 2892 mm/mempolicy.c new->v.nodes = nodes; new 2894 mm/mempolicy.c new->v.preferred_node = first_node(nodes); new 2896 mm/mempolicy.c new->flags |= MPOL_F_LOCAL; new 2902 mm/mempolicy.c new->w.user_nodemask = nodes; new 2913 mm/mempolicy.c *mpol = new; new 213 mm/migrate.c struct page *new; new 220 mm/migrate.c new = page; new 222 mm/migrate.c new = page - pvmw.page->index + new 229 mm/migrate.c remove_migration_pmd(&pvmw, new); new 234 mm/migrate.c get_page(new); new 235 mm/migrate.c pte = pte_mkold(mk_pte(new, READ_ONCE(vma->vm_page_prot))); new 246 mm/migrate.c if (unlikely(is_zone_device_page(new))) { new 247 mm/migrate.c if (is_device_private_page(new)) { new 248 mm/migrate.c entry = make_device_private_entry(new, pte_write(pte)); new 254 mm/migrate.c if (PageHuge(new)) { new 256 mm/migrate.c pte = arch_make_huge_pte(pte, vma, new, 0); new 258 mm/migrate.c if (PageAnon(new)) new 259 mm/migrate.c hugepage_add_anon_rmap(new, vma, pvmw.address); new 261 mm/migrate.c page_dup_rmap(new, true); new 267 mm/migrate.c if (PageAnon(new)) new 268 mm/migrate.c page_add_anon_rmap(new, vma, pvmw.address, false); new 270 mm/migrate.c page_add_file_rmap(new, false); new 272 mm/migrate.c if (vma->vm_flags & VM_LOCKED && !PageTransCompound(new)) new 273 mm/migrate.c mlock_vma_page(new); new 289 mm/migrate.c void remove_migration_ptes(struct page *old, struct page *new, bool locked) new 297 mm/migrate.c rmap_walk_locked(new, &rwc); new 299 mm/migrate.c rmap_walk(new, &rwc); new 2659 mm/mmap.c struct vm_area_struct *new; new 2668 mm/mmap.c new = vm_area_dup(vma); new 2669 mm/mmap.c if (!new) new 2673 mm/mmap.c new->vm_end = addr; new 2675 mm/mmap.c new->vm_start = addr; new 2676 mm/mmap.c new->vm_pgoff += ((addr - vma->vm_start) >> PAGE_SHIFT); new 2679 mm/mmap.c err = vma_dup_policy(vma, new); new 2683 mm/mmap.c err = anon_vma_clone(new, vma); new 2687 mm/mmap.c if (new->vm_file) new 2688 mm/mmap.c get_file(new->vm_file); new 2690 mm/mmap.c if (new->vm_ops && new->vm_ops->open) new 2691 mm/mmap.c new->vm_ops->open(new); new 2695 mm/mmap.c ((addr - new->vm_start) >> PAGE_SHIFT), new); new 2697 mm/mmap.c err = vma_adjust(vma, vma->vm_start, addr, vma->vm_pgoff, new); new 2704 mm/mmap.c if (new->vm_ops && new->vm_ops->close) new 2705 mm/mmap.c new->vm_ops->close(new); new 2706 mm/mmap.c if (new->vm_file) new 2707 mm/mmap.c fput(new->vm_file); new 2708 mm/mmap.c unlink_anon_vmas(new); new 2710 mm/mmap.c mpol_put(vma_policy(new)); new 2712 mm/mmap.c vm_area_free(new); new 1386 mm/nommu.c struct vm_area_struct *new; new 1402 mm/nommu.c new = vm_area_dup(vma); new 1403 mm/nommu.c if (!new) { new 1410 mm/nommu.c new->vm_region = region; new 1415 mm/nommu.c region->vm_top = region->vm_end = new->vm_end = addr; new 1417 mm/nommu.c region->vm_start = new->vm_start = addr; new 1418 mm/nommu.c region->vm_pgoff = new->vm_pgoff += npages; new 1421 mm/nommu.c if (new->vm_ops && new->vm_ops->open) new 1422 mm/nommu.c new->vm_ops->open(new); new 1435 mm/nommu.c add_nommu_region(new->vm_region); new 1438 mm/nommu.c add_vma_to_mm(mm, new); new 57 mm/page_counter.c long new; new 59 mm/page_counter.c new = atomic_long_sub_return(nr_pages, &counter->usage); new 60 mm/page_counter.c propagate_protected_usage(counter, new); new 62 mm/page_counter.c WARN_ON_ONCE(new < 0); new 77 mm/page_counter.c long new; new 79 mm/page_counter.c new = atomic_long_add_return(nr_pages, &c->usage); new 80 mm/page_counter.c propagate_protected_usage(counter, new); new 85 mm/page_counter.c if (new > c->watermark) new 86 mm/page_counter.c c->watermark = new; new 106 mm/page_counter.c long new; new 121 mm/page_counter.c new = atomic_long_add_return(nr_pages, &c->usage); new 122 mm/page_counter.c if (new > c->max) { new 124 mm/page_counter.c propagate_protected_usage(counter, new); new 133 mm/page_counter.c propagate_protected_usage(counter, new); new 138 mm/page_counter.c if (new > c->watermark) new 139 mm/page_counter.c c->watermark = new; new 200 mm/slab_common.c struct memcg_cache_array *old, *new; new 202 mm/slab_common.c new = kvzalloc(sizeof(struct memcg_cache_array) + new 204 mm/slab_common.c if (!new) new 210 mm/slab_common.c memcpy(new->entries, old->entries, new 213 mm/slab_common.c rcu_assign_pointer(s->memcg_params.memcg_caches, new); new 1795 mm/slub.c struct page new; new 1806 mm/slub.c new.counters = counters; new 1807 mm/slub.c *objects = new.objects - new.inuse; new 1809 mm/slub.c new.inuse = page->objects; new 1810 mm/slub.c new.freelist = NULL; new 1812 mm/slub.c new.freelist = freelist; new 1815 mm/slub.c VM_BUG_ON(new.frozen); new 1816 mm/slub.c new.frozen = 1; new 1820 mm/slub.c new.freelist, new.counters, new 2045 mm/slub.c struct page new; new 2069 mm/slub.c new.counters = counters; new 2070 mm/slub.c new.inuse--; new 2071 mm/slub.c VM_BUG_ON(!new.frozen); new 2075 mm/slub.c freelist, new.counters, new 2102 mm/slub.c new.counters = old.counters; new 2104 mm/slub.c new.inuse--; new 2106 mm/slub.c new.freelist = freelist; new 2108 mm/slub.c new.freelist = old.freelist; new 2110 mm/slub.c new.frozen = 0; new 2112 mm/slub.c if (!new.inuse && n->nr_partial >= s->min_partial) new 2114 mm/slub.c else if (new.freelist) { new 2153 mm/slub.c new.freelist, new.counters, new 2189 mm/slub.c struct page new; new 2209 mm/slub.c new.counters = old.counters; new 2210 mm/slub.c new.freelist = old.freelist; new 2212 mm/slub.c new.frozen = 0; new 2216 mm/slub.c new.freelist, new.counters, new 2219 mm/slub.c if (unlikely(!new.inuse && n->nr_partial >= s->min_partial)) { new 2496 mm/slub.c struct page new; new 2504 mm/slub.c new.counters = counters; new 2505 mm/slub.c VM_BUG_ON(!new.frozen); new 2507 mm/slub.c new.inuse = page->objects; new 2508 mm/slub.c new.frozen = freelist != NULL; new 2512 mm/slub.c NULL, new.counters, new 2843 mm/slub.c struct page new; new 2862 mm/slub.c new.counters = counters; new 2863 mm/slub.c was_frozen = new.frozen; new 2864 mm/slub.c new.inuse -= cnt; new 2865 mm/slub.c if ((!new.inuse || !prior) && !was_frozen) { new 2875 mm/slub.c new.frozen = 1; new 2895 mm/slub.c head, new.counters, new 2904 mm/slub.c if (new.frozen && !was_frozen) { new 2917 mm/slub.c if (unlikely(!new.inuse && n->nr_partial >= s->min_partial)) new 98 mm/swap_cgroup.c unsigned short old, unsigned short new) new 110 mm/swap_cgroup.c sc->id = new; new 21 net/6lowpan/nhc.c struct rb_node **new = &rb_root.rb_node, *parent = NULL; new 24 net/6lowpan/nhc.c while (*new) { new 25 net/6lowpan/nhc.c struct lowpan_nhc *this = rb_entry(*new, struct lowpan_nhc, new 40 net/6lowpan/nhc.c parent = *new; new 42 net/6lowpan/nhc.c new = &((*new)->rb_left); new 44 net/6lowpan/nhc.c new = &((*new)->rb_right); new 50 net/6lowpan/nhc.c rb_link_node(&nhc->node, parent, new); new 373 net/batman-adv/multicast.c struct batadv_hw_addr *new; new 405 net/batman-adv/multicast.c new = kmalloc(sizeof(*new), GFP_ATOMIC); new 406 net/batman-adv/multicast.c if (!new) { new 411 net/batman-adv/multicast.c ether_addr_copy(new->addr, mcast_addr); new 412 net/batman-adv/multicast.c hlist_add_head(&new->list, mcast_list); new 440 net/batman-adv/multicast.c struct batadv_hw_addr *new; new 477 net/batman-adv/multicast.c new = kmalloc(sizeof(*new), GFP_ATOMIC); new 478 net/batman-adv/multicast.c if (!new) { new 483 net/batman-adv/multicast.c ether_addr_copy(new->addr, mcast_addr); new 484 net/batman-adv/multicast.c hlist_add_head(&new->list, mcast_list); new 594 net/batman-adv/multicast.c struct batadv_hw_addr *new; new 639 net/batman-adv/multicast.c new = kmalloc(sizeof(*new), GFP_ATOMIC); new 640 net/batman-adv/multicast.c if (!new) { new 645 net/batman-adv/multicast.c ether_addr_copy(new->addr, mcast_addr); new 646 net/batman-adv/multicast.c hlist_add_head(&new->list, mcast_list); new 1230 net/batman-adv/tp_meter.c struct batadv_tp_unacked *un, *new; new 1234 net/batman-adv/tp_meter.c new = kmalloc(sizeof(*new), GFP_ATOMIC); new 1235 net/batman-adv/tp_meter.c if (unlikely(!new)) new 1240 net/batman-adv/tp_meter.c new->seqno = ntohl(icmp->seqno); new 1242 net/batman-adv/tp_meter.c new->len = payload_len; new 1247 net/batman-adv/tp_meter.c list_add(&new->list, &tp_vars->unacked_list); new 1260 net/batman-adv/tp_meter.c if (new->seqno == un->seqno) { new 1261 net/batman-adv/tp_meter.c if (new->len > un->len) new 1262 net/batman-adv/tp_meter.c un->len = new->len; new 1263 net/batman-adv/tp_meter.c kfree(new); new 1269 net/batman-adv/tp_meter.c if (batadv_seq_before(new->seqno, un->seqno)) new 1276 net/batman-adv/tp_meter.c list_add_tail(&new->list, &un->list); new 1283 net/batman-adv/tp_meter.c list_add(&new->list, &tp_vars->unacked_list); new 858 net/bluetooth/rfcomm/tty.c struct ktermios *new = &tty->termios; new 860 net/bluetooth/rfcomm/tty.c int new_baud_rate = tty_termios_baud_rate(new); new 873 net/bluetooth/rfcomm/tty.c if ((old->c_cflag & CRTSCTS) && !(new->c_cflag & CRTSCTS)) new 877 net/bluetooth/rfcomm/tty.c if (((old->c_cflag & PARENB) != (new->c_cflag & PARENB)) || new 878 net/bluetooth/rfcomm/tty.c ((old->c_cflag & PARODD) != (new->c_cflag & PARODD))) { new 884 net/bluetooth/rfcomm/tty.c if (new->c_cflag & PARENB) { new 885 net/bluetooth/rfcomm/tty.c if (new->c_cflag & PARODD) { new 898 net/bluetooth/rfcomm/tty.c if (old->c_cc[VSTOP] != new->c_cc[VSTOP]) { new 900 net/bluetooth/rfcomm/tty.c x_on = new->c_cc[VSTOP]; new 907 net/bluetooth/rfcomm/tty.c if (old->c_cc[VSTART] != new->c_cc[VSTART]) { new 909 net/bluetooth/rfcomm/tty.c x_off = new->c_cc[VSTART]; new 917 net/bluetooth/rfcomm/tty.c if ((old->c_cflag & CSTOPB) != (new->c_cflag & CSTOPB)) new 923 net/bluetooth/rfcomm/tty.c if (new->c_cflag & CSTOPB) new 929 net/bluetooth/rfcomm/tty.c if ((old->c_cflag & CSIZE) != (new->c_cflag & CSIZE)) new 932 net/bluetooth/rfcomm/tty.c switch (new->c_cflag & CSIZE) { new 639 net/ceph/osdmap.c static int __insert_pg_pool(struct rb_root *root, struct ceph_pg_pool_info *new) new 648 net/ceph/osdmap.c if (new->id < pi->id) new 650 net/ceph/osdmap.c else if (new->id > pi->id) new 656 net/ceph/osdmap.c rb_link_node(&new->node, parent, p); new 657 net/ceph/osdmap.c rb_insert_color(&new->node, root); new 5075 net/core/dev.c struct bpf_prog *new = xdp->prog; new 5080 net/core/dev.c rcu_assign_pointer(dev->xdp_prog, new); new 5084 net/core/dev.c if (old && !new) { new 5086 net/core/dev.c } else if (new && !old) { new 5929 net/core/dev.c unsigned long val, new; new 5935 net/core/dev.c new = val | NAPIF_STATE_SCHED; new 5943 net/core/dev.c new |= (val & NAPIF_STATE_SCHED) / NAPIF_STATE_SCHED * new 5945 net/core/dev.c } while (cmpxchg(&n->state, val, new) != val); new 5965 net/core/dev.c unsigned long flags, val, new; new 6007 net/core/dev.c new = val & ~(NAPIF_STATE_MISSED | NAPIF_STATE_SCHED); new 6013 net/core/dev.c new |= (val & NAPIF_STATE_MISSED) / NAPIF_STATE_MISSED * new 6015 net/core/dev.c } while (cmpxchg(&n->state, val, new) != val); new 205 net/core/dst.c unsigned long prev, new; new 210 net/core/dst.c new = (unsigned long) p; new 211 net/core/dst.c prev = cmpxchg(&dst->_metrics, old, new); new 231 net/core/dst.c unsigned long prev, new; new 233 net/core/dst.c new = ((unsigned long) &dst_default_metrics) | DST_METRICS_READ_ONLY; new 234 net/core/dst.c prev = cmpxchg(&dst->_metrics, old, new); new 2349 net/core/filter.c u32 new, i = 0, l = 0, space, copy = 0, offset = 0; new 2423 net/core/filter.c new = i; new 2455 net/core/filter.c msg->sg.copy[new] = false; new 2456 net/core/filter.c sg_set_page(&msg->sg.data[new], page, len + copy, 0); new 2459 net/core/filter.c sk_msg_iter_var_next(new); new 2460 net/core/filter.c msg->sg.data[new] = rsge; new 1227 net/core/neighbour.c u8 new, u32 flags, u32 nlmsg_pid, new 1237 net/core/neighbour.c trace_neigh_update(neigh, lladdr, new, flags, nlmsg_pid); new 1255 net/core/neighbour.c if (!(new & NUD_VALID)) { new 1259 net/core/neighbour.c neigh->nud_state = new; new 1263 net/core/neighbour.c (new & NUD_FAILED)) { new 1298 net/core/neighbour.c if (new & NUD_CONNECTED) new 1312 net/core/neighbour.c new = NUD_STALE; new 1316 net/core/neighbour.c if (lladdr == neigh->ha && new == NUD_STALE && new 1318 net/core/neighbour.c new = old; new 1326 net/core/neighbour.c if (new != old || lladdr != neigh->ha) new 1329 net/core/neighbour.c if (new != old) { new 1331 net/core/neighbour.c if (new & NUD_PROBE) new 1333 net/core/neighbour.c if (new & NUD_IN_TIMER) new 1335 net/core/neighbour.c ((new & NUD_REACHABLE) ? new 1338 net/core/neighbour.c neigh->nud_state = new; new 1347 net/core/neighbour.c if (!(new & NUD_CONNECTED)) new 1352 net/core/neighbour.c if (new == old) new 1354 net/core/neighbour.c if (new & NUD_CONNECTED) new 1399 net/core/neighbour.c if (((new ^ old) & NUD_PERMANENT) || ext_learn_change) new 1410 net/core/neighbour.c int neigh_update(struct neighbour *neigh, const u8 *lladdr, u8 new, new 1413 net/core/neighbour.c return __neigh_update(neigh, lladdr, new, flags, nlmsg_pid, NULL); new 82 net/core/net-sysfs.c unsigned long new; new 88 net/core/net-sysfs.c ret = kstrtoul(buf, 0, &new); new 96 net/core/net-sysfs.c ret = (*set)(netdev, new); new 43 net/core/netprio_cgroup.c struct netprio_map *old, *new; new 59 net/core/netprio_cgroup.c sizeof(new->priomap[0]); new 69 net/core/netprio_cgroup.c new = kzalloc(new_sz, GFP_KERNEL); new 70 net/core/netprio_cgroup.c if (!new) new 74 net/core/netprio_cgroup.c memcpy(new->priomap, old->priomap, new 77 net/core/netprio_cgroup.c new->priomap_len = new_len; new 80 net/core/netprio_cgroup.c rcu_assign_pointer(dev->priomap, new); new 929 net/core/skbuff.c static void __copy_skb_header(struct sk_buff *new, const struct sk_buff *old) new 931 net/core/skbuff.c new->tstamp = old->tstamp; new 933 net/core/skbuff.c new->dev = old->dev; new 934 net/core/skbuff.c memcpy(new->cb, old->cb, sizeof(old->cb)); new 935 net/core/skbuff.c skb_dst_copy(new, old); new 936 net/core/skbuff.c __skb_ext_copy(new, old); new 937 net/core/skbuff.c __nf_copy(new, old, false); new 942 net/core/skbuff.c new->queue_mapping = old->queue_mapping; new 944 net/core/skbuff.c memcpy(&new->headers_start, &old->headers_start, new 1475 net/core/skbuff.c void skb_copy_header(struct sk_buff *new, const struct sk_buff *old) new 1477 net/core/skbuff.c __copy_skb_header(new, old); new 1479 net/core/skbuff.c skb_shinfo(new)->gso_size = skb_shinfo(old)->gso_size; new 1480 net/core/skbuff.c skb_shinfo(new)->gso_segs = skb_shinfo(old)->gso_segs; new 1481 net/core/skbuff.c skb_shinfo(new)->gso_type = skb_shinfo(old)->gso_type; new 5981 net/core/skbuff.c struct skb_ext *new = kmem_cache_alloc(skbuff_ext_cache, GFP_ATOMIC); new 5983 net/core/skbuff.c if (new) { new 5984 net/core/skbuff.c memset(new->offset, 0, sizeof(new->offset)); new 5985 net/core/skbuff.c refcount_set(&new->refcnt, 1); new 5988 net/core/skbuff.c return new; new 5994 net/core/skbuff.c struct skb_ext *new; new 5999 net/core/skbuff.c new = kmem_cache_alloc(skbuff_ext_cache, GFP_ATOMIC); new 6000 net/core/skbuff.c if (!new) new 6003 net/core/skbuff.c memcpy(new, old, old->chunks * SKB_EXT_ALIGN_VALUE); new 6004 net/core/skbuff.c refcount_set(&new->refcnt, 1); new 6016 net/core/skbuff.c return new; new 6035 net/core/skbuff.c struct skb_ext *new, *old = NULL; new 6041 net/core/skbuff.c new = skb_ext_maybe_cow(old, skb->active_extensions); new 6042 net/core/skbuff.c if (!new) new 6045 net/core/skbuff.c if (__skb_ext_exist(new, id)) new 6048 net/core/skbuff.c newoff = new->chunks; new 6050 net/core/skbuff.c newoff = SKB_EXT_CHUNKSIZEOF(*new); new 6052 net/core/skbuff.c new = skb_ext_alloc(); new 6053 net/core/skbuff.c if (!new) new 6058 net/core/skbuff.c new->chunks = newlen; new 6059 net/core/skbuff.c new->offset[id] = newoff; new 6061 net/core/skbuff.c skb->extensions = new; new 6063 net/core/skbuff.c return skb_ext_get_ptr(new, id); new 642 net/core/sock_map.c struct bpf_htab_elem *new; new 651 net/core/sock_map.c new = kmalloc_node(htab->elem_size, GFP_ATOMIC | __GFP_NOWARN, new 653 net/core/sock_map.c if (!new) { new 657 net/core/sock_map.c memcpy(new->key, key, key_size); new 658 net/core/sock_map.c new->sk = sk; new 659 net/core/sock_map.c new->hash = hash; new 660 net/core/sock_map.c return new; new 1849 net/dcb/dcbnl.c int dcb_setapp(struct net_device *dev, struct dcb_app *new) new 1856 net/dcb/dcbnl.c memcpy(&event.app, new, sizeof(event.app)); new 1862 net/dcb/dcbnl.c itr = dcb_app_lookup(new, dev->ifindex, -1); new 1864 net/dcb/dcbnl.c if (new->priority) new 1865 net/dcb/dcbnl.c itr->app.priority = new->priority; new 1873 net/dcb/dcbnl.c if (new->priority) new 1874 net/dcb/dcbnl.c err = dcb_app_add(new, dev->ifindex); new 1913 net/dcb/dcbnl.c int dcb_ieee_setapp(struct net_device *dev, struct dcb_app *new) new 1919 net/dcb/dcbnl.c memcpy(&event.app, new, sizeof(event.app)); new 1925 net/dcb/dcbnl.c if (dcb_app_lookup(new, dev->ifindex, new->priority)) { new 1930 net/dcb/dcbnl.c err = dcb_app_add(new, dev->ifindex); new 349 net/dccp/ackvec.c struct dccp_ackvec_parsed *new = kmalloc(sizeof(*new), GFP_ATOMIC); new 351 net/dccp/ackvec.c if (new == NULL) new 353 net/dccp/ackvec.c new->vec = vec; new 354 net/dccp/ackvec.c new->len = len; new 355 net/dccp/ackvec.c new->nonce = nonce; new 357 net/dccp/ackvec.c list_add_tail(&new->node, head); new 138 net/dccp/ccids/lib/loss_interval.c struct tfrc_loss_interval *cur = tfrc_lh_peek(lh), *new; new 143 net/dccp/ccids/lib/loss_interval.c new = tfrc_lh_demand_next(lh); new 144 net/dccp/ccids/lib/loss_interval.c if (unlikely(new == NULL)) { new 149 net/dccp/ccids/lib/loss_interval.c new->li_seqno = tfrc_rx_hist_loss_prev(rh)->tfrchrx_seqno; new 150 net/dccp/ccids/lib/loss_interval.c new->li_ccval = tfrc_rx_hist_loss_prev(rh)->tfrchrx_ccval; new 151 net/dccp/ccids/lib/loss_interval.c new->li_is_closed = 0; new 154 net/dccp/ccids/lib/loss_interval.c lh->i_mean = new->li_length = (*calc_first_li)(sk); new 156 net/dccp/ccids/lib/loss_interval.c cur->li_length = dccp_delta_seqno(cur->li_seqno, new->li_seqno); new 157 net/dccp/ccids/lib/loss_interval.c new->li_length = dccp_delta_seqno(new->li_seqno, new 390 net/dccp/feat.c struct dccp_feat_entry *new; new 396 net/dccp/feat.c new = kmemdup(original, sizeof(struct dccp_feat_entry), gfp_any()); new 397 net/dccp/feat.c if (new == NULL) new 400 net/dccp/feat.c if (type == FEAT_SP && dccp_feat_clone_sp_val(&new->val, new 403 net/dccp/feat.c kfree(new); new 406 net/dccp/feat.c return new; new 482 net/dccp/feat.c struct dccp_feat_entry *new = dccp_feat_entry_new(fn_list, feat, local); new 484 net/dccp/feat.c if (new == NULL) new 487 net/dccp/feat.c new->feat_num = feat; new 488 net/dccp/feat.c new->is_local = local; new 489 net/dccp/feat.c new->state = FEAT_INITIALISING; new 490 net/dccp/feat.c new->needs_confirm = false; new 491 net/dccp/feat.c new->empty_confirm = false; new 492 net/dccp/feat.c new->val = *fval; new 493 net/dccp/feat.c new->needs_mandatory = mandatory; new 510 net/dccp/feat.c struct dccp_feat_entry *new = dccp_feat_entry_new(fn_list, feat, local); new 512 net/dccp/feat.c if (new == NULL) new 515 net/dccp/feat.c new->feat_num = feat; new 516 net/dccp/feat.c new->is_local = local; new 517 net/dccp/feat.c new->state = FEAT_STABLE; /* transition in 6.6.2 */ new 518 net/dccp/feat.c new->needs_confirm = true; new 519 net/dccp/feat.c new->empty_confirm = (fval == NULL); new 520 net/dccp/feat.c new->val.nn = 0; /* zeroes the whole structure */ new 521 net/dccp/feat.c if (!new->empty_confirm) new 522 net/dccp/feat.c new->val = *fval; new 523 net/dccp/feat.c new->needs_mandatory = false; new 552 net/dccp/feat.c struct dccp_feat_entry *entry, *new; new 556 net/dccp/feat.c new = dccp_feat_clone_entry(entry); new 557 net/dccp/feat.c if (new == NULL) new 559 net/dccp/feat.c list_add_tail(&new->node, to); new 142 net/ipv4/fib_frontend.c struct fib_table *new) new 145 net/ipv4/fib_frontend.c switch (new->tb_id) { new 147 net/ipv4/fib_frontend.c rcu_assign_pointer(net->ipv4.fib_main, new); new 150 net/ipv4/fib_frontend.c rcu_assign_pointer(net->ipv4.fib_default, new); new 158 net/ipv4/fib_frontend.c hlist_replace_rcu(&old->tb_hlist, &new->tb_hlist); new 163 net/ipv4/fib_frontend.c struct fib_table *old, *new, *main_table; new 170 net/ipv4/fib_frontend.c new = fib_trie_unmerge(old); new 171 net/ipv4/fib_frontend.c if (!new) new 175 net/ipv4/fib_frontend.c if (new == old) new 179 net/ipv4/fib_frontend.c fib_replace_table(net, old, new); new 1874 net/ipv4/fib_semantics.c void fib_nhc_update_mtu(struct fib_nh_common *nhc, u32 new, u32 orig) new 1890 net/ipv4/fib_semantics.c if (new <= fnhe->fnhe_pmtu) { new 1891 net/ipv4/fib_semantics.c fnhe->fnhe_pmtu = new; new 1894 net/ipv4/fib_semantics.c } else if (new < fnhe->fnhe_pmtu || new 1896 net/ipv4/fib_semantics.c fnhe->fnhe_pmtu = new; new 1015 net/ipv4/fib_trie.c struct fib_alias *new, t_key key) new 1019 net/ipv4/fib_trie.c l = leaf_new(key, new); new 1052 net/ipv4/fib_trie.c node_push_suffix(tp, new->fa_slen); new 1068 net/ipv4/fib_trie.c struct key_vector *l, struct fib_alias *new, new 1072 net/ipv4/fib_trie.c return fib_insert_node(t, tp, new, key); new 1075 net/ipv4/fib_trie.c hlist_add_before_rcu(&new->fa_list, &fa->fa_list); new 1080 net/ipv4/fib_trie.c if (new->fa_slen < last->fa_slen) new 1082 net/ipv4/fib_trie.c if ((new->fa_slen == last->fa_slen) && new 1083 net/ipv4/fib_trie.c (new->tb_id > last->tb_id)) new 1089 net/ipv4/fib_trie.c hlist_add_behind_rcu(&new->fa_list, &fa->fa_list); new 1091 net/ipv4/fib_trie.c hlist_add_head_rcu(&new->fa_list, &l->leaf); new 1095 net/ipv4/fib_trie.c if (l->slen < new->fa_slen) { new 1096 net/ipv4/fib_trie.c l->slen = new->fa_slen; new 1097 net/ipv4/fib_trie.c node_push_suffix(tp, new->fa_slen); new 394 net/ipv4/netfilter/nf_nat_h323.c static void ip_nat_q931_expect(struct nf_conn *new, new 400 net/ipv4/netfilter/nf_nat_h323.c nf_nat_follow_master(new, this); new 405 net/ipv4/netfilter/nf_nat_h323.c BUG_ON(new->status & IPS_NAT_DONE_MASK); new 410 net/ipv4/netfilter/nf_nat_h323.c new->tuplehash[!this->dir].tuple.src.u3; new 411 net/ipv4/netfilter/nf_nat_h323.c nf_nat_setup_info(new, &range, NF_NAT_MANIP_SRC); new 417 net/ipv4/netfilter/nf_nat_h323.c new->master->tuplehash[!this->dir].tuple.src.u3; new 418 net/ipv4/netfilter/nf_nat_h323.c nf_nat_setup_info(new, &range, NF_NAT_MANIP_DST); new 496 net/ipv4/netfilter/nf_nat_h323.c static void ip_nat_callforwarding_expect(struct nf_conn *new, new 502 net/ipv4/netfilter/nf_nat_h323.c BUG_ON(new->status & IPS_NAT_DONE_MASK); new 507 net/ipv4/netfilter/nf_nat_h323.c new->tuplehash[!this->dir].tuple.src.u3; new 508 net/ipv4/netfilter/nf_nat_h323.c nf_nat_setup_info(new, &range, NF_NAT_MANIP_SRC); new 514 net/ipv4/netfilter/nf_nat_h323.c nf_nat_setup_info(new, &range, NF_NAT_MANIP_DST); new 607 net/ipv4/nexthop.c static int fib6_check_nh_list(struct nexthop *old, struct nexthop *new, new 620 net/ipv4/nexthop.c return fib6_check_nexthop(new, NULL, extack); new 671 net/ipv4/nexthop.c static int fib_check_nh_list(struct nexthop *old, struct nexthop *new, new 679 net/ipv4/nexthop.c err = fib_check_nexthop(new, fi->fib_scope, extack); new 854 net/ipv4/nexthop.c struct nexthop *new, new 860 net/ipv4/nexthop.c if (!new->is_group) { new 866 net/ipv4/nexthop.c newg = rtnl_dereference(new->nh_grp); new 875 net/ipv4/nexthop.c oldg->nh_entries[i].nh_parent = new; new 877 net/ipv4/nexthop.c rcu_assign_pointer(new->nh_grp, oldg); new 883 net/ipv4/nexthop.c struct nexthop *new, new 888 net/ipv4/nexthop.c if (new->is_group) { new 894 net/ipv4/nexthop.c newi = rtnl_dereference(new->nh_info); new 897 net/ipv4/nexthop.c oldi->nh_parent = new; new 899 net/ipv4/nexthop.c old->protocol = new->protocol; new 900 net/ipv4/nexthop.c old->nh_flags = new->nh_flags; new 903 net/ipv4/nexthop.c rcu_assign_pointer(new->nh_info, oldi); new 949 net/ipv4/nexthop.c struct nexthop *new, struct netlink_ext_ack *extack) new 958 net/ipv4/nexthop.c err = fib_check_nh_list(old, new, extack); new 962 net/ipv4/nexthop.c err = fib6_check_nh_list(old, new, extack); new 966 net/ipv4/nexthop.c if (!new->is_group) { new 967 net/ipv4/nexthop.c struct nh_info *nhi = rtnl_dereference(new->nh_info); new 982 net/ipv4/nexthop.c err = fib_check_nh_list(nhge->nh_parent, new, extack); new 986 net/ipv4/nexthop.c err = fib6_check_nh_list(nhge->nh_parent, new, extack); new 992 net/ipv4/nexthop.c err = replace_nexthop_grp(net, old, new, extack); new 994 net/ipv4/nexthop.c err = replace_nexthop_single(net, old, new, extack); new 999 net/ipv4/nexthop.c __remove_nexthop(net, new, NULL); new 1000 net/ipv4/nexthop.c nexthop_put(new); new 1097 net/ipv4/route.c bool new = false; new 1120 net/ipv4/route.c new = true; new 1126 net/ipv4/route.c if (new) new 1133 net/ipv4/route.c new = true; new 1136 net/ipv4/route.c if (new) new 2685 net/ipv4/route.c struct dst_entry *new = &rt->dst; new 2687 net/ipv4/route.c new->__use = 1; new 2688 net/ipv4/route.c new->input = dst_discard; new 2689 net/ipv4/route.c new->output = dst_discard_out; new 2691 net/ipv4/route.c new->dev = net->loopback_dev; new 2692 net/ipv4/route.c if (new->dev) new 2693 net/ipv4/route.c dev_hold(new->dev); new 2309 net/ipv4/tcp_output.c static void tcp_chrono_set(struct tcp_sock *tp, const enum tcp_chrono new) new 2317 net/ipv4/tcp_output.c tp->chrono_type = new; new 6187 net/ipv6/addrconf.c int old, new; new 6191 net/ipv6/addrconf.c new = *valp; new 6193 net/ipv6/addrconf.c if (write && old != new) { new 917 net/ipv6/calipso.c struct ipv6_opt_hdr *new; new 932 net/ipv6/calipso.c new = kzalloc(buf_len, GFP_ATOMIC); new 933 net/ipv6/calipso.c if (!new) new 937 net/ipv6/calipso.c memcpy(new, hop, start); new 938 net/ipv6/calipso.c ret_val = calipso_genopt((unsigned char *)new, start, buf_len, doi_def, new 941 net/ipv6/calipso.c kfree(new); new 948 net/ipv6/calipso.c calipso_pad_write((unsigned char *)new, buf_len, pad); new 952 net/ipv6/calipso.c memcpy((char *)new + buf_len, (char *)hop + end, hop_len - end); new 955 net/ipv6/calipso.c new->nexthdr = 0; new 956 net/ipv6/calipso.c new->hdrlen = buf_len / 8 - 1; new 958 net/ipv6/calipso.c return new; new 976 net/ipv6/calipso.c struct ipv6_opt_hdr **new) new 988 net/ipv6/calipso.c *new = NULL; new 993 net/ipv6/calipso.c *new = kzalloc(hop_len - delta, GFP_ATOMIC); new 994 net/ipv6/calipso.c if (!*new) new 997 net/ipv6/calipso.c memcpy(*new, hop, start); new 998 net/ipv6/calipso.c (*new)->hdrlen -= delta / 8; new 1000 net/ipv6/calipso.c calipso_pad_write((unsigned char *)*new, start, pad); new 1002 net/ipv6/calipso.c memcpy((char *)*new + start + pad, (char *)hop + end, new 1134 net/ipv6/calipso.c struct ipv6_opt_hdr *old, *new; new 1141 net/ipv6/calipso.c new = calipso_opt_insert(old, doi_def, secattr); new 1143 net/ipv6/calipso.c if (IS_ERR(new)) new 1144 net/ipv6/calipso.c return PTR_ERR(new); new 1146 net/ipv6/calipso.c ret_val = calipso_opt_update(sk, new); new 1148 net/ipv6/calipso.c kfree(new); new 1199 net/ipv6/calipso.c struct ipv6_opt_hdr *old, *new; new 1207 net/ipv6/calipso.c new = calipso_opt_insert(old, doi_def, secattr); new 1208 net/ipv6/calipso.c if (IS_ERR(new)) new 1209 net/ipv6/calipso.c return PTR_ERR(new); new 1211 net/ipv6/calipso.c txopts = ipv6_renew_options(sk, req_inet->ipv6_opt, IPV6_HOPOPTS, new); new 1213 net/ipv6/calipso.c kfree(new); new 1238 net/ipv6/calipso.c struct ipv6_opt_hdr *new; new 1245 net/ipv6/calipso.c if (calipso_opt_del(req_inet->ipv6_opt->hopopt, &new)) new 1248 net/ipv6/calipso.c txopts = ipv6_renew_options(sk, req_inet->ipv6_opt, IPV6_HOPOPTS, new); new 1257 net/ipv6/calipso.c kfree(new); new 1017 net/ipv6/exthdrs.c struct ipv6_opt_hdr *new, new 1022 net/ipv6/exthdrs.c src = (renewtype == newtype ? new : old); new 92 net/ipv6/ip6_fib.c int new, old; new 96 net/ipv6/ip6_fib.c new = old < INT_MAX ? old + 1 : 1; new 98 net/ipv6/ip6_fib.c old, new) != old); new 99 net/ipv6/ip6_fib.c return new; new 405 net/ipv6/ipv6_sockglue.c struct ipv6_opt_hdr *new = NULL; new 423 net/ipv6/ipv6_sockglue.c new = memdup_user(optval, optlen); new 424 net/ipv6/ipv6_sockglue.c if (IS_ERR(new)) { new 425 net/ipv6/ipv6_sockglue.c retv = PTR_ERR(new); new 428 net/ipv6/ipv6_sockglue.c if (unlikely(ipv6_optlen(new) > optlen)) { new 429 net/ipv6/ipv6_sockglue.c kfree(new); new 436 net/ipv6/ipv6_sockglue.c opt = ipv6_renew_options(sk, opt, optname, new); new 437 net/ipv6/ipv6_sockglue.c kfree(new); new 759 net/ipv6/ndisc.c const u8 *lladdr, u8 new, u32 flags, u8 icmp6_type, new 762 net/ipv6/ndisc.c neigh_update(neigh, lladdr, new, flags, 0); new 2542 net/ipv6/route.c struct dst_entry *new = NULL; new 2550 net/ipv6/route.c new = &rt->dst; new 2551 net/ipv6/route.c new->__use = 1; new 2552 net/ipv6/route.c new->input = dst_discard; new 2553 net/ipv6/route.c new->output = dst_discard_out; new 2555 net/ipv6/route.c dst_copy_metrics(new, &ort->dst); new 2568 net/ipv6/route.c return new ? new : ERR_PTR(-ENOMEM); new 4108 net/ipv6/route.c netevent.new = &nrt->dst; new 312 net/key/af_key.c static inline void pfkey_hdr_dup(struct sadb_msg *new, new 315 net/key/af_key.c *new = *orig; new 803 net/mac80211/cfg.c struct probe_resp *new, *old; new 810 net/mac80211/cfg.c new = kzalloc(sizeof(struct probe_resp) + resp_len, GFP_KERNEL); new 811 net/mac80211/cfg.c if (!new) new 814 net/mac80211/cfg.c new->len = resp_len; new 815 net/mac80211/cfg.c memcpy(new->data, resp, resp_len); new 818 net/mac80211/cfg.c memcpy(new->csa_counter_offsets, csa->counter_offsets_presp, new 820 net/mac80211/cfg.c sizeof(new->csa_counter_offsets[0])); new 822 net/mac80211/cfg.c rcu_assign_pointer(sdata->u.ap.probe_resp, new); new 834 net/mac80211/cfg.c struct ieee80211_ftm_responder_params *new, *old; new 846 net/mac80211/cfg.c new = kzalloc(sizeof(*new) + len, GFP_KERNEL); new 847 net/mac80211/cfg.c if (!new) new 850 net/mac80211/cfg.c pos = (u8 *)(new + 1); new 852 net/mac80211/cfg.c new->lci_len = lci_len; new 853 net/mac80211/cfg.c new->lci = pos; new 859 net/mac80211/cfg.c new->civicloc_len = civicloc_len; new 860 net/mac80211/cfg.c new->civicloc = pos; new 865 net/mac80211/cfg.c bss_conf->ftmr_params = new; new 875 net/mac80211/cfg.c struct beacon_data *new, *old; new 900 net/mac80211/cfg.c size = sizeof(*new) + new_head_len + new_tail_len; new 902 net/mac80211/cfg.c new = kzalloc(size, GFP_KERNEL); new 903 net/mac80211/cfg.c if (!new) new 912 net/mac80211/cfg.c new->head = ((u8 *) new) + sizeof(*new); new 913 net/mac80211/cfg.c new->tail = new->head + new_head_len; new 914 net/mac80211/cfg.c new->head_len = new_head_len; new 915 net/mac80211/cfg.c new->tail_len = new_tail_len; new 918 net/mac80211/cfg.c new->csa_current_counter = csa->count; new 919 net/mac80211/cfg.c memcpy(new->csa_counter_offsets, csa->counter_offsets_beacon, new 921 net/mac80211/cfg.c sizeof(new->csa_counter_offsets[0])); new 926 net/mac80211/cfg.c memcpy(new->head, params->head, new_head_len); new 928 net/mac80211/cfg.c memcpy(new->head, old->head, new_head_len); new 932 net/mac80211/cfg.c memcpy(new->tail, params->tail, new_tail_len); new 935 net/mac80211/cfg.c memcpy(new->tail, old->tail, new_tail_len); new 940 net/mac80211/cfg.c kfree(new); new 955 net/mac80211/cfg.c kfree(new); new 962 net/mac80211/cfg.c rcu_assign_pointer(sdata->u.ap.beacon, new); new 157 net/mac80211/iface.c u64 new, mask, tmp; new 165 net/mac80211/iface.c new = ((u64)m[0] << 5*8) | ((u64)m[1] << 4*8) | new 191 net/mac80211/iface.c if ((new & ~mask) != (tmp & ~mask)) { new 289 net/mac80211/key.c struct ieee80211_key *new) new 291 net/mac80211/key.c struct ieee80211_local *local = new->local; new 292 net/mac80211/key.c struct sta_info *sta = new->sta; new 297 net/mac80211/key.c if (new->conf.flags & IEEE80211_KEY_FLAG_NO_AUTO_TX) { new 406 net/mac80211/key.c struct ieee80211_key *new) new 413 net/mac80211/key.c if (WARN_ON(!new && !old)) new 416 net/mac80211/key.c if (new) new 417 net/mac80211/key.c list_add_tail_rcu(&new->list, &sdata->key_list); new 419 net/mac80211/key.c WARN_ON(new && old && new->conf.keyidx != old->conf.keyidx); new 421 net/mac80211/key.c if (new && sta && pairwise) { new 425 net/mac80211/key.c ieee80211_pairwise_rekey(old, new); new 434 net/mac80211/key.c if (new) new 435 net/mac80211/key.c ret = ieee80211_key_enable_hw_accel(new); new 439 net/mac80211/key.c idx = new->conf.keyidx; new 440 net/mac80211/key.c if (!new->local->wowlan) new 441 net/mac80211/key.c ret = ieee80211_key_enable_hw_accel(new); new 449 net/mac80211/key.c rcu_assign_pointer(sta->ptk[idx], new); new 450 net/mac80211/key.c if (new && new 451 net/mac80211/key.c !(new->conf.flags & IEEE80211_KEY_FLAG_NO_AUTO_TX)) new 452 net/mac80211/key.c _ieee80211_set_tx_key(new, true); new 454 net/mac80211/key.c rcu_assign_pointer(sta->gtk[idx], new); new 460 net/mac80211/key.c if (new && !old) new 473 net/mac80211/key.c if (defunikey && !new) new 475 net/mac80211/key.c if (defmultikey && !new) new 477 net/mac80211/key.c if (defmgmtkey && !new) new 480 net/mac80211/key.c rcu_assign_pointer(sdata->keys[idx], new); new 481 net/mac80211/key.c if (defunikey && new) new 482 net/mac80211/key.c __ieee80211_set_default_key(sdata, new->conf.keyidx, new 484 net/mac80211/key.c if (defmultikey && new) new 485 net/mac80211/key.c __ieee80211_set_default_key(sdata, new->conf.keyidx, new 487 net/mac80211/key.c if (defmgmtkey && new) new 489 net/mac80211/key.c new->conf.keyidx); new 732 net/mac80211/key.c struct ieee80211_key *new) new 737 net/mac80211/key.c if (!old || new->conf.keylen != old->conf.keylen) new 741 net/mac80211/key.c tk_new = new->conf.key; new 749 net/mac80211/key.c new->conf.cipher == WLAN_CIPHER_SUITE_TKIP && new 750 net/mac80211/key.c new->conf.keylen == WLAN_KEY_LEN_TKIP && new 751 net/mac80211/key.c !(new->conf.flags & IEEE80211_KEY_FLAG_PAIRWISE)) { new 760 net/mac80211/key.c return !crypto_memneq(tk_old, tk_new, new->conf.keylen); new 25 net/mac80211/rc80211_minstrel.h minstrel_ewma(int old, int new, int weight) new 29 net/mac80211/rc80211_minstrel.h diff = new - old; new 3973 net/mac80211/rx.c }, *old, *new = NULL; new 4084 net/mac80211/rx.c new = kmemdup(&fastrx, sizeof(fastrx), GFP_KERNEL); new 4088 net/mac80211/rx.c rcu_assign_pointer(sta->fast_rx, new); new 2992 net/mac80211/util.c struct cfg80211_chan_def new = *chandef; new 3032 net/mac80211/util.c new.width = NL80211_CHAN_WIDTH_80; new 3033 net/mac80211/util.c new.center_freq1 = cf0; new 3040 net/mac80211/util.c new.width = NL80211_CHAN_WIDTH_160; new 3041 net/mac80211/util.c new.center_freq1 = cf1; new 3043 net/mac80211/util.c new.width = NL80211_CHAN_WIDTH_80P80; new 3044 net/mac80211/util.c new.center_freq2 = cf1; new 3050 net/mac80211/util.c new.width = NL80211_CHAN_WIDTH_160; new 3051 net/mac80211/util.c new.center_freq1 = cf0; new 3055 net/mac80211/util.c new.width = NL80211_CHAN_WIDTH_80P80; new 3056 net/mac80211/util.c new.center_freq1 = cf0; new 3057 net/mac80211/util.c new.center_freq2 = cf1; new 3063 net/mac80211/util.c if (!cfg80211_chandef_valid(&new)) new 3066 net/mac80211/util.c *chandef = new; new 217 net/mac802154/llsec.c struct ieee802154_llsec_key_entry *pos, *new; new 244 net/mac802154/llsec.c new = kzalloc(sizeof(*new), GFP_KERNEL); new 245 net/mac802154/llsec.c if (!new) new 256 net/mac802154/llsec.c new->id = *id; new 257 net/mac802154/llsec.c new->key = &mkey->key; new 259 net/mac802154/llsec.c list_add_rcu(&new->list, &sec->table.keys); new 264 net/mac802154/llsec.c kzfree(new); new 537 net/mpls/af_mpls.c struct mpls_route *old, struct mpls_route *new, new 542 net/mpls/af_mpls.c int event = new ? RTM_NEWROUTE : RTM_DELROUTE; new 543 net/mpls/af_mpls.c struct mpls_route *rt = new ? new : old; new 544 net/mpls/af_mpls.c unsigned nlm_flags = (old && new) ? NLM_F_REPLACE : 0; new 551 net/mpls/af_mpls.c struct mpls_route *new, new 561 net/mpls/af_mpls.c rcu_assign_pointer(platform_label[index], new); new 563 net/mpls/af_mpls.c mpls_notify_route(net, index, rt, new, info); new 110 net/netfilter/core.c struct nf_hook_entries *new; new 128 net/netfilter/core.c new = allocate_hook_entries_size(alloc_entries); new 129 net/netfilter/core.c if (!new) new 132 net/netfilter/core.c new_ops = nf_hook_entries_get_hook_ops(new); new 144 net/netfilter/core.c new->hooks[nhooks] = old->hooks[i]; new 148 net/netfilter/core.c new->hooks[nhooks].hook = reg->hook; new 149 net/netfilter/core.c new->hooks[nhooks].priv = reg->priv; new 157 net/netfilter/core.c new->hooks[nhooks].hook = reg->hook; new 158 net/netfilter/core.c new->hooks[nhooks].priv = reg->priv; new 161 net/netfilter/core.c return new; new 226 net/netfilter/core.c struct nf_hook_entries *new = NULL; new 248 net/netfilter/core.c new = allocate_hook_entries_size(hook_entries); new 249 net/netfilter/core.c if (!new) new 252 net/netfilter/core.c new_ops = nf_hook_entries_get_hook_ops(new); new 256 net/netfilter/core.c new->hooks[j] = old->hooks[i]; new 260 net/netfilter/core.c hooks_validate(new); new 262 net/netfilter/core.c rcu_assign_pointer(*pp, new); new 559 net/netfilter/core.c void nf_ct_attach(struct sk_buff *new, const struct sk_buff *skb) new 567 net/netfilter/core.c attach(new, skb); new 261 net/netfilter/nf_conntrack_ecache.c struct nf_ct_event_notifier *new) new 273 net/netfilter/nf_conntrack_ecache.c rcu_assign_pointer(net->ct.nf_conntrack_event_cb, new); new 283 net/netfilter/nf_conntrack_ecache.c struct nf_ct_event_notifier *new) new 290 net/netfilter/nf_conntrack_ecache.c BUG_ON(notify != new); new 298 net/netfilter/nf_conntrack_ecache.c struct nf_exp_event_notifier *new) new 310 net/netfilter/nf_conntrack_ecache.c rcu_assign_pointer(net->ct.nf_expect_event_cb, new); new 320 net/netfilter/nf_conntrack_ecache.c struct nf_exp_event_notifier *new) new 327 net/netfilter/nf_conntrack_ecache.c BUG_ON(notify != new); new 284 net/netfilter/nf_conntrack_expect.c struct nf_conntrack_expect *new; new 286 net/netfilter/nf_conntrack_expect.c new = kmem_cache_alloc(nf_ct_expect_cachep, GFP_ATOMIC); new 287 net/netfilter/nf_conntrack_expect.c if (!new) new 290 net/netfilter/nf_conntrack_expect.c new->master = me; new 291 net/netfilter/nf_conntrack_expect.c refcount_set(&new->use, 1); new 292 net/netfilter/nf_conntrack_expect.c return new; new 399 net/netfilter/nf_conntrack_expect.c struct nf_conntrack_expect *new) new 405 net/netfilter/nf_conntrack_expect.c if (exp->class == new->class) new 43 net/netfilter/nf_conntrack_extend.c struct nf_ct_ext *old, *new; new 56 net/netfilter/nf_conntrack_extend.c oldlen = sizeof(*new); new 72 net/netfilter/nf_conntrack_extend.c new = __krealloc(old, alloc, gfp); new 73 net/netfilter/nf_conntrack_extend.c if (!new) new 77 net/netfilter/nf_conntrack_extend.c memset(new->offset, 0, sizeof(new->offset)); new 78 net/netfilter/nf_conntrack_extend.c ct->ext = new; new 79 net/netfilter/nf_conntrack_extend.c } else if (new != old) { new 81 net/netfilter/nf_conntrack_extend.c rcu_assign_pointer(ct->ext, new); new 84 net/netfilter/nf_conntrack_extend.c new->offset[id] = newoff; new 85 net/netfilter/nf_conntrack_extend.c new->len = newlen; new 86 net/netfilter/nf_conntrack_extend.c memset((void *)new + newoff, 0, newlen - newoff); new 87 net/netfilter/nf_conntrack_extend.c return (void *)new + newoff; new 16 net/netfilter/nf_conntrack_labels.c static int replace_u32(u32 *address, u32 mask, u32 new) new 22 net/netfilter/nf_conntrack_labels.c tmp = (old & mask) ^ new; new 243 net/netfilter/nft_exthdr.c } old, new; new 268 net/netfilter/nft_exthdr.c new.v16 = src; new 273 net/netfilter/nft_exthdr.c if (ntohs(old.v16) <= ntohs(new.v16)) new 278 net/netfilter/nft_exthdr.c if (old.v16 == new.v16) new 281 net/netfilter/nft_exthdr.c put_unaligned(new.v16, (u16*)(opt + offset)); new 283 net/netfilter/nft_exthdr.c old.v16, new.v16, false); new 286 net/netfilter/nft_exthdr.c new.v32 = src; new 289 net/netfilter/nft_exthdr.c if (old.v32 == new.v32) new 292 net/netfilter/nft_exthdr.c put_unaligned(new.v32, (u32*)(opt + offset)); new 294 net/netfilter/nft_exthdr.c old.v32, new.v32, false); new 128 net/netfilter/nft_set_bitmap.c struct nft_bitmap_elem *new = elem->priv, *be; new 132 net/netfilter/nft_set_bitmap.c be = nft_bitmap_elem_find(set, new, genmask); new 138 net/netfilter/nft_set_bitmap.c nft_bitmap_location(set, nft_set_ext_key(&new->ext), &idx, &off); new 141 net/netfilter/nft_set_bitmap.c list_add_tail_rcu(&new->head, &priv->list); new 114 net/netfilter/nft_set_hash.c void *(*new)(struct nft_set *, new 133 net/netfilter/nft_set_hash.c he = new(set, expr, regs); new 218 net/netfilter/nft_set_rbtree.c struct nft_rbtree_elem *new, new 233 net/netfilter/nft_set_rbtree.c nft_set_ext_key(&new->ext), new 241 net/netfilter/nft_set_rbtree.c nft_rbtree_interval_start(new)) { new 244 net/netfilter/nft_set_rbtree.c nft_rbtree_interval_end(new)) { new 254 net/netfilter/nft_set_rbtree.c rb_link_node_rcu(&new->node, parent, p); new 255 net/netfilter/nft_set_rbtree.c rb_insert_color(&new->node, &priv->root); new 158 net/netlink/af_netlink.c struct sk_buff *new; new 160 net/netlink/af_netlink.c new = alloc_skb(len, gfp_mask); new 161 net/netlink/af_netlink.c if (new == NULL) new 164 net/netlink/af_netlink.c NETLINK_CB(new).portid = NETLINK_CB(skb).portid; new 165 net/netlink/af_netlink.c NETLINK_CB(new).dst_group = NETLINK_CB(skb).dst_group; new 166 net/netlink/af_netlink.c NETLINK_CB(new).creds = NETLINK_CB(skb).creds; new 168 net/netlink/af_netlink.c skb_put_data(new, skb->data, len); new 169 net/netlink/af_netlink.c return new; new 1610 net/netlink/af_netlink.c int old, new = !!is_new, subscriptions; new 1613 net/netlink/af_netlink.c subscriptions = nlk->subscriptions - old + new; new 1614 net/netlink/af_netlink.c if (new) new 2112 net/netlink/af_netlink.c struct listeners *new, *old; new 2119 net/netlink/af_netlink.c new = kzalloc(sizeof(*new) + NLGRPSZ(groups), GFP_ATOMIC); new 2120 net/netlink/af_netlink.c if (!new) new 2123 net/netlink/af_netlink.c memcpy(new->masks, old->masks, NLGRPSZ(tbl->groups)); new 2124 net/netlink/af_netlink.c rcu_assign_pointer(tbl->listeners, new); new 301 net/openvswitch/flow_table.c struct table_instance *new, bool ufid) new 307 net/openvswitch/flow_table.c new->node_ver = !old_ver; new 317 net/openvswitch/flow_table.c ufid_table_instance_insert(new, flow); new 321 net/openvswitch/flow_table.c table_instance_insert(new, flow); new 623 net/openvswitch/flow_table.c const struct sw_flow_mask *new) new 626 net/openvswitch/flow_table.c mask = flow_mask_find(tbl, new); new 632 net/openvswitch/flow_table.c mask->key = new->key; new 633 net/openvswitch/flow_table.c mask->range = new->range; new 1523 net/packet/af_packet.c static void __fanout_set_data_bpf(struct packet_fanout *f, struct bpf_prog *new) new 1529 net/packet/af_packet.c rcu_assign_pointer(f->bpf_prog, new); new 1541 net/packet/af_packet.c struct bpf_prog *new; new 1552 net/packet/af_packet.c ret = bpf_prog_create_from_user(&new, &fprog, NULL, false); new 1556 net/packet/af_packet.c __fanout_set_data_bpf(po->fanout, new); new 1563 net/packet/af_packet.c struct bpf_prog *new; new 1573 net/packet/af_packet.c new = bpf_prog_get_type(fd, BPF_PROG_TYPE_SOCKET_FILTER); new 1574 net/packet/af_packet.c if (IS_ERR(new)) new 1575 net/packet/af_packet.c return PTR_ERR(new); new 1577 net/packet/af_packet.c __fanout_set_data_bpf(po->fanout, new); new 799 net/rds/rds.h rds_conn_path_transition(struct rds_conn_path *cp, int old, int new) new 801 net/rds/rds.h return atomic_cmpxchg(&cp->cp_state, old, new) == old; new 805 net/rds/rds.h rds_conn_transition(struct rds_connection *conn, int old, int new) new 808 net/rds/rds.h return rds_conn_path_transition(&conn->c_path[0], old, new); new 60 net/sched/act_police.c struct tcf_police_params *new; new 145 net/sched/act_police.c new = kzalloc(sizeof(*new), GFP_KERNEL); new 146 net/sched/act_police.c if (unlikely(!new)) { new 152 net/sched/act_police.c new->tcfp_result = tcfp_result; new 153 net/sched/act_police.c new->tcfp_mtu = parm->mtu; new 154 net/sched/act_police.c if (!new->tcfp_mtu) { new 155 net/sched/act_police.c new->tcfp_mtu = ~0; new 157 net/sched/act_police.c new->tcfp_mtu = 255 << R_tab->rate.cell_log; new 160 net/sched/act_police.c new->rate_present = true; new 163 net/sched/act_police.c psched_ratecfg_precompute(&new->rate, &R_tab->rate, rate64); new 166 net/sched/act_police.c new->rate_present = false; new 169 net/sched/act_police.c new->peak_present = true; new 172 net/sched/act_police.c psched_ratecfg_precompute(&new->peak, &P_tab->rate, prate64); new 175 net/sched/act_police.c new->peak_present = false; new 178 net/sched/act_police.c new->tcfp_burst = PSCHED_TICKS2NS(parm->burst); new 179 net/sched/act_police.c if (new->peak_present) new 180 net/sched/act_police.c new->tcfp_mtu_ptoks = (s64)psched_l2t_ns(&new->peak, new 181 net/sched/act_police.c new->tcfp_mtu); new 184 net/sched/act_police.c new->tcfp_ewma_rate = nla_get_u32(tb[TCA_POLICE_AVRATE]); new 189 net/sched/act_police.c police->tcfp_toks = new->tcfp_burst; new 190 net/sched/act_police.c if (new->peak_present) new 191 net/sched/act_police.c police->tcfp_ptoks = new->tcfp_mtu_ptoks; new 195 net/sched/act_police.c new, new 201 net/sched/act_police.c if (new) new 202 net/sched/act_police.c kfree_rcu(new, rcu); new 84 net/sched/cls_cgroup.c struct cls_cgroup_head *new; new 96 net/sched/cls_cgroup.c new = kzalloc(sizeof(*head), GFP_KERNEL); new 97 net/sched/cls_cgroup.c if (!new) new 100 net/sched/cls_cgroup.c err = tcf_exts_init(&new->exts, net, TCA_CGROUP_ACT, TCA_CGROUP_POLICE); new 103 net/sched/cls_cgroup.c new->handle = handle; new 104 net/sched/cls_cgroup.c new->tp = tp; new 111 net/sched/cls_cgroup.c err = tcf_exts_validate(net, tp, tb, tca[TCA_RATE], &new->exts, ovr, new 116 net/sched/cls_cgroup.c err = tcf_em_tree_validate(tp, tb[TCA_CGROUP_EMATCHES], &new->ematches); new 120 net/sched/cls_cgroup.c rcu_assign_pointer(tp->root, new); new 127 net/sched/cls_cgroup.c tcf_exts_destroy(&new->exts); new 128 net/sched/cls_cgroup.c kfree(new); new 191 net/sched/cls_matchall.c struct cls_mall_head *new; new 212 net/sched/cls_matchall.c new = kzalloc(sizeof(*new), GFP_KERNEL); new 213 net/sched/cls_matchall.c if (!new) new 216 net/sched/cls_matchall.c err = tcf_exts_init(&new->exts, net, TCA_MATCHALL_ACT, 0); new 222 net/sched/cls_matchall.c new->handle = handle; new 223 net/sched/cls_matchall.c new->flags = flags; new 224 net/sched/cls_matchall.c new->pf = alloc_percpu(struct tc_matchall_pcnt); new 225 net/sched/cls_matchall.c if (!new->pf) { new 230 net/sched/cls_matchall.c err = mall_set_parms(net, tp, new, base, tb, tca[TCA_RATE], ovr, new 235 net/sched/cls_matchall.c if (!tc_skip_hw(new->flags)) { new 236 net/sched/cls_matchall.c err = mall_replace_hw_filter(tp, new, (unsigned long)new, new 242 net/sched/cls_matchall.c if (!tc_in_hw(new->flags)) new 243 net/sched/cls_matchall.c new->flags |= TCA_CLS_FLAGS_NOT_IN_HW; new 246 net/sched/cls_matchall.c rcu_assign_pointer(tp->root, new); new 251 net/sched/cls_matchall.c free_percpu(new->pf); new 253 net/sched/cls_matchall.c tcf_exts_destroy(&new->exts); new 255 net/sched/cls_matchall.c kfree(new); new 384 net/sched/cls_route.c struct nlattr **tb, struct nlattr *est, int new, new 398 net/sched/cls_route.c if (new && handle & 0x8000) new 421 net/sched/cls_route.c if (handle && new) { new 478 net/sched/cls_route.c bool new = true; new 509 net/sched/cls_route.c new = false; new 513 net/sched/cls_route.c tca[TCA_RATE], new, ovr, extack); new 797 net/sched/cls_u32.c struct tc_u_knode *new; new 799 net/sched/cls_u32.c new = kzalloc(sizeof(*n) + s->nkeys*sizeof(struct tc_u32_key), new 802 net/sched/cls_u32.c if (!new) new 805 net/sched/cls_u32.c RCU_INIT_POINTER(new->next, n->next); new 806 net/sched/cls_u32.c new->handle = n->handle; new 807 net/sched/cls_u32.c RCU_INIT_POINTER(new->ht_up, n->ht_up); new 809 net/sched/cls_u32.c new->ifindex = n->ifindex; new 810 net/sched/cls_u32.c new->fshift = n->fshift; new 811 net/sched/cls_u32.c new->res = n->res; new 812 net/sched/cls_u32.c new->flags = n->flags; new 813 net/sched/cls_u32.c RCU_INIT_POINTER(new->ht_down, ht); new 824 net/sched/cls_u32.c new->pf = n->pf; new 828 net/sched/cls_u32.c new->val = n->val; new 829 net/sched/cls_u32.c new->mask = n->mask; new 831 net/sched/cls_u32.c new->pcpu_success = n->pcpu_success; new 833 net/sched/cls_u32.c memcpy(&new->sel, s, struct_size(s, keys, s->nkeys)); new 835 net/sched/cls_u32.c if (tcf_exts_init(&new->exts, net, TCA_U32_ACT, TCA_U32_POLICE)) { new 836 net/sched/cls_u32.c kfree(new); new 840 net/sched/cls_u32.c return new; new 885 net/sched/cls_u32.c struct tc_u_knode *new; new 898 net/sched/cls_u32.c new = u32_init_knode(net, tp, n); new 899 net/sched/cls_u32.c if (!new) new 902 net/sched/cls_u32.c err = u32_set_parms(net, tp, base, new, tb, new 906 net/sched/cls_u32.c u32_destroy_key(new, false); new 910 net/sched/cls_u32.c err = u32_replace_hw_knode(tp, new, flags, extack); new 912 net/sched/cls_u32.c u32_destroy_key(new, false); new 916 net/sched/cls_u32.c if (!tc_in_hw(new->flags)) new 917 net/sched/cls_u32.c new->flags |= TCA_CLS_FLAGS_NOT_IN_HW; new 919 net/sched/cls_u32.c u32_replace_knode(tp, tp_c, new); new 824 net/sched/sch_api.c struct Qdisc *new, struct Qdisc *old, new 837 net/sched/sch_api.c if (!err || !new || new == &noop_qdisc) new 843 net/sched/sch_api.c any_qdisc_is_offloaded = new->flags & TCQ_F_OFFLOADED; new 853 net/sched/sch_api.c struct Qdisc *new, struct Qdisc *old, new 858 net/sched/sch_api.c .handle = new ? new->handle : 0, new 859 net/sched/sch_api.c .ingress = (new && new->flags & TCQ_F_INGRESS) || new 863 net/sched/sch_api.c qdisc_offload_graft_helper(dev, NULL, new, old, new 958 net/sched/sch_api.c struct Qdisc *old, struct Qdisc *new) new 972 net/sched/sch_api.c if (new && !tc_qdisc_dump_ignore(new, false)) { new 973 net/sched/sch_api.c if (tc_fill_qdisc(skb, new, clid, portid, n->nlmsg_seq, new 989 net/sched/sch_api.c struct Qdisc *old, struct Qdisc *new) new 991 net/sched/sch_api.c if (new || old) new 992 net/sched/sch_api.c qdisc_notify(net, skb, n, clid, old, new); new 1022 net/sched/sch_api.c struct Qdisc *new, struct Qdisc *old, new 1034 net/sched/sch_api.c (new && new->flags & TCQ_F_INGRESS)) { new 1046 net/sched/sch_api.c qdisc_offload_graft_root(dev, new, old, extack); new 1048 net/sched/sch_api.c if (new && new->ops->attach) new 1057 net/sched/sch_api.c old = dev_graft_qdisc(dev_queue, new); new 1058 net/sched/sch_api.c if (new && i > 0) new 1059 net/sched/sch_api.c qdisc_refcount_inc(new); new 1068 net/sched/sch_api.c dev->qdisc, new); new 1069 net/sched/sch_api.c if (new && !new->ops->attach) new 1070 net/sched/sch_api.c qdisc_refcount_inc(new); new 1071 net/sched/sch_api.c dev->qdisc = new ? : &noop_qdisc; new 1073 net/sched/sch_api.c if (new && new->ops->attach) new 1074 net/sched/sch_api.c new->ops->attach(new); new 1076 net/sched/sch_api.c notify_and_destroy(net, skb, n, classid, old, new); new 1087 net/sched/sch_api.c if (new && (new->flags & TCQ_F_NOLOCK) && new 1089 net/sched/sch_api.c qdisc_clear_nolock(new); new 1100 net/sched/sch_api.c err = cops->graft(parent, cl, new, &old, extack); new 1103 net/sched/sch_api.c notify_and_destroy(net, skb, n, classid, old, new); new 86 net/sched/sch_atm.c struct Qdisc *new, struct Qdisc **old, new 93 net/sched/sch_atm.c sch, p, flow, new, old); new 96 net/sched/sch_atm.c if (!new) new 97 net/sched/sch_atm.c new = &noop_qdisc; new 99 net/sched/sch_atm.c flow->q = new; new 498 net/sched/sch_atm.c struct sk_buff *new; new 500 net/sched/sch_atm.c new = skb_realloc_headroom(skb, flow->hdr_len); new 502 net/sched/sch_atm.c if (!new) new 504 net/sched/sch_atm.c skb = new; new 184 net/sched/sch_cbq.c struct cbq_class *new = cl->defaults[TC_PRIO_BESTEFFORT]; new 186 net/sched/sch_cbq.c if (new != NULL && new != this) new 187 net/sched/sch_cbq.c return new; new 1396 net/sched/sch_cbq.c static int cbq_graft(struct Qdisc *sch, unsigned long arg, struct Qdisc *new, new 1401 net/sched/sch_cbq.c if (new == NULL) { new 1402 net/sched/sch_cbq.c new = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops, new 1404 net/sched/sch_cbq.c if (new == NULL) new 1408 net/sched/sch_cbq.c *old = qdisc_replace(sch, new, &cl->q); new 491 net/sched/sch_cbs.c static int cbs_graft(struct Qdisc *sch, unsigned long arg, struct Qdisc *new, new 496 net/sched/sch_cbs.c if (!new) { new 497 net/sched/sch_cbs.c new = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops, new 499 net/sched/sch_cbs.c if (!new) new 500 net/sched/sch_cbs.c new = &noop_qdisc; new 503 net/sched/sch_cbs.c *old = qdisc_replace(sch, new, &q->qdisc); new 205 net/sched/sch_drr.c struct Qdisc *new, struct Qdisc **old, new 210 net/sched/sch_drr.c if (new == NULL) { new 211 net/sched/sch_drr.c new = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops, new 213 net/sched/sch_drr.c if (new == NULL) new 214 net/sched/sch_drr.c new = &noop_qdisc; new 217 net/sched/sch_drr.c *old = qdisc_replace(sch, new, &cl->qdisc); new 65 net/sched/sch_dsmark.c struct Qdisc *new, struct Qdisc **old, new 71 net/sched/sch_dsmark.c __func__, sch, p, new, old); new 73 net/sched/sch_dsmark.c if (new == NULL) { new 74 net/sched/sch_dsmark.c new = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops, new 76 net/sched/sch_dsmark.c if (new == NULL) new 77 net/sched/sch_dsmark.c new = &noop_qdisc; new 80 net/sched/sch_dsmark.c *old = qdisc_replace(sch, new, &p->q); new 1170 net/sched/sch_hfsc.c hfsc_graft_class(struct Qdisc *sch, unsigned long arg, struct Qdisc *new, new 1177 net/sched/sch_hfsc.c if (new == NULL) { new 1178 net/sched/sch_hfsc.c new = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops, new 1180 net/sched/sch_hfsc.c if (new == NULL) new 1181 net/sched/sch_hfsc.c new = &noop_qdisc; new 1184 net/sched/sch_hfsc.c *old = qdisc_replace(sch, new, &cl->qdisc); new 1147 net/sched/sch_htb.c static int htb_graft(struct Qdisc *sch, unsigned long arg, struct Qdisc *new, new 1154 net/sched/sch_htb.c if (new == NULL && new 1155 net/sched/sch_htb.c (new = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops, new 1159 net/sched/sch_htb.c *old = qdisc_replace(sch, new, &cl->leaf.q); new 190 net/sched/sch_mq.c static int mq_graft(struct Qdisc *sch, unsigned long cl, struct Qdisc *new, new 200 net/sched/sch_mq.c *old = dev_graft_qdisc(dev_queue, new); new 201 net/sched/sch_mq.c if (new) new 202 net/sched/sch_mq.c new->flags |= TCQ_F_ONETXQUEUE | TCQ_F_NOPARENT; new 208 net/sched/sch_mq.c graft_offload.graft_params.child_handle = new ? new->handle : 0; new 211 net/sched/sch_mq.c qdisc_offload_graft_helper(qdisc_dev(sch), sch, new, *old, new 320 net/sched/sch_mqprio.c static int mqprio_graft(struct Qdisc *sch, unsigned long cl, struct Qdisc *new, new 332 net/sched/sch_mqprio.c *old = dev_graft_qdisc(dev_queue, new); new 334 net/sched/sch_mqprio.c if (new) new 335 net/sched/sch_mqprio.c new->flags |= TCQ_F_ONETXQUEUE | TCQ_F_NOPARENT; new 281 net/sched/sch_multiq.c static int multiq_graft(struct Qdisc *sch, unsigned long arg, struct Qdisc *new, new 287 net/sched/sch_multiq.c if (new == NULL) new 288 net/sched/sch_multiq.c new = &noop_qdisc; new 290 net/sched/sch_multiq.c *old = qdisc_replace(sch, new, &q->queues[band]); new 1227 net/sched/sch_netem.c static int netem_graft(struct Qdisc *sch, unsigned long arg, struct Qdisc *new, new 1232 net/sched/sch_netem.c *old = qdisc_replace(sch, new, &q->qdisc); new 288 net/sched/sch_prio.c static int prio_graft(struct Qdisc *sch, unsigned long arg, struct Qdisc *new, new 295 net/sched/sch_prio.c if (!new) { new 296 net/sched/sch_prio.c new = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops, new 298 net/sched/sch_prio.c if (!new) new 299 net/sched/sch_prio.c new = &noop_qdisc; new 301 net/sched/sch_prio.c qdisc_hash_add(new, true); new 304 net/sched/sch_prio.c *old = qdisc_replace(sch, new, &q->queues[band]); new 309 net/sched/sch_prio.c graft_offload.graft_params.child_handle = new->handle; new 312 net/sched/sch_prio.c qdisc_offload_graft_helper(qdisc_dev(sch), sch, new, *old, new 586 net/sched/sch_qfq.c struct Qdisc *new, struct Qdisc **old, new 591 net/sched/sch_qfq.c if (new == NULL) { new 592 net/sched/sch_qfq.c new = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops, new 594 net/sched/sch_qfq.c if (new == NULL) new 595 net/sched/sch_qfq.c new = &noop_qdisc; new 598 net/sched/sch_qfq.c *old = qdisc_replace(sch, new, &cl->qdisc); new 368 net/sched/sch_red.c struct Qdisc *new, struct Qdisc *old, new 374 net/sched/sch_red.c .child_handle = new->handle, new 378 net/sched/sch_red.c qdisc_offload_graft_helper(qdisc_dev(sch), sch, new, old, new 382 net/sched/sch_red.c static int red_graft(struct Qdisc *sch, unsigned long arg, struct Qdisc *new, new 387 net/sched/sch_red.c if (new == NULL) new 388 net/sched/sch_red.c new = &noop_qdisc; new 390 net/sched/sch_red.c *old = qdisc_replace(sch, new, &q->qdisc); new 392 net/sched/sch_red.c red_graft_offload(sch, new, *old, extack); new 617 net/sched/sch_sfb.c static int sfb_graft(struct Qdisc *sch, unsigned long arg, struct Qdisc *new, new 622 net/sched/sch_sfb.c if (new == NULL) new 623 net/sched/sch_sfb.c new = &noop_qdisc; new 625 net/sched/sch_sfb.c *old = qdisc_replace(sch, new, &q->qdisc); new 866 net/sched/sch_taprio.c struct sched_gate_list *new, new 877 net/sched/sch_taprio.c new->base_time = nla_get_s64(tb[TCA_TAPRIO_ATTR_SCHED_BASE_TIME]); new 880 net/sched/sch_taprio.c new->cycle_time_extension = nla_get_s64(tb[TCA_TAPRIO_ATTR_SCHED_CYCLE_TIME_EXTENSION]); new 883 net/sched/sch_taprio.c new->cycle_time = nla_get_s64(tb[TCA_TAPRIO_ATTR_SCHED_CYCLE_TIME]); new 887 net/sched/sch_taprio.c tb[TCA_TAPRIO_ATTR_SCHED_ENTRY_LIST], new, extack); new 891 net/sched/sch_taprio.c if (!new->cycle_time) { new 895 net/sched/sch_taprio.c list_for_each_entry(entry, &new->entries, list) new 897 net/sched/sch_taprio.c new->cycle_time = cycle; new 1023 net/sched/sch_taprio.c ktime_t start, struct sched_gate_list *new) new 1388 net/sched/sch_taprio.c u32 new = 0; new 1391 net/sched/sch_taprio.c new = nla_get_u32(attr); new 1393 net/sched/sch_taprio.c if (old != TAPRIO_FLAGS_INVALID && old != new) { new 1398 net/sched/sch_taprio.c if (!taprio_flags_valid(new)) { new 1403 net/sched/sch_taprio.c return new; new 1686 net/sched/sch_taprio.c struct Qdisc *new, struct Qdisc **old, new 1700 net/sched/sch_taprio.c q->qdiscs[cl - 1] = new; new 1702 net/sched/sch_taprio.c if (new) new 1703 net/sched/sch_taprio.c new->flags |= TCQ_F_ONETXQUEUE | TCQ_F_NOPARENT; new 489 net/sched/sch_tbf.c static int tbf_graft(struct Qdisc *sch, unsigned long arg, struct Qdisc *new, new 494 net/sched/sch_tbf.c if (new == NULL) new 495 net/sched/sch_tbf.c new = &noop_qdisc; new 497 net/sched/sch_tbf.c *old = qdisc_replace(sch, new, &q->qdisc); new 1101 net/sctp/associola.c struct sctp_association *new) new 1107 net/sctp/associola.c asoc->c = new->c; new 1108 net/sctp/associola.c asoc->peer.rwnd = new->peer.rwnd; new 1109 net/sctp/associola.c asoc->peer.sack_needed = new->peer.sack_needed; new 1110 net/sctp/associola.c asoc->peer.auth_capable = new->peer.auth_capable; new 1111 net/sctp/associola.c asoc->peer.i = new->peer.i; new 1120 net/sctp/associola.c if (!sctp_assoc_lookup_paddr(new, &trans->ipaddr)) { new 1135 net/sctp/associola.c asoc->next_tsn = new->next_tsn; new 1136 net/sctp/associola.c asoc->ctsn_ack_point = new->ctsn_ack_point; new 1137 net/sctp/associola.c asoc->adv_peer_ack_point = new->adv_peer_ack_point; new 1158 net/sctp/associola.c list_for_each_entry(trans, &new->peer.transport_addr_list, new 1169 net/sctp/associola.c sctp_stream_update(&asoc->stream, &new->stream); new 1180 net/sctp/associola.c asoc->peer.peer_random = new->peer.peer_random; new 1181 net/sctp/associola.c new->peer.peer_random = NULL; new 1184 net/sctp/associola.c asoc->peer.peer_chunks = new->peer.peer_chunks; new 1185 net/sctp/associola.c new->peer.peer_chunks = NULL; new 1188 net/sctp/associola.c asoc->peer.peer_hmacs = new->peer.peer_hmacs; new 1189 net/sctp/associola.c new->peer.peer_hmacs = NULL; new 81 net/sctp/auth.c struct sctp_shared_key *new; new 84 net/sctp/auth.c new = kzalloc(sizeof(struct sctp_shared_key), gfp); new 85 net/sctp/auth.c if (!new) new 88 net/sctp/auth.c INIT_LIST_HEAD(&new->key_list); new 89 net/sctp/auth.c refcount_set(&new->refcnt, 1); new 90 net/sctp/auth.c new->key_id = key_id; new 92 net/sctp/auth.c return new; new 190 net/sctp/auth.c struct sctp_auth_bytes *new; new 202 net/sctp/auth.c new = sctp_auth_create_key(len, gfp); new 203 net/sctp/auth.c if (!new) new 206 net/sctp/auth.c memcpy(new->data, random, random_len); new 210 net/sctp/auth.c memcpy(new->data + offset, chunks, chunks_len); new 214 net/sctp/auth.c memcpy(new->data + offset, hmacs, hmacs_len); new 216 net/sctp/auth.c return new; new 358 net/sctp/auth.c struct sctp_shared_key *new; new 363 net/sctp/auth.c new = sctp_auth_shkey_create(sh_key->key_id, gfp); new 364 net/sctp/auth.c if (!new) new 367 net/sctp/auth.c new->key = sh_key->key; new 368 net/sctp/auth.c sctp_auth_key_hold(new->key); new 369 net/sctp/auth.c list_add(&new->key_list, &asoc->endpoint_shared_keys); new 138 net/sctp/bind_addr.c int sctp_add_bind_addr(struct sctp_bind_addr *bp, union sctp_addr *new, new 148 net/sctp/bind_addr.c memcpy(&addr->a, new, min_t(size_t, sizeof(*new), new_size)); new 318 net/sctp/outqueue.c static void sctp_insert_list(struct list_head *head, struct list_head *new) new 325 net/sctp/outqueue.c nchunk = list_entry(new, struct sctp_chunk, transmitted_list); new 332 net/sctp/outqueue.c list_add(new, pos->prev); new 338 net/sctp/outqueue.c list_add_tail(new, head); new 830 net/sctp/sm_sideeffect.c struct sctp_association *new) new 835 net/sctp/sm_sideeffect.c if (!sctp_assoc_update(asoc, new)) new 1952 net/sctp/socket.c bool new = false; new 2011 net/sctp/socket.c new = true; new 2033 net/sctp/socket.c if (err < 0 && err != -ESRCH && new) new 30 net/sctp/stream.c struct sctp_stream *new, __u16 outcnt) new 60 net/sctp/stream.c if (new) { new 66 net/sctp/stream.c kfree(SCTP_SO(new, i)->ext); new 67 net/sctp/stream.c SCTP_SO(new, i)->ext = SCTP_SO(stream, i)->ext; new 201 net/sctp/stream.c void sctp_stream_update(struct sctp_stream *stream, struct sctp_stream *new) new 206 net/sctp/stream.c sctp_stream_outq_migrate(stream, new, new->outcnt); new 209 net/sctp/stream.c stream->out = new->out; new 210 net/sctp/stream.c stream->in = new->in; new 211 net/sctp/stream.c stream->outcnt = new->outcnt; new 212 net/sctp/stream.c stream->incnt = new->incnt; new 216 net/sctp/stream.c new->out.tree.root = NULL; new 217 net/sctp/stream.c new->in.tree.root = NULL; new 218 net/sctp/stream.c new->outcnt = 0; new 219 net/sctp/stream.c new->incnt = 0; new 343 net/sctp/tsnmap.c unsigned long *new; new 353 net/sctp/tsnmap.c new = kzalloc(len>>3, GFP_ATOMIC); new 354 net/sctp/tsnmap.c if (!new) new 357 net/sctp/tsnmap.c bitmap_copy(new, map->tsn_map, new 360 net/sctp/tsnmap.c map->tsn_map = new; new 324 net/sctp/ulpqueue.c struct sk_buff *new = NULL; new 351 net/sctp/ulpqueue.c new = skb_copy(f_frag, GFP_ATOMIC); new 352 net/sctp/ulpqueue.c if (!new) new 355 net/sctp/ulpqueue.c sctp_skb_set_owner_r(new, f_frag->sk); new 357 net/sctp/ulpqueue.c skb_shinfo(new)->frag_list = pos; new 366 net/sctp/ulpqueue.c if (new) { new 368 net/sctp/ulpqueue.c f_frag = new; new 168 net/smc/smc_cdc.h union smc_host_cursor *new) new 170 net/smc/smc_cdc.h if (old->wrap != new->wrap) new 172 net/smc/smc_cdc.h ((size - old->count) + new->count)); new 174 net/smc/smc_cdc.h return max_t(int, 0, (new->count - old->count)); new 182 net/smc/smc_cdc.h union smc_host_cursor *new) new 184 net/smc/smc_cdc.h if (old->wrap > new->wrap || new 185 net/smc/smc_cdc.h (old->wrap == new->wrap && old->count > new->count)) new 186 net/smc/smc_cdc.h return -smc_curs_diff(size, new, old); new 187 net/smc/smc_cdc.h return smc_curs_diff(size, old, new); new 195 net/smc/smc_cdc.h union smc_host_cursor *new) new 197 net/smc/smc_cdc.h if (old->wrap < new->wrap) new 199 net/smc/smc_cdc.h (size - old->count) + new->count + new 200 net/smc/smc_cdc.h (new->wrap - old->wrap - 1) * size, new 203 net/smc/smc_cdc.h if (old->wrap > new->wrap) /* wrap has switched from 0xffff to 0x0000 */ new 205 net/smc/smc_cdc.h (size - old->count) + new->count + new 206 net/smc/smc_cdc.h (new->wrap + 0xffff - old->wrap) * size, new 209 net/smc/smc_cdc.h return max_t(int, 0, (new->count - old->count)); new 342 net/sunrpc/auth.c struct rpc_cred_cache *new; new 345 net/sunrpc/auth.c new = kmalloc(sizeof(*new), GFP_KERNEL); new 346 net/sunrpc/auth.c if (!new) new 348 net/sunrpc/auth.c new->hashbits = auth_hashbits; new 349 net/sunrpc/auth.c hashsize = 1U << new->hashbits; new 350 net/sunrpc/auth.c new->hashtable = kcalloc(hashsize, sizeof(new->hashtable[0]), GFP_KERNEL); new 351 net/sunrpc/auth.c if (!new->hashtable) new 353 net/sunrpc/auth.c spin_lock_init(&new->lock); new 354 net/sunrpc/auth.c auth->au_credcache = new; new 357 net/sunrpc/auth.c kfree(new); new 572 net/sunrpc/auth.c *entry, *new; new 590 net/sunrpc/auth.c new = auth->au_ops->crcreate(auth, acred, flags, gfp); new 591 net/sunrpc/auth.c if (IS_ERR(new)) { new 592 net/sunrpc/auth.c cred = new; new 605 net/sunrpc/auth.c cred = new; new 610 net/sunrpc/auth.c list_add_tail(&new->cr_lru, &free); new 698 net/sunrpc/auth.c struct rpc_cred *new = NULL; new 709 net/sunrpc/auth.c new = get_rpccred(task->tk_op_cred); new 711 net/sunrpc/auth.c new = auth->au_ops->lookup_cred(auth, &acred, lookupflags); new 713 net/sunrpc/auth.c new = rpcauth_bind_machine_cred(task, lookupflags); new 716 net/sunrpc/auth.c if (new) new 719 net/sunrpc/auth.c new = rpcauth_bind_root_cred(task, lookupflags); new 721 net/sunrpc/auth.c new = authnull_ops.lookup_cred(NULL, NULL, 0); new 723 net/sunrpc/auth.c new = rpcauth_bind_new_cred(task, lookupflags); new 724 net/sunrpc/auth.c if (IS_ERR(new)) new 725 net/sunrpc/auth.c return PTR_ERR(new); new 727 net/sunrpc/auth.c req->rq_cred = new; new 1167 net/sunrpc/auth_gss/auth_gss.c struct gss_auth *new) new 1193 net/sunrpc/auth_gss/auth_gss.c if (new) new 1194 net/sunrpc/auth_gss/auth_gss.c hash_add(gss_auth_hash_table, &new->hash, hashval); new 1195 net/sunrpc/auth_gss/auth_gss.c gss_auth = new; new 1206 net/sunrpc/auth_gss/auth_gss.c struct gss_auth *new; new 1211 net/sunrpc/auth_gss/auth_gss.c new = gss_create_new(args, clnt); new 1212 net/sunrpc/auth_gss/auth_gss.c if (IS_ERR(new)) new 1213 net/sunrpc/auth_gss/auth_gss.c return new; new 1214 net/sunrpc/auth_gss/auth_gss.c gss_auth = gss_auth_find_or_add_hashed(args, clnt, new); new 1215 net/sunrpc/auth_gss/auth_gss.c if (gss_auth != new) new 1216 net/sunrpc/auth_gss/auth_gss.c gss_destroy(&new->rpc_auth); new 1244 net/sunrpc/auth_gss/auth_gss.c struct gss_cred *new; new 1247 net/sunrpc/auth_gss/auth_gss.c new = kzalloc(sizeof(*gss_cred), GFP_NOFS); new 1248 net/sunrpc/auth_gss/auth_gss.c if (new) { new 1255 net/sunrpc/auth_gss/auth_gss.c rpcauth_init_cred(&new->gc_base, &acred, new 1258 net/sunrpc/auth_gss/auth_gss.c new->gc_base.cr_flags = 1UL << RPCAUTH_CRED_UPTODATE; new 1259 net/sunrpc/auth_gss/auth_gss.c new->gc_service = gss_cred->gc_service; new 1260 net/sunrpc/auth_gss/auth_gss.c new->gc_principal = gss_cred->gc_principal; new 1262 net/sunrpc/auth_gss/auth_gss.c rcu_assign_pointer(new->gc_ctx, ctx); new 1265 net/sunrpc/auth_gss/auth_gss.c return new; new 1280 net/sunrpc/auth_gss/auth_gss.c struct gss_cred *new; new 1283 net/sunrpc/auth_gss/auth_gss.c new = gss_dup_cred(gss_auth, gss_cred); new 1284 net/sunrpc/auth_gss/auth_gss.c if (new) { new 1287 net/sunrpc/auth_gss/auth_gss.c task = rpc_call_null(gss_auth->client, &new->gc_base, new 1292 net/sunrpc/auth_gss/auth_gss.c put_rpccred(&new->gc_base); new 1611 net/sunrpc/auth_gss/auth_gss.c struct rpc_cred *new; new 1613 net/sunrpc/auth_gss/auth_gss.c new = gss_lookup_cred(auth, &acred, RPCAUTH_LOOKUP_NEW); new 1614 net/sunrpc/auth_gss/auth_gss.c if (IS_ERR(new)) new 1615 net/sunrpc/auth_gss/auth_gss.c return PTR_ERR(new); new 1616 net/sunrpc/auth_gss/auth_gss.c task->tk_rqstp->rq_cred = new; new 2075 net/sunrpc/auth_gss/auth_gss.c gss_seq_is_newer(u32 new, u32 old) new 2077 net/sunrpc/auth_gss/auth_gss.c return (s32)(new - old) > 0; new 48 net/sunrpc/auth_gss/gss_mech_switch.c char *new; new 50 net/sunrpc/auth_gss/gss_mech_switch.c new = kmalloc(strlen(name) + strlen(prefix) + 1, GFP_KERNEL); new 51 net/sunrpc/auth_gss/gss_mech_switch.c if (new) { new 52 net/sunrpc/auth_gss/gss_mech_switch.c strcpy(new, prefix); new 53 net/sunrpc/auth_gss/gss_mech_switch.c strcat(new, name); new 55 net/sunrpc/auth_gss/gss_mech_switch.c return new; new 83 net/sunrpc/auth_gss/svcauth_gss.c static struct rsi *rsi_update(struct cache_detail *cd, struct rsi *new, struct rsi *old); new 139 net/sunrpc/auth_gss/svcauth_gss.c struct rsi *new = container_of(cnew, struct rsi, h); new 142 net/sunrpc/auth_gss/svcauth_gss.c new->out_handle.data = NULL; new 143 net/sunrpc/auth_gss/svcauth_gss.c new->out_handle.len = 0; new 144 net/sunrpc/auth_gss/svcauth_gss.c new->out_token.data = NULL; new 145 net/sunrpc/auth_gss/svcauth_gss.c new->out_token.len = 0; new 146 net/sunrpc/auth_gss/svcauth_gss.c new->in_handle.len = item->in_handle.len; new 148 net/sunrpc/auth_gss/svcauth_gss.c new->in_token.len = item->in_token.len; new 150 net/sunrpc/auth_gss/svcauth_gss.c new->in_handle.data = item->in_handle.data; new 152 net/sunrpc/auth_gss/svcauth_gss.c new->in_token.data = item->in_token.data; new 158 net/sunrpc/auth_gss/svcauth_gss.c struct rsi *new = container_of(cnew, struct rsi, h); new 161 net/sunrpc/auth_gss/svcauth_gss.c BUG_ON(new->out_handle.data || new->out_token.data); new 162 net/sunrpc/auth_gss/svcauth_gss.c new->out_handle.len = item->out_handle.len; new 164 net/sunrpc/auth_gss/svcauth_gss.c new->out_token.len = item->out_token.len; new 166 net/sunrpc/auth_gss/svcauth_gss.c new->out_handle.data = item->out_handle.data; new 168 net/sunrpc/auth_gss/svcauth_gss.c new->out_token.data = item->out_token.data; new 171 net/sunrpc/auth_gss/svcauth_gss.c new->major_status = item->major_status; new 172 net/sunrpc/auth_gss/svcauth_gss.c new->minor_status = item->minor_status; new 302 net/sunrpc/auth_gss/svcauth_gss.c static struct rsi *rsi_update(struct cache_detail *cd, struct rsi *new, struct rsi *old) new 305 net/sunrpc/auth_gss/svcauth_gss.c int hash = rsi_hash(new); new 307 net/sunrpc/auth_gss/svcauth_gss.c ch = sunrpc_cache_update(cd, &new->h, new 346 net/sunrpc/auth_gss/svcauth_gss.c static struct rsc *rsc_update(struct cache_detail *cd, struct rsc *new, struct rsc *old); new 384 net/sunrpc/auth_gss/svcauth_gss.c struct rsc *new = container_of(a, struct rsc, h); new 387 net/sunrpc/auth_gss/svcauth_gss.c return netobj_equal(&new->handle, &tmp->handle); new 393 net/sunrpc/auth_gss/svcauth_gss.c struct rsc *new = container_of(cnew, struct rsc, h); new 396 net/sunrpc/auth_gss/svcauth_gss.c new->handle.len = tmp->handle.len; new 398 net/sunrpc/auth_gss/svcauth_gss.c new->handle.data = tmp->handle.data; new 400 net/sunrpc/auth_gss/svcauth_gss.c new->mechctx = NULL; new 401 net/sunrpc/auth_gss/svcauth_gss.c init_svc_cred(&new->cred); new 407 net/sunrpc/auth_gss/svcauth_gss.c struct rsc *new = container_of(cnew, struct rsc, h); new 410 net/sunrpc/auth_gss/svcauth_gss.c new->mechctx = tmp->mechctx; new 412 net/sunrpc/auth_gss/svcauth_gss.c memset(&new->seqdata, 0, sizeof(new->seqdata)); new 413 net/sunrpc/auth_gss/svcauth_gss.c spin_lock_init(&new->seqdata.sd_lock); new 414 net/sunrpc/auth_gss/svcauth_gss.c new->cred = tmp->cred; new 573 net/sunrpc/auth_gss/svcauth_gss.c static struct rsc *rsc_update(struct cache_detail *cd, struct rsc *new, struct rsc *old) new 576 net/sunrpc/auth_gss/svcauth_gss.c int hash = rsc_hash(new); new 578 net/sunrpc/auth_gss/svcauth_gss.c ch = sunrpc_cache_update(cd, &new->h, new 806 net/sunrpc/auth_gss/svcauth_gss.c struct gss_domain *new; new 810 net/sunrpc/auth_gss/svcauth_gss.c new = kmalloc(sizeof(*new), GFP_KERNEL); new 811 net/sunrpc/auth_gss/svcauth_gss.c if (!new) new 813 net/sunrpc/auth_gss/svcauth_gss.c kref_init(&new->h.ref); new 814 net/sunrpc/auth_gss/svcauth_gss.c new->h.name = kstrdup(name, GFP_KERNEL); new 815 net/sunrpc/auth_gss/svcauth_gss.c if (!new->h.name) new 817 net/sunrpc/auth_gss/svcauth_gss.c new->h.flavour = &svcauthops_gss; new 818 net/sunrpc/auth_gss/svcauth_gss.c new->pseudoflavor = pseudoflavor; new 821 net/sunrpc/auth_gss/svcauth_gss.c test = auth_domain_lookup(name, &new->h); new 822 net/sunrpc/auth_gss/svcauth_gss.c if (test != &new->h) { /* Duplicate registration */ new 824 net/sunrpc/auth_gss/svcauth_gss.c kfree(new->h.name); new 830 net/sunrpc/auth_gss/svcauth_gss.c kfree(new); new 241 net/sunrpc/backchannel_rqst.c struct rpc_rqst *new) new 247 net/sunrpc/backchannel_rqst.c if (!new) new 251 net/sunrpc/backchannel_rqst.c list_add_tail(&new->rq_bc_pa_list, &xprt->bc_pa_list); new 327 net/sunrpc/backchannel_rqst.c struct rpc_rqst *req, *new = NULL; new 337 net/sunrpc/backchannel_rqst.c req = xprt_get_bc_request(xprt, xid, new); new 340 net/sunrpc/backchannel_rqst.c if (new) { new 341 net/sunrpc/backchannel_rqst.c if (req != new) new 342 net/sunrpc/backchannel_rqst.c xprt_free_allocation(new); new 346 net/sunrpc/backchannel_rqst.c new = xprt_alloc_bc_req(xprt, GFP_KERNEL); new 347 net/sunrpc/backchannel_rqst.c } while (new); new 84 net/sunrpc/cache.c struct cache_head *new, *tmp, *freeme = NULL; new 87 net/sunrpc/cache.c new = detail->alloc(); new 88 net/sunrpc/cache.c if (!new) new 94 net/sunrpc/cache.c cache_init(new, detail); new 95 net/sunrpc/cache.c detail->init(new, key); new 110 net/sunrpc/cache.c cache_put(new, detail); new 115 net/sunrpc/cache.c hlist_add_head_rcu(&new->cache_list, head); new 117 net/sunrpc/cache.c cache_get(new); new 124 net/sunrpc/cache.c return new; new 165 net/sunrpc/cache.c struct cache_head *new, struct cache_head *old, int hash) new 176 net/sunrpc/cache.c if (test_bit(CACHE_NEGATIVE, &new->flags)) new 179 net/sunrpc/cache.c detail->update(old, new); new 180 net/sunrpc/cache.c cache_fresh_locked(old, new->expiry_time, detail); new 197 net/sunrpc/cache.c if (test_bit(CACHE_NEGATIVE, &new->flags)) new 200 net/sunrpc/cache.c detail->update(tmp, new); new 204 net/sunrpc/cache.c cache_fresh_locked(tmp, new->expiry_time, detail); new 617 net/sunrpc/clnt.c struct rpc_clnt *new; new 633 net/sunrpc/clnt.c new = rpc_new_client(args, xps, xprt, clnt); new 634 net/sunrpc/clnt.c if (IS_ERR(new)) { new 635 net/sunrpc/clnt.c err = PTR_ERR(new); new 640 net/sunrpc/clnt.c new->cl_autobind = 0; new 641 net/sunrpc/clnt.c new->cl_softrtry = clnt->cl_softrtry; new 642 net/sunrpc/clnt.c new->cl_softerr = clnt->cl_softerr; new 643 net/sunrpc/clnt.c new->cl_noretranstimeo = clnt->cl_noretranstimeo; new 644 net/sunrpc/clnt.c new->cl_discrtry = clnt->cl_discrtry; new 645 net/sunrpc/clnt.c new->cl_chatty = clnt->cl_chatty; new 646 net/sunrpc/clnt.c new->cl_principal = clnt->cl_principal; new 647 net/sunrpc/clnt.c return new; new 246 net/sunrpc/svc_xprt.c void svc_add_new_perm_xprt(struct svc_serv *serv, struct svc_xprt *new) new 248 net/sunrpc/svc_xprt.c clear_bit(XPT_TEMP, &new->xpt_flags); new 250 net/sunrpc/svc_xprt.c list_add(&new->xpt_list, &serv->sv_permsocks); new 252 net/sunrpc/svc_xprt.c svc_xprt_received(new); new 163 net/sunrpc/svcauth.c auth_domain_lookup(char *name, struct auth_domain *new) new 179 net/sunrpc/svcauth.c if (new) new 180 net/sunrpc/svcauth.c hlist_add_head_rcu(&new->hash, head); new 182 net/sunrpc/svcauth.c return new; new 58 net/sunrpc/svcauth_unix.c struct unix_domain *new = NULL; new 63 net/sunrpc/svcauth_unix.c if (new && rv != &new->h) new 64 net/sunrpc/svcauth_unix.c svcauth_unix_domain_release(&new->h); new 73 net/sunrpc/svcauth_unix.c new = kmalloc(sizeof(*new), GFP_KERNEL); new 74 net/sunrpc/svcauth_unix.c if (new == NULL) new 76 net/sunrpc/svcauth_unix.c kref_init(&new->h.ref); new 77 net/sunrpc/svcauth_unix.c new->h.name = kstrdup(name, GFP_KERNEL); new 78 net/sunrpc/svcauth_unix.c if (new->h.name == NULL) { new 79 net/sunrpc/svcauth_unix.c kfree(new); new 82 net/sunrpc/svcauth_unix.c new->h.flavour = &svcauth_unix; new 83 net/sunrpc/svcauth_unix.c rv = auth_domain_lookup(name, &new->h); new 122 net/sunrpc/svcauth_unix.c struct ip_map *new = container_of(cnew, struct ip_map, h); new 123 net/sunrpc/svcauth_unix.c return strcmp(orig->m_class, new->m_class) == 0 && new 124 net/sunrpc/svcauth_unix.c ipv6_addr_equal(&orig->m_addr, &new->m_addr); new 128 net/sunrpc/svcauth_unix.c struct ip_map *new = container_of(cnew, struct ip_map, h); new 131 net/sunrpc/svcauth_unix.c strcpy(new->m_class, item->m_class); new 132 net/sunrpc/svcauth_unix.c new->m_addr = item->m_addr; new 136 net/sunrpc/svcauth_unix.c struct ip_map *new = container_of(cnew, struct ip_map, h); new 140 net/sunrpc/svcauth_unix.c new->m_client = item->m_client; new 444 net/sunrpc/svcauth_unix.c struct unix_gid *new = container_of(cnew, struct unix_gid, h); new 445 net/sunrpc/svcauth_unix.c return uid_eq(orig->uid, new->uid); new 449 net/sunrpc/svcauth_unix.c struct unix_gid *new = container_of(cnew, struct unix_gid, h); new 451 net/sunrpc/svcauth_unix.c new->uid = item->uid; new 455 net/sunrpc/svcauth_unix.c struct unix_gid *new = container_of(cnew, struct unix_gid, h); new 459 net/sunrpc/svcauth_unix.c new->gi = item->gi; new 676 net/sunrpc/xdr.c int new; new 698 net/sunrpc/xdr.c new = buf->page_base + buf->page_len; new 700 net/sunrpc/xdr.c xdr->page_ptr = buf->pages + (new >> PAGE_SHIFT); new 705 net/sunrpc/xdr.c xdr->p = (void *)xdr->p + (new % PAGE_SIZE); new 926 net/sunrpc/xprt.c xprt_request_rb_insert(struct rpc_xprt *xprt, struct rpc_rqst *new) new 935 net/sunrpc/xprt.c switch(xprt_xid_cmp(new->rq_xid, req->rq_xid)) { new 943 net/sunrpc/xprt.c WARN_ON_ONCE(new != req); new 947 net/sunrpc/xprt.c rb_link_node(&new->rq_recv, n, p); new 948 net/sunrpc/xprt.c rb_insert_color(&new->rq_recv, &xprt->recv_queue); new 2858 net/sunrpc/xprtsock.c struct sock_xprt *new; new 2865 net/sunrpc/xprtsock.c xprt = xprt_alloc(args->net, sizeof(*new), slot_table_size, new 2873 net/sunrpc/xprtsock.c new = container_of(xprt, struct sock_xprt, xprt); new 2874 net/sunrpc/xprtsock.c mutex_init(&new->recv_mutex); new 2878 net/sunrpc/xprtsock.c memcpy(&new->srcaddr, args->srcaddr, args->addrlen); new 2882 net/sunrpc/xprtsock.c (struct sockaddr *)&new->srcaddr); new 555 net/tls/tls_sw.c struct tls_rec *new; new 558 net/tls/tls_sw.c new = tls_get_rec(sk); new 559 net/tls/tls_sw.c if (!new) new 561 net/tls/tls_sw.c ret = sk_msg_alloc(sk, &new->msg_encrypted, msg_opl->sg.size + new 564 net/tls/tls_sw.c tls_free_rec(sk, new); new 598 net/tls/tls_sw.c msg_npl = &new->msg_plaintext; new 626 net/tls/tls_sw.c *to = new; new 1423 net/unix/af_unix.c struct socket *new) new 1426 net/unix/af_unix.c set_bit(SOCK_PASSCRED, &new->flags); new 1428 net/unix/af_unix.c set_bit(SOCK_PASSSEC, &new->flags); new 288 net/vmw_vsock/hyperv_transport.c struct sock *sk, *new = NULL; new 316 net/vmw_vsock/hyperv_transport.c new = __vsock_create(sock_net(sk), NULL, sk, GFP_KERNEL, new 318 net/vmw_vsock/hyperv_transport.c if (!new) new 321 net/vmw_vsock/hyperv_transport.c new->sk_state = TCP_SYN_SENT; new 322 net/vmw_vsock/hyperv_transport.c vnew = vsock_sk(new); new 365 net/vmw_vsock/hyperv_transport.c conn_from_host ? new : sk); new 369 net/vmw_vsock/hyperv_transport.c sock_put(new); new 376 net/vmw_vsock/hyperv_transport.c set_per_channel_state(chan, conn_from_host ? new : sk); new 379 net/vmw_vsock/hyperv_transport.c sock_hold(conn_from_host ? new : sk); new 390 net/vmw_vsock/hyperv_transport.c new->sk_state = TCP_ESTABLISHED; new 399 net/vmw_vsock/hyperv_transport.c vsock_enqueue_accept(sk, new); new 1016 net/wireless/scan.c struct cfg80211_internal_bss *new) new 1025 net/wireless/scan.c ies = rcu_access_pointer(new->pub.beacon_ies); new 1053 net/wireless/scan.c if (!ether_addr_equal(bss->pub.bssid, new->pub.bssid)) new 1055 net/wireless/scan.c if (bss->pub.channel != new->pub.channel) new 1057 net/wireless/scan.c if (bss->pub.scan_width != new->pub.scan_width) new 1074 net/wireless/scan.c list_add(&bss->hidden_list, &new->hidden_list); new 1075 net/wireless/scan.c bss->pub.hidden_beacon_bss = &new->pub; new 1076 net/wireless/scan.c new->refcount += bss->refcount; new 1078 net/wireless/scan.c new->pub.beacon_ies); new 1097 net/wireless/scan.c struct cfg80211_internal_bss *new, new 1103 net/wireless/scan.c if (rcu_access_pointer(new->pub.proberesp_ies)) { new 1109 net/wireless/scan.c new->pub.proberesp_ies); new 1112 net/wireless/scan.c new->pub.proberesp_ies); new 1115 net/wireless/scan.c } else if (rcu_access_pointer(new->pub.beacon_ies)) { new 1132 net/wireless/scan.c f = rcu_access_pointer(new->pub.beacon_ies); new 1139 net/wireless/scan.c rcu_assign_pointer(known->pub.beacon_ies, new->pub.beacon_ies); new 1143 net/wireless/scan.c rcu_assign_pointer(known->pub.ies, new->pub.beacon_ies); new 1153 net/wireless/scan.c new->pub.beacon_ies); new 1160 net/wireless/scan.c known->pub.beacon_interval = new->pub.beacon_interval; new 1166 net/wireless/scan.c known->pub.signal = new->pub.signal; new 1167 net/wireless/scan.c known->pub.capability = new->pub.capability; new 1168 net/wireless/scan.c known->ts = new->ts; new 1169 net/wireless/scan.c known->ts_boottime = new->ts_boottime; new 1170 net/wireless/scan.c known->parent_tsf = new->parent_tsf; new 1171 net/wireless/scan.c known->pub.chains = new->pub.chains; new 1172 net/wireless/scan.c memcpy(known->pub.chain_signal, new->pub.chain_signal, new 1174 net/wireless/scan.c ether_addr_copy(known->parent_bssid, new->parent_bssid); new 1175 net/wireless/scan.c known->pub.max_bssid_indicator = new->pub.max_bssid_indicator; new 1176 net/wireless/scan.c known->pub.bssid_index = new->pub.bssid_index; new 1207 net/wireless/scan.c struct cfg80211_internal_bss *new; new 1216 net/wireless/scan.c new = kzalloc(sizeof(*new) + rdev->wiphy.bss_priv_size, new 1218 net/wireless/scan.c if (!new) { new 1227 net/wireless/scan.c memcpy(new, tmp, sizeof(*new)); new 1228 net/wireless/scan.c new->refcount = 1; new 1229 net/wireless/scan.c INIT_LIST_HEAD(&new->hidden_list); new 1230 net/wireless/scan.c INIT_LIST_HEAD(&new->pub.nontrans_list); new 1238 net/wireless/scan.c new->pub.hidden_beacon_bss = &hidden->pub; new 1239 net/wireless/scan.c list_add(&new->hidden_list, new 1242 net/wireless/scan.c rcu_assign_pointer(new->pub.beacon_ies, new 1252 net/wireless/scan.c if (!cfg80211_combine_bsses(rdev, new)) { new 1253 net/wireless/scan.c kfree(new); new 1260 net/wireless/scan.c kfree(new); new 1271 net/wireless/scan.c new->pub.transmitted_bss = tmp->pub.transmitted_bss; new 1275 net/wireless/scan.c list_add_tail(&new->list, &rdev->bss_list); new 1277 net/wireless/scan.c rb_insert_bss(rdev, new); new 1278 net/wireless/scan.c found = new; new 2018 net/wireless/scan.c struct cfg80211_internal_bss *new = NULL; new 2050 net/wireless/scan.c new = bss; new 2055 net/wireless/scan.c if (new) { new 2057 net/wireless/scan.c if (cfg80211_update_known_bss(rdev, cbss, new, false)) { new 2058 net/wireless/scan.c new->pub.proberesp_ies = NULL; new 2059 net/wireless/scan.c new->pub.beacon_ies = NULL; new 2063 net/wireless/scan.c &new->pub.nontrans_list, new 2071 net/wireless/scan.c WARN_ON(atomic_read(&new->hold)); new 2072 net/wireless/scan.c if (!WARN_ON(!__cfg80211_unlink_bss(rdev, new))) new 266 net/x25/x25_facilities.c struct x25_facilities *new, struct x25_dte_facilities *dte) new 274 net/x25/x25_facilities.c memcpy(new, ours, sizeof(*new)); new 289 net/x25/x25_facilities.c new->reverse = theirs.reverse; new 298 net/x25/x25_facilities.c new->throughput = (new->throughput & 0xf0) | theirs_in; new 303 net/x25/x25_facilities.c new->throughput = (new->throughput & 0x0f) | theirs_out; new 310 net/x25/x25_facilities.c new->pacsize_in = theirs.pacsize_in; new 314 net/x25/x25_facilities.c new->pacsize_out = theirs.pacsize_out; new 321 net/x25/x25_facilities.c new->winsize_in = theirs.winsize_in; new 325 net/x25/x25_facilities.c new->winsize_out = theirs.winsize_out; new 890 net/xfrm/xfrm_policy.c struct rb_root *new, new 900 net/xfrm/xfrm_policy.c p = &new->rb_node; new 934 net/xfrm/xfrm_policy.c rb_erase(*p, new); new 942 net/xfrm/xfrm_policy.c rb_insert_color(&n->node, new); new 1407 net/xfrm/xfrm_policy.c struct xfrm_policy *new) new 1423 net/xfrm/xfrm_policy.c pq = &new->polq; new 1429 net/xfrm/xfrm_policy.c xfrm_pol_hold(new); new 51 samples/bpf/test_lru_dist.c static inline void __list_add(struct list_head *new, new 55 samples/bpf/test_lru_dist.c next->prev = new; new 56 samples/bpf/test_lru_dist.c new->next = next; new 57 samples/bpf/test_lru_dist.c new->prev = prev; new 58 samples/bpf/test_lru_dist.c prev->next = new; new 61 samples/bpf/test_lru_dist.c static inline void list_add(struct list_head *new, struct list_head *head) new 63 samples/bpf/test_lru_dist.c __list_add(new, head, head->next); new 500 samples/bpf/xdp_redirect_cpu_user.c __u32 avail_idx, bool new) new 530 samples/bpf/xdp_redirect_cpu_user.c if (new) { new 541 samples/bpf/xdp_redirect_cpu_user.c new ? "Add-new":"Replace", cpu, avail_idx, new 253 scripts/dtc/dtc.h struct reserve_info *new); new 687 scripts/dtc/flattree.c struct reserve_info *new; new 705 scripts/dtc/flattree.c new = build_reserve_entry(address, size); new 706 scripts/dtc/flattree.c reservelist = add_reserve_entry(reservelist, new); new 119 scripts/dtc/libfdt/fdt_rw.c char *new; new 130 scripts/dtc/libfdt/fdt_rw.c new = strtab + fdt_size_dt_strings(fdt); new 137 scripts/dtc/libfdt/fdt_rw.c memcpy(new, s, len); new 138 scripts/dtc/libfdt/fdt_rw.c return (new - strtab); new 378 scripts/dtc/libfdt/fdt_rw.c static void fdt_packblocks_(const char *old, char *new, new 387 scripts/dtc/libfdt/fdt_rw.c memmove(new + mem_rsv_off, old + fdt_off_mem_rsvmap(old), mem_rsv_size); new 388 scripts/dtc/libfdt/fdt_rw.c fdt_set_off_mem_rsvmap(new, mem_rsv_off); new 390 scripts/dtc/libfdt/fdt_rw.c memmove(new + struct_off, old + fdt_off_dt_struct(old), struct_size); new 391 scripts/dtc/libfdt/fdt_rw.c fdt_set_off_dt_struct(new, struct_off); new 392 scripts/dtc/libfdt/fdt_rw.c fdt_set_size_dt_struct(new, struct_size); new 394 scripts/dtc/libfdt/fdt_rw.c memmove(new + strings_off, old + fdt_off_dt_strings(old), new 396 scripts/dtc/libfdt/fdt_rw.c fdt_set_off_dt_strings(new, strings_off); new 397 scripts/dtc/libfdt/fdt_rw.c fdt_set_size_dt_strings(new, fdt_size_dt_strings(old)); new 15 scripts/dtc/livetree.c struct label *new; new 18 scripts/dtc/livetree.c for_each_label_withdel(*labels, new) new 19 scripts/dtc/livetree.c if (streq(new->label, label)) { new 20 scripts/dtc/livetree.c new->deleted = 0; new 24 scripts/dtc/livetree.c new = xmalloc(sizeof(*new)); new 25 scripts/dtc/livetree.c memset(new, 0, sizeof(*new)); new 26 scripts/dtc/livetree.c new->label = label; new 27 scripts/dtc/livetree.c new->next = *labels; new 28 scripts/dtc/livetree.c *labels = new; new 42 scripts/dtc/livetree.c struct property *new = xmalloc(sizeof(*new)); new 44 scripts/dtc/livetree.c memset(new, 0, sizeof(*new)); new 46 scripts/dtc/livetree.c new->name = name; new 47 scripts/dtc/livetree.c new->val = val; new 48 scripts/dtc/livetree.c new->srcpos = srcpos_copy(srcpos); new 50 scripts/dtc/livetree.c return new; new 55 scripts/dtc/livetree.c struct property *new = xmalloc(sizeof(*new)); new 57 scripts/dtc/livetree.c memset(new, 0, sizeof(*new)); new 59 scripts/dtc/livetree.c new->name = name; new 60 scripts/dtc/livetree.c new->deleted = 1; new 62 scripts/dtc/livetree.c return new; new 91 scripts/dtc/livetree.c struct node *new = xmalloc(sizeof(*new)); new 94 scripts/dtc/livetree.c memset(new, 0, sizeof(*new)); new 96 scripts/dtc/livetree.c new->proplist = reverse_properties(proplist); new 97 scripts/dtc/livetree.c new->children = children; new 98 scripts/dtc/livetree.c new->srcpos = srcpos_copy(srcpos); new 100 scripts/dtc/livetree.c for_each_child(new, child) { new 101 scripts/dtc/livetree.c child->parent = new; new 104 scripts/dtc/livetree.c return new; new 109 scripts/dtc/livetree.c struct node *new = xmalloc(sizeof(*new)); new 111 scripts/dtc/livetree.c memset(new, 0, sizeof(*new)); new 113 scripts/dtc/livetree.c new->deleted = 1; new 114 scripts/dtc/livetree.c new->srcpos = srcpos_copy(srcpos); new 116 scripts/dtc/livetree.c return new; new 360 scripts/dtc/livetree.c struct reserve_info *new = xmalloc(sizeof(*new)); new 362 scripts/dtc/livetree.c memset(new, 0, sizeof(*new)); new 364 scripts/dtc/livetree.c new->address = address; new 365 scripts/dtc/livetree.c new->size = size; new 367 scripts/dtc/livetree.c return new; new 380 scripts/dtc/livetree.c struct reserve_info *new) new 384 scripts/dtc/livetree.c new->next = NULL; new 387 scripts/dtc/livetree.c return new; new 392 scripts/dtc/livetree.c last->next = new; new 40 scripts/dtc/util.h void *new = malloc(len); new 42 scripts/dtc/util.h if (!new) new 45 scripts/dtc/util.h return new; new 50 scripts/dtc/util.h void *new = realloc(p, len); new 52 scripts/dtc/util.h if (!new) new 55 scripts/dtc/util.h return new; new 123 scripts/gcc-plugins/gcc-generate-gimple-pass.h virtual opt_pass * clone () { return new _PASS_NAME_PASS(); } new 137 scripts/gcc-plugins/gcc-generate-gimple-pass.h return new _PASS_NAME_PASS(); new 212 scripts/gcc-plugins/gcc-generate-ipa-pass.h virtual opt_pass *clone() { return new _PASS_NAME_PASS(); } new 226 scripts/gcc-plugins/gcc-generate-ipa-pass.h return new _PASS_NAME_PASS(); new 123 scripts/gcc-plugins/gcc-generate-rtl-pass.h virtual opt_pass *clone() { return new _PASS_NAME_PASS(); } new 137 scripts/gcc-plugins/gcc-generate-rtl-pass.h return new _PASS_NAME_PASS(); new 123 scripts/gcc-plugins/gcc-generate-simple_ipa-pass.h virtual opt_pass *clone() { return new _PASS_NAME_PASS(); } new 137 scripts/gcc-plugins/gcc-generate-simple_ipa-pass.h return new _PASS_NAME_PASS(); new 217 scripts/mod/modpost.c struct symbol *new; new 220 scripts/mod/modpost.c new = symbolhash[hash] = alloc_symbol(name, 0, symbolhash[hash]); new 221 scripts/mod/modpost.c new->module = module; new 222 scripts/mod/modpost.c new->export = export; new 223 scripts/mod/modpost.c return new; new 845 security/apparmor/apparmorfs.c struct multi_transaction *new, size_t n) new 851 security/apparmor/apparmorfs.c new->size = n; new 854 security/apparmor/apparmorfs.c file->private_data = new; new 1445 security/apparmor/apparmorfs.c struct aa_profile *new) new 1450 security/apparmor/apparmorfs.c AA_BUG(!new); new 1454 security/apparmor/apparmorfs.c new->dents[i] = old->dents[i]; new 1455 security/apparmor/apparmorfs.c if (new->dents[i]) new 1456 security/apparmor/apparmorfs.c new->dents[i]->d_inode->i_mtime = current_time(new->dents[i]->d_inode); new 556 security/apparmor/domain.c struct aa_label *new = NULL; new 571 security/apparmor/domain.c new = x_table_lookup(profile, xindex, lookupname); new 579 security/apparmor/domain.c new = find_attach(bprm, ns, &profile->base.profiles, new 583 security/apparmor/domain.c new = find_attach(bprm, ns, &ns->base.profiles, new 589 security/apparmor/domain.c if (!new) { new 596 security/apparmor/domain.c new = aa_get_newest_label(&profile->label); new 598 security/apparmor/domain.c new = aa_get_newest_label(ns_unconfined(profile->ns)); new 603 security/apparmor/domain.c if (new && stack) { new 605 security/apparmor/domain.c struct aa_label *base = new; new 607 security/apparmor/domain.c new = aa_label_parse(base, stack, GFP_ATOMIC, true, false); new 608 security/apparmor/domain.c if (IS_ERR(new)) new 609 security/apparmor/domain.c new = NULL; new 614 security/apparmor/domain.c return new; new 622 security/apparmor/domain.c struct aa_label *new = NULL; new 642 security/apparmor/domain.c new = aa_get_newest_label(&profile->label); new 649 security/apparmor/domain.c new = find_attach(bprm, profile->ns, new 651 security/apparmor/domain.c if (new) { new 653 security/apparmor/domain.c return new; new 663 security/apparmor/domain.c new = x_to_label(profile, bprm, name, perms.xindex, &target, new 665 security/apparmor/domain.c if (new && new->proxy == profile->label.proxy && info) { new 668 security/apparmor/domain.c } else if (!new) { new 677 security/apparmor/domain.c label_for_each(i, new, component) { new 683 security/apparmor/domain.c aa_put_label(new); new 684 security/apparmor/domain.c new = NULL; new 711 security/apparmor/domain.c new = &new_profile->label; new 718 security/apparmor/domain.c if (!new) new 726 security/apparmor/domain.c aa_label_printk(new, GFP_ATOMIC); new 733 security/apparmor/domain.c aa_audit_file(profile, &perms, OP_EXEC, MAY_EXEC, name, target, new, new 735 security/apparmor/domain.c if (!new || nonewprivs) { new 736 security/apparmor/domain.c aa_put_label(new); new 740 security/apparmor/domain.c return new; new 822 security/apparmor/domain.c struct aa_label *new; new 836 security/apparmor/domain.c new = fn_label_build_in_ns(label, profile, GFP_ATOMIC, new 848 security/apparmor/domain.c new = fn_label_build_in_ns(label, profile, GFP_ATOMIC, new 855 security/apparmor/domain.c if (new) new 856 security/apparmor/domain.c return new; new 878 security/apparmor/domain.c struct aa_label *label, *new = NULL; new 913 security/apparmor/domain.c new = handle_onexec(label, ctx->onexec, ctx->token, new 916 security/apparmor/domain.c new = fn_label_build(label, profile, GFP_ATOMIC, new 920 security/apparmor/domain.c AA_BUG(!new); new 921 security/apparmor/domain.c if (IS_ERR(new)) { new 922 security/apparmor/domain.c error = PTR_ERR(new); new 924 security/apparmor/domain.c } else if (!new) { new 938 security/apparmor/domain.c !unconfined(label) && !aa_label_is_subset(new, ctx->nnp)) { new 951 security/apparmor/domain.c error = may_change_ptraced_domain(new, &info); new 960 security/apparmor/domain.c aa_label_printk(new, GFP_ATOMIC); new 966 security/apparmor/domain.c if (label->proxy != new->proxy) { new 971 security/apparmor/domain.c aa_label_printk(new, GFP_ATOMIC); new 978 security/apparmor/domain.c set_cred_label(bprm->cred, new); new 989 security/apparmor/domain.c bprm->filename, NULL, new, new 992 security/apparmor/domain.c aa_put_label(new); new 1057 security/apparmor/domain.c struct aa_label *new; new 1137 security/apparmor/domain.c new = fn_label_build_in_ns(label, profile, GFP_KERNEL, new 1140 security/apparmor/domain.c if (!new) { new 1146 security/apparmor/domain.c return new; new 1170 security/apparmor/domain.c struct aa_label *label, *previous, *new = NULL, *target = NULL; new 1198 security/apparmor/domain.c new = change_hat(label, hats, count, flags); new 1199 security/apparmor/domain.c AA_BUG(!new); new 1200 security/apparmor/domain.c if (IS_ERR(new)) { new 1201 security/apparmor/domain.c error = PTR_ERR(new); new 1202 security/apparmor/domain.c new = NULL; new 1207 security/apparmor/domain.c error = may_change_ptraced_domain(new, &info); new 1216 security/apparmor/domain.c !aa_label_is_subset(new, ctx->nnp)) { new 1226 security/apparmor/domain.c target = new; new 1227 security/apparmor/domain.c error = aa_set_current_hat(new, token); new 1257 security/apparmor/domain.c aa_put_label(new); new 1313 security/apparmor/domain.c struct aa_label *label, *new = NULL, *target = NULL; new 1424 security/apparmor/domain.c new = fn_label_build_in_ns(label, profile, GFP_KERNEL, new 1432 security/apparmor/domain.c !aa_label_is_subset(new, ctx->nnp)) { new 1443 security/apparmor/domain.c new = aa_label_merge(label, target, GFP_KERNEL); new 1444 security/apparmor/domain.c if (IS_ERR_OR_NULL(new)) { new 1446 security/apparmor/domain.c if (!new) new 1449 security/apparmor/domain.c error = PTR_ERR(new); new 1450 security/apparmor/domain.c new = NULL; new 1454 security/apparmor/domain.c error = aa_replace_current_label(new); new 1456 security/apparmor/domain.c if (new) { new 1457 security/apparmor/domain.c aa_put_label(new); new 1458 security/apparmor/domain.c new = NULL; new 1468 security/apparmor/domain.c NULL, new ? new : target, new 1472 security/apparmor/domain.c aa_put_label(new); new 194 security/apparmor/file.c u32 new = old & 0xf; new 196 security/apparmor/file.c new |= AA_MAY_GETATTR | AA_MAY_OPEN; new 198 security/apparmor/file.c new |= AA_MAY_SETATTR | AA_MAY_CREATE | AA_MAY_DELETE | new 201 security/apparmor/file.c new |= AA_MAY_LINK; new 206 security/apparmor/file.c new |= AA_MAY_LOCK | AA_LINK_SUBSET; new 208 security/apparmor/file.c new |= AA_EXEC_MMAP; new 210 security/apparmor/file.c return new; new 110 security/apparmor/include/apparmorfs.h struct aa_profile *new); new 289 security/apparmor/include/label.h bool aa_label_replace(struct aa_label *old, struct aa_label *new); new 291 security/apparmor/include/label.h struct aa_label *new); new 464 security/apparmor/include/label.h void __aa_proxy_redirect(struct aa_label *orig, struct aa_label *new); new 21 security/apparmor/include/policy_unpack.h struct aa_profile *new; new 39 security/apparmor/include/resource.h void __aa_transition_rlimits(struct aa_label *old, struct aa_label *new); new 56 security/apparmor/include/task.h static inline void aa_dup_task_ctx(struct aa_task_ctx *new, new 59 security/apparmor/include/task.h *new = *old; new 60 security/apparmor/include/task.h aa_get_label(new->nnp); new 61 security/apparmor/include/task.h aa_get_label(new->previous); new 62 security/apparmor/include/task.h aa_get_label(new->onexec); new 62 security/apparmor/label.c struct aa_proxy *new; new 64 security/apparmor/label.c new = kzalloc(sizeof(struct aa_proxy), gfp); new 65 security/apparmor/label.c if (new) { new 66 security/apparmor/label.c kref_init(&new->count); new 67 security/apparmor/label.c rcu_assign_pointer(new->label, aa_get_label(label)); new 69 security/apparmor/label.c return new; new 73 security/apparmor/label.c void __aa_proxy_redirect(struct aa_label *orig, struct aa_label *new) new 78 security/apparmor/label.c AA_BUG(!new); new 83 security/apparmor/label.c rcu_assign_pointer(orig->proxy->label, aa_get_label(new)); new 88 security/apparmor/label.c static void __proxy_share(struct aa_label *old, struct aa_label *new) new 90 security/apparmor/label.c struct aa_proxy *proxy = new->proxy; new 92 security/apparmor/label.c new->proxy = aa_get_proxy(old->proxy); new 93 security/apparmor/label.c __aa_proxy_redirect(old, new); new 392 security/apparmor/label.c static void label_free_or_put_new(struct aa_label *label, struct aa_label *new) new 394 security/apparmor/label.c if (label != new) new 396 security/apparmor/label.c aa_label_free(new); new 398 security/apparmor/label.c aa_put_label(new); new 428 security/apparmor/label.c struct aa_label *new; new 433 security/apparmor/label.c new = kzalloc(sizeof(*new) + sizeof(struct aa_profile *) * (size + 1), new 435 security/apparmor/label.c AA_DEBUG("%s (%p)\n", __func__, new); new 436 security/apparmor/label.c if (!new) new 439 security/apparmor/label.c if (!aa_label_init(new, size, gfp)) new 443 security/apparmor/label.c proxy = aa_alloc_proxy(new, gfp); new 449 security/apparmor/label.c new->proxy = proxy; new 451 security/apparmor/label.c return new; new 454 security/apparmor/label.c kfree(new); new 563 security/apparmor/label.c static bool __label_remove(struct aa_label *label, struct aa_label *new) new 571 security/apparmor/label.c if (new) new 572 security/apparmor/label.c __aa_proxy_redirect(label, new); new 599 security/apparmor/label.c static bool __label_replace(struct aa_label *old, struct aa_label *new) new 605 security/apparmor/label.c AA_BUG(!new); new 607 security/apparmor/label.c AA_BUG(new->flags & FLAG_IN_TREE); new 613 security/apparmor/label.c rb_replace_node(&old->node, &new->node, &ls->root); new 615 security/apparmor/label.c new->flags |= FLAG_IN_TREE; new 638 security/apparmor/label.c struct rb_node **new, *parent = NULL; new 647 security/apparmor/label.c new = &ls->root.rb_node; new 648 security/apparmor/label.c while (*new) { new 649 security/apparmor/label.c struct aa_label *this = rb_entry(*new, struct aa_label, node); new 652 security/apparmor/label.c parent = *new; new 667 security/apparmor/label.c new = &((*new)->rb_left); new 669 security/apparmor/label.c new = &((*new)->rb_right); new 673 security/apparmor/label.c rb_link_node(&label->node, parent, new); new 765 security/apparmor/label.c bool aa_label_replace(struct aa_label *old, struct aa_label *new) new 770 security/apparmor/label.c if (name_is_shared(old, new) && labels_ns(old) == labels_ns(new)) { new 772 security/apparmor/label.c if (old->proxy != new->proxy) new 773 security/apparmor/label.c __proxy_share(old, new); new 775 security/apparmor/label.c __aa_proxy_redirect(old, new); new 776 security/apparmor/label.c res = __label_replace(old, new); new 783 security/apparmor/label.c res = __label_remove(old, new); new 784 security/apparmor/label.c if (labels_ns(old) != labels_ns(new)) { new 786 security/apparmor/label.c ls = labels_set(new); new 789 security/apparmor/label.c l = __label_insert(ls, new, true); new 790 security/apparmor/label.c res = (l == new); new 831 security/apparmor/label.c struct aa_label *new; new 844 security/apparmor/label.c new = aa_label_alloc(len, NULL, gfp); new 845 security/apparmor/label.c if (!new) new 849 security/apparmor/label.c new->vec[i] = aa_get_profile(vec[i]); new 852 security/apparmor/label.c label = __label_insert(ls, new, false); new 854 security/apparmor/label.c label_free_or_put_new(label, new); new 1019 security/apparmor/label.c static struct aa_label *label_merge_insert(struct aa_label *new, new 1035 security/apparmor/label.c AA_BUG(!new); new 1036 security/apparmor/label.c AA_BUG(new->size < a->size + b->size); new 1041 security/apparmor/label.c new->vec[k] = aa_get_newest_profile(next); new 1042 security/apparmor/label.c AA_BUG(!new->vec[k]->label.proxy); new 1043 security/apparmor/label.c AA_BUG(!new->vec[k]->label.proxy->label); new 1044 security/apparmor/label.c if (next->label.proxy != new->vec[k]->label.proxy) new 1049 security/apparmor/label.c new->vec[k++] = aa_get_profile(next); new 1052 security/apparmor/label.c new->size = k; new 1053 security/apparmor/label.c new->vec[k] = NULL; new 1056 security/apparmor/label.c new->size -= aa_vec_unique(&new->vec[0], new->size, new 1059 security/apparmor/label.c if (new->size == 1) { new 1060 security/apparmor/label.c label = aa_get_label(&new->vec[0]->label); new 1073 security/apparmor/label.c if (vec_unconfined(new->vec, new->size)) new 1074 security/apparmor/label.c new->flags |= FLAG_UNCONFINED; new 1075 security/apparmor/label.c ls = labels_set(new); new 1077 security/apparmor/label.c label = __label_insert(labels_set(new), new, false); new 1208 security/apparmor/label.c struct aa_label *new; new 1216 security/apparmor/label.c new = aa_label_alloc(a->size + b->size, NULL, gfp); new 1217 security/apparmor/label.c if (!new) new 1220 security/apparmor/label.c label = label_merge_insert(new, a, b); new 1221 security/apparmor/label.c label_free_or_put_new(label, new); new 2023 security/apparmor/label.c struct aa_label *new, *tmp; new 2031 security/apparmor/label.c new = aa_label_alloc(label->size, label->proxy, GFP_KERNEL); new 2032 security/apparmor/label.c if (!new) new 2043 security/apparmor/label.c new->vec[i] = aa_get_newest_profile(label->vec[i]); new 2044 security/apparmor/label.c AA_BUG(!new->vec[i]); new 2045 security/apparmor/label.c AA_BUG(!new->vec[i]->label.proxy); new 2046 security/apparmor/label.c AA_BUG(!new->vec[i]->label.proxy->label); new 2047 security/apparmor/label.c if (new->vec[i]->label.proxy != label->vec[i]->label.proxy) new 2053 security/apparmor/label.c new->size -= aa_vec_unique(&new->vec[0], new->size, new 2056 security/apparmor/label.c if (new->size == 1) { new 2057 security/apparmor/label.c tmp = aa_get_label(&new->vec[0]->label); new 2061 security/apparmor/label.c if (labels_set(label) != labels_set(new)) { new 2063 security/apparmor/label.c tmp = aa_label_insert(labels_set(new), new); new 2068 security/apparmor/label.c AA_BUG(labels_ns(label) != labels_ns(new)); new 2070 security/apparmor/label.c tmp = __label_insert(labels_set(label), new, true); new 2075 security/apparmor/label.c label_free_or_put_new(tmp, new); new 74 security/apparmor/lsm.c static int apparmor_cred_prepare(struct cred *new, const struct cred *old, new 77 security/apparmor/lsm.c set_cred_label(new, aa_get_newest_label(cred_label(old))); new 84 security/apparmor/lsm.c static void apparmor_cred_transfer(struct cred *new, const struct cred *old) new 86 security/apparmor/lsm.c set_cred_label(new, aa_get_newest_label(cred_label(old))); new 98 security/apparmor/lsm.c struct aa_task_ctx *new = task_ctx(task); new 100 security/apparmor/lsm.c aa_dup_task_ctx(new, task_ctx(current)); new 792 security/apparmor/lsm.c struct aa_sk_ctx *new = SK_CTX(newsk); new 794 security/apparmor/lsm.c new->label = aa_get_label(ctx->label); new 795 security/apparmor/lsm.c new->peer = aa_get_label(ctx->peer); new 717 security/apparmor/policy.c if (ent->new == profile) new 719 security/apparmor/policy.c if (strncmp(ent->new->base.hname, profile->base.hname, len) == new 720 security/apparmor/policy.c 0 && ent->new->base.hname[len] == 0) new 721 security/apparmor/policy.c return ent->new; new 740 security/apparmor/policy.c static void __replace_profile(struct aa_profile *old, struct aa_profile *new) new 752 security/apparmor/policy.c p = __find_child(&new->base.profiles, child->base.name); new 763 security/apparmor/policy.c rcu_assign_pointer(child->parent, aa_get_profile(new)); new 764 security/apparmor/policy.c list_add_rcu(&child->base.list, &new->base.profiles); new 769 security/apparmor/policy.c if (!rcu_access_pointer(new->parent)) { new 771 security/apparmor/policy.c rcu_assign_pointer(new->parent, aa_get_profile(parent)); new 773 security/apparmor/policy.c aa_label_replace(&old->label, &new->label); new 775 security/apparmor/policy.c __aafs_profile_migrate_dents(old, new); new 777 security/apparmor/policy.c if (list_empty(&new->base.list)) { new 779 security/apparmor/policy.c list_replace_rcu(&old->base.list, &new->base.list); new 780 security/apparmor/policy.c aa_get_profile(new); new 812 security/apparmor/policy.c static void share_name(struct aa_profile *old, struct aa_profile *new) new 814 security/apparmor/policy.c aa_put_str(new->base.hname); new 816 security/apparmor/policy.c new->base.hname = old->base.hname; new 817 security/apparmor/policy.c new->base.name = old->base.name; new 818 security/apparmor/policy.c new->label.hname = old->label.hname; new 824 security/apparmor/policy.c static struct aa_profile *update_to_newest_parent(struct aa_profile *new) new 828 security/apparmor/policy.c parent = rcu_dereference_protected(new->parent, new 829 security/apparmor/policy.c mutex_is_locked(&new->ns->lock)); new 835 security/apparmor/policy.c rcu_assign_pointer(new->parent, newest); new 931 security/apparmor/policy.c ent->new->rawdata = aa_get_loaddata(udata); new 932 security/apparmor/policy.c error = __lookup_replace(ns, ent->new->base.hname, new 938 security/apparmor/policy.c if (ent->new->rename) { new 939 security/apparmor/policy.c error = __lookup_replace(ns, ent->new->rename, new 947 security/apparmor/policy.c ent->new->ns = aa_get_ns(ns); new 953 security/apparmor/policy.c policy = __lookup_parent(ns, ent->new->base.hname); new 956 security/apparmor/policy.c p = __list_lookup_parent(&lh, ent->new); new 962 security/apparmor/policy.c rcu_assign_pointer(ent->new->parent, aa_get_profile(p)); new 966 security/apparmor/policy.c rcu_assign_pointer(ent->new->parent, aa_get_profile(p)); new 982 security/apparmor/policy.c if (rcu_access_pointer(ent->new->parent)) { new 984 security/apparmor/policy.c p = aa_deref_parent(ent->new); new 987 security/apparmor/policy.c parent = ns_subprofs_dir(ent->new->ns); new 988 security/apparmor/policy.c error = __aafs_profile_mkdir(ent->new, parent); new 1004 security/apparmor/policy.c if (ent->old && ent->old->rawdata == ent->new->rawdata) { new 1006 security/apparmor/policy.c audit_policy(label, op, ns_name, ent->new->base.hname, new 1010 security/apparmor/policy.c aa_put_proxy(ent->new->label.proxy); new 1011 security/apparmor/policy.c ent->new->label.proxy = NULL; new 1019 security/apparmor/policy.c audit_policy(label, op, ns_name, ent->new->base.hname, NULL, new 1023 security/apparmor/policy.c share_name(ent->old, ent->new); new 1024 security/apparmor/policy.c __replace_profile(ent->old, ent->new); new 1028 security/apparmor/policy.c if (rcu_access_pointer(ent->new->parent)) { new 1031 security/apparmor/policy.c parent = update_to_newest_parent(ent->new); new 1035 security/apparmor/policy.c __add_profile(lh, ent->new); new 1057 security/apparmor/policy.c audit_policy(label, op, ns_name, ent ? ent->new->base.hname : NULL, new 1068 security/apparmor/policy.c audit_policy(label, op, ns_name, tmp->new->base.hname, info, new 104 security/apparmor/policy_unpack.c static int audit_iface(struct aa_profile *new, const char *ns_name, new 113 security/apparmor/policy_unpack.c if (new) new 114 security/apparmor/policy_unpack.c aad(&sa)->name = new->base.hname; new 1028 security/apparmor/policy_unpack.c aa_put_profile(ent->new); new 1097 security/apparmor/policy_unpack.c ent->new = profile; new 145 security/apparmor/resource.c struct aa_profile *old, *new; new 149 security/apparmor/resource.c new = labels_profile(new_l); new 171 security/apparmor/resource.c label_for_each_confined(i, new_l, new) { new 174 security/apparmor/resource.c if (!new->rlimits.mask) new 177 security/apparmor/resource.c if (!(new->rlimits.mask & mask)) new 182 security/apparmor/resource.c new->rlimits.limits[j].rlim_max); new 45 security/apparmor/task.c struct cred *new; new 55 security/apparmor/task.c new = prepare_creds(); new 56 security/apparmor/task.c if (!new) new 79 security/apparmor/task.c aa_put_label(cred_label(new)); new 80 security/apparmor/task.c set_cred_label(new, label); new 82 security/apparmor/task.c commit_creds(new); new 118 security/apparmor/task.c struct cred *new; new 120 security/apparmor/task.c new = prepare_creds(); new 121 security/apparmor/task.c if (!new) new 127 security/apparmor/task.c ctx->previous = cred_label(new); new 130 security/apparmor/task.c aa_put_label(cred_label(new)); new 133 security/apparmor/task.c abort_creds(new); new 137 security/apparmor/task.c set_cred_label(new, aa_get_newest_label(label)); new 142 security/apparmor/task.c commit_creds(new); new 158 security/apparmor/task.c struct cred *new; new 166 security/apparmor/task.c new = prepare_creds(); new 167 security/apparmor/task.c if (!new) new 170 security/apparmor/task.c aa_put_label(cred_label(new)); new 171 security/apparmor/task.c set_cred_label(new, aa_get_newest_label(ctx->previous)); new 172 security/apparmor/task.c AA_BUG(!cred_label(new)); new 176 security/apparmor/task.c commit_creds(new); new 241 security/commoncap.c int cap_capset(struct cred *new, new 268 security/commoncap.c new->cap_effective = *effective; new 269 security/commoncap.c new->cap_inheritable = *inheritable; new 270 security/commoncap.c new->cap_permitted = *permitted; new 276 security/commoncap.c new->cap_ambient = cap_intersect(new->cap_ambient, new 279 security/commoncap.c if (WARN_ON(!cap_ambient_invariant_ok(new))) new 533 security/commoncap.c struct cred *new = bprm->cred; new 551 security/commoncap.c new->cap_permitted.cap[i] = new 552 security/commoncap.c (new->cap_bset.cap[i] & permitted) | new 553 security/commoncap.c (new->cap_inheritable.cap[i] & inheritable); new 555 security/commoncap.c if (permitted & ~new->cap_permitted.cap[i]) new 717 security/commoncap.c struct cred *new = bprm->cred; new 726 security/commoncap.c if (has_fcap && __is_suid(root_uid, new)) { new 735 security/commoncap.c if (__is_eff(root_uid, new) || __is_real(root_uid, new)) { new 737 security/commoncap.c new->cap_permitted = cap_combine(old->cap_bset, new 743 security/commoncap.c if (__is_eff(root_uid, new)) new 754 security/commoncap.c static inline bool __is_setuid(struct cred *new, const struct cred *old) new 755 security/commoncap.c { return !uid_eq(new->euid, old->uid); } new 757 security/commoncap.c static inline bool __is_setgid(struct cred *new, const struct cred *old) new 758 security/commoncap.c { return !gid_eq(new->egid, old->gid); } new 777 security/commoncap.c static inline bool nonroot_raised_pE(struct cred *new, const struct cred *old, new 782 security/commoncap.c if ((__cap_grew(effective, ambient, new) && new 783 security/commoncap.c !(__cap_full(effective, new) && new 784 security/commoncap.c (__is_eff(root, new) || __is_real(root, new)) && new 787 security/commoncap.c __is_suid(root, new) && new 788 security/commoncap.c !__cap_full(effective, new)) || new 789 security/commoncap.c (!__is_setuid(new, old) && new 791 security/commoncap.c __cap_gained(permitted, new, old)) || new 792 security/commoncap.c __cap_gained(ambient, new, old)))) new 810 security/commoncap.c struct cred *new = bprm->cred; new 815 security/commoncap.c new->cap_ambient = old->cap_ambient; new 823 security/commoncap.c root_uid = make_kuid(new->user_ns, 0); new 828 security/commoncap.c if (__cap_gained(permitted, new, old)) new 836 security/commoncap.c is_setid = __is_setuid(new, old) || __is_setgid(new, old); new 838 security/commoncap.c if ((is_setid || __cap_gained(permitted, new, old)) && new 840 security/commoncap.c !ptracer_capable(current, new->user_ns))) { new 842 security/commoncap.c if (!ns_capable(new->user_ns, CAP_SETUID) || new 844 security/commoncap.c new->euid = new->uid; new 845 security/commoncap.c new->egid = new->gid; new 847 security/commoncap.c new->cap_permitted = cap_intersect(new->cap_permitted, new 851 security/commoncap.c new->suid = new->fsuid = new->euid; new 852 security/commoncap.c new->sgid = new->fsgid = new->egid; new 856 security/commoncap.c cap_clear(new->cap_ambient); new 862 security/commoncap.c new->cap_permitted = cap_combine(new->cap_permitted, new->cap_ambient); new 869 security/commoncap.c new->cap_effective = new->cap_permitted; new 871 security/commoncap.c new->cap_effective = new->cap_ambient; new 873 security/commoncap.c if (WARN_ON(!cap_ambient_invariant_ok(new))) new 876 security/commoncap.c if (nonroot_raised_pE(new, old, root_uid, has_fcap)) { new 877 security/commoncap.c ret = audit_log_bprm_fcaps(bprm, new, old); new 882 security/commoncap.c new->securebits &= ~issecure_mask(SECURE_KEEP_CAPS); new 884 security/commoncap.c if (WARN_ON(!cap_ambient_invariant_ok(new))) new 890 security/commoncap.c (!__is_real(root_uid, new) && new 892 security/commoncap.c __cap_grew(permitted, ambient, new)))) new 998 security/commoncap.c static inline void cap_emulate_setxuid(struct cred *new, const struct cred *old) new 1005 security/commoncap.c (!uid_eq(new->uid, root_uid) && new 1006 security/commoncap.c !uid_eq(new->euid, root_uid) && new 1007 security/commoncap.c !uid_eq(new->suid, root_uid))) { new 1009 security/commoncap.c cap_clear(new->cap_permitted); new 1010 security/commoncap.c cap_clear(new->cap_effective); new 1018 security/commoncap.c cap_clear(new->cap_ambient); new 1020 security/commoncap.c if (uid_eq(old->euid, root_uid) && !uid_eq(new->euid, root_uid)) new 1021 security/commoncap.c cap_clear(new->cap_effective); new 1022 security/commoncap.c if (!uid_eq(old->euid, root_uid) && uid_eq(new->euid, root_uid)) new 1023 security/commoncap.c new->cap_effective = new->cap_permitted; new 1035 security/commoncap.c int cap_task_fix_setuid(struct cred *new, const struct cred *old, int flags) new 1044 security/commoncap.c cap_emulate_setxuid(new, old); new 1056 security/commoncap.c if (uid_eq(old->fsuid, root_uid) && !uid_eq(new->fsuid, root_uid)) new 1057 security/commoncap.c new->cap_effective = new 1058 security/commoncap.c cap_drop_fs_set(new->cap_effective); new 1060 security/commoncap.c if (!uid_eq(old->fsuid, root_uid) && uid_eq(new->fsuid, root_uid)) new 1061 security/commoncap.c new->cap_effective = new 1062 security/commoncap.c cap_raise_fs_set(new->cap_effective, new 1063 security/commoncap.c new->cap_permitted); new 1142 security/commoncap.c struct cred *new; new 1149 security/commoncap.c new = prepare_creds(); new 1150 security/commoncap.c if (!new) new 1152 security/commoncap.c cap_lower(new->cap_bset, cap); new 1153 security/commoncap.c return commit_creds(new); new 1172 security/commoncap.c struct cred *new; new 1222 security/commoncap.c new = prepare_creds(); new 1223 security/commoncap.c if (!new) new 1225 security/commoncap.c new->securebits = arg2; new 1226 security/commoncap.c return commit_creds(new); new 1240 security/commoncap.c new = prepare_creds(); new 1241 security/commoncap.c if (!new) new 1244 security/commoncap.c new->securebits |= issecure_mask(SECURE_KEEP_CAPS); new 1246 security/commoncap.c new->securebits &= ~issecure_mask(SECURE_KEEP_CAPS); new 1247 security/commoncap.c return commit_creds(new); new 1254 security/commoncap.c new = prepare_creds(); new 1255 security/commoncap.c if (!new) new 1257 security/commoncap.c cap_clear(new->cap_ambient); new 1258 security/commoncap.c return commit_creds(new); new 1277 security/commoncap.c new = prepare_creds(); new 1278 security/commoncap.c if (!new) new 1281 security/commoncap.c cap_raise(new->cap_ambient, arg3); new 1283 security/commoncap.c cap_lower(new->cap_ambient, arg3); new 1284 security/commoncap.c return commit_creds(new); new 61 security/device_cgroup.c struct dev_exception_item *ex, *tmp, *new; new 66 security/device_cgroup.c new = kmemdup(ex, sizeof(*ex), GFP_KERNEL); new 67 security/device_cgroup.c if (!new) new 69 security/device_cgroup.c list_add_tail(&new->list, dest); new 1141 security/keys/keyctl.c struct cred *new; new 1143 security/keys/keyctl.c new = prepare_creds(); new 1144 security/keys/keyctl.c if (!new) new 1147 security/keys/keyctl.c key_put(new->request_key_auth); new 1148 security/keys/keyctl.c new->request_key_auth = key_get(key); new 1150 security/keys/keyctl.c return commit_creds(new); new 1385 security/keys/keyctl.c struct cred *new; new 1393 security/keys/keyctl.c new = prepare_creds(); new 1394 security/keys/keyctl.c if (!new) new 1399 security/keys/keyctl.c ret = install_thread_keyring_to_cred(new); new 1405 security/keys/keyctl.c ret = install_process_keyring_to_cred(new); new 1425 security/keys/keyctl.c new->jit_keyring = reqkey_defl; new 1426 security/keys/keyctl.c commit_creds(new); new 1429 security/keys/keyctl.c abort_creds(new); new 221 security/keys/process_keys.c int install_thread_keyring_to_cred(struct cred *new) new 225 security/keys/process_keys.c if (new->thread_keyring) new 228 security/keys/process_keys.c keyring = keyring_alloc("_tid", new->uid, new->gid, new, new 235 security/keys/process_keys.c new->thread_keyring = keyring; new 246 security/keys/process_keys.c struct cred *new; new 249 security/keys/process_keys.c new = prepare_creds(); new 250 security/keys/process_keys.c if (!new) new 253 security/keys/process_keys.c ret = install_thread_keyring_to_cred(new); new 255 security/keys/process_keys.c abort_creds(new); new 259 security/keys/process_keys.c return commit_creds(new); new 268 security/keys/process_keys.c int install_process_keyring_to_cred(struct cred *new) new 272 security/keys/process_keys.c if (new->process_keyring) new 275 security/keys/process_keys.c keyring = keyring_alloc("_pid", new->uid, new->gid, new, new 282 security/keys/process_keys.c new->process_keyring = keyring; new 293 security/keys/process_keys.c struct cred *new; new 296 security/keys/process_keys.c new = prepare_creds(); new 297 security/keys/process_keys.c if (!new) new 300 security/keys/process_keys.c ret = install_process_keyring_to_cred(new); new 302 security/keys/process_keys.c abort_creds(new); new 306 security/keys/process_keys.c return commit_creds(new); new 358 security/keys/process_keys.c struct cred *new; new 361 security/keys/process_keys.c new = prepare_creds(); new 362 security/keys/process_keys.c if (!new) new 365 security/keys/process_keys.c ret = install_session_keyring_to_cred(new, keyring); new 367 security/keys/process_keys.c abort_creds(new); new 371 security/keys/process_keys.c return commit_creds(new); new 841 security/keys/process_keys.c struct cred *new; new 845 security/keys/process_keys.c new = prepare_creds(); new 846 security/keys/process_keys.c if (!new) new 852 security/keys/process_keys.c ret = install_session_keyring_to_cred(new, NULL); new 856 security/keys/process_keys.c serial = new->session_keyring->serial; new 857 security/keys/process_keys.c ret = commit_creds(new); new 881 security/keys/process_keys.c } else if (keyring == new->session_keyring) { new 887 security/keys/process_keys.c ret = install_session_keyring_to_cred(new, keyring); new 891 security/keys/process_keys.c commit_creds(new); new 904 security/keys/process_keys.c abort_creds(new); new 915 security/keys/process_keys.c struct cred *new = container_of(twork, struct cred, rcu); new 918 security/keys/process_keys.c put_cred(new); new 922 security/keys/process_keys.c new-> uid = old-> uid; new 923 security/keys/process_keys.c new-> euid = old-> euid; new 924 security/keys/process_keys.c new-> suid = old-> suid; new 925 security/keys/process_keys.c new->fsuid = old->fsuid; new 926 security/keys/process_keys.c new-> gid = old-> gid; new 927 security/keys/process_keys.c new-> egid = old-> egid; new 928 security/keys/process_keys.c new-> sgid = old-> sgid; new 929 security/keys/process_keys.c new->fsgid = old->fsgid; new 930 security/keys/process_keys.c new->user = get_uid(old->user); new 931 security/keys/process_keys.c new->user_ns = get_user_ns(old->user_ns); new 932 security/keys/process_keys.c new->group_info = get_group_info(old->group_info); new 934 security/keys/process_keys.c new->securebits = old->securebits; new 935 security/keys/process_keys.c new->cap_inheritable = old->cap_inheritable; new 936 security/keys/process_keys.c new->cap_permitted = old->cap_permitted; new 937 security/keys/process_keys.c new->cap_effective = old->cap_effective; new 938 security/keys/process_keys.c new->cap_ambient = old->cap_ambient; new 939 security/keys/process_keys.c new->cap_bset = old->cap_bset; new 941 security/keys/process_keys.c new->jit_keyring = old->jit_keyring; new 942 security/keys/process_keys.c new->thread_keyring = key_get(old->thread_keyring); new 943 security/keys/process_keys.c new->process_keyring = key_get(old->process_keyring); new 945 security/keys/process_keys.c security_transfer_creds(new, old); new 947 security/keys/process_keys.c commit_creds(new); new 128 security/safesetid/lsm.c static int safesetid_task_fix_setuid(struct cred *new, new 137 security/safesetid/lsm.c if (uid_permitted_for_cred(old, new->uid) && new 138 security/safesetid/lsm.c uid_permitted_for_cred(old, new->euid) && new 139 security/safesetid/lsm.c uid_permitted_for_cred(old, new->suid) && new 140 security/safesetid/lsm.c uid_permitted_for_cred(old, new->fsuid)) new 281 security/security.c static int lsm_append(const char *new, char **result); new 414 security/security.c static int lsm_append(const char *new, char **result) new 419 security/security.c *result = kstrdup(new, GFP_KERNEL); new 424 security/security.c if (match_last_lsm(*result, new)) new 426 security/security.c cp = kasprintf(GFP_KERNEL, "%s,%s", *result, new); new 715 security/security.c int security_capset(struct cred *new, const struct cred *old, new 720 security/security.c return call_int_hook(capset, 0, new, old, new 962 security/security.c const struct cred *old, struct cred *new) new 965 security/security.c name, old, new); new 1353 security/security.c int security_inode_copy_up(struct dentry *src, struct cred **new) new 1355 security/security.c return call_int_hook(inode_copy_up, 0, src, new); new 1552 security/security.c int security_prepare_creds(struct cred *new, const struct cred *old, gfp_t gfp) new 1554 security/security.c int rc = lsm_cred_alloc(new, gfp); new 1559 security/security.c rc = call_int_hook(cred_prepare, 0, new, old, gfp); new 1561 security/security.c security_cred_free(new); new 1565 security/security.c void security_transfer_creds(struct cred *new, const struct cred *old) new 1567 security/security.c call_void_hook(cred_transfer, new, old); new 1577 security/security.c int security_kernel_act_as(struct cred *new, u32 secid) new 1579 security/security.c return call_int_hook(kernel_act_as, 0, new, secid); new 1582 security/security.c int security_kernel_create_files_as(struct cred *new, struct inode *inode) new 1584 security/security.c return call_int_hook(kernel_create_files_as, 0, new, inode); new 1631 security/security.c int security_task_fix_setuid(struct cred *new, const struct cred *old, new 1634 security/security.c return call_int_hook(task_fix_setuid, 0, new, old, flags); new 454 security/selinux/avc.c struct avc_node *new, struct avc_node *old) new 456 security/selinux/avc.c hlist_replace_rcu(&old->list, &new->list); new 888 security/selinux/hooks.c struct superblock_security_struct *new = newsb->s_security; new 890 security/selinux/hooks.c char newflags = new->flags & SE_MNTMASK; new 894 security/selinux/hooks.c if ((oldflags & FSCONTEXT_MNT) && old->sid != new->sid) new 896 security/selinux/hooks.c if ((oldflags & CONTEXT_MNT) && old->mntpoint_sid != new->mntpoint_sid) new 898 security/selinux/hooks.c if ((oldflags & DEFCONTEXT_MNT) && old->def_sid != new->def_sid) new 2155 security/selinux/hooks.c static int selinux_capset(struct cred *new, const struct cred *old, new 2161 security/selinux/hooks.c cred_sid(old), cred_sid(new), SECCLASS_PROCESS, new 2888 security/selinux/hooks.c struct cred *new) new 2901 security/selinux/hooks.c tsec = selinux_cred(new); new 3424 security/selinux/hooks.c static int selinux_inode_copy_up(struct dentry *src, struct cred **new) new 3428 security/selinux/hooks.c struct cred *new_creds = *new; new 3440 security/selinux/hooks.c *new = new_creds; new 3895 security/selinux/hooks.c static int selinux_cred_prepare(struct cred *new, const struct cred *old, new 3899 security/selinux/hooks.c struct task_security_struct *tsec = selinux_cred(new); new 3908 security/selinux/hooks.c static void selinux_cred_transfer(struct cred *new, const struct cred *old) new 3911 security/selinux/hooks.c struct task_security_struct *tsec = selinux_cred(new); new 3925 security/selinux/hooks.c static int selinux_kernel_act_as(struct cred *new, u32 secid) new 3927 security/selinux/hooks.c struct task_security_struct *tsec = selinux_cred(new); new 3949 security/selinux/hooks.c static int selinux_kernel_create_files_as(struct cred *new, struct inode *inode) new 3952 security/selinux/hooks.c struct task_security_struct *tsec = selinux_cred(new); new 6347 security/selinux/hooks.c struct cred *new; new 6416 security/selinux/hooks.c new = prepare_creds(); new 6417 security/selinux/hooks.c if (!new) new 6426 security/selinux/hooks.c tsec = selinux_cred(new); new 6479 security/selinux/hooks.c commit_creds(new); new 6483 security/selinux/hooks.c abort_creds(new); new 134 security/selinux/ibpkey.c struct sel_ib_pkey *new = NULL; new 153 security/selinux/ibpkey.c new = kzalloc(sizeof(*new), GFP_ATOMIC); new 154 security/selinux/ibpkey.c if (!new) new 157 security/selinux/ibpkey.c new->psec.subnet_prefix = subnet_prefix; new 158 security/selinux/ibpkey.c new->psec.pkey = pkey_num; new 159 security/selinux/ibpkey.c new->psec.sid = *sid; new 160 security/selinux/ibpkey.c sel_ib_pkey_insert(new); new 137 security/selinux/netif.c struct sel_netif *new; new 160 security/selinux/netif.c new = kzalloc(sizeof(*new), GFP_ATOMIC); new 161 security/selinux/netif.c if (new) { new 162 security/selinux/netif.c new->nsec.ns = ns; new 163 security/selinux/netif.c new->nsec.ifindex = ifindex; new 164 security/selinux/netif.c new->nsec.sid = *sid; new 165 security/selinux/netif.c if (sel_netif_insert(new)) new 166 security/selinux/netif.c kfree(new); new 194 security/selinux/netnode.c struct sel_netnode *new; new 204 security/selinux/netnode.c new = kzalloc(sizeof(*new), GFP_ATOMIC); new 209 security/selinux/netnode.c if (new) new 210 security/selinux/netnode.c new->nsec.addr.ipv4 = *(__be32 *)addr; new 215 security/selinux/netnode.c if (new) new 216 security/selinux/netnode.c new->nsec.addr.ipv6 = *(struct in6_addr *)addr; new 222 security/selinux/netnode.c if (ret == 0 && new) { new 223 security/selinux/netnode.c new->nsec.family = family; new 224 security/selinux/netnode.c new->nsec.sid = *sid; new 225 security/selinux/netnode.c sel_netnode_insert(new); new 227 security/selinux/netnode.c kfree(new); new 142 security/selinux/netport.c struct sel_netport *new; new 155 security/selinux/netport.c new = kzalloc(sizeof(*new), GFP_ATOMIC); new 156 security/selinux/netport.c if (new) { new 157 security/selinux/netport.c new->psec.port = pnum; new 158 security/selinux/netport.c new->psec.protocol = protocol; new 159 security/selinux/netport.c new->psec.sid = *sid; new 160 security/selinux/netport.c sel_netport_insert(new); new 54 security/selinux/ss/ebitmap.c struct ebitmap_node *n, *new, *prev; new 60 security/selinux/ss/ebitmap.c new = kmem_cache_zalloc(ebitmap_node_cachep, GFP_ATOMIC); new 61 security/selinux/ss/ebitmap.c if (!new) { new 65 security/selinux/ss/ebitmap.c new->startbit = n->startbit; new 66 security/selinux/ss/ebitmap.c memcpy(new->maps, n->maps, EBITMAP_SIZE / 8); new 67 security/selinux/ss/ebitmap.c new->next = NULL; new 69 security/selinux/ss/ebitmap.c prev->next = new; new 71 security/selinux/ss/ebitmap.c dst->node = new; new 72 security/selinux/ss/ebitmap.c prev = new; new 261 security/selinux/ss/ebitmap.c struct ebitmap_node *n, *prev, *new; new 305 security/selinux/ss/ebitmap.c new = kmem_cache_zalloc(ebitmap_node_cachep, GFP_ATOMIC); new 306 security/selinux/ss/ebitmap.c if (!new) new 309 security/selinux/ss/ebitmap.c new->startbit = bit - (bit % EBITMAP_SIZE); new 310 security/selinux/ss/ebitmap.c ebitmap_node_set_bit(new, bit); new 314 security/selinux/ss/ebitmap.c e->highbit = new->startbit + EBITMAP_SIZE; new 317 security/selinux/ss/ebitmap.c new->next = prev->next; new 318 security/selinux/ss/ebitmap.c prev->next = new; new 320 security/selinux/ss/ebitmap.c new->next = e->node; new 321 security/selinux/ss/ebitmap.c e->node = new; new 1943 security/smack/smack_lsm.c static int smack_cred_prepare(struct cred *new, const struct cred *old, new 1947 security/smack/smack_lsm.c struct task_smack *new_tsp = smack_cred(new); new 1968 security/smack/smack_lsm.c static void smack_cred_transfer(struct cred *new, const struct cred *old) new 1971 security/smack/smack_lsm.c struct task_smack *new_tsp = smack_cred(new); new 2005 security/smack/smack_lsm.c static int smack_kernel_act_as(struct cred *new, u32 secid) new 2007 security/smack/smack_lsm.c struct task_smack *new_tsp = smack_cred(new); new 2021 security/smack/smack_lsm.c static int smack_kernel_create_files_as(struct cred *new, new 2025 security/smack/smack_lsm.c struct task_smack *tsp = smack_cred(new); new 3512 security/smack/smack_lsm.c struct cred *new; new 3548 security/smack/smack_lsm.c new = prepare_creds(); new 3549 security/smack/smack_lsm.c if (new == NULL) new 3552 security/smack/smack_lsm.c tsp = smack_cred(new); new 3559 security/smack/smack_lsm.c commit_creds(new); new 4499 security/smack/smack_lsm.c static int smack_inode_copy_up(struct dentry *dentry, struct cred **new) new 4505 security/smack/smack_lsm.c struct cred *new_creds = *new; new 4521 security/smack/smack_lsm.c *new = new_creds; new 4539 security/smack/smack_lsm.c struct cred *new) new 4542 security/smack/smack_lsm.c struct task_smack *ntsp = smack_cred(new); new 1088 security/smack/smackfs.c static void smk_net4addr_insert(struct smk_net4addr *new) new 1094 security/smack/smackfs.c list_add_rcu(&new->list, &smk_net4addr_list); new 1102 security/smack/smackfs.c if (new->smk_masks > m->smk_masks) { new 1103 security/smack/smackfs.c list_add_rcu(&new->list, &smk_net4addr_list); new 1109 security/smack/smackfs.c list_add_rcu(&new->list, &m->list); new 1114 security/smack/smackfs.c if (new->smk_masks > m_next->smk_masks) { new 1115 security/smack/smackfs.c list_add_rcu(&new->list, &m->list); new 1352 security/smack/smackfs.c static void smk_net6addr_insert(struct smk_net6addr *new) new 1358 security/smack/smackfs.c list_add_rcu(&new->list, &smk_net6addr_list); new 1365 security/smack/smackfs.c if (new->smk_masks > m->smk_masks) { new 1366 security/smack/smackfs.c list_add_rcu(&new->list, &smk_net6addr_list); new 1372 security/smack/smackfs.c list_add_rcu(&new->list, &m->list); new 1377 security/smack/smackfs.c if (new->smk_masks > m_next->smk_masks) { new 1378 security/smack/smackfs.c list_add_rcu(&new->list, &m->list); new 36 security/tomoyo/tomoyo.c static int tomoyo_cred_prepare(struct cred *new, const struct cred *old, new 505 security/tomoyo/tomoyo.c struct tomoyo_task *new = tomoyo_task(task); new 507 security/tomoyo/tomoyo.c new->domain_info = old->domain_info; new 508 security/tomoyo/tomoyo.c atomic_inc(&new->domain_info->users); new 509 security/tomoyo/tomoyo.c new->old_domain_info = NULL; new 1120 sound/core/pcm_lib.c struct snd_pcm_hw_rule *new; new 1122 sound/core/pcm_lib.c new = krealloc(constrs->rules, new_rules * sizeof(*c), new 1124 sound/core/pcm_lib.c if (!new) { new 1128 sound/core/pcm_lib.c constrs->rules = new; new 234 sound/isa/es18xx.c unsigned char old, new, oval; new 253 sound/isa/es18xx.c new = (old & ~mask) | (val & mask); new 254 sound/isa/es18xx.c ret = snd_es18xx_dsp_command(chip, new); new 259 sound/isa/es18xx.c reg, old, new, ret); new 299 sound/isa/es18xx.c unsigned char old, new, oval; new 306 sound/isa/es18xx.c new = (old & ~mask) | (val & mask); new 307 sound/isa/es18xx.c outb(new, chip->port + 0x05); new 310 sound/isa/es18xx.c reg, old, new); new 320 sound/isa/es18xx.c int old, expected, new; new 327 sound/isa/es18xx.c new = inb(chip->port + 0x05); new 331 sound/isa/es18xx.c reg, old, expected, new); new 333 sound/isa/es18xx.c return expected == new; new 222 sound/mips/hal2.c u32 old, new; new 231 sound/mips/hal2.c new = old & ~(H2I_C2_L_ATT_M | H2I_C2_R_ATT_M | H2I_C2_MUTE); new 235 sound/mips/hal2.c new |= (l << H2I_C2_L_ATT_SHIFT); new 236 sound/mips/hal2.c new |= (r << H2I_C2_R_ATT_SHIFT); new 238 sound/mips/hal2.c new |= H2I_C2_L_ATT_M | H2I_C2_R_ATT_M | H2I_C2_MUTE; new 239 sound/mips/hal2.c hal2_i_write32(hal2, H2I_DAC_C2, new); new 243 sound/mips/hal2.c new = old & ~(H2I_C2_L_GAIN_M | H2I_C2_R_GAIN_M); new 244 sound/mips/hal2.c new |= (l << H2I_C2_L_GAIN_SHIFT); new 245 sound/mips/hal2.c new |= (r << H2I_C2_R_GAIN_SHIFT); new 246 sound/mips/hal2.c hal2_i_write32(hal2, H2I_ADC_C2, new); new 251 sound/mips/hal2.c return old != new; new 402 sound/pci/ac97/ac97_codec.c unsigned short old, new; new 405 sound/pci/ac97/ac97_codec.c new = (old & ~mask) | (value & mask); new 406 sound/pci/ac97/ac97_codec.c change = old != new; new 408 sound/pci/ac97/ac97_codec.c ac97->regs[reg] = new; new 409 sound/pci/ac97/ac97_codec.c ac97->bus->ops->write(ac97, reg, new); new 418 sound/pci/ac97/ac97_codec.c unsigned short old, new, cfg; new 422 sound/pci/ac97/ac97_codec.c new = (old & ~mask) | (value & mask); new 423 sound/pci/ac97/ac97_codec.c change = old != new; new 427 sound/pci/ac97/ac97_codec.c ac97->spec.ad18xx.pcmreg[codec] = new; new 433 sound/pci/ac97/ac97_codec.c ac97->bus->ops->write(ac97, AC97_PCM, new); new 732 sound/pci/ac97/ac97_codec.c unsigned int new = 0; new 736 sound/pci/ac97/ac97_codec.c new = val = ucontrol->value.iec958.status[0] & (IEC958_AES0_PROFESSIONAL|IEC958_AES0_NONAUDIO); new 738 sound/pci/ac97/ac97_codec.c new |= ucontrol->value.iec958.status[0] & (IEC958_AES0_PRO_FS|IEC958_AES0_PRO_EMPHASIS_5015); new 739 sound/pci/ac97/ac97_codec.c switch (new & IEC958_AES0_PRO_FS) { new 745 sound/pci/ac97/ac97_codec.c if ((new & IEC958_AES0_PRO_EMPHASIS) == IEC958_AES0_PRO_EMPHASIS_5015) new 748 sound/pci/ac97/ac97_codec.c new |= ucontrol->value.iec958.status[0] & (IEC958_AES0_CON_EMPHASIS_5015|IEC958_AES0_CON_NOT_COPYRIGHT); new 749 sound/pci/ac97/ac97_codec.c new |= ((ucontrol->value.iec958.status[1] & (IEC958_AES1_CON_CATEGORY|IEC958_AES1_CON_ORIGINAL)) << 8); new 750 sound/pci/ac97/ac97_codec.c new |= ((ucontrol->value.iec958.status[3] & IEC958_AES3_CON_FS) << 24); new 751 sound/pci/ac97/ac97_codec.c if ((new & IEC958_AES0_CON_EMPHASIS) == IEC958_AES0_CON_EMPHASIS_5015) new 753 sound/pci/ac97/ac97_codec.c if (!(new & IEC958_AES0_CON_NOT_COPYRIGHT)) new 755 sound/pci/ac97/ac97_codec.c val |= ((new >> 8) & 0xff) << 4; // category + original new 756 sound/pci/ac97/ac97_codec.c switch ((new >> 24) & 0xff) { new 765 sound/pci/ac97/ac97_codec.c change = ac97->spdif_status != new; new 766 sound/pci/ac97/ac97_codec.c ac97->spdif_status = new; new 778 sound/pci/ac97/ac97_codec.c v = new & (IEC958_AES0_CON_EMPHASIS_5015|IEC958_AES0_CON_NOT_COPYRIGHT) ? 0 : AC97_CXR_COPYRGT; new 779 sound/pci/ac97/ac97_codec.c v |= new & IEC958_AES0_NONAUDIO ? AC97_CXR_SPDIF_AC3 : AC97_CXR_SPDIF_PCM; new 810 sound/pci/ac97/ac97_codec.c unsigned short value, old, new; new 819 sound/pci/ac97/ac97_codec.c new = (old & ~mask) | value; new 820 sound/pci/ac97/ac97_codec.c change = old != new; new 325 sound/pci/azt3328.c u8 prev = inb(reg), new; new 327 sound/pci/azt3328.c new = (do_set) ? (prev|mask) : (prev & ~mask); new 330 sound/pci/azt3328.c outb(new, reg); new 331 sound/pci/azt3328.c if (new != prev) new 706 sound/pci/ctxfi/ctmixer.c ct_mixer_kcontrol_new(struct ct_mixer *mixer, struct snd_kcontrol_new *new) new 711 sound/pci/ctxfi/ctmixer.c kctl = snd_ctl_new1(new, mixer->atc); new 722 sound/pci/ctxfi/ctmixer.c switch (new->private_value) { new 276 sound/pci/es1938.c unsigned char old, new, oval; new 282 sound/pci/es1938.c new = (old & ~mask) | (val & mask); new 283 sound/pci/es1938.c outb(new, SLSB_REG(chip, MIXERDATA)); new 286 sound/pci/es1938.c reg, old, new); new 359 sound/pci/es1938.c unsigned char old, new, oval; new 367 sound/pci/es1938.c new = (old & ~mask) | (val & mask); new 368 sound/pci/es1938.c snd_es1938_write_cmd(chip, new); new 370 sound/pci/es1938.c reg, old, new); new 760 sound/pci/es1938.c size_t old, new; new 763 sound/pci/es1938.c while ((new = inw(SLDM_REG(chip, DMACOUNT))) != old) new 764 sound/pci/es1938.c old = new; new 765 sound/pci/es1938.c ptr = chip->dma1_size - 1 - new; new 801 sound/pci/es1938.c size_t old, new; new 805 sound/pci/es1938.c while ((new = inw(SLDM_REG(chip, DMACOUNT))) != old) new 806 sound/pci/es1938.c old = new; new 807 sound/pci/es1938.c ptr = chip->dma1_size - 1 - new; new 260 sound/pci/fm801.c unsigned short old, new; new 264 sound/pci/fm801.c new = (old & ~mask) | value; new 265 sound/pci/fm801.c change = old != new; new 267 sound/pci/fm801.c fm801_iowrite16(chip, reg, new); new 999 sound/pci/ice1712/aureon.c unsigned short new, old; new 1005 sound/pci/ice1712/aureon.c new = (~ucontrol->value.integer.value[i]<<5&0x20) | (old&~0x20); new 1006 sound/pci/ice1712/aureon.c if (new != old) { new 1007 sound/pci/ice1712/aureon.c wm_put(ice, WM_ADC_GAIN + i, new); new 518 sound/pci/ice1712/juli.c unsigned int old, new; new 522 sound/pci/ice1712/juli.c new = (old & ~GPIO_RATE_MASK) | get_gpio_val(rate); new 527 sound/pci/ice1712/juli.c ice->gpio.set_data(ice, new); new 81 sound/pci/ice1712/prodigy192.c unsigned char new, old; new 84 sound/pci/ice1712/prodigy192.c new = (~mute << 7 & 0x80) | (old & ~0x80); new 85 sound/pci/ice1712/prodigy192.c change = (new != old); new 88 sound/pci/ice1712/prodigy192.c stac9460_put(ice, idx, new); new 206 sound/pci/ice1712/prodigy192.c unsigned char new, old; new 213 sound/pci/ice1712/prodigy192.c new = (~ucontrol->value.integer.value[i]<<7&0x80) | (old&~0x80); new 214 sound/pci/ice1712/prodigy192.c change = (new != old); new 216 sound/pci/ice1712/prodigy192.c stac9460_put(ice, reg, new); new 292 sound/pci/ice1712/prodigy192.c unsigned char new, old; new 295 sound/pci/ice1712/prodigy192.c new = (ucontrol->value.enumerated.item[0] << 7 & 0x80) | (old & ~0x80); new 296 sound/pci/ice1712/prodigy192.c change = (new != old); new 298 sound/pci/ice1712/prodigy192.c stac9460_put(ice, STAC946X_GENERAL_PURPOSE, new); new 306 sound/pci/ice1712/prodigy192.c unsigned char old, new; new 314 sound/pci/ice1712/prodigy192.c new = 0x08; /* 256x, base rate mode */ new 316 sound/pci/ice1712/prodigy192.c new = 0x11; /* 256x, mid rate mode */ new 318 sound/pci/ice1712/prodigy192.c new = 0x12; /* 128x, high rate mode */ new 320 sound/pci/ice1712/prodigy192.c if (old == new) new 330 sound/pci/ice1712/prodigy192.c stac9460_put(ice, STAC946X_MASTER_CLOCKING, new); new 566 sound/pci/ice1712/prodigy192.c unsigned char new, old, itemvalue; new 573 sound/pci/ice1712/prodigy192.c new = (itemvalue & AK4114_IPS0) | (old & ~AK4114_IPS0); new 574 sound/pci/ice1712/prodigy192.c change = (new != old); new 576 sound/pci/ice1712/prodigy192.c prodigy192_ak4114_write(ice, AK4114_REG_IO1, new); new 507 sound/pci/ice1712/quartet.c unsigned int old, new, smute; new 511 sound/pci/ice1712/quartet.c new = 0; new 516 sound/pci/ice1712/quartet.c new = SCR_MUTE; new 520 sound/pci/ice1712/quartet.c if (old != new) { new 522 sound/pci/ice1712/quartet.c set_scr(ice, (get_scr(ice) & ~SCR_MUTE) | new); new 568 sound/pci/ice1712/quartet.c unsigned int old, new, tmp, masked_old; new 569 sound/pci/ice1712/quartet.c old = new = get_scr(ice); new 579 sound/pci/ice1712/quartet.c new = old & ~(SCR_AIN12_SEL1 | SCR_AIN12_SEL0); new 580 sound/pci/ice1712/quartet.c set_scr(ice, new); new 582 sound/pci/ice1712/quartet.c new &= ~SCR_RELAY; new 583 sound/pci/ice1712/quartet.c set_scr(ice, new); new 587 sound/pci/ice1712/quartet.c new = old | SCR_RELAY; new 588 sound/pci/ice1712/quartet.c set_scr(ice, new); new 589 sound/pci/ice1712/quartet.c new = (new & ~SCR_AIN12_SEL1) | SCR_AIN12_SEL0; new 590 sound/pci/ice1712/quartet.c set_scr(ice, new); new 594 sound/pci/ice1712/quartet.c new = old | SCR_RELAY; new 595 sound/pci/ice1712/quartet.c set_scr(ice, new); new 596 sound/pci/ice1712/quartet.c new |= SCR_AIN12_SEL1 | SCR_AIN12_SEL0; new 597 sound/pci/ice1712/quartet.c set_scr(ice, new); new 623 sound/pci/ice1712/quartet.c unsigned int old, new; new 624 sound/pci/ice1712/quartet.c old = new = get_scr(ice); new 629 sound/pci/ice1712/quartet.c new = old | SCR_PHP_V; new 630 sound/pci/ice1712/quartet.c set_scr(ice, new); new 632 sound/pci/ice1712/quartet.c new &= ~SCR_PHP; new 633 sound/pci/ice1712/quartet.c set_scr(ice, new); new 638 sound/pci/ice1712/quartet.c new = old & ~SCR_PHP_V; new 639 sound/pci/ice1712/quartet.c set_scr(ice, new); new 641 sound/pci/ice1712/quartet.c new |= SCR_PHP; new 642 sound/pci/ice1712/quartet.c set_scr(ice, new); new 644 sound/pci/ice1712/quartet.c if (old != new) new 699 sound/pci/ice1712/quartet.c unsigned int old, new; new 702 sound/pci/ice1712/quartet.c new = old | private.bit; new 704 sound/pci/ice1712/quartet.c new = old & ~private.bit; new 705 sound/pci/ice1712/quartet.c if (old != new) { new 706 sound/pci/ice1712/quartet.c private.set_register(ice, new); new 848 sound/pci/ice1712/quartet.c unsigned int new; new 854 sound/pci/ice1712/quartet.c new = (get_cpld(ice) & ~CPLD_CKS_MASK) | get_cks_val(rate); new 856 sound/pci/ice1712/quartet.c new &= ~CPLD_SYNC_SEL; new 859 sound/pci/ice1712/quartet.c set_cpld(ice, new); new 872 sound/pci/ice1712/quartet.c unsigned int old, new; new 874 sound/pci/ice1712/quartet.c old = new = get_cpld(ice); new 875 sound/pci/ice1712/quartet.c new &= ~(CPLD_CKS_MASK | CPLD_WORD_SEL); new 878 sound/pci/ice1712/quartet.c new |= CPLD_EXT_SPDIF; new 881 sound/pci/ice1712/quartet.c new |= CPLD_EXT_WORDCLOCK_1FS; new 884 sound/pci/ice1712/quartet.c new |= CPLD_EXT_WORDCLOCK_256FS; new 889 sound/pci/ice1712/quartet.c if (old != new) { new 890 sound/pci/ice1712/quartet.c set_cpld(ice, new); new 67 sound/pci/ice1712/wtm.c unsigned char new, old; new 78 sound/pci/ice1712/wtm.c new = (~mute << 7 & 0x80) | (old & ~0x80); new 79 sound/pci/ice1712/wtm.c change = (new != old); new 81 sound/pci/ice1712/wtm.c stac9460_put(ice, idx, new); new 97 sound/pci/ice1712/wtm.c new = (~mute << 7 & 0x80) | (old & ~0x80); new 98 sound/pci/ice1712/wtm.c change = (new != old); new 100 sound/pci/ice1712/wtm.c stac9460_2_put(ice, idx, new); new 144 sound/pci/ice1712/wtm.c unsigned char new, old; new 151 sound/pci/ice1712/wtm.c new = (~ucontrol->value.integer.value[0] << 7 & 0x80) | new 153 sound/pci/ice1712/wtm.c change = (new != old); new 155 sound/pci/ice1712/wtm.c stac9460_put(ice, idx, new); new 156 sound/pci/ice1712/wtm.c stac9460_2_put(ice, idx, new); new 165 sound/pci/ice1712/wtm.c new = (~ucontrol->value.integer.value[0] << 7 & 0x80) | new 167 sound/pci/ice1712/wtm.c change = (new != old); new 170 sound/pci/ice1712/wtm.c stac9460_put(ice, idx, new); new 172 sound/pci/ice1712/wtm.c stac9460_2_put(ice, idx - 6, new); new 284 sound/pci/ice1712/wtm.c unsigned char new, old; new 293 sound/pci/ice1712/wtm.c new = (~ucontrol->value.integer.value[i]<<7&0x80) | new 295 sound/pci/ice1712/wtm.c change = (new != old); new 297 sound/pci/ice1712/wtm.c stac9460_put(ice, reg, new); new 303 sound/pci/ice1712/wtm.c new = (~ucontrol->value.integer.value[i]<<7&0x80) | new 305 sound/pci/ice1712/wtm.c change = (new != old); new 307 sound/pci/ice1712/wtm.c stac9460_2_put(ice, reg, new); new 415 sound/pci/ice1712/wtm.c unsigned char new, old; new 423 sound/pci/ice1712/wtm.c new = (ucontrol->value.enumerated.item[0] << 7 & 0x80) | (old & ~0x80); new 424 sound/pci/ice1712/wtm.c change = (new != old); new 427 sound/pci/ice1712/wtm.c stac9460_put(ice, STAC946X_GENERAL_PURPOSE, new); new 429 sound/pci/ice1712/wtm.c stac9460_2_put(ice, STAC946X_GENERAL_PURPOSE, new); new 440 sound/pci/ice1712/wtm.c unsigned char old, new; new 447 sound/pci/ice1712/wtm.c new = 0x08; /* 256x, base rate mode */ new 449 sound/pci/ice1712/wtm.c new = 0x11; /* 256x, mid rate mode */ new 451 sound/pci/ice1712/wtm.c new = 0x12; /* 128x, high rate mode */ new 454 sound/pci/ice1712/wtm.c if (old == new) new 463 sound/pci/ice1712/wtm.c stac9460_put(ice, STAC946X_MASTER_CLOCKING, new); new 464 sound/pci/ice1712/wtm.c stac9460_2_put(ice, STAC946X_MASTER_CLOCKING, new); new 76 sound/pci/oxygen/xonar_dg_mixer.c unsigned int new = value->value.enumerated.item[0]; new 81 sound/pci/oxygen/xonar_dg_mixer.c if (data->output_sel != new) { new 82 sound/pci/oxygen/xonar_dg_mixer.c data->output_sel = new; new 963 sound/soc/codecs/madera.c unsigned int cur, new; new 984 sound/soc/codecs/madera.c new = priv->adsp_rate_cache[dsp->num - 1] << MADERA_DSP_RATE_SHIFT; new 986 sound/soc/codecs/madera.c if (new == cur) { new 213 sound/soc/codecs/wm8994.c int change, new; new 232 sound/soc/codecs/wm8994.c new = WM8994_SYSCLK_SRC; new 234 sound/soc/codecs/wm8994.c new = 0; new 237 sound/soc/codecs/wm8994.c WM8994_SYSCLK_SRC, new); new 723 sound/soc/codecs/wm8995.c int change, new; new 743 sound/soc/codecs/wm8995.c new = WM8995_SYSCLK_SRC; new 745 sound/soc/codecs/wm8995.c new = 0; new 748 sound/soc/codecs/wm8995.c WM8995_SYSCLK_SRC_MASK, new); new 157 sound/soc/intel/common/sst-dsp.c unsigned int old, new; new 163 sound/soc/intel/common/sst-dsp.c new = (old & (~mask)) | (value & mask); new 165 sound/soc/intel/common/sst-dsp.c change = (old != new); new 167 sound/soc/intel/common/sst-dsp.c sst_dsp_shim_write_unlocked(sst, offset, new); new 177 sound/soc/intel/common/sst-dsp.c u64 old, new; new 181 sound/soc/intel/common/sst-dsp.c new = (old & (~mask)) | (value & mask); new 183 sound/soc/intel/common/sst-dsp.c change = (old != new); new 185 sound/soc/intel/common/sst-dsp.c sst_dsp_shim_write64_unlocked(sst, offset, new); new 195 sound/soc/intel/common/sst-dsp.c unsigned int old, new; new 201 sound/soc/intel/common/sst-dsp.c new = (old & (~mask)) | (value & mask); new 203 sound/soc/intel/common/sst-dsp.c sst_dsp_shim_write_unlocked(sst, offset, new); new 3212 sound/soc/soc-dapm.c if (w->new) new 3256 sound/soc/soc-dapm.c w->new = 1; new 79 sound/soc/soc-io.c unsigned int old, new; new 88 sound/soc/soc-io.c new = (old & ~mask) | (val & mask); new 89 sound/soc/soc-io.c *change = old != new; new 91 sound/soc/soc-io.c ret = snd_soc_component_write(component, reg, new); new 193 sound/soc/soc-io.c unsigned int old, new; new 199 sound/soc/soc-io.c new = (old & ~mask) | value; new 200 sound/soc/soc-io.c return old != new; new 1435 sound/soc/soc-pcm.c int i, new = 0, err; new 1480 sound/soc/soc-pcm.c new++; new 1483 sound/soc/soc-pcm.c dev_dbg(fe->dev, "ASoC: found %d new BE paths\n", new); new 1484 sound/soc/soc-pcm.c return new; new 1492 sound/soc/soc-pcm.c int stream, struct snd_soc_dapm_widget_list **list, int new) new 1494 sound/soc/soc-pcm.c if (new) new 2669 sound/soc/soc-pcm.c static int soc_dpcm_fe_runtime_update(struct snd_soc_pcm_runtime *fe, int new) new 2683 sound/soc/soc-pcm.c new ? "new" : "old", fe->dai_link->name); new 2702 sound/soc/soc-pcm.c count = dpcm_process_paths(fe, SNDRV_PCM_STREAM_PLAYBACK, &list, new); new 2704 sound/soc/soc-pcm.c if (new) new 2733 sound/soc/soc-pcm.c count = dpcm_process_paths(fe, SNDRV_PCM_STREAM_CAPTURE, &list, new); new 2735 sound/soc/soc-pcm.c if (new) new 19 sound/soc/sof/ops.c unsigned int old, new; new 26 sound/soc/sof/ops.c new = (old & ~mask) | (value & mask); new 28 sound/soc/sof/ops.c if (old == new) new 31 sound/soc/sof/ops.c pci_write_config_dword(pci, offset, new); new 54 sound/soc/sof/ops.c unsigned int old, new; new 60 sound/soc/sof/ops.c new = (old & ~mask) | (value & mask); new 62 sound/soc/sof/ops.c if (old == new) new 65 sound/soc/sof/ops.c snd_sof_dsp_write(sdev, bar, offset, new); new 74 sound/soc/sof/ops.c u64 old, new; new 78 sound/soc/sof/ops.c new = (old & ~mask) | (value & mask); new 80 sound/soc/sof/ops.c if (old == new) new 83 sound/soc/sof/ops.c snd_sof_dsp_write64(sdev, bar, offset, new); new 122 sound/soc/sof/ops.c unsigned int old, new; new 128 sound/soc/sof/ops.c new = (old & ~mask) | (value & mask); new 130 sound/soc/sof/ops.c snd_sof_dsp_write(sdev, bar, offset, new); new 68 tools/arch/x86/include/asm/atomic.h static __always_inline int atomic_cmpxchg(atomic_t *v, int old, int new) new 70 tools/arch/x86/include/asm/atomic.h return cmpxchg(&v->counter, old, new); new 35 tools/arch/x86/include/asm/cmpxchg.h #define __raw_cmpxchg(ptr, old, new, size, lock) \ new 39 tools/arch/x86/include/asm/cmpxchg.h __typeof__(*(ptr)) __new = (new); \ new 83 tools/arch/x86/include/asm/cmpxchg.h #define __cmpxchg(ptr, old, new, size) \ new 84 tools/arch/x86/include/asm/cmpxchg.h __raw_cmpxchg((ptr), (old), (new), (size), LOCK_PREFIX) new 86 tools/arch/x86/include/asm/cmpxchg.h #define cmpxchg(ptr, old, new) \ new 87 tools/arch/x86/include/asm/cmpxchg.h __cmpxchg(ptr, old, new, sizeof(*(ptr))) new 8 tools/build/feature/test-sync-compare-and-swap.c uint64_t old, new = argc; new 13 tools/build/feature/test-sync-compare-and-swap.c } while (!__sync_bool_compare_and_swap(&x, old, new)); new 14 tools/build/feature/test-sync-compare-and-swap.c return old == new; new 38 tools/include/linux/list.h static inline void __list_add(struct list_head *new, new 42 tools/include/linux/list.h next->prev = new; new 43 tools/include/linux/list.h new->next = next; new 44 tools/include/linux/list.h new->prev = prev; new 45 tools/include/linux/list.h prev->next = new; new 48 tools/include/linux/list.h extern void __list_add(struct list_head *new, new 61 tools/include/linux/list.h static inline void list_add(struct list_head *new, struct list_head *head) new 63 tools/include/linux/list.h __list_add(new, head, head->next); new 75 tools/include/linux/list.h static inline void list_add_tail(struct list_head *new, struct list_head *head) new 77 tools/include/linux/list.h __list_add(new, head->prev, head); new 124 tools/include/linux/list.h struct list_head *new) new 126 tools/include/linux/list.h new->next = old->next; new 127 tools/include/linux/list.h new->next->prev = new; new 128 tools/include/linux/list.h new->prev = old->prev; new 129 tools/include/linux/list.h new->prev->next = new; new 133 tools/include/linux/list.h struct list_head *new) new 135 tools/include/linux/list.h list_replace(old, new); new 686 tools/include/linux/list.h struct hlist_head *new) new 688 tools/include/linux/list.h new->first = old->first; new 689 tools/include/linux/list.h if (new->first) new 690 tools/include/linux/list.h new->first->pprev = &new->first; new 63 tools/include/linux/rbtree.h extern void rb_replace_node(struct rb_node *victim, struct rb_node *new, new 147 tools/include/linux/rbtree.h struct rb_node *new, new 151 tools/include/linux/rbtree.h root->rb_leftmost = new; new 152 tools/include/linux/rbtree.h rb_replace_node(victim, new, &root->rb_root); new 31 tools/include/linux/rbtree_augmented.h void (*copy)(struct rb_node *old, struct rb_node *new); new 32 tools/include/linux/rbtree_augmented.h void (*rotate)(struct rb_node *old, struct rb_node *new); new 36 tools/include/linux/rbtree_augmented.h void (*augment_rotate)(struct rb_node *old, struct rb_node *new)); new 92 tools/include/linux/rbtree_augmented.h RBSTRUCT *new = rb_entry(rb_new, RBSTRUCT, RBFIELD); \ new 93 tools/include/linux/rbtree_augmented.h new->RBAUGMENTED = old->RBAUGMENTED; \ new 99 tools/include/linux/rbtree_augmented.h RBSTRUCT *new = rb_entry(rb_new, RBSTRUCT, RBFIELD); \ new 100 tools/include/linux/rbtree_augmented.h new->RBAUGMENTED = old->RBAUGMENTED; \ new 171 tools/include/linux/rbtree_augmented.h __rb_change_child(struct rb_node *old, struct rb_node *new, new 176 tools/include/linux/rbtree_augmented.h WRITE_ONCE(parent->rb_left, new); new 178 tools/include/linux/rbtree_augmented.h WRITE_ONCE(parent->rb_right, new); new 180 tools/include/linux/rbtree_augmented.h WRITE_ONCE(root->rb_node, new); new 184 tools/include/linux/rbtree_augmented.h void (*augment_rotate)(struct rb_node *old, struct rb_node *new)); new 78 tools/include/linux/refcount.h unsigned int old, new, val = atomic_read(&r->refs); new 81 tools/include/linux/refcount.h new = val + 1; new 86 tools/include/linux/refcount.h if (unlikely(!new)) new 89 tools/include/linux/refcount.h old = atomic_cmpxchg_relaxed(&r->refs, val, new); new 96 tools/include/linux/refcount.h REFCOUNT_WARN(new == UINT_MAX, "refcount_t: saturated; leaking memory.\n"); new 123 tools/include/linux/refcount.h unsigned int old, new, val = atomic_read(&r->refs); new 129 tools/include/linux/refcount.h new = val - i; new 130 tools/include/linux/refcount.h if (new > val) { new 131 tools/include/linux/refcount.h REFCOUNT_WARN(new > val, "refcount_t: underflow; use-after-free.\n"); new 135 tools/include/linux/refcount.h old = atomic_cmpxchg_release(&r->refs, val, new); new 142 tools/include/linux/refcount.h return !new; new 33 tools/include/linux/string.h char *strreplace(char *s, char old, char new); new 1506 tools/include/nolibc/nolibc.h int sys_dup2(int old, int new) new 1508 tools/include/nolibc/nolibc.h return my_syscall2(__NR_dup2, old, new); new 1566 tools/include/nolibc/nolibc.h int sys_link(const char *old, const char *new) new 1569 tools/include/nolibc/nolibc.h return my_syscall5(__NR_linkat, AT_FDCWD, old, AT_FDCWD, new, 0); new 1571 tools/include/nolibc/nolibc.h return my_syscall2(__NR_link, old, new); new 1619 tools/include/nolibc/nolibc.h int sys_pivot_root(const char *new, const char *old) new 1621 tools/include/nolibc/nolibc.h return my_syscall2(__NR_pivot_root, new, old); new 1716 tools/include/nolibc/nolibc.h int sys_symlink(const char *old, const char *new) new 1719 tools/include/nolibc/nolibc.h return my_syscall3(__NR_symlinkat, old, AT_FDCWD, new); new 1721 tools/include/nolibc/nolibc.h return my_syscall2(__NR_symlink, old, new); new 1856 tools/include/nolibc/nolibc.h int dup2(int old, int new) new 1858 tools/include/nolibc/nolibc.h int ret = sys_dup2(old, new); new 1976 tools/include/nolibc/nolibc.h int link(const char *old, const char *new) new 1978 tools/include/nolibc/nolibc.h int ret = sys_link(old, new); new 2050 tools/include/nolibc/nolibc.h int pivot_root(const char *new, const char *old) new 2052 tools/include/nolibc/nolibc.h int ret = sys_pivot_root(new, old); new 2182 tools/include/nolibc/nolibc.h int symlink(const char *old, const char *new) new 2184 tools/include/nolibc/nolibc.h int ret = sys_symlink(old, new); new 75 tools/lib/rbtree.c __rb_rotate_set_parents(struct rb_node *old, struct rb_node *new, new 79 tools/lib/rbtree.c new->__rb_parent_color = old->__rb_parent_color; new 80 tools/lib/rbtree.c rb_set_parent_color(old, new, color); new 81 tools/lib/rbtree.c __rb_change_child(old, new, parent, root); new 86 tools/lib/rbtree.c void (*augment_rotate)(struct rb_node *old, struct rb_node *new)) new 228 tools/lib/rbtree.c void (*augment_rotate)(struct rb_node *old, struct rb_node *new)) new 411 tools/lib/rbtree.c void (*augment_rotate)(struct rb_node *old, struct rb_node *new)) new 424 tools/lib/rbtree.c static inline void dummy_copy(struct rb_node *old, struct rb_node *new) {} new 425 tools/lib/rbtree.c static inline void dummy_rotate(struct rb_node *old, struct rb_node *new) {} new 454 tools/lib/rbtree.c void (*augment_rotate)(struct rb_node *old, struct rb_node *new)) new 545 tools/lib/rbtree.c void rb_replace_node(struct rb_node *victim, struct rb_node *new, new 551 tools/lib/rbtree.c *new = *victim; new 555 tools/lib/rbtree.c rb_set_parent(victim->rb_left, new); new 557 tools/lib/rbtree.c rb_set_parent(victim->rb_right, new); new 558 tools/lib/rbtree.c __rb_change_child(victim, new, parent, root); new 164 tools/lib/string.c char *strreplace(char *s, char old, char new) new 168 tools/lib/string.c *s = new; new 39 tools/objtool/special.c unsigned char size, orig, new; new 51 tools/objtool/special.c .new = ALT_NEW_OFFSET, new 60 tools/objtool/special.c .new = JUMP_NEW_OFFSET, new 66 tools/objtool/special.c .new = EX_NEW_OFFSET, new 136 tools/objtool/special.c new_rela = find_rela_by_dest(sec, offset + entry->new); new 139 tools/objtool/special.c sec, offset + entry->new); new 176 tools/perf/builtin-c2c.c .new = c2c_he_zalloc, new 377 tools/perf/builtin-diff.c .new = block_hist_zalloc, new 151 tools/perf/builtin-help.c char *new = strdup(path); new 152 tools/perf/builtin-help.c char *dest = strrchr(new, '/'); new 156 tools/perf/builtin-help.c path = new; new 211 tools/perf/builtin-help.c struct man_viewer_info_list *new = zalloc(sizeof(*new) + len + 1); new 213 tools/perf/builtin-help.c strncpy(new->name, name, len); new 214 tools/perf/builtin-help.c new->info = strdup(value); new 215 tools/perf/builtin-help.c new->next = man_viewer_info_list; new 216 tools/perf/builtin-help.c man_viewer_info_list = new; new 683 tools/perf/builtin-kmem.c char *new; new 691 tools/perf/builtin-kmem.c new = realloc(new_flags, len + strlen(cpt) + 2); new 692 tools/perf/builtin-kmem.c if (new == NULL) { new 698 tools/perf/builtin-kmem.c new_flags = new; new 762 tools/perf/builtin-kmem.c struct gfp_flag *new; new 764 tools/perf/builtin-kmem.c new = realloc(gfps, (nr_gfps + 1) * sizeof(*gfps)); new 765 tools/perf/builtin-kmem.c if (new == NULL) new 768 tools/perf/builtin-kmem.c gfps = new; new 769 tools/perf/builtin-kmem.c new += nr_gfps++; new 771 tools/perf/builtin-kmem.c new->flags = gfp_flags; new 772 tools/perf/builtin-kmem.c new->human_readable = strdup(str + 10); new 773 tools/perf/builtin-kmem.c new->compact_str = compact_gfp_flags(str + 10); new 774 tools/perf/builtin-kmem.c if (!new->human_readable || !new->compact_str) new 1256 tools/perf/builtin-kmem.c struct rb_node **new = &(root->rb_node); new 1260 tools/perf/builtin-kmem.c while (*new) { new 1264 tools/perf/builtin-kmem.c this = rb_entry(*new, struct alloc_stat, node); new 1265 tools/perf/builtin-kmem.c parent = *new; new 1274 tools/perf/builtin-kmem.c new = &((*new)->rb_left); new 1276 tools/perf/builtin-kmem.c new = &((*new)->rb_right); new 1279 tools/perf/builtin-kmem.c rb_link_node(&data->node, parent, new); new 1303 tools/perf/builtin-kmem.c struct rb_node **new = &root->rb_node; new 1307 tools/perf/builtin-kmem.c while (*new) { new 1311 tools/perf/builtin-kmem.c this = rb_entry(*new, struct page_stat, node); new 1312 tools/perf/builtin-kmem.c parent = *new; new 1321 tools/perf/builtin-kmem.c new = &parent->rb_left; new 1323 tools/perf/builtin-kmem.c new = &parent->rb_right; new 1326 tools/perf/builtin-kmem.c rb_link_node(&data->node, parent, new); new 142 tools/perf/builtin-lock.c static void thread_stat_insert(struct thread_stat *new) new 152 tools/perf/builtin-lock.c if (new->tid < p->tid) new 154 tools/perf/builtin-lock.c else if (new->tid > p->tid) new 160 tools/perf/builtin-lock.c rb_link_node(&new->rb, parent, rb); new 161 tools/perf/builtin-lock.c rb_insert_color(&new->rb, &thread_stats); new 321 tools/perf/builtin-lock.c struct lock_stat *ret, *new; new 328 tools/perf/builtin-lock.c new = zalloc(sizeof(struct lock_stat)); new 329 tools/perf/builtin-lock.c if (!new) new 332 tools/perf/builtin-lock.c new->addr = addr; new 333 tools/perf/builtin-lock.c new->name = zalloc(sizeof(char) * strlen(name) + 1); new 334 tools/perf/builtin-lock.c if (!new->name) { new 335 tools/perf/builtin-lock.c free(new); new 339 tools/perf/builtin-lock.c strcpy(new->name, name); new 340 tools/perf/builtin-lock.c new->wait_time_min = ULLONG_MAX; new 342 tools/perf/builtin-lock.c list_add(&new->hash_entry, entry); new 343 tools/perf/builtin-lock.c return new; new 989 tools/perf/builtin-sched.c struct rb_node **new = &(root->rb_root.rb_node), *parent = NULL; new 992 tools/perf/builtin-sched.c while (*new) { new 996 tools/perf/builtin-sched.c this = container_of(*new, struct work_atoms, node); new 997 tools/perf/builtin-sched.c parent = *new; new 1002 tools/perf/builtin-sched.c new = &((*new)->rb_left); new 1004 tools/perf/builtin-sched.c new = &((*new)->rb_right); new 1009 tools/perf/builtin-sched.c rb_link_node(&data->node, parent, new); new 3085 tools/perf/builtin-sched.c struct rb_node **new = &(root->rb_root.rb_node), *parent = NULL; new 3090 tools/perf/builtin-sched.c while (*new) { new 3093 tools/perf/builtin-sched.c this = container_of(*new, struct work_atoms, node); new 3094 tools/perf/builtin-sched.c parent = *new; new 3099 tools/perf/builtin-sched.c new = &((*new)->rb_left); new 3101 tools/perf/builtin-sched.c new = &((*new)->rb_right); new 3119 tools/perf/builtin-sched.c rb_link_node(&data->node, parent, new); new 941 tools/perf/ui/browsers/hists.c char *new; new 946 tools/perf/ui/browsers/hists.c if (asprintf(&new, "%s%s%s", old_str, new 948 tools/perf/ui/browsers/hists.c new = NULL; new 951 tools/perf/ui/browsers/hists.c if (asprintf(&new, "%s %s", value_str, str) < 0) new 952 tools/perf/ui/browsers/hists.c new = NULL; new 954 tools/perf/ui/browsers/hists.c if (asprintf(&new, "%s", str) < 0) new 955 tools/perf/ui/browsers/hists.c new = NULL; new 958 tools/perf/ui/browsers/hists.c return new; new 528 tools/perf/util/callchain.c struct callchain_node *new; new 530 tools/perf/util/callchain.c new = zalloc(sizeof(*new)); new 531 tools/perf/util/callchain.c if (!new) { new 535 tools/perf/util/callchain.c new->parent = parent; new 536 tools/perf/util/callchain.c INIT_LIST_HEAD(&new->val); new 537 tools/perf/util/callchain.c INIT_LIST_HEAD(&new->parent_val); new 543 tools/perf/util/callchain.c new->rb_root_in = parent->rb_root_in; new 546 tools/perf/util/callchain.c n = rb_first(&new->rb_root_in); new 549 tools/perf/util/callchain.c child->parent = new; new 554 tools/perf/util/callchain.c rb_link_node(&new->rb_node_in, NULL, &parent->rb_root_in.rb_node); new 555 tools/perf/util/callchain.c rb_insert_color(&new->rb_node_in, &parent->rb_root_in); new 558 tools/perf/util/callchain.c return new; new 634 tools/perf/util/callchain.c struct callchain_node *new; new 636 tools/perf/util/callchain.c new = create_child(parent, false); new 637 tools/perf/util/callchain.c if (new == NULL) new 640 tools/perf/util/callchain.c if (fill_node(new, cursor) < 0) { new 643 tools/perf/util/callchain.c list_for_each_entry_safe(call, tmp, &new->val, list) { new 648 tools/perf/util/callchain.c free(new); new 652 tools/perf/util/callchain.c new->children_hit = 0; new 653 tools/perf/util/callchain.c new->hit = period; new 654 tools/perf/util/callchain.c new->children_count = 0; new 655 tools/perf/util/callchain.c new->count = 1; new 656 tools/perf/util/callchain.c return new; new 794 tools/perf/util/callchain.c struct callchain_node *new; new 799 tools/perf/util/callchain.c new = create_child(parent, true); new 800 tools/perf/util/callchain.c if (new == NULL) new 806 tools/perf/util/callchain.c new->val.next = &to_split->list; new 807 tools/perf/util/callchain.c new->val.prev = old_tail; new 808 tools/perf/util/callchain.c to_split->list.prev = &new->val; new 809 tools/perf/util/callchain.c old_tail->next = &new->val; new 812 tools/perf/util/callchain.c new->hit = parent->hit; new 813 tools/perf/util/callchain.c new->children_hit = parent->children_hit; new 814 tools/perf/util/callchain.c parent->children_hit = callchain_cumul_hits(new); new 815 tools/perf/util/callchain.c new->val_nr = parent->val_nr - idx_local; new 817 tools/perf/util/callchain.c new->count = parent->count; new 818 tools/perf/util/callchain.c new->children_count = parent->children_count; new 819 tools/perf/util/callchain.c parent->children_count = callchain_cumul_counts(new); new 834 tools/perf/util/callchain.c new = add_child(parent, cursor, period); new 835 tools/perf/util/callchain.c if (new == NULL) new 852 tools/perf/util/callchain.c rb_link_node(&new->rb_node_in, p, pp); new 853 tools/perf/util/callchain.c rb_insert_color(&new->rb_node_in, &parent->rb_root_in); new 1522 tools/perf/util/callchain.c struct callchain_list *chain, *new; new 1527 tools/perf/util/callchain.c new = malloc(sizeof(*new)); new 1528 tools/perf/util/callchain.c if (new == NULL) new 1530 tools/perf/util/callchain.c *new = *chain; new 1531 tools/perf/util/callchain.c new->has_children = false; new 1532 tools/perf/util/callchain.c map__get(new->ms.map); new 1533 tools/perf/util/callchain.c list_add_tail(&new->list, &head); new 1538 tools/perf/util/callchain.c list_for_each_entry_safe_reverse(chain, new, &head, list) new 1551 tools/perf/util/callchain.c list_for_each_entry_safe(chain, new, &head, list) { new 65 tools/perf/util/comm.c struct comm_str *iter, *new; new 87 tools/perf/util/comm.c new = comm_str__alloc(str); new 88 tools/perf/util/comm.c if (!new) new 91 tools/perf/util/comm.c rb_link_node(&new->rb_node, parent, p); new 92 tools/perf/util/comm.c rb_insert_color(&new->rb_node, root); new 94 tools/perf/util/comm.c return new; new 129 tools/perf/util/comm.c struct comm_str *new, *old = comm->comm_str; new 131 tools/perf/util/comm.c new = comm_str__findnew(str, &comm_str_root); new 132 tools/perf/util/comm.c if (!new) new 136 tools/perf/util/comm.c comm->comm_str = new; new 797 tools/perf/util/dso.c dso_cache__insert(struct dso *dso, struct dso_cache *new) new 803 tools/perf/util/dso.c u64 offset = new->offset; new 821 tools/perf/util/dso.c rb_link_node(&new->rb_node, parent, p); new 822 tools/perf/util/dso.c rb_insert_color(&new->rb_node, root); new 517 tools/perf/util/hist.c .new = hist_entry__zalloc, new 535 tools/perf/util/hist.c he = ops->new(callchain_size); new 1402 tools/perf/util/hist.c struct hist_entry *iter, *new; new 1431 tools/perf/util/hist.c new = hist_entry__new(he, true); new 1432 tools/perf/util/hist.c if (new == NULL) new 1438 tools/perf/util/hist.c new->hpp_list = hpp_list; new 1439 tools/perf/util/hist.c new->parent_he = parent_he; new 1441 tools/perf/util/hist.c hist_entry__apply_hierarchy_filters(new); new 1448 tools/perf/util/hist.c new->trace_output = NULL; new 1453 tools/perf/util/hist.c new->srcline = NULL; new 1458 tools/perf/util/hist.c new->srcfile = NULL; new 1461 tools/perf/util/hist.c rb_link_node(&new->rb_node_in, parent, p); new 1462 tools/perf/util/hist.c rb_insert_color_cached(&new->rb_node_in, root, leftmost); new 1463 tools/perf/util/hist.c return new; new 893 tools/perf/util/map.c struct map *new = map__clone(map); new 894 tools/perf/util/map.c if (new == NULL) new 897 tools/perf/util/map.c err = unwind__prepare_access(mg, new, NULL); new 901 tools/perf/util/map.c map_groups__insert(mg, new); new 902 tools/perf/util/map.c map__put(new); new 18 tools/perf/util/ordered-events.c static void queue_event(struct ordered_events *oe, struct ordered_event *new) new 21 tools/perf/util/ordered-events.c u64 timestamp = new->timestamp; new 25 tools/perf/util/ordered-events.c oe->last = new; new 30 tools/perf/util/ordered-events.c list_add(&new->list, &oe->events); new 44 tools/perf/util/ordered-events.c list_add_tail(&new->list, &oe->events); new 50 tools/perf/util/ordered-events.c list_add_tail(&new->list, &last->list); new 55 tools/perf/util/ordered-events.c list_add(&new->list, &oe->events); new 60 tools/perf/util/ordered-events.c list_add(&new->list, &last->list); new 103 tools/perf/util/ordered-events.c struct ordered_event *new = NULL; new 138 tools/perf/util/ordered-events.c size = sizeof(*oe->buffer) + MAX_SAMPLE_BUFFER * sizeof(*new); new 141 tools/perf/util/ordered-events.c new = list_entry(cache->next, struct ordered_event, list); new 142 tools/perf/util/ordered-events.c list_del_init(&new->list); new 144 tools/perf/util/ordered-events.c new = &oe->buffer->event[oe->buffer_idx]; new 161 tools/perf/util/ordered-events.c new = &oe->buffer->event[0]; new 167 tools/perf/util/ordered-events.c new->event = new_event; new 168 tools/perf/util/ordered-events.c return new; new 175 tools/perf/util/ordered-events.c struct ordered_event *new; new 177 tools/perf/util/ordered-events.c new = alloc_event(oe, event); new 178 tools/perf/util/ordered-events.c if (new) { new 179 tools/perf/util/ordered-events.c new->timestamp = timestamp; new 180 tools/perf/util/ordered-events.c queue_event(oe, new); new 183 tools/perf/util/ordered-events.c return new; new 2745 tools/perf/util/parse-events.c int parse_events_term__clone(struct parse_events_term **new, new 2756 tools/perf/util/parse-events.c return new_term(new, &temp, term->val.str, term->val.num); new 2760 tools/perf/util/parse-events.c struct list_head **new) new 2766 tools/perf/util/parse-events.c *new = NULL; new 2770 tools/perf/util/parse-events.c *new = malloc(sizeof(struct list_head)); new 2771 tools/perf/util/parse-events.c if (!*new) new 2773 tools/perf/util/parse-events.c INIT_LIST_HEAD(*new); new 2779 tools/perf/util/parse-events.c list_add_tail(&n->list, *new); new 138 tools/perf/util/parse-events.h int parse_events_term__clone(struct parse_events_term **new, new 186 tools/perf/util/parse-events.h struct list_head **new); new 1895 tools/perf/util/session.c void *new = realloc(buf, size); new 1896 tools/perf/util/session.c if (!new) { new 1900 tools/perf/util/session.c buf = new; new 77 tools/perf/util/sort.h void *(*new)(size_t size); new 174 tools/perf/util/thread.c struct namespaces *new, *curr = __thread__namespaces(thread); new 176 tools/perf/util/thread.c new = namespaces__new(event); new 177 tools/perf/util/thread.c if (!new) new 180 tools/perf/util/thread.c list_add(&new->list, &thread->namespaces_list); new 188 tools/perf/util/thread.c curr = list_next_entry(new, list); new 240 tools/perf/util/thread.c struct comm *new, *curr = thread__comm(thread); new 248 tools/perf/util/thread.c new = comm__new(str, timestamp, exec); new 249 tools/perf/util/thread.c if (!new) new 251 tools/perf/util/thread.c list_add(&new->list, &thread->comm_list); new 1212 tools/power/x86/turbostat/turbostat.c #define DELTA_WRAP32(new, old) \ new 1213 tools/power/x86/turbostat/turbostat.c if (new > old) { \ new 1214 tools/power/x86/turbostat/turbostat.c old = new - old; \ new 1216 tools/power/x86/turbostat/turbostat.c old = 0x100000000 + new - old; \ new 1220 tools/power/x86/turbostat/turbostat.c delta_package(struct pkg_data *new, struct pkg_data *old) new 1227 tools/power/x86/turbostat/turbostat.c old->pkg_wtd_core_c0 = new->pkg_wtd_core_c0 - old->pkg_wtd_core_c0; new 1229 tools/power/x86/turbostat/turbostat.c old->pkg_any_core_c0 = new->pkg_any_core_c0 - old->pkg_any_core_c0; new 1231 tools/power/x86/turbostat/turbostat.c old->pkg_any_gfxe_c0 = new->pkg_any_gfxe_c0 - old->pkg_any_gfxe_c0; new 1233 tools/power/x86/turbostat/turbostat.c old->pkg_both_core_gfxe_c0 = new->pkg_both_core_gfxe_c0 - old->pkg_both_core_gfxe_c0; new 1235 tools/power/x86/turbostat/turbostat.c old->pc2 = new->pc2 - old->pc2; new 1237 tools/power/x86/turbostat/turbostat.c old->pc3 = new->pc3 - old->pc3; new 1239 tools/power/x86/turbostat/turbostat.c old->pc6 = new->pc6 - old->pc6; new 1241 tools/power/x86/turbostat/turbostat.c old->pc7 = new->pc7 - old->pc7; new 1242 tools/power/x86/turbostat/turbostat.c old->pc8 = new->pc8 - old->pc8; new 1243 tools/power/x86/turbostat/turbostat.c old->pc9 = new->pc9 - old->pc9; new 1244 tools/power/x86/turbostat/turbostat.c old->pc10 = new->pc10 - old->pc10; new 1245 tools/power/x86/turbostat/turbostat.c old->cpu_lpi = new->cpu_lpi - old->cpu_lpi; new 1246 tools/power/x86/turbostat/turbostat.c old->sys_lpi = new->sys_lpi - old->sys_lpi; new 1247 tools/power/x86/turbostat/turbostat.c old->pkg_temp_c = new->pkg_temp_c; new 1250 tools/power/x86/turbostat/turbostat.c if (old->gfx_rc6_ms > new->gfx_rc6_ms) new 1253 tools/power/x86/turbostat/turbostat.c old->gfx_rc6_ms = new->gfx_rc6_ms - old->gfx_rc6_ms; new 1255 tools/power/x86/turbostat/turbostat.c old->gfx_mhz = new->gfx_mhz; new 1257 tools/power/x86/turbostat/turbostat.c DELTA_WRAP32(new->energy_pkg, old->energy_pkg); new 1258 tools/power/x86/turbostat/turbostat.c DELTA_WRAP32(new->energy_cores, old->energy_cores); new 1259 tools/power/x86/turbostat/turbostat.c DELTA_WRAP32(new->energy_gfx, old->energy_gfx); new 1260 tools/power/x86/turbostat/turbostat.c DELTA_WRAP32(new->energy_dram, old->energy_dram); new 1261 tools/power/x86/turbostat/turbostat.c DELTA_WRAP32(new->rapl_pkg_perf_status, old->rapl_pkg_perf_status); new 1262 tools/power/x86/turbostat/turbostat.c DELTA_WRAP32(new->rapl_dram_perf_status, old->rapl_dram_perf_status); new 1266 tools/power/x86/turbostat/turbostat.c old->counter[i] = new->counter[i]; new 1268 tools/power/x86/turbostat/turbostat.c old->counter[i] = new->counter[i] - old->counter[i]; new 1275 tools/power/x86/turbostat/turbostat.c delta_core(struct core_data *new, struct core_data *old) new 1280 tools/power/x86/turbostat/turbostat.c old->c3 = new->c3 - old->c3; new 1281 tools/power/x86/turbostat/turbostat.c old->c6 = new->c6 - old->c6; new 1282 tools/power/x86/turbostat/turbostat.c old->c7 = new->c7 - old->c7; new 1283 tools/power/x86/turbostat/turbostat.c old->core_temp_c = new->core_temp_c; new 1284 tools/power/x86/turbostat/turbostat.c old->mc6_us = new->mc6_us - old->mc6_us; new 1286 tools/power/x86/turbostat/turbostat.c DELTA_WRAP32(new->core_energy, old->core_energy); new 1290 tools/power/x86/turbostat/turbostat.c old->counter[i] = new->counter[i]; new 1292 tools/power/x86/turbostat/turbostat.c old->counter[i] = new->counter[i] - old->counter[i]; new 1308 tools/power/x86/turbostat/turbostat.c delta_thread(struct thread_data *new, struct thread_data *old, new 1316 tools/power/x86/turbostat/turbostat.c new->apic_id = old->apic_id; new 1318 tools/power/x86/turbostat/turbostat.c new->x2apic_id = old->x2apic_id; new 1326 tools/power/x86/turbostat/turbostat.c timersub(&new->tv_begin, &old->tv_begin, &old->tv_delta); new 1327 tools/power/x86/turbostat/turbostat.c old->tv_begin = new->tv_begin; new 1328 tools/power/x86/turbostat/turbostat.c old->tv_end = new->tv_end; new 1330 tools/power/x86/turbostat/turbostat.c old->tsc = new->tsc - old->tsc; new 1338 tools/power/x86/turbostat/turbostat.c old->c1 = new->c1 - old->c1; new 1342 tools/power/x86/turbostat/turbostat.c if ((new->aperf > old->aperf) && (new->mperf > old->mperf)) { new 1343 tools/power/x86/turbostat/turbostat.c old->aperf = new->aperf - old->aperf; new 1344 tools/power/x86/turbostat/turbostat.c old->mperf = new->mperf - old->mperf; new 1378 tools/power/x86/turbostat/turbostat.c old->irq_count = new->irq_count - old->irq_count; new 1381 tools/power/x86/turbostat/turbostat.c old->smi_count = new->smi_count - old->smi_count; new 1385 tools/power/x86/turbostat/turbostat.c old->counter[i] = new->counter[i]; new 1387 tools/power/x86/turbostat/turbostat.c old->counter[i] = new->counter[i] - old->counter[i]; new 9 tools/testing/radix-tree/linux/percpu.h #define this_cpu_cmpxchg(var, old, new) uatomic_cmpxchg(&var, old, new) new 42 tools/testing/selftests/powerpc/vphn/vphn.c u16 new = be16_to_cpup(field++); new 50 tools/testing/selftests/powerpc/vphn/vphn.c cpu_to_be32(last << 16 | new); new 52 tools/testing/selftests/powerpc/vphn/vphn.c } else if (new == VPHN_FIELD_UNUSED) new 55 tools/testing/selftests/powerpc/vphn/vphn.c else if (new & VPHN_FIELD_MSB) { new 58 tools/testing/selftests/powerpc/vphn/vphn.c cpu_to_be32(new & VPHN_FIELD_MASK); new 64 tools/testing/selftests/powerpc/vphn/vphn.c last = new; new 117 tools/testing/selftests/rtc/rtctest.c time_t secs, new; new 161 tools/testing/selftests/rtc/rtctest.c new = timegm((struct tm *)&tm); new 162 tools/testing/selftests/rtc/rtctest.c ASSERT_EQ(new, secs); new 171 tools/testing/selftests/rtc/rtctest.c time_t secs, new; new 210 tools/testing/selftests/rtc/rtctest.c new = timegm((struct tm *)&tm); new 211 tools/testing/selftests/rtc/rtctest.c ASSERT_EQ(new, secs); new 219 tools/testing/selftests/rtc/rtctest.c time_t secs, new; new 263 tools/testing/selftests/rtc/rtctest.c new = timegm((struct tm *)&tm); new 264 tools/testing/selftests/rtc/rtctest.c ASSERT_EQ(new, secs); new 273 tools/testing/selftests/rtc/rtctest.c time_t secs, new; new 312 tools/testing/selftests/rtc/rtctest.c new = timegm((struct tm *)&tm); new 313 tools/testing/selftests/rtc/rtctest.c ASSERT_EQ(new, secs); new 24 tools/testing/selftests/rtc/setdate.c struct rtc_time new, current; new 47 tools/testing/selftests/rtc/setdate.c sscanf(date, "%d-%d-%d", &new.tm_mday, &new.tm_mon, &new.tm_year); new 48 tools/testing/selftests/rtc/setdate.c new.tm_mon -= 1; new 49 tools/testing/selftests/rtc/setdate.c new.tm_year -= 1900; new 50 tools/testing/selftests/rtc/setdate.c sscanf(time, "%d:%d:%d", &new.tm_hour, &new.tm_min, &new.tm_sec); new 53 tools/testing/selftests/rtc/setdate.c new.tm_mday, new.tm_mon + 1, new.tm_year + 1900, new 54 tools/testing/selftests/rtc/setdate.c new.tm_hour, new.tm_min, new.tm_sec); new 57 tools/testing/selftests/rtc/setdate.c retval = ioctl(fd, RTC_SET_TIME, &new); new 27 tools/testing/selftests/vm/mlock-random-test.c struct rlimit new; new 30 tools/testing/selftests/vm/mlock-random-test.c new.rlim_cur = max; new 31 tools/testing/selftests/vm/mlock-random-test.c new.rlim_max = max; new 32 tools/testing/selftests/vm/mlock-random-test.c if (setrlimit(RLIMIT_MEMLOCK, &new)) { new 40 tools/usb/usbip/libsrc/list.h static inline void __list_add(struct list_head *new, new 44 tools/usb/usbip/libsrc/list.h next->prev = new; new 45 tools/usb/usbip/libsrc/list.h new->next = next; new 46 tools/usb/usbip/libsrc/list.h new->prev = prev; new 47 tools/usb/usbip/libsrc/list.h prev->next = new; new 58 tools/usb/usbip/libsrc/list.h static inline void list_add(struct list_head *new, struct list_head *head) new 60 tools/usb/usbip/libsrc/list.h __list_add(new, head, head->next); new 77 virt/kvm/arm/aarch32.c unsigned long old, new; new 80 virt/kvm/arm/aarch32.c new = 0; new 82 virt/kvm/arm/aarch32.c new |= (old & PSR_AA32_N_BIT); new 83 virt/kvm/arm/aarch32.c new |= (old & PSR_AA32_Z_BIT); new 84 virt/kvm/arm/aarch32.c new |= (old & PSR_AA32_C_BIT); new 85 virt/kvm/arm/aarch32.c new |= (old & PSR_AA32_V_BIT); new 86 virt/kvm/arm/aarch32.c new |= (old & PSR_AA32_Q_BIT); new 92 virt/kvm/arm/aarch32.c new |= (old & PSR_AA32_DIT_BIT); new 97 virt/kvm/arm/aarch32.c new |= PSR_AA32_SSBS_BIT; new 102 virt/kvm/arm/aarch32.c new |= (old & PSR_AA32_PAN_BIT); new 104 virt/kvm/arm/aarch32.c new |= PSR_AA32_PAN_BIT; new 111 virt/kvm/arm/aarch32.c new |= (old & PSR_AA32_GE_MASK); new 120 virt/kvm/arm/aarch32.c new |= PSR_AA32_E_BIT; new 126 virt/kvm/arm/aarch32.c new |= (old & PSR_AA32_A_BIT); new 128 virt/kvm/arm/aarch32.c new |= PSR_AA32_A_BIT; new 133 virt/kvm/arm/aarch32.c new |= PSR_AA32_I_BIT; new 139 virt/kvm/arm/aarch32.c new |= (old & PSR_AA32_F_BIT); new 141 virt/kvm/arm/aarch32.c new |= PSR_AA32_F_BIT; new 147 virt/kvm/arm/aarch32.c new |= PSR_AA32_T_BIT; new 149 virt/kvm/arm/aarch32.c new |= mode; new 151 virt/kvm/arm/aarch32.c return new; new 2264 virt/kvm/arm/mmu.c const struct kvm_memory_slot *new, new 174 virt/kvm/irqchip.c struct kvm_irq_routing_table *new, *old; new 187 virt/kvm/irqchip.c new = kzalloc(struct_size(new, map, nr_rt_entries), GFP_KERNEL_ACCOUNT); new 188 virt/kvm/irqchip.c if (!new) new 191 virt/kvm/irqchip.c new->nr_rt_entries = nr_rt_entries; new 194 virt/kvm/irqchip.c new->chip[i][j] = -1; new 213 virt/kvm/irqchip.c r = setup_routing_entry(kvm, new, e, ue); new 221 virt/kvm/irqchip.c rcu_assign_pointer(kvm->irq_routing, new); new 230 virt/kvm/irqchip.c new = old; new 237 virt/kvm/irqchip.c free_irq_routing_table(new); new 880 virt/kvm/kvm_main.c struct kvm_memory_slot *new, new 883 virt/kvm/kvm_main.c int id = new->id; new 891 virt/kvm/kvm_main.c WARN_ON(mslots[i].npages || !new->npages); new 895 virt/kvm/kvm_main.c WARN_ON(new->npages || !mslots[i].npages); new 902 virt/kvm/kvm_main.c new->base_gfn <= mslots[i + 1].base_gfn) { new 919 virt/kvm/kvm_main.c if (new->npages) { new 921 virt/kvm/kvm_main.c new->base_gfn >= mslots[i - 1].base_gfn) { new 929 virt/kvm/kvm_main.c mslots[i] = *new; new 998 virt/kvm/kvm_main.c struct kvm_memory_slot old, new; new 1034 virt/kvm/kvm_main.c new = old = *slot; new 1036 virt/kvm/kvm_main.c new.id = id; new 1037 virt/kvm/kvm_main.c new.base_gfn = base_gfn; new 1038 virt/kvm/kvm_main.c new.npages = npages; new 1039 virt/kvm/kvm_main.c new.flags = mem->flags; new 1047 virt/kvm/kvm_main.c ((new.flags ^ old.flags) & KVM_MEM_READONLY)) new 1052 virt/kvm/kvm_main.c else if (new.flags != old.flags) new 1064 virt/kvm/kvm_main.c new.base_gfn = 0; new 1065 virt/kvm/kvm_main.c new.flags = 0; new 1081 virt/kvm/kvm_main.c if (!(new.flags & KVM_MEM_LOG_DIRTY_PAGES)) new 1082 virt/kvm/kvm_main.c new.dirty_bitmap = NULL; new 1086 virt/kvm/kvm_main.c new.userspace_addr = mem->userspace_addr; new 1088 virt/kvm/kvm_main.c if (kvm_arch_create_memslot(kvm, &new, npages)) new 1093 virt/kvm/kvm_main.c if ((new.flags & KVM_MEM_LOG_DIRTY_PAGES) && !new.dirty_bitmap) { new 1094 virt/kvm/kvm_main.c if (kvm_create_dirty_bitmap(&new) < 0) new 1126 virt/kvm/kvm_main.c r = kvm_arch_prepare_memory_region(kvm, &new, mem, change); new 1132 virt/kvm/kvm_main.c new.dirty_bitmap = NULL; new 1133 virt/kvm/kvm_main.c memset(&new.arch, 0, sizeof(new.arch)); new 1136 virt/kvm/kvm_main.c update_memslots(slots, &new, change); new 1139 virt/kvm/kvm_main.c kvm_arch_commit_memory_region(kvm, mem, &old, &new, change); new 1141 virt/kvm/kvm_main.c kvm_free_memslot(kvm, &old, &new); new 1148 virt/kvm/kvm_main.c kvm_free_memslot(kvm, &new, &old);