READ_ONCE          30 arch/alpha/include/asm/atomic.h #define atomic_read(v)		READ_ONCE((v)->counter)
READ_ONCE          31 arch/alpha/include/asm/atomic.h #define atomic64_read(v)	READ_ONCE((v)->counter)
READ_ONCE          21 arch/arc/include/asm/atomic.h #define atomic_read(v)  READ_ONCE((v)->counter)
READ_ONCE         280 arch/arc/kernel/smp.c 		new = old = READ_ONCE(*ipi_data_ptr);
READ_ONCE          27 arch/arm/include/asm/atomic.h #define atomic_read(v)	READ_ONCE((v)->counter)
READ_ONCE          75 arch/arm/include/asm/spinlock.h 		lockval.tickets.owner = READ_ONCE(lock->tickets.owner);
READ_ONCE         121 arch/arm/include/asm/spinlock.h 	return !arch_spin_value_unlocked(READ_ONCE(*lock));
READ_ONCE         126 arch/arm/include/asm/spinlock.h 	struct __raw_tickets tickets = READ_ONCE(lock->tickets);
READ_ONCE          65 arch/arm/kvm/reset.c 	if (READ_ONCE(vcpu->arch.reset_state.reset)) {
READ_ONCE          26 arch/arm/vdso/vgettimeofday.c 	seq = READ_ONCE(vdata->seq_count);
READ_ONCE         104 arch/arm64/include/asm/atomic.h #define arch_atomic_read(v)			READ_ONCE((v)->counter)
READ_ONCE         135 arch/arm64/include/asm/barrier.h 		VAL = READ_ONCE(*__PTR);				\
READ_ONCE          23 arch/arm64/include/asm/hugetlb.h 	return READ_ONCE(*ptep);
READ_ONCE         228 arch/arm64/include/asm/kvm_mmu.h 	pteval = READ_ONCE(pte_val(*ptep));
READ_ONCE         239 arch/arm64/include/asm/kvm_mmu.h 	return (READ_ONCE(pte_val(*ptep)) & PTE_S2_RDWR) == PTE_S2_RDONLY;
READ_ONCE         244 arch/arm64/include/asm/kvm_mmu.h 	return !(READ_ONCE(pte_val(*ptep)) & PTE_S2_XN);
READ_ONCE         259 arch/arm64/include/asm/kvm_mmu.h 	return !(READ_ONCE(pmd_val(*pmdp)) & PMD_S2_XN);
READ_ONCE         274 arch/arm64/include/asm/kvm_mmu.h 	return !(READ_ONCE(pud_val(*pudp)) & PUD_S2_XN);
READ_ONCE          43 arch/arm64/include/asm/percpu.h 	return READ_ONCE(*(u##sz *)ptr);				\
READ_ONCE         251 arch/arm64/include/asm/pgtable.h 	old_pte = READ_ONCE(*ptep);
READ_ONCE         501 arch/arm64/include/asm/pgtable.h #define pte_offset_phys(dir,addr)	(pmd_page_paddr(READ_ONCE(*(dir))) + pte_index(addr) * sizeof(pte_t))
READ_ONCE         560 arch/arm64/include/asm/pgtable.h #define pmd_offset_phys(dir, addr)	(pud_page_paddr(READ_ONCE(*(dir))) + pmd_index(addr) * sizeof(pmd_t))
READ_ONCE         618 arch/arm64/include/asm/pgtable.h #define pud_offset_phys(dir, addr)	(pgd_page_paddr(READ_ONCE(*(dir))) + pud_index(addr) * sizeof(pud_t))
READ_ONCE         707 arch/arm64/include/asm/pgtable.h 	pte = READ_ONCE(*ptep);
READ_ONCE         781 arch/arm64/include/asm/pgtable.h 	pte = READ_ONCE(*ptep);
READ_ONCE          12 arch/arm64/include/asm/preempt.h 	return READ_ONCE(current_thread_info()->preempt.count);
READ_ONCE          46 arch/arm64/include/asm/preempt.h 	u32 pc = READ_ONCE(current_thread_info()->preempt.count);
READ_ONCE          53 arch/arm64/include/asm/preempt.h 	u32 pc = READ_ONCE(current_thread_info()->preempt.count);
READ_ONCE          61 arch/arm64/include/asm/preempt.h 	u64 pc = READ_ONCE(ti->preempt_count);
READ_ONCE          73 arch/arm64/include/asm/preempt.h 	return !pc || !READ_ONCE(ti->preempt_count);
READ_ONCE          78 arch/arm64/include/asm/preempt.h 	u64 pc = READ_ONCE(current_thread_info()->preempt_count);
READ_ONCE         134 arch/arm64/include/asm/uaccess.h 	ttbr0 = READ_ONCE(current_thread_info()->ttbr0);
READ_ONCE         220 arch/arm64/kernel/alternative.c 		while (!READ_ONCE(all_alternatives_applied))
READ_ONCE          88 arch/arm64/kernel/efi.c 	pte_t pte = READ_ONCE(*ptep);
READ_ONCE         226 arch/arm64/kernel/hibernate.c 	if (pgd_none(READ_ONCE(*pgdp))) {
READ_ONCE         236 arch/arm64/kernel/hibernate.c 	if (pud_none(READ_ONCE(*pudp))) {
READ_ONCE         246 arch/arm64/kernel/hibernate.c 	if (pmd_none(READ_ONCE(*pmdp))) {
READ_ONCE         344 arch/arm64/kernel/hibernate.c 	pte_t pte = READ_ONCE(*src_ptep);
READ_ONCE         398 arch/arm64/kernel/hibernate.c 	if (pud_none(READ_ONCE(*dst_pudp))) {
READ_ONCE         408 arch/arm64/kernel/hibernate.c 		pmd_t pmd = READ_ONCE(*src_pmdp);
READ_ONCE         433 arch/arm64/kernel/hibernate.c 	if (pgd_none(READ_ONCE(*dst_pgdp))) {
READ_ONCE         443 arch/arm64/kernel/hibernate.c 		pud_t pud = READ_ONCE(*src_pudp);
READ_ONCE         470 arch/arm64/kernel/hibernate.c 		if (pgd_none(READ_ONCE(*src_pgdp)))
READ_ONCE        1832 arch/arm64/kernel/ptrace.c 	unsigned long flags = READ_ONCE(current_thread_info()->flags);
READ_ONCE         938 arch/arm64/kernel/signal.c 		thread_flags = READ_ONCE(current_thread_info()->flags);
READ_ONCE         140 arch/arm64/kernel/smp.c 	status = READ_ONCE(secondary_data.status);
READ_ONCE         144 arch/arm64/kernel/smp.c 			status = READ_ONCE(__early_cpu_boot_status);
READ_ONCE         302 arch/arm64/mm/dump.c 		note_page(st, addr, 4, READ_ONCE(pte_val(*ptep)));
READ_ONCE         313 arch/arm64/mm/dump.c 		pmd_t pmd = READ_ONCE(*pmdp);
READ_ONCE         332 arch/arm64/mm/dump.c 		pud_t pud = READ_ONCE(*pudp);
READ_ONCE         352 arch/arm64/mm/dump.c 		pgd_t pgd = READ_ONCE(*pgdp);
READ_ONCE         155 arch/arm64/mm/fault.c 	pgd = READ_ONCE(*pgdp);
READ_ONCE         167 arch/arm64/mm/fault.c 		pud = READ_ONCE(*pudp);
READ_ONCE         173 arch/arm64/mm/fault.c 		pmd = READ_ONCE(*pmdp);
READ_ONCE         179 arch/arm64/mm/fault.c 		pte = READ_ONCE(*ptep);
READ_ONCE         202 arch/arm64/mm/fault.c 	pte_t pte = READ_ONCE(*ptep);
READ_ONCE         247 arch/arm64/mm/hugetlbpage.c 		    pud_none(READ_ONCE(*pudp)))
READ_ONCE         268 arch/arm64/mm/hugetlbpage.c 	if (!pgd_present(READ_ONCE(*pgdp)))
READ_ONCE         272 arch/arm64/mm/hugetlbpage.c 	pud = READ_ONCE(*pudp);
READ_ONCE         284 arch/arm64/mm/hugetlbpage.c 	pmd = READ_ONCE(*pmdp);
READ_ONCE         413 arch/arm64/mm/hugetlbpage.c 	if (!pte_cont(READ_ONCE(*ptep))) {
READ_ONCE         437 arch/arm64/mm/hugetlbpage.c 	if (!pte_cont(READ_ONCE(*ptep))) {
READ_ONCE          63 arch/arm64/mm/kasan_init.c 	if (pmd_none(READ_ONCE(*pmdp))) {
READ_ONCE          77 arch/arm64/mm/kasan_init.c 	if (pud_none(READ_ONCE(*pudp))) {
READ_ONCE          90 arch/arm64/mm/kasan_init.c 	if (pgd_none(READ_ONCE(*pgdp))) {
READ_ONCE         114 arch/arm64/mm/kasan_init.c 	} while (ptep++, addr = next, addr != end && pte_none(READ_ONCE(*ptep)));
READ_ONCE         126 arch/arm64/mm/kasan_init.c 	} while (pmdp++, addr = next, addr != end && pmd_none(READ_ONCE(*pmdp)));
READ_ONCE         138 arch/arm64/mm/kasan_init.c 	} while (pudp++, addr = next, addr != end && pud_none(READ_ONCE(*pudp)));
READ_ONCE         184 arch/arm64/mm/kasan_init.c 		set_pgd(pgdp_new, READ_ONCE(*pgdp));
READ_ONCE         147 arch/arm64/mm/mmu.c 		pte_t old_pte = READ_ONCE(*ptep);
READ_ONCE         156 arch/arm64/mm/mmu.c 					      READ_ONCE(pte_val(*ptep))));
READ_ONCE         171 arch/arm64/mm/mmu.c 	pmd_t pmd = READ_ONCE(*pmdp);
READ_ONCE         179 arch/arm64/mm/mmu.c 		pmd = READ_ONCE(*pmdp);
READ_ONCE         208 arch/arm64/mm/mmu.c 		pmd_t old_pmd = READ_ONCE(*pmdp);
READ_ONCE         222 arch/arm64/mm/mmu.c 						      READ_ONCE(pmd_val(*pmdp))));
READ_ONCE         228 arch/arm64/mm/mmu.c 			       pmd_val(old_pmd) != READ_ONCE(pmd_val(*pmdp)));
READ_ONCE         242 arch/arm64/mm/mmu.c 	pud_t pud = READ_ONCE(*pudp);
READ_ONCE         253 arch/arm64/mm/mmu.c 		pud = READ_ONCE(*pudp);
READ_ONCE         292 arch/arm64/mm/mmu.c 	pgd_t pgd = READ_ONCE(*pgdp);
READ_ONCE         299 arch/arm64/mm/mmu.c 		pgd = READ_ONCE(*pgdp);
READ_ONCE         305 arch/arm64/mm/mmu.c 		pud_t old_pud = READ_ONCE(*pudp);
READ_ONCE         321 arch/arm64/mm/mmu.c 						      READ_ONCE(pud_val(*pudp))));
READ_ONCE         327 arch/arm64/mm/mmu.c 			       pud_val(old_pud) != READ_ONCE(pud_val(*pudp)));
READ_ONCE         641 arch/arm64/mm/mmu.c 	if (!READ_ONCE(pgd_val(*pgd_offset_raw(pgdp, FIXADDR_START)))) {
READ_ONCE         648 arch/arm64/mm/mmu.c 			READ_ONCE(*pgd_offset_k(FIXADDR_START)));
READ_ONCE         702 arch/arm64/mm/mmu.c 	if (pgd_none(READ_ONCE(*pgdp)))
READ_ONCE         706 arch/arm64/mm/mmu.c 	pud = READ_ONCE(*pudp);
READ_ONCE         714 arch/arm64/mm/mmu.c 	pmd = READ_ONCE(*pmdp);
READ_ONCE         722 arch/arm64/mm/mmu.c 	pte = READ_ONCE(*ptep);
READ_ONCE         757 arch/arm64/mm/mmu.c 		if (pmd_none(READ_ONCE(*pmdp))) {
READ_ONCE         781 arch/arm64/mm/mmu.c 	pgd_t pgd = READ_ONCE(*pgdp);
READ_ONCE         791 arch/arm64/mm/mmu.c 	pud_t pud = READ_ONCE(*pudp);
READ_ONCE         817 arch/arm64/mm/mmu.c 	pgd = READ_ONCE(*pgdp);
READ_ONCE         832 arch/arm64/mm/mmu.c 	if (pud_none(READ_ONCE(*pudp)))
READ_ONCE         961 arch/arm64/mm/mmu.c 	if (!pgattr_change_is_safe(READ_ONCE(pud_val(*pudp)),
READ_ONCE         975 arch/arm64/mm/mmu.c 	if (!pgattr_change_is_safe(READ_ONCE(pmd_val(*pmdp)),
READ_ONCE         986 arch/arm64/mm/mmu.c 	if (!pud_sect(READ_ONCE(*pudp)))
READ_ONCE         994 arch/arm64/mm/mmu.c 	if (!pmd_sect(READ_ONCE(*pmdp)))
READ_ONCE        1005 arch/arm64/mm/mmu.c 	pmd = READ_ONCE(*pmdp);
READ_ONCE        1026 arch/arm64/mm/mmu.c 	pud = READ_ONCE(*pudp);
READ_ONCE          25 arch/arm64/mm/pageattr.c 	pte_t pte = READ_ONCE(*ptep);
READ_ONCE         210 arch/arm64/mm/pageattr.c 	if (pgd_none(READ_ONCE(*pgdp)))
READ_ONCE         214 arch/arm64/mm/pageattr.c 	pud = READ_ONCE(*pudp);
READ_ONCE         221 arch/arm64/mm/pageattr.c 	pmd = READ_ONCE(*pmdp);
READ_ONCE         228 arch/arm64/mm/pageattr.c 	return pte_valid(READ_ONCE(*ptep));
READ_ONCE          32 arch/csky/include/asm/spinlock.h 		lockval.tickets.owner = READ_ONCE(lock->tickets.owner);
READ_ONCE          78 arch/csky/include/asm/spinlock.h 	return !arch_spin_value_unlocked(READ_ONCE(*lock));
READ_ONCE          83 arch/csky/include/asm/spinlock.h 	struct __raw_tickets tickets = READ_ONCE(lock->tickets);
READ_ONCE         145 arch/csky/include/asm/spinlock.h #define arch_spin_is_locked(x)	(READ_ONCE((x)->lock) != 0)
READ_ONCE          17 arch/h8300/include/asm/atomic.h #define atomic_read(v)		READ_ONCE((v)->counter)
READ_ONCE          39 arch/hexagon/include/asm/atomic.h #define atomic_read(v)		READ_ONCE((v)->counter)
READ_ONCE          25 arch/ia64/include/asm/atomic.h #define atomic_read(v)		READ_ONCE((v)->counter)
READ_ONCE          26 arch/ia64/include/asm/atomic.h #define atomic64_read(v)	READ_ONCE((v)->counter)
READ_ONCE          65 arch/ia64/include/asm/barrier.h 	typeof(*p) ___p1 = READ_ONCE(*p);				\
READ_ONCE          65 arch/ia64/include/asm/spinlock.h 	int tmp = READ_ONCE(lock->lock);
READ_ONCE          84 arch/ia64/include/asm/spinlock.h 	long tmp = READ_ONCE(lock->lock);
READ_ONCE          91 arch/ia64/include/asm/spinlock.h 	long tmp = READ_ONCE(lock->lock);
READ_ONCE          21 arch/m68k/include/asm/atomic.h #define atomic_read(v)		READ_ONCE((v)->counter)
READ_ONCE         308 arch/mips/cavium-octeon/executive/cvmx-l2c.c 		READ_ONCE(*ptr);
READ_ONCE          44 arch/mips/include/asm/atomic.h #define atomic_read(v)		READ_ONCE((v)->counter)
READ_ONCE         248 arch/mips/include/asm/atomic.h #define atomic64_read(v)	READ_ONCE((v)->counter)
READ_ONCE         210 arch/mips/mm/context.c 	old_active_mmid = READ_ONCE(cpu_data[cpu].asid_cache);
READ_ONCE          26 arch/nds32/kernel/vdso/gettimeofday.c 	seq = READ_ONCE(vdata->seq_count);
READ_ONCE          73 arch/openrisc/include/asm/cmpxchg.h 	load32 = READ_ONCE(*p);
READ_ONCE         106 arch/openrisc/include/asm/cmpxchg.h 		oldv = READ_ONCE(*p);
READ_ONCE          73 arch/parisc/include/asm/atomic.h 	return READ_ONCE((v)->counter);
READ_ONCE         218 arch/parisc/include/asm/atomic.h 	return READ_ONCE((v)->counter);
READ_ONCE          75 arch/powerpc/include/asm/barrier.h 	typeof(*p) ___p1 = READ_ONCE(*p);				\
READ_ONCE         428 arch/powerpc/include/asm/kvm_book3s_64.h 		old_pte = READ_ONCE(*ptep);
READ_ONCE          22 arch/powerpc/include/asm/local.h 	return READ_ONCE(l->v);
READ_ONCE        3282 arch/powerpc/kernel/ptrace.c 	flags = READ_ONCE(current_thread_info()->flags) &
READ_ONCE         426 arch/powerpc/kernel/smp.c 		fn = READ_ONCE(nmi_ipi_function);
READ_ONCE        1272 arch/powerpc/kvm/book3s_64_mmu_radix.c 		pgd = READ_ONCE(*pgdp);
READ_ONCE        1279 arch/powerpc/kvm/book3s_64_mmu_radix.c 		pud = READ_ONCE(*pudp);
READ_ONCE        1291 arch/powerpc/kvm/book3s_64_mmu_radix.c 		pmd = READ_ONCE(*pmdp);
READ_ONCE        1303 arch/powerpc/kvm/book3s_64_mmu_radix.c 		pte = pte_val(READ_ONCE(*ptep));
READ_ONCE         172 arch/powerpc/kvm/book3s_hv.c 		vcpu = READ_ONCE(vc->runnable_threads[i]);
READ_ONCE         240 arch/powerpc/kvm/book3s_hv.c 	cpu = READ_ONCE(vcpu->arch.thread_cpu);
READ_ONCE        1179 arch/powerpc/kvm/book3s_hv.c 		pcpu = READ_ONCE(v->cpu);
READ_ONCE          89 arch/powerpc/kvm/book3s_hv_rm_xics.c 		old = new = READ_ONCE(rm_core[core].rm_state);
READ_ONCE         253 arch/powerpc/kvm/book3s_hv_rm_xics.c 		old_state = new_state = READ_ONCE(icp->state);
READ_ONCE         452 arch/powerpc/kvm/book3s_hv_rm_xics.c 		old_state = new_state = READ_ONCE(icp->state);
READ_ONCE         511 arch/powerpc/kvm/book3s_hv_rm_xics.c 		old_state = new_state = READ_ONCE(icp->state);
READ_ONCE         577 arch/powerpc/kvm/book3s_hv_rm_xics.c 		old_state = new_state = READ_ONCE(icp->state);
READ_ONCE         652 arch/powerpc/kvm/book3s_hv_rm_xics.c 		old_state = new_state = READ_ONCE(icp->state);
READ_ONCE         346 arch/powerpc/kvm/book3s_xics.c 		old_state = new_state = READ_ONCE(icp->state);
READ_ONCE         549 arch/powerpc/kvm/book3s_xics.c 		old_state = new_state = READ_ONCE(icp->state);
READ_ONCE         603 arch/powerpc/kvm/book3s_xics.c 		old_state = new_state = READ_ONCE(icp->state);
READ_ONCE         670 arch/powerpc/kvm/book3s_xics.c 		old_state = new_state = READ_ONCE(icp->state);
READ_ONCE         715 arch/powerpc/kvm/book3s_xics.c 	state = READ_ONCE(icp->state);
READ_ONCE         756 arch/powerpc/kvm/book3s_xics.c 		old_state = new_state = READ_ONCE(icp->state);
READ_ONCE         971 arch/powerpc/kvm/book3s_xics.c 		state.raw = READ_ONCE(icp->state.raw);
READ_ONCE        1168 arch/powerpc/kvm/book3s_xics.c 		old_state = READ_ONCE(icp->state);
READ_ONCE         150 arch/powerpc/kvm/book3s_xive_template.c 		qpage = READ_ONCE(q->qpage);
READ_ONCE         390 arch/powerpc/kvm/book3s_xive_template.c 		qpage = READ_ONCE(q->qpage);
READ_ONCE         479 arch/powerpc/kvm/e500_mmu_host.c 		pte_t pte = READ_ONCE(*ptep);
READ_ONCE          34 arch/powerpc/mm/book3s64/hash_4k.c 		pte_t pte = READ_ONCE(*ptep);
READ_ONCE          52 arch/powerpc/mm/book3s64/hash_64k.c 		pte_t pte = READ_ONCE(*ptep);
READ_ONCE         238 arch/powerpc/mm/book3s64/hash_64k.c 		pte_t pte = READ_ONCE(*ptep);
READ_ONCE          36 arch/powerpc/mm/book3s64/hash_hugepage.c 		pmd_t pmd = READ_ONCE(*pmdp);
READ_ONCE        1670 arch/powerpc/mm/book3s64/hash_utils.c 		pkey = pte_to_pkey_bits(pte_val(READ_ONCE(*ptep)));
READ_ONCE         329 arch/powerpc/mm/pgtable.c 	pgd  = READ_ONCE(*pgdp);
READ_ONCE         356 arch/powerpc/mm/pgtable.c 	pud  = READ_ONCE(*pudp);
READ_ONCE         373 arch/powerpc/mm/pgtable.c 	pmd  = READ_ONCE(*pmdp);
READ_ONCE         136 arch/powerpc/perf/callchain.c 	pte = READ_ONCE(*ptep);
READ_ONCE         973 arch/powerpc/perf/imc-pmu.c 	data = be64_to_cpu(READ_ONCE(*addr));
READ_ONCE        1164 arch/powerpc/perf/imc-pmu.c 	if (be64_to_cpu(READ_ONCE(mem->tb1)) > *prev_tb)
READ_ONCE        1165 arch/powerpc/perf/imc-pmu.c 		*prev_tb = be64_to_cpu(READ_ONCE(mem->tb1));
READ_ONCE        1169 arch/powerpc/perf/imc-pmu.c 	if ((be64_to_cpu(READ_ONCE(mem->tb1)) & IMC_TRACE_RECORD_TB1_MASK) !=
READ_ONCE        1170 arch/powerpc/perf/imc-pmu.c 			 be64_to_cpu(READ_ONCE(mem->tb2)))
READ_ONCE        1174 arch/powerpc/perf/imc-pmu.c 	data->ip =  be64_to_cpu(READ_ONCE(mem->ip));
READ_ONCE         261 arch/powerpc/platforms/powernv/idle.c 	u64 s = READ_ONCE(*state);
READ_ONCE          46 arch/powerpc/platforms/powernv/opal-irqchip.c 	e = READ_ONCE(last_outstanding_events) & opal_event_irqchip.mask;
READ_ONCE          42 arch/powerpc/platforms/powernv/opal-msglog.c 	out_pos = be32_to_cpu(READ_ONCE(mc->out_pos));
READ_ONCE          64 arch/powerpc/platforms/powernv/pci-ioda-tce.c 		unsigned long oldtce, tce = be64_to_cpu(READ_ONCE(tmp[n]));
READ_ONCE          32 arch/riscv/include/asm/atomic.h 	return READ_ONCE(v->counter);
READ_ONCE          43 arch/riscv/include/asm/atomic.h 	return READ_ONCE(v->counter);
READ_ONCE          39 arch/riscv/include/asm/barrier.h 	typeof(*p) ___p1 = READ_ONCE(*p);				\
READ_ONCE          20 arch/riscv/include/asm/spinlock.h #define arch_spin_is_locked(x)	(READ_ONCE((x)->lock) != 0)
READ_ONCE          43 arch/s390/include/asm/barrier.h 	typeof(*p) ___p1 = READ_ONCE(*p);				\
READ_ONCE        1247 arch/s390/include/asm/pgtable.h #define pgd_offset(mm, address) pgd_offset_raw(READ_ONCE((mm)->pgd), address)
READ_ONCE          17 arch/s390/include/asm/preempt.h 	return READ_ONCE(S390_lowcore.preempt_count) & ~PREEMPT_NEED_RESCHED;
READ_ONCE          25 arch/s390/include/asm/preempt.h 		old = READ_ONCE(S390_lowcore.preempt_count);
READ_ONCE          50 arch/s390/include/asm/preempt.h 	return !(READ_ONCE(S390_lowcore.preempt_count) & PREEMPT_NEED_RESCHED);
READ_ONCE          73 arch/s390/include/asm/preempt.h 	return unlikely(READ_ONCE(S390_lowcore.preempt_count) ==
READ_ONCE          83 arch/s390/include/asm/preempt.h 	return READ_ONCE(S390_lowcore.preempt_count);
READ_ONCE          55 arch/s390/include/asm/spinlock.h 	return READ_ONCE(lp->lock) != 0;
READ_ONCE         142 arch/s390/include/asm/spinlock.h 	old = READ_ONCE(rw->cnts);
READ_ONCE         151 arch/s390/include/asm/spinlock.h 	old = READ_ONCE(rw->cnts);
READ_ONCE          65 arch/s390/include/asm/tlbflush.h 	gmap_asce = READ_ONCE(mm->context.gmap_asce);
READ_ONCE          61 arch/s390/kernel/idle.c 		idle_count = READ_ONCE(idle->idle_count);
READ_ONCE          62 arch/s390/kernel/idle.c 		if (READ_ONCE(idle->clock_idle_enter))
READ_ONCE          78 arch/s390/kernel/idle.c 		idle_time = READ_ONCE(idle->idle_time);
READ_ONCE          79 arch/s390/kernel/idle.c 		idle_enter = READ_ONCE(idle->clock_idle_enter);
READ_ONCE          80 arch/s390/kernel/idle.c 		idle_exit = READ_ONCE(idle->clock_idle_exit);
READ_ONCE         104 arch/s390/kernel/idle.c 		idle_enter = READ_ONCE(idle->clock_idle_enter);
READ_ONCE         105 arch/s390/kernel/idle.c 		idle_exit = READ_ONCE(idle->clock_idle_exit);
READ_ONCE         627 arch/s390/kernel/time.c 		while (READ_ONCE(sync->in_sync) == 0)
READ_ONCE         155 arch/s390/kernel/vtime.c 				READ_ONCE(S390_lowcore.user_timer));
READ_ONCE         157 arch/s390/kernel/vtime.c 				 READ_ONCE(S390_lowcore.guest_timer));
READ_ONCE         159 arch/s390/kernel/vtime.c 				  READ_ONCE(S390_lowcore.system_timer));
READ_ONCE         161 arch/s390/kernel/vtime.c 				   READ_ONCE(S390_lowcore.hardirq_timer));
READ_ONCE         163 arch/s390/kernel/vtime.c 				   READ_ONCE(S390_lowcore.softirq_timer));
READ_ONCE         289 arch/s390/kvm/gaccess.c 		old = READ_ONCE(*ic);
READ_ONCE         314 arch/s390/kvm/gaccess.c 		old = READ_ONCE(*ic);
READ_ONCE         332 arch/s390/kvm/gaccess.c 		old = READ_ONCE(*ic);
READ_ONCE         352 arch/s390/kvm/gaccess.c 		old = READ_ONCE(*ic);
READ_ONCE         245 arch/s390/kvm/interrupt.c 		word = READ_ONCE(gisa->u64.word[0]);
READ_ONCE         268 arch/s390/kvm/interrupt.c 		word = READ_ONCE(gisa->u64.word[0]);
READ_ONCE         289 arch/s390/kvm/interrupt.c 		word = READ_ONCE(gi->origin->u64.word[0]);
READ_ONCE         290 arch/s390/kvm/interrupt.c 		alert_mask = READ_ONCE(gi->alert.mask);
READ_ONCE         302 arch/s390/kvm/interrupt.c 	return READ_ONCE(gisa->next_alert) != (u32)(u64)gisa;
READ_ONCE         312 arch/s390/kvm/interrupt.c 	return READ_ONCE(gisa->ipm);
READ_ONCE        1850 arch/s390/kvm/interrupt.c 	u64 type = READ_ONCE(inti->type);
READ_ONCE         287 arch/s390/kvm/vsie.c 	const uint32_t crycbd_o = READ_ONCE(scb_o->crycbd);
READ_ONCE         364 arch/s390/kvm/vsie.c 	const uint32_t new_ibc = READ_ONCE(__new_ibc) & 0x0fffU;
READ_ONCE         443 arch/s390/kvm/vsie.c 	const uint32_t new_prefix = READ_ONCE(__new_prefix);
READ_ONCE         444 arch/s390/kvm/vsie.c 	const bool wants_tx = READ_ONCE(scb_o->ecb) & ECB_TE;
READ_ONCE         492 arch/s390/kvm/vsie.c 		new_mso = READ_ONCE(scb_o->mso) & 0xfffffffffff00000UL;
READ_ONCE         580 arch/s390/kvm/vsie.c 		page = READ_ONCE(kvm->arch.vsie.pages[i]);
READ_ONCE         584 arch/s390/kvm/vsie.c 		if (READ_ONCE(cur->gmap) != gmap)
READ_ONCE         727 arch/s390/kvm/vsie.c 	gpa = READ_ONCE(scb_o->scaol) & ~0xfUL;
READ_ONCE         729 arch/s390/kvm/vsie.c 		gpa |= (u64) READ_ONCE(scb_o->scaoh) << 32;
READ_ONCE         750 arch/s390/kvm/vsie.c 	gpa = READ_ONCE(scb_o->itdba) & ~0xffUL;
READ_ONCE         766 arch/s390/kvm/vsie.c 	gpa = READ_ONCE(scb_o->gvrd) & ~0x1ffUL;
READ_ONCE         785 arch/s390/kvm/vsie.c 	gpa = READ_ONCE(scb_o->riccbd) & ~0x3fUL;
READ_ONCE         805 arch/s390/kvm/vsie.c 		gpa = READ_ONCE(scb_o->sdnxo) & ~0xfUL;
READ_ONCE         806 arch/s390/kvm/vsie.c 		sdnxc = READ_ONCE(scb_o->sdnxo) & 0xfUL;
READ_ONCE         972 arch/s390/kvm/vsie.c 	__u32 fac = READ_ONCE(vsie_page->scb_o->fac) & 0x7ffffff8U;
READ_ONCE        1368 arch/s390/kvm/vsie.c 	struct kvm_s390_sie_block *scb = READ_ONCE(vcpu->arch.vsie_block);
READ_ONCE         131 arch/s390/lib/spinlock.c 		old = READ_ONCE(lp->lock);
READ_ONCE         168 arch/s390/lib/spinlock.c 		while (READ_ONCE(node->prev) != NULL) {
READ_ONCE         182 arch/s390/lib/spinlock.c 		old = READ_ONCE(lp->lock);
READ_ONCE         202 arch/s390/lib/spinlock.c 		while ((next = READ_ONCE(node->next)) == NULL)
READ_ONCE         218 arch/s390/lib/spinlock.c 	owner = arch_spin_yield_target(READ_ONCE(lp->lock), NULL);
READ_ONCE         259 arch/s390/lib/spinlock.c 		owner = READ_ONCE(lp->lock);
READ_ONCE         273 arch/s390/lib/spinlock.c 		while (READ_ONCE(rw->cnts) & 0x10000)
READ_ONCE         285 arch/s390/lib/spinlock.c 	while (READ_ONCE(rw->cnts) & 0x10000)
READ_ONCE         302 arch/s390/lib/spinlock.c 		old = READ_ONCE(rw->cnts);
READ_ONCE         318 arch/s390/lib/spinlock.c 	cpu = READ_ONCE(lp->lock) & _Q_LOCK_CPU_MASK;
READ_ONCE          36 arch/s390/mm/pgtable.c 		asce = READ_ONCE(mm->context.gmap_asce);
READ_ONCE          56 arch/s390/mm/pgtable.c 		asce = READ_ONCE(mm->context.gmap_asce);
READ_ONCE          24 arch/sh/include/asm/atomic.h #define atomic_read(v)		READ_ONCE((v)->counter)
READ_ONCE          32 arch/sh/include/asm/cmpxchg-xchg.h 		oldv = READ_ONCE(*p);
READ_ONCE          37 arch/sparc/include/asm/atomic_32.h #define atomic_read(v)          READ_ONCE((v)->counter)
READ_ONCE          18 arch/sparc/include/asm/atomic_64.h #define atomic_read(v)		READ_ONCE((v)->counter)
READ_ONCE          19 arch/sparc/include/asm/atomic_64.h #define atomic64_read(v)	READ_ONCE((v)->counter)
READ_ONCE          50 arch/sparc/include/asm/barrier_64.h 	typeof(*p) ___p1 = READ_ONCE(*p);				\
READ_ONCE          45 arch/sparc/include/asm/vvar.h 	ret = READ_ONCE(s->seq);
READ_ONCE          80 arch/x86/entry/common.c 	work = READ_ONCE(ti->flags);
READ_ONCE         173 arch/x86/entry/common.c 		cached_flags = READ_ONCE(current_thread_info()->flags);
READ_ONCE         191 arch/x86/entry/common.c 	cached_flags = READ_ONCE(ti->flags);
READ_ONCE         197 arch/x86/entry/common.c 	cached_flags = READ_ONCE(ti->flags);
READ_ONCE         256 arch/x86/entry/common.c 	u32 cached_flags = READ_ONCE(ti->flags);
READ_ONCE         285 arch/x86/entry/common.c 	if (READ_ONCE(ti->flags) & _TIF_WORK_SYSCALL_ENTRY)
READ_ONCE         320 arch/x86/entry/common.c 	if (READ_ONCE(ti->flags) & _TIF_WORK_SYSCALL_ENTRY) {
READ_ONCE         957 arch/x86/events/core.c 		    READ_ONCE(cpuc->excl_cntrs->exclusive_present))
READ_ONCE        2113 arch/x86/events/core.c 	if (READ_ONCE(x86_pmu.attr_rdpmc) &&
READ_ONCE        2407 arch/x86/events/core.c 		ldt = READ_ONCE(current->active_mm->context.ldt);
READ_ONCE         304 arch/x86/events/intel/bts.c 	int state = READ_ONCE(bts->state);
READ_ONCE         336 arch/x86/events/intel/bts.c 	int state = READ_ONCE(bts->state);
READ_ONCE         361 arch/x86/events/intel/bts.c 	if (READ_ONCE(bts->state) != BTS_STATE_ACTIVE)
READ_ONCE         464 arch/x86/events/intel/bts.c 	if (READ_ONCE(bts->state) == BTS_STATE_STOPPED)
READ_ONCE         504 arch/x86/events/intel/pt.c 	if (READ_ONCE(pt->vmx_on))
READ_ONCE         513 arch/x86/events/intel/pt.c 	u64 ctl = READ_ONCE(event->hw.config);
READ_ONCE         520 arch/x86/events/intel/pt.c 	if (!READ_ONCE(pt->vmx_on))
READ_ONCE        1345 arch/x86/events/intel/pt.c 	if (!READ_ONCE(pt->handle_nmi))
READ_ONCE          51 arch/x86/hyperv/hv_spinlock.c 	if (READ_ONCE(*byte) == val)
READ_ONCE          31 arch/x86/include/asm/atomic.h 	return READ_ONCE((v)->counter);
READ_ONCE          22 arch/x86/include/asm/atomic64_64.h 	return READ_ONCE((v)->counter);
READ_ONCE          75 arch/x86/include/asm/barrier.h 	typeof(*p) ___p1 = READ_ONCE(*p);				\
READ_ONCE         106 arch/x86/include/asm/mmu_context.h 	ldt = READ_ONCE(mm->context.ldt);
READ_ONCE          32 arch/x86/include/asm/switch_to.h 	READ_ONCE(*(unsigned char *)next->thread.sp);
READ_ONCE         111 arch/x86/include/asm/unwind.h 		val = READ_ONCE(x);			\
READ_ONCE          33 arch/x86/include/asm/vdso/vsyscall.h 	WRITE_ONCE(vclocks_used, READ_ONCE(vclocks_used) | (1 << vclock_mode));
READ_ONCE          21 arch/x86/include/asm/vgtod.h 	return READ_ONCE(vclocks_used) & (1 << vclock);
READ_ONCE         913 arch/x86/kernel/apic/apic.c 	jif_start = READ_ONCE(jiffies);
READ_ONCE         936 arch/x86/kernel/apic/apic.c 				unsigned long jif_now = READ_ONCE(jiffies);
READ_ONCE         245 arch/x86/kernel/cpu/mce/dev-mcelog.c 	if (READ_ONCE(mcelog.next))
READ_ONCE         879 arch/x86/kernel/cpu/resctrl/pseudo_lock.c 	mem_r = READ_ONCE(plr->kmem);
READ_ONCE         981 arch/x86/kernel/cpu/resctrl/pseudo_lock.c 	line_size = READ_ONCE(plr->line_size);
READ_ONCE         982 arch/x86/kernel/cpu/resctrl/pseudo_lock.c 	mem_r = READ_ONCE(plr->kmem);
READ_ONCE         983 arch/x86/kernel/cpu/resctrl/pseudo_lock.c 	size = READ_ONCE(plr->size);
READ_ONCE          41 arch/x86/kernel/cpu/umwait.c 	wrmsr(MSR_IA32_UMWAIT_CONTROL, READ_ONCE(umwait_control_cached), 0);
READ_ONCE         134 arch/x86/kernel/cpu/umwait.c 	u32 ctrl = READ_ONCE(umwait_control_cached);
READ_ONCE         153 arch/x86/kernel/cpu/umwait.c 	ctrl = READ_ONCE(umwait_control_cached);
READ_ONCE         166 arch/x86/kernel/cpu/umwait.c 	u32 ctrl = READ_ONCE(umwait_control_cached);
READ_ONCE         188 arch/x86/kernel/cpu/umwait.c 	ctrl = READ_ONCE(umwait_control_cached);
READ_ONCE         150 arch/x86/kernel/espfix_64.c 	stack_page = READ_ONCE(espfix_pages[page]);
READ_ONCE         157 arch/x86/kernel/espfix_64.c 	stack_page = READ_ONCE(espfix_pages[page]);
READ_ONCE        1254 arch/x86/kernel/fpu/xstate.c 	unsigned long timestamp = READ_ONCE(task->thread.fpu.avx512_timestamp);
READ_ONCE         699 arch/x86/kernel/hpet.c 	old.lockval = READ_ONCE(hpet.lockval);
READ_ONCE         732 arch/x86/kernel/hpet.c 		new.lockval = READ_ONCE(hpet.lockval);
READ_ONCE         598 arch/x86/kernel/kvm.c 		state = READ_ONCE(src->preempted);
READ_ONCE         779 arch/x86/kernel/kvm.c 	if (READ_ONCE(*ptr) != val)
READ_ONCE         109 arch/x86/kernel/nmi.c 	u64 whole_msecs = READ_ONCE(action->max_duration);
READ_ONCE         498 arch/x86/kernel/process.c 	tifn = READ_ONCE(task_thread_info(next_p)->flags);
READ_ONCE         499 arch/x86/kernel/process.c 	tifp = READ_ONCE(task_thread_info(prev_p)->flags);
READ_ONCE         837 arch/x86/kernel/process.c 	sp = READ_ONCE(p->thread.sp);
READ_ONCE        1531 arch/x86/kvm/hyperv.c 	return READ_ONCE(kvm->arch.hyperv.hv_hypercall) & HV_X64_MSR_HYPERCALL_ENABLE;
READ_ONCE         395 arch/x86/kvm/lapic.c 		pir_val = READ_ONCE(pir[i]);
READ_ONCE         389 arch/x86/kvm/mmu.c 	return READ_ONCE(nx_huge_pages);
READ_ONCE         663 arch/x86/kvm/mmu.c 	return READ_ONCE(*sptep);
READ_ONCE        5446 arch/x86/kvm/mmu.c 	if (!READ_ONCE(vcpu->kvm->arch.indirect_shadow_pages))
READ_ONCE        6408 arch/x86/kvm/mmu.c 	if (READ_ONCE(nx_huge_pages) &&
READ_ONCE        6434 arch/x86/kvm/mmu.c 	ratio = READ_ONCE(nx_huge_pages_recovery_ratio);
READ_ONCE        6462 arch/x86/kvm/mmu.c 	return READ_ONCE(nx_huge_pages) && READ_ONCE(nx_huge_pages_recovery_ratio)
READ_ONCE         159 arch/x86/kvm/page_track.c 	return !!READ_ONCE(slot->arch.gfn_track[mode][index]);
READ_ONCE         498 arch/x86/kvm/svm.c 	return (READ_ONCE(*entry) & AVIC_PHYSICAL_ID_ENTRY_IS_RUNNING_MASK);
READ_ONCE        2093 arch/x86/kvm/svm.c 	entry = READ_ONCE(*(svm->avic_physical_id_cache));
READ_ONCE        2116 arch/x86/kvm/svm.c 	entry = READ_ONCE(*(svm->avic_physical_id_cache));
READ_ONCE        4622 arch/x86/kvm/svm.c 	new_entry = READ_ONCE(*entry);
READ_ONCE        7446 arch/x86/kvm/x86.c 	if (target && READ_ONCE(target->ready))
READ_ONCE        9993 arch/x86/kvm/x86.c 	if (READ_ONCE(vcpu->arch.pv.pv_unhalted))
READ_ONCE         136 arch/x86/mm/pkeys.c 	u32 init_pkru_value_snapshot = READ_ONCE(init_pkru_value);
READ_ONCE         559 arch/x86/xen/p2m.c 		mid_mfn = READ_ONCE(p2m_top_mfn_p[topidx]);
READ_ONCE         589 arch/x86/xen/p2m.c 	p2m_pfn = pte_pfn(READ_ONCE(*ptep));
READ_ONCE          52 arch/x86/xen/spinlock.c 	} else if (READ_ONCE(*byte) == val) {
READ_ONCE          48 arch/xtensa/include/asm/atomic.h #define atomic_read(v)		READ_ONCE((v)->counter)
READ_ONCE         188 arch/xtensa/include/asm/cmpxchg.h 		oldv = READ_ONCE(*p);
READ_ONCE         224 arch/xtensa/kernel/smp.c 			ccount = READ_ONCE(cpu_start_ccount);
READ_ONCE         315 arch/xtensa/kernel/smp.c 		if (READ_ONCE(cpu_start_id) == -cpu) {
READ_ONCE          88 arch/xtensa/platforms/iss/simdisk.c 		READ_ONCE(*buffer);
READ_ONCE        1759 block/bio.c    	stamp = READ_ONCE(part->stamp);
READ_ONCE         985 block/blk-iocost.c 		u32 active_sum = READ_ONCE(parent->child_active_sum);
READ_ONCE         986 block/blk-iocost.c 		u32 inuse_sum = READ_ONCE(parent->child_inuse_sum);
READ_ONCE         987 block/blk-iocost.c 		u32 active = READ_ONCE(child->active);
READ_ONCE         988 block/blk-iocost.c 		u32 inuse = READ_ONCE(child->inuse);
READ_ONCE        1289 block/blk-iocost.c 			u32 this_met = READ_ONCE(stat->missed[rw].nr_met);
READ_ONCE        1290 block/blk-iocost.c 			u32 this_missed = READ_ONCE(stat->missed[rw].nr_missed);
READ_ONCE        1298 block/blk-iocost.c 		this_rq_wait_ns = READ_ONCE(stat->rq_wait_ns);
READ_ONCE         410 block/blk-iolatency.c 	scale_lat = READ_ONCE(lat_info->scale_lat);
READ_ONCE         138 block/blk-mq-sched.c 	struct blk_mq_ctx *ctx = READ_ONCE(hctx->dispatch_from);
READ_ONCE         878 block/blk-mq.c 	deadline = READ_ONCE(rq->deadline);
READ_ONCE         138 block/blk-mq.h 	return READ_ONCE(rq->state);
READ_ONCE         218 block/blk-wbt.c 	u64 now, issue = READ_ONCE(rwb->sync_issue);
READ_ONCE         694 block/kyber-iosched.c 	struct blk_mq_hw_ctx *hctx = READ_ONCE(wqe->private);
READ_ONCE        6022 drivers/android/binder.c 	int debug_id = READ_ONCE(e->debug_id_done);
READ_ONCE        6041 drivers/android/binder.c 	seq_printf(m, debug_id && debug_id == READ_ONCE(e->debug_id_done) ?
READ_ONCE         951 drivers/base/core.c 	drv = READ_ONCE(dev->driver);
READ_ONCE         276 drivers/base/power/domain_governor.c 			next_hrtimer = READ_ONCE(dev->next_hrtimer);
READ_ONCE         270 drivers/base/power/main.c 		if (READ_ONCE(link->status) != DL_STATE_DORMANT)
READ_ONCE         327 drivers/base/power/main.c 		if (READ_ONCE(link->status) != DL_STATE_DORMANT)
READ_ONCE         171 drivers/base/power/runtime.c 	autosuspend_delay = READ_ONCE(dev->power.autosuspend_delay);
READ_ONCE         175 drivers/base/power/runtime.c 	expires  = READ_ONCE(dev->power.last_busy);
READ_ONCE         295 drivers/base/power/runtime.c 		    READ_ONCE(link->status) == DL_STATE_SUPPLIER_UNBIND)
READ_ONCE         315 drivers/base/power/runtime.c 		if (READ_ONCE(link->status) == DL_STATE_SUPPLIER_UNBIND)
READ_ONCE         559 drivers/base/power/runtime.c 				u64 slack = (u64)READ_ONCE(dev->power.autosuspend_delay) *
READ_ONCE         989 drivers/block/xen-blkback/blkback.c 		first_sect = READ_ONCE(segments[i].first_sect);
READ_ONCE         990 drivers/block/xen-blkback/blkback.c 		last_sect = READ_ONCE(segments[i].last_sect);
READ_ONCE         399 drivers/block/xen-blkback/common.h 	dst->operation = READ_ONCE(src->operation);
READ_ONCE         447 drivers/block/xen-blkback/common.h 	dst->operation = READ_ONCE(src->operation);
READ_ONCE         764 drivers/char/ipmi/ipmi_msghandler.c 		int intf_num = READ_ONCE(intf->intf_num);
READ_ONCE        1957 drivers/char/ipmi/ipmi_msghandler.c 	chans = READ_ONCE(intf->channel_list)->c;
READ_ONCE        2096 drivers/char/ipmi/ipmi_msghandler.c 	chans = READ_ONCE(intf->channel_list)->c;
READ_ONCE        4265 drivers/char/ipmi/ipmi_msghandler.c 			chans = READ_ONCE(intf->channel_list)->c;
READ_ONCE        4358 drivers/char/ipmi/ipmi_msghandler.c 		chans = READ_ONCE(intf->channel_list)->c;
READ_ONCE         722 drivers/char/random.c 	entropy_count = orig = READ_ONCE(r->entropy_count);
READ_ONCE        1300 drivers/char/random.c 	idx = READ_ONCE(f->reg_idx);
READ_ONCE        1457 drivers/char/random.c 	entropy_count = orig = READ_ONCE(r->entropy_count);
READ_ONCE        1683 drivers/char/random.c 	    (previous && (caller == READ_ONCE(*previous))))
READ_ONCE          61 drivers/clocksource/bcm2835_timer.c 		event_handler = READ_ONCE(timer->evt.event_handler);
READ_ONCE         293 drivers/cpufreq/cpufreq_governor.c 	lst = READ_ONCE(policy_dbs->last_sample_time);
READ_ONCE         311 drivers/cpufreq/cpufreq_governor.c 		if (unlikely(lst != READ_ONCE(policy_dbs->last_sample_time))) {
READ_ONCE        1597 drivers/cpufreq/intel_pstate.c 	u64 hwp_req = READ_ONCE(cpu->hwp_req_cached);
READ_ONCE        1087 drivers/crypto/amcc/crypto4xx_core.c 		     ((READ_ONCE(pd->pd_ctl.w) &
READ_ONCE         205 drivers/crypto/caam/jr.c 		head = READ_ONCE(jrp->head);
READ_ONCE         374 drivers/crypto/caam/jr.c 	tail = READ_ONCE(jrp->tail);
READ_ONCE         510 drivers/crypto/cavium/nitrox/nitrox_reqmgr.c 	u64 orh = READ_ONCE(*sr->resp.orh);
READ_ONCE         516 drivers/crypto/cavium/nitrox/nitrox_reqmgr.c 	while (READ_ONCE(*sr->resp.completion) == PENDING_SIG) {
READ_ONCE         559 drivers/crypto/cavium/nitrox/nitrox_reqmgr.c 					    READ_ONCE(*sr->resp.orh));
READ_ONCE         568 drivers/crypto/cavium/nitrox/nitrox_reqmgr.c 		err = READ_ONCE(*sr->resp.orh) & 0xff;
READ_ONCE         186 drivers/crypto/nx/nx-842-powernv.c 	while (!(READ_ONCE(csb->flags) & CSB_V)) {
READ_ONCE         489 drivers/devfreq/tegra30-devfreq.c 	cur_freq = READ_ONCE(tegra->cur_freq);
READ_ONCE         485 drivers/dma-buf/st-dma-fence.c 		if (!READ_ONCE(cb.seen)) {
READ_ONCE         568 drivers/dma/mediatek/mtk-hsdma.c 		desc2 = READ_ONCE(rxd->desc2);
READ_ONCE         172 drivers/edac/altera_edac.c 	reg = READ_ONCE(ptemp[0]);
READ_ONCE         173 drivers/edac/altera_edac.c 	read_reg = READ_ONCE(ptemp[1]);
READ_ONCE         678 drivers/edac/altera_edac.c 		if (READ_ONCE(ptemp[i]))
READ_ONCE         695 drivers/edac/altera_edac.c 		if (READ_ONCE(ptemp[i]) != i)
READ_ONCE         724 drivers/firewire/ohci.c 	res_count = READ_ONCE(ctx->descriptors[i].res_count);
READ_ONCE         732 drivers/firewire/ohci.c 		next_res_count = READ_ONCE(ctx->descriptors[next_i].res_count);
READ_ONCE         748 drivers/firewire/ohci.c 				next_res_count = READ_ONCE(ctx->descriptors[next_i].res_count);
READ_ONCE        2806 drivers/firewire/ohci.c 	res_count = le16_to_cpu(READ_ONCE(last->res_count));
READ_ONCE         547 drivers/firmware/efi/efi.c 			size = READ_ONCE(seed->size);
READ_ONCE          80 drivers/firmware/qemu_fw_cfg.c 		u32 ctrl = be32_to_cpu(READ_ONCE(d->control));
READ_ONCE         119 drivers/firmware/qemu_fw_cfg.c 	if (be32_to_cpu(READ_ONCE(d->control)) & FW_CFG_DMA_CTL_ERROR) {
READ_ONCE          97 drivers/firmware/tegra/ivc.c 	u32 tx = READ_ONCE(header->tx.count);
READ_ONCE          98 drivers/firmware/tegra/ivc.c 	u32 rx = READ_ONCE(header->rx.count);
READ_ONCE         119 drivers/firmware/tegra/ivc.c 	u32 tx = READ_ONCE(header->tx.count);
READ_ONCE         120 drivers/firmware/tegra/ivc.c 	u32 rx = READ_ONCE(header->rx.count);
READ_ONCE         132 drivers/firmware/tegra/ivc.c 	u32 tx = READ_ONCE(header->tx.count);
READ_ONCE         133 drivers/firmware/tegra/ivc.c 	u32 rx = READ_ONCE(header->rx.count);
READ_ONCE         147 drivers/firmware/tegra/ivc.c 		   READ_ONCE(ivc->tx.channel->tx.count) + 1);
READ_ONCE         158 drivers/firmware/tegra/ivc.c 		   READ_ONCE(ivc->rx.channel->rx.count) + 1);
READ_ONCE         423 drivers/firmware/tegra/ivc.c 	state = READ_ONCE(ivc->rx.channel->tx.state);
READ_ONCE         309 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c 	uint64_t seq = READ_ONCE(ring->fence_drv.sync_seq);
READ_ONCE         373 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c 	emitted += READ_ONCE(ring->fence_drv.sync_seq);
READ_ONCE         821 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c 	pin_count = READ_ONCE(bo->pin_count);
READ_ONCE         825 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c 	dma_buf = READ_ONCE(bo->tbo.base.dma_buf);
READ_ONCE         826 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c 	attachment = READ_ONCE(bo->tbo.base.import_attach);
READ_ONCE         566 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 		wptr = READ_ONCE(*((u64 *)&adev->wb.wb[ring->wptr_offs]));
READ_ONCE         635 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 		wptr = READ_ONCE(*((u64 *)&adev->wb.wb[ring->wptr_offs]));
READ_ONCE         864 drivers/gpu/drm/drm_dp_mst_topology.c 	state = READ_ONCE(txmsg->state);
READ_ONCE         246 drivers/gpu/drm/drm_fb_helper.c 	if (READ_ONCE(fb_helper->deferred_setup))
READ_ONCE         390 drivers/gpu/drm/drm_lease.c 	bool universal_planes = READ_ONCE(lessor_priv->universal_planes);
READ_ONCE         159 drivers/gpu/drm/drm_sysfs.c 	status = READ_ONCE(connector->status);
READ_ONCE         172 drivers/gpu/drm/drm_sysfs.c 	dpms = READ_ONCE(connector->dpms);
READ_ONCE         185 drivers/gpu/drm/drm_sysfs.c 	enabled = READ_ONCE(connector->encoder);
READ_ONCE         421 drivers/gpu/drm/drm_vblank.c 		WARN_ON(READ_ONCE(vblank->enabled) &&
READ_ONCE        1249 drivers/gpu/drm/drm_vblank.c 	WARN_ON(!READ_ONCE(vblank->inmodeset));
READ_ONCE        1487 drivers/gpu/drm/drm_vblank.c 	if (!READ_ONCE(vblank->enabled)) {
READ_ONCE        1631 drivers/gpu/drm/drm_vblank.c 	    READ_ONCE(vblank->enabled)) {
READ_ONCE        1678 drivers/gpu/drm/drm_vblank.c 				      !READ_ONCE(vblank->enabled),
READ_ONCE        1850 drivers/gpu/drm/drm_vblank.c 	vblank_enabled = dev->vblank_disable_immediate && READ_ONCE(vblank->enabled);
READ_ONCE        1947 drivers/gpu/drm/drm_vblank.c 	if (!READ_ONCE(vblank->enabled)) {
READ_ONCE         324 drivers/gpu/drm/etnaviv/etnaviv_buffer.c 	unsigned int new_flush_seq = READ_ONCE(gpu->mmu_context->flush_seq);
READ_ONCE        1240 drivers/gpu/drm/i915/display/intel_fbc.c 	if (READ_ONCE(fbc->underrun_detected))
READ_ONCE        2154 drivers/gpu/drm/i915/display/intel_hdmi.c 		READ_ONCE(to_intel_digital_connector_state(connector->state)->force_audio) == HDMI_AUDIO_OFF_DVI;
READ_ONCE         609 drivers/gpu/drm/i915/display/intel_hotplug.c 	enabled = READ_ONCE(dev_priv->hotplug.poll_enabled);
READ_ONCE         901 drivers/gpu/drm/i915/display/intel_psr.c 	if (!CAN_PSR(dev_priv) || READ_ONCE(psr->dp) != intel_dp)
READ_ONCE         956 drivers/gpu/drm/i915/display/intel_psr.c 	if (READ_ONCE(dev_priv->psr.psr2_enabled))
READ_ONCE        1100 drivers/gpu/drm/i915/display/intel_psr.c 	if (READ_ONCE(dev_priv->psr.irq_aux_error))
READ_ONCE        1978 drivers/gpu/drm/i915/gem/i915_gem_context.c 		vm = READ_ONCE(src->vm);
READ_ONCE        2000 drivers/gpu/drm/i915/gem/i915_gem_context.c 		if (vm == READ_ONCE(src->vm))
READ_ONCE          30 drivers/gpu/drm/i915/gem/i915_gem_domain.c 	if (!READ_ONCE(obj->pin_global))
READ_ONCE         618 drivers/gpu/drm/i915/gem/i915_gem_domain.c 	if (READ_ONCE(obj->write_domain) == read_domains) {
READ_ONCE         170 drivers/gpu/drm/i915/gem/i915_gem_object.h 	return READ_ONCE(obj->frontbuffer);
READ_ONCE         256 drivers/gpu/drm/i915/gem/i915_gem_object.h 	return !IS_ERR_OR_NULL(READ_ONCE(obj->mm.pages));
READ_ONCE         395 drivers/gpu/drm/i915/gem/i915_gem_pages.c 	if (n < READ_ONCE(iter->sg_idx))
READ_ONCE          81 drivers/gpu/drm/i915/gem/i915_gem_shrinker.c 	if (READ_ONCE(obj->pin_global))
READ_ONCE         322 drivers/gpu/drm/i915/gem/i915_gem_shrinker.c 	count = READ_ONCE(i915->mm.shrink_memory) >> PAGE_SHIFT;
READ_ONCE         323 drivers/gpu/drm/i915/gem/i915_gem_shrinker.c 	num_objects = READ_ONCE(i915->mm.shrink_count);
READ_ONCE         409 drivers/gpu/drm/i915/gem/i915_gem_tiling.c 			READ_ONCE(obj->tiling_and_stride) & TILING_MASK;
READ_ONCE         136 drivers/gpu/drm/i915/gt/intel_engine.h 	return READ_ONCE(*execlists->active);
READ_ONCE         160 drivers/gpu/drm/i915/gt/intel_engine.h 	return READ_ONCE(engine->status_page.addr[reg]);
READ_ONCE         449 drivers/gpu/drm/i915/gt/intel_engine.h 	if (READ_ONCE(engine->stats.enabled) == 0)
READ_ONCE         467 drivers/gpu/drm/i915/gt/intel_engine.h 	if (READ_ONCE(engine->stats.enabled) == 0)
READ_ONCE        1093 drivers/gpu/drm/i915/gt/intel_engine_cs.c 	if (!READ_ONCE(gt->awake))
READ_ONCE        1263 drivers/gpu/drm/i915/gt/intel_engine_cs.c 		write = READ_ONCE(*execlists->csb_write);
READ_ONCE         267 drivers/gpu/drm/i915/gt/intel_hangcheck.c 	if (!READ_ONCE(gt->awake))
READ_ONCE         363 drivers/gpu/drm/i915/gt/intel_lrc.c 			next = READ_ONCE(ve->request);
READ_ONCE         582 drivers/gpu/drm/i915/gt/intel_lrc.c 	old = READ_ONCE(ce->inflight);
READ_ONCE         597 drivers/gpu/drm/i915/gt/intel_lrc.c 	struct i915_request *next = READ_ONCE(ve->request);
READ_ONCE         636 drivers/gpu/drm/i915/gt/intel_lrc.c 	old = READ_ONCE(ce->inflight);
READ_ONCE         928 drivers/gpu/drm/i915/gt/intel_lrc.c 	struct i915_request * const *last = READ_ONCE(execlists->active);
READ_ONCE        1072 drivers/gpu/drm/i915/gt/intel_lrc.c 		struct i915_request *rq = READ_ONCE(ve->request);
READ_ONCE        1220 drivers/gpu/drm/i915/gt/intel_lrc.c 				GEM_BUG_ON(READ_ONCE(ve->context.inflight));
READ_ONCE        1506 drivers/gpu/drm/i915/gt/intel_lrc.c 	tail = READ_ONCE(*execlists->csb_write);
READ_ONCE        1638 drivers/gpu/drm/i915/gt/intel_lrc.c 	if (!READ_ONCE(engine->execlists.pending[0])) {
READ_ONCE        3530 drivers/gpu/drm/i915/gt/intel_lrc.c 	rq = READ_ONCE(ve->request);
READ_ONCE        3564 drivers/gpu/drm/i915/gt/intel_lrc.c 	for (n = 0; READ_ONCE(ve->request) && n < ve->num_siblings; n++) {
READ_ONCE        3698 drivers/gpu/drm/i915/gt/intel_lrc.c 	exec = READ_ONCE(rq->execution_mask);
READ_ONCE        3976 drivers/gpu/drm/i915/gt/intel_lrc.c 		struct i915_request *rq = READ_ONCE(ve->request);
READ_ONCE         632 drivers/gpu/drm/i915/gt/intel_reset.c 		vma = READ_ONCE(gt->ggtt->fence_regs[i].vma);
READ_ONCE         272 drivers/gpu/drm/i915/gt/selftest_hangcheck.c 	return READ_ONCE(h->seqno[rq->fence.context % (PAGE_SIZE/sizeof(u32))]);
READ_ONCE         412 drivers/gpu/drm/i915/gt/selftest_lrc.c 		if (wait_for(READ_ONCE(*map), 10)) {
READ_ONCE         459 drivers/gpu/drm/i915/gt/selftest_lrc.c 		GEM_BUG_ON(READ_ONCE(*map));
READ_ONCE         384 drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c #define done (READ_ONCE(desc->fence) == fence)
READ_ONCE         430 drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c #define done INTEL_GUC_MSG_IS_RESPONSE(READ_ONCE(req->status))
READ_ONCE         416 drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c 	wq_off = READ_ONCE(desc->tail);
READ_ONCE         417 drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c 	GEM_BUG_ON(CIRC_SPACE(wq_off, READ_ONCE(desc->head),
READ_ONCE         456 drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c 	cookie = READ_ONCE(db->cookie);
READ_ONCE          69 drivers/gpu/drm/i915/gt/uc/selftest_guc.c 	err = wait_for(READ_ONCE(desc->head) == READ_ONCE(desc->tail), 10);
READ_ONCE          58 drivers/gpu/drm/i915/i915_active.c 	return (struct intel_engine_cs *)READ_ONCE(node->base.link.prev);
READ_ONCE         195 drivers/gpu/drm/i915/i915_active.c 	node = READ_ONCE(ref->cache);
READ_ONCE          97 drivers/gpu/drm/i915/i915_debugfs.c 	return READ_ONCE(obj->userfault_count) ? 'g' : ' ';
READ_ONCE        1151 drivers/gpu/drm/i915/i915_debugfs.c 			   READ_ONCE(fw_domain->wake_count));
READ_ONCE        1735 drivers/gpu/drm/i915/i915_debugfs.c 	seq_printf(m, "Interactive? %d\n", READ_ONCE(rps->power.interactive));
READ_ONCE        2280 drivers/gpu/drm/i915/i915_debugfs.c 	*val = READ_ONCE(dev_priv->psr.debug);
READ_ONCE         485 drivers/gpu/drm/i915/i915_gem_fence_reg.c 		struct i915_vma *vma = READ_ONCE(reg->vma);
READ_ONCE         798 drivers/gpu/drm/i915/i915_gpu_error.c 	if (READ_ONCE(error->sgl))
READ_ONCE         843 drivers/gpu/drm/i915/i915_gpu_error.c 	sg = READ_ONCE(error->fit);
READ_ONCE        1681 drivers/gpu/drm/i915/i915_gpu_error.c 	error = READ_ONCE(i915->gpu_error.first_error);
READ_ONCE        1745 drivers/gpu/drm/i915/i915_gpu_error.c 	if (READ_ONCE(i915->gpu_error.first_error))
READ_ONCE         366 drivers/gpu/drm/i915/i915_irq.c 	if (READ_ONCE(rps->interrupts_enabled))
READ_ONCE         393 drivers/gpu/drm/i915/i915_irq.c 	if (!READ_ONCE(rps->interrupts_enabled))
READ_ONCE        3090 drivers/gpu/drm/i915/i915_irq.c 			      READ_ONCE(dev_priv->i945gm_vblank.enabled) ?
READ_ONCE         262 drivers/gpu/drm/i915/i915_pmu.c 	if (!READ_ONCE(pmu->timer_enabled))
READ_ONCE         172 drivers/gpu/drm/i915/i915_request.c 	file_priv = READ_ONCE(request->file_priv);
READ_ONCE         207 drivers/gpu/drm/i915/i915_request.c 	locked = READ_ONCE(rq->engine);
READ_ONCE         209 drivers/gpu/drm/i915/i915_request.c 	while (unlikely(locked != (engine = READ_ONCE(rq->engine)))) {
READ_ONCE         579 drivers/gpu/drm/i915/i915_request.c 		if (!(READ_ONCE(rq->sched.attr.priority) & I915_PRIORITY_NOSEMAPHORE)) {
READ_ONCE         340 drivers/gpu/drm/i915/i915_request.h 	return READ_ONCE(*rq->hwsp_seqno);
READ_ONCE         164 drivers/gpu/drm/i915/i915_scheduler.c 	while (locked != (engine = READ_ONCE(rq->engine))) {
READ_ONCE         244 drivers/gpu/drm/i915/i915_scheduler.c 	if (prio <= READ_ONCE(node->attr.priority))
READ_ONCE         290 drivers/gpu/drm/i915/i915_scheduler.c 			if (prio > READ_ONCE(p->signaler->attr.priority))
READ_ONCE         380 drivers/gpu/drm/i915/i915_scheduler.c 	if (READ_ONCE(rq->sched.attr.priority) & bump)
READ_ONCE        6692 drivers/gpu/drm/i915/intel_pm.c 		if (!rps->power.interactive++ && READ_ONCE(i915->gt.awake))
READ_ONCE        6898 drivers/gpu/drm/i915/intel_pm.c 	if (READ_ONCE(rps->cur_freq) < rps->boost_freq)
READ_ONCE         150 drivers/gpu/drm/i915/intel_runtime_pm.c 		stack = READ_ONCE(rpm->debug.last_release);
READ_ONCE         790 drivers/gpu/drm/i915/intel_uncore.c 		unsigned int actual = READ_ONCE(domain->wake_count);
READ_ONCE         163 drivers/gpu/drm/i915/intel_wakeref.h 	return READ_ONCE(wf->wakeref);
READ_ONCE         548 drivers/gpu/drm/i915/selftests/i915_sw_fence.c 	if (READ_ONCE(ipc.value)) {
READ_ONCE         556 drivers/gpu/drm/i915/selftests/i915_sw_fence.c 	if (!READ_ONCE(ipc.value)) {
READ_ONCE         175 drivers/gpu/drm/i915/selftests/igt_spinner.c 	return READ_ONCE(*seqno);
READ_ONCE         448 drivers/gpu/drm/nouveau/nouveau_fbcon.c 	int state = READ_ONCE(drm->fbcon_new_state);
READ_ONCE         460 drivers/gpu/drm/radeon/radeon_gem.c 	cur_placement = READ_ONCE(robj->tbo.mem.mem_type);
READ_ONCE         490 drivers/gpu/drm/radeon/radeon_gem.c 	cur_placement = READ_ONCE(robj->tbo.mem.mem_type);
READ_ONCE         119 drivers/gpu/drm/scheduler/sched_entity.c 	if (READ_ONCE(entity->dependency))
READ_ONCE         492 drivers/gpu/drm/scheduler/sched_entity.c 	fence = READ_ONCE(entity->last_scheduled);
READ_ONCE        1235 drivers/gpu/drm/vmwgfx/vmwgfx_drv.c 		if (READ_ONCE(dev_priv->suspend_locked)) {
READ_ONCE        1475 drivers/gpu/drm/vmwgfx/vmwgfx_drv.h 	return READ_ONCE(*addr);
READ_ONCE         190 drivers/gpu/drm/vmwgfx/vmwgfx_fb.c 	if (!READ_ONCE(par->dirty.active))
READ_ONCE          86 drivers/gpu/drm/vmwgfx/vmwgfx_irq.c 	masked_status = status & READ_ONCE(dev_priv->irq_mask);
READ_ONCE         795 drivers/hv/channel_mgmt.c 			message_type = READ_ONCE(msg->header.message_type);
READ_ONCE        1071 drivers/hv/channel_mgmt.c 	while (READ_ONCE(channel->probe_done) == false) {
READ_ONCE        1134 drivers/hv/channel_mgmt.c 	while (!READ_ONCE(channel->probe_done) || !READ_ONCE(channel->rescind))
READ_ONCE         371 drivers/hv/connection.c 		callback_fn = READ_ONCE(channel->onchannel_callback);
READ_ONCE          55 drivers/hv/ring_buffer.c 	if (READ_ONCE(rbi->ring_buffer->interrupt_mask))
READ_ONCE          64 drivers/hv/ring_buffer.c 	if (old_write == READ_ONCE(rbi->ring_buffer->read_index)) {
READ_ONCE         146 drivers/hv/ring_buffer.c 	read_loc = READ_ONCE(rbi->ring_buffer->read_index);
READ_ONCE         147 drivers/hv/ring_buffer.c 	write_loc = READ_ONCE(rbi->ring_buffer->write_index);
READ_ONCE         381 drivers/hv/ring_buffer.c 	u32 write_loc = READ_ONCE(rbi->ring_buffer->write_index);
READ_ONCE         503 drivers/hv/ring_buffer.c 	pending_sz = READ_ONCE(rbi->ring_buffer->pending_send_sz);
READ_ONCE        1162 drivers/hv/vmbus_drv.c 	callback_fn = READ_ONCE(channel->onchannel_callback);
READ_ONCE         123 drivers/hwmon/xgene-hwmon.c 	val = le16_to_cpu(READ_ONCE(*addr));
READ_ONCE         151 drivers/hwmon/xgene-hwmon.c 	val = le16_to_cpu(READ_ONCE(generic_comm_base->status));
READ_ONCE        1139 drivers/hwtracing/coresight/coresight-tmc-etr.c 	sysfs_buf = READ_ONCE(drvdata->sysfs_buf);
READ_ONCE        1171 drivers/hwtracing/coresight/coresight-tmc-etr.c 	sysfs_buf = READ_ONCE(drvdata->sysfs_buf);
READ_ONCE         131 drivers/i2c/busses/i2c-xgene-slimpro.c 	val = le16_to_cpu(READ_ONCE(*addr));
READ_ONCE         191 drivers/i2c/busses/i2c-xgene-slimpro.c 	status = le16_to_cpu(READ_ONCE(generic_comm_base->status));
READ_ONCE         261 drivers/i2c/muxes/i2c-mux-pca954x.c 	idle_state = READ_ONCE(data->idle_state);
READ_ONCE         286 drivers/i2c/muxes/i2c-mux-pca954x.c 	return sprintf(buf, "%d\n", READ_ONCE(data->idle_state));
READ_ONCE         504 drivers/infiniband/core/addr.c 	struct net_device *ndev = READ_ONCE(dst->dev);
READ_ONCE        1294 drivers/infiniband/core/cache.c 		    (ndev && ((READ_ONCE(ndev->flags) & IFF_UP) == 0)))
READ_ONCE          93 drivers/infiniband/core/netlink.c 	cb_table = READ_ONCE(rdma_nl_types[type].cb_table);
READ_ONCE         104 drivers/infiniband/core/netlink.c 		cb_table = READ_ONCE(rdma_nl_types[type].cb_table);
READ_ONCE         115 drivers/infiniband/core/netlink.c 	    WARN_ON(READ_ONCE(rdma_nl_types[index].cb_table)))
READ_ONCE         744 drivers/infiniband/core/security.c 		if (!READ_ONCE(map->agent.smp_allowed))
READ_ONCE         106 drivers/infiniband/hw/efa/efa_com.c 		if (READ_ONCE(read_resp->req_id) == mmio_read->seq_num)
READ_ONCE         466 drivers/infiniband/hw/efa/efa_com.c 	while ((READ_ONCE(cqe->acq_common_descriptor.flags) &
READ_ONCE         863 drivers/infiniband/hw/efa/efa_com.c 	while ((READ_ONCE(aenq_common->flags) &
READ_ONCE        1661 drivers/infiniband/hw/hfi1/file_ops.c 			!(READ_ONCE(dd->flags) & HFI1_FROZEN),
READ_ONCE        1475 drivers/infiniband/hw/hfi1/pio.c 		sc->alloc_free = READ_ONCE(sc->free);
READ_ONCE        1482 drivers/infiniband/hw/hfi1/pio.c 			sc->alloc_free = READ_ONCE(sc->free);
READ_ONCE        1722 drivers/infiniband/hw/hfi1/pio.c 	head = READ_ONCE(sc->sr_head);	/* snapshot the head */
READ_ONCE         482 drivers/infiniband/hw/hfi1/rc.c 		if (qp->s_last == READ_ONCE(qp->s_head))
READ_ONCE         526 drivers/infiniband/hw/hfi1/rc.c 			if (qp->s_tail == READ_ONCE(qp->s_head)) {
READ_ONCE        2370 drivers/infiniband/hw/hfi1/rc.c 	if (cmp_psn(psn, READ_ONCE(qp->s_next_psn)) >= 0)
READ_ONCE        1728 drivers/infiniband/hw/hfi1/sdma.c 		swtail = READ_ONCE(sde->descq_tail) & sde->sdma_mask;
READ_ONCE        1873 drivers/infiniband/hw/hfi1/sdma.c 		swtail = READ_ONCE(sde->descq_tail) & sde->sdma_mask;
READ_ONCE        2221 drivers/infiniband/hw/hfi1/sdma.c 	tail = READ_ONCE(sde->descq_tail) & sde->sdma_mask;
READ_ONCE        3300 drivers/infiniband/hw/hfi1/sdma.c 		nr = ffz(READ_ONCE(sde->ahg_bits));
READ_ONCE         441 drivers/infiniband/hw/hfi1/sdma.h 		 READ_ONCE(sde->descq_head)) - 1;
READ_ONCE          82 drivers/infiniband/hw/hfi1/uc.c 		if (qp->s_last == READ_ONCE(qp->s_head))
READ_ONCE         121 drivers/infiniband/hw/hfi1/uc.c 		if (qp->s_cur == READ_ONCE(qp->s_head)) {
READ_ONCE         494 drivers/infiniband/hw/hfi1/ud.c 		if (qp->s_last == READ_ONCE(qp->s_head))
READ_ONCE         507 drivers/infiniband/hw/hfi1/ud.c 	if (qp->s_cur == READ_ONCE(qp->s_head))
READ_ONCE         793 drivers/infiniband/hw/hfi1/user_sdma.c 	if (READ_ONCE(req->has_error))
READ_ONCE         817 drivers/infiniband/hw/hfi1/user_sdma.c 		if (READ_ONCE(req->has_error))
READ_ONCE         843 drivers/infiniband/hw/hfi1/user_sdma.c 			if (READ_ONCE(iovec->offset) == iovec->iov.iov_len) {
READ_ONCE         176 drivers/infiniband/hw/hfi1/vnic_sdma.c 	if (unlikely(READ_ONCE(vnic_sdma->state) != HFI1_VNIC_SDMA_Q_ACTIVE))
READ_ONCE         274 drivers/infiniband/hw/hfi1/vnic_sdma.c 	return (READ_ONCE(vnic_sdma->state) == HFI1_VNIC_SDMA_Q_ACTIVE);
READ_ONCE         653 drivers/infiniband/hw/mlx5/odp.c 	current_seq = READ_ONCE(odp->notifiers_seq);
READ_ONCE         990 drivers/infiniband/hw/qedr/verbs.c 	while (oparams.num_cq_notif != READ_ONCE(cq->cnq_notif) && iter) {
READ_ONCE         996 drivers/infiniband/hw/qedr/verbs.c 	while (oparams.num_cq_notif != READ_ONCE(cq->cnq_notif) && iter) {
READ_ONCE         244 drivers/infiniband/hw/qib/qib_rc.c 		if (qp->s_last == READ_ONCE(qp->s_head))
READ_ONCE         290 drivers/infiniband/hw/qib/qib_rc.c 			if (qp->s_tail == READ_ONCE(qp->s_head))
READ_ONCE        1312 drivers/infiniband/hw/qib/qib_rc.c 	if (qib_cmp24(psn, READ_ONCE(qp->s_next_psn)) >= 0)
READ_ONCE          63 drivers/infiniband/hw/qib/qib_uc.c 		if (qp->s_last == READ_ONCE(qp->s_head))
READ_ONCE          92 drivers/infiniband/hw/qib/qib_uc.c 		if (qp->s_cur == READ_ONCE(qp->s_head))
READ_ONCE         250 drivers/infiniband/hw/qib/qib_ud.c 		if (qp->s_last == READ_ONCE(qp->s_head))
READ_ONCE         263 drivers/infiniband/hw/qib/qib_ud.c 	if (qp->s_cur == READ_ONCE(qp->s_head))
READ_ONCE         955 drivers/infiniband/sw/rdmavt/mr.c 	if (!READ_ONCE(mr->lkey_published))
READ_ONCE        1064 drivers/infiniband/sw/rdmavt/mr.c 	if (!READ_ONCE(mr->lkey_published))
READ_ONCE        1354 drivers/infiniband/sw/rdmavt/qp.c 	if (READ_ONCE(qp->s_last) != qp->s_head)
READ_ONCE        1850 drivers/infiniband/sw/rdmavt/qp.c 		if (next == READ_ONCE(wq->tail)) {
READ_ONCE        2208 drivers/infiniband/sw/rdmavt/qp.c 	call_send = qp->s_head == READ_ONCE(qp->s_last) && !wr->next;
READ_ONCE        2265 drivers/infiniband/sw/rdmavt/qp.c 		if (next == READ_ONCE(wq->tail)) {
READ_ONCE        3002 drivers/infiniband/sw/rdmavt/qp.c 	if (sqp->s_last == READ_ONCE(sqp->s_head))
READ_ONCE         110 drivers/infiniband/sw/rdmavt/rc.c 		credits = READ_ONCE(qp->r_rq.kwq->count);
READ_ONCE         117 drivers/infiniband/sw/rdmavt/rc.c 				head = READ_ONCE(qp->r_rq.kwq->head);
READ_ONCE         118 drivers/infiniband/sw/rdmavt/rc.c 				tail = READ_ONCE(qp->r_rq.kwq->tail);
READ_ONCE         643 drivers/infiniband/sw/siw/siw.h 	return READ_ONCE(sqe->flags) == 0;
READ_ONCE         650 drivers/infiniband/sw/siw/siw.h 	if (READ_ONCE(sqe->flags) & SIW_WQE_VALID)
READ_ONCE         670 drivers/infiniband/sw/siw/siw.h 	if (orq_e && READ_ONCE(orq_e->flags) == 0)
READ_ONCE         685 drivers/infiniband/sw/siw/siw.h 	if (READ_ONCE(irq_e->flags) == 0) {
READ_ONCE          56 drivers/infiniband/sw/siw/siw_cq.c 	if (READ_ONCE(cqe->flags) & SIW_WQE_VALID) {
READ_ONCE        1029 drivers/infiniband/sw/siw/siw_qp.c 	cq_notify = READ_ONCE(cq->notify->flags);
READ_ONCE        1064 drivers/infiniband/sw/siw/siw_qp.c 		if (!READ_ONCE(cqe->flags)) {
READ_ONCE        1121 drivers/infiniband/sw/siw/siw_qp.c 		if (!READ_ONCE(cqe->flags)) {
READ_ONCE        1186 drivers/infiniband/sw/siw/siw_qp.c 		if (!READ_ONCE(sqe->flags))
READ_ONCE        1222 drivers/infiniband/sw/siw/siw_qp.c 		if (!READ_ONCE(sqe->flags))
READ_ONCE        1287 drivers/infiniband/sw/siw/siw_qp.c 		if (!READ_ONCE(rqe->flags))
READ_ONCE         746 drivers/infiniband/sw/siw/siw_qp_rx.c 	if (READ_ONCE(orqe->flags) & SIW_WQE_VALID) {
READ_ONCE          72 drivers/input/misc/pwm-beeper.c 	unsigned long period = READ_ONCE(beeper->period);
READ_ONCE         233 drivers/input/misc/regulator-haptic.c 	magnitude = READ_ONCE(haptic->magnitude);
READ_ONCE        1170 drivers/iommu/arm-smmu-v3.c 		llq->val = READ_ONCE(cmdq->q.llq.val);
READ_ONCE        1176 drivers/iommu/arm-smmu-v3.c 		llq->val = READ_ONCE(smmu->cmdq.q.llq.val);
READ_ONCE        1223 drivers/iommu/arm-smmu-v3.c 	llq->val = READ_ONCE(smmu->cmdq.q.llq.val);
READ_ONCE        1322 drivers/iommu/arm-smmu-v3.c 	llq.val = READ_ONCE(cmdq->q.llq.val);
READ_ONCE         289 drivers/iommu/intel-pasid.c 	old = READ_ONCE(*ptr);
READ_ONCE         309 drivers/iommu/intel-pasid.c 	return (u16)(READ_ONCE(pe->val[1]) & GENMASK_ULL(15, 0));
READ_ONCE          59 drivers/iommu/intel-pasid.h 	return READ_ONCE(pde->val) & PASID_PTE_PRESENT;
READ_ONCE          69 drivers/iommu/intel-pasid.h 	return phys_to_virt(READ_ONCE(pde->val) & PDE_PFN_MASK);
READ_ONCE          75 drivers/iommu/intel-pasid.h 	return READ_ONCE(pte->val[0]) & PASID_PTE_PRESENT;
READ_ONCE         494 drivers/iommu/io-pgtable-arm-v7s.c 	pte = READ_ONCE(*ptep);
READ_ONCE         654 drivers/iommu/io-pgtable-arm-v7s.c 		pte[i] = READ_ONCE(ptep[i]);
READ_ONCE         738 drivers/iommu/io-pgtable-arm-v7s.c 		pte = READ_ONCE(*ptep);
READ_ONCE         403 drivers/iommu/io-pgtable-arm.c 	pte = READ_ONCE(*ptep);
READ_ONCE         611 drivers/iommu/io-pgtable-arm.c 	pte = READ_ONCE(*ptep);
READ_ONCE         678 drivers/iommu/io-pgtable-arm.c 		pte = READ_ONCE(*ptep);
READ_ONCE         174 drivers/lightnvm/pblk-rb.c 	flags = READ_ONCE(w_ctx->flags);
READ_ONCE         194 drivers/lightnvm/pblk-rb.c 	unsigned int mem = READ_ONCE(rb->mem);
READ_ONCE         195 drivers/lightnvm/pblk-rb.c 	unsigned int sync = READ_ONCE(rb->sync);
READ_ONCE         212 drivers/lightnvm/pblk-rb.c 	unsigned int mem = READ_ONCE(rb->mem);
READ_ONCE         213 drivers/lightnvm/pblk-rb.c 	unsigned int subm = READ_ONCE(rb->subm);
READ_ONCE         220 drivers/lightnvm/pblk-rb.c 	unsigned int mem = READ_ONCE(rb->mem);
READ_ONCE         221 drivers/lightnvm/pblk-rb.c 	unsigned int sync = READ_ONCE(rb->sync);
READ_ONCE         230 drivers/lightnvm/pblk-rb.c 	subm = READ_ONCE(rb->subm);
READ_ONCE         251 drivers/lightnvm/pblk-rb.c 		flags = READ_ONCE(entry->w_ctx.flags);
READ_ONCE         345 drivers/lightnvm/pblk-rb.c 	flags = READ_ONCE(entry->w_ctx.flags);
READ_ONCE         369 drivers/lightnvm/pblk-rb.c 	flags = READ_ONCE(entry->w_ctx.flags);
READ_ONCE         393 drivers/lightnvm/pblk-rb.c 	sync = READ_ONCE(rb->sync);
READ_ONCE         425 drivers/lightnvm/pblk-rb.c 	sync = READ_ONCE(rb->sync);
READ_ONCE         426 drivers/lightnvm/pblk-rb.c 	mem = READ_ONCE(rb->mem);
READ_ONCE         455 drivers/lightnvm/pblk-rb.c 	unsigned int mem = READ_ONCE(rb->mem);
READ_ONCE         583 drivers/lightnvm/pblk-rb.c 		flags = READ_ONCE(entry->w_ctx.flags);
READ_ONCE         663 drivers/lightnvm/pblk-rb.c 	flags = READ_ONCE(w_ctx->flags);
READ_ONCE         718 drivers/lightnvm/pblk-rb.c 	sync = READ_ONCE(rb->sync);
READ_ONCE         719 drivers/lightnvm/pblk-rb.c 	flush_point = READ_ONCE(rb->flush_point);
READ_ONCE         754 drivers/lightnvm/pblk-rb.c 	subm = READ_ONCE(rb->subm);
READ_ONCE          64 drivers/lightnvm/pblk-rl.c 	rb_user_active = READ_ONCE(rl->rb_user_active);
READ_ONCE          80 drivers/lightnvm/pblk-sysfs.c 				READ_ONCE(pblk->rl.rb_user_active));
READ_ONCE          36 drivers/lightnvm/pblk-write.c 		flags = READ_ONCE(w_ctx->flags);
READ_ONCE         171 drivers/lightnvm/pblk-write.c 		flags = READ_ONCE(w_ctx->flags);
READ_ONCE         112 drivers/md/bcache/closure.c 	p = READ_ONCE(s->task);
READ_ONCE         342 drivers/md/dm-bufio.c 	dm_bufio_cache_size_latch = READ_ONCE(dm_bufio_cache_size);
READ_ONCE        1544 drivers/md/dm-bufio.c 	unsigned long retain_bytes = READ_ONCE(dm_bufio_retain_bytes);
READ_ONCE        1595 drivers/md/dm-bufio.c 	unsigned long count = READ_ONCE(c->n_buffers[LIST_CLEAN]) +
READ_ONCE        1596 drivers/md/dm-bufio.c 			      READ_ONCE(c->n_buffers[LIST_DIRTY]);
READ_ONCE        1777 drivers/md/dm-bufio.c 	unsigned max_age = READ_ONCE(dm_bufio_max_age);
READ_ONCE         161 drivers/md/dm-clone-target.c 	return READ_ONCE(clone->mode);
READ_ONCE         984 drivers/md/dm-clone-target.c 	unsigned int max_batch_size = READ_ONCE(clone->hydration_batch_size);
READ_ONCE        1103 drivers/md/dm-clone-target.c 		if (current_volume > READ_ONCE(clone->hydration_threshold))
READ_ONCE        1420 drivers/md/dm-clone-target.c 	       READ_ONCE(clone->hydration_threshold),
READ_ONCE        1421 drivers/md/dm-clone-target.c 	       READ_ONCE(clone->hydration_batch_size));
READ_ONCE         381 drivers/md/dm-integrity.c 	return READ_ONCE(ic->failed);
READ_ONCE        1826 drivers/md/dm-integrity.c 		if (READ_ONCE(ic->free_sectors) <= ic->free_sectors_threshold) {
READ_ONCE        2120 drivers/md/dm-integrity.c 	if (READ_ONCE(ic->free_sectors) <= ic->free_sectors_threshold)
READ_ONCE         131 drivers/md/dm-kcopyd.c 	throttle = READ_ONCE(t->throttle);
READ_ONCE         181 drivers/md/dm-kcopyd.c 	if (likely(READ_ONCE(t->throttle) >= 100))
READ_ONCE         357 drivers/md/dm-mpath.c 	if (unlikely(READ_ONCE(m->current_pg) != pg)) {
READ_ONCE         381 drivers/md/dm-mpath.c 	if (READ_ONCE(m->next_pg)) {
READ_ONCE         397 drivers/md/dm-mpath.c 	pg = READ_ONCE(m->current_pg);
READ_ONCE         470 drivers/md/dm-mpath.c 	unsigned long flags = READ_ONCE(m->flags);
READ_ONCE         476 drivers/md/dm-mpath.c 	unsigned long flags = READ_ONCE(m->flags);
READ_ONCE         496 drivers/md/dm-mpath.c 	pgpath = READ_ONCE(m->current_pgpath);
READ_ONCE         578 drivers/md/dm-mpath.c 	pgpath = READ_ONCE(m->current_pgpath);
READ_ONCE        1858 drivers/md/dm-mpath.c 	current_pgpath = READ_ONCE(m->current_pgpath);
READ_ONCE        1879 drivers/md/dm-mpath.c 		if (!READ_ONCE(m->current_pg)) {
READ_ONCE        1948 drivers/md/dm-mpath.c 	pg = READ_ONCE(m->current_pg);
READ_ONCE        1949 drivers/md/dm-mpath.c 	next_pg = READ_ONCE(m->next_pg);
READ_ONCE        1950 drivers/md/dm-mpath.c 	if (unlikely(!READ_ONCE(m->current_pgpath) && next_pg))
READ_ONCE         644 drivers/md/dm-stats.c 			(bi_sector == (READ_ONCE(last->last_sector) &&
READ_ONCE         646 drivers/md/dm-stats.c 					(READ_ONCE(last->last_rw) == WRITE))
READ_ONCE         698 drivers/md/dm-stats.c 		shared->tmp.sectors[READ] += READ_ONCE(p->sectors[READ]);
READ_ONCE         699 drivers/md/dm-stats.c 		shared->tmp.sectors[WRITE] += READ_ONCE(p->sectors[WRITE]);
READ_ONCE         700 drivers/md/dm-stats.c 		shared->tmp.ios[READ] += READ_ONCE(p->ios[READ]);
READ_ONCE         701 drivers/md/dm-stats.c 		shared->tmp.ios[WRITE] += READ_ONCE(p->ios[WRITE]);
READ_ONCE         702 drivers/md/dm-stats.c 		shared->tmp.merges[READ] += READ_ONCE(p->merges[READ]);
READ_ONCE         703 drivers/md/dm-stats.c 		shared->tmp.merges[WRITE] += READ_ONCE(p->merges[WRITE]);
READ_ONCE         704 drivers/md/dm-stats.c 		shared->tmp.ticks[READ] += READ_ONCE(p->ticks[READ]);
READ_ONCE         705 drivers/md/dm-stats.c 		shared->tmp.ticks[WRITE] += READ_ONCE(p->ticks[WRITE]);
READ_ONCE         706 drivers/md/dm-stats.c 		shared->tmp.io_ticks[READ] += READ_ONCE(p->io_ticks[READ]);
READ_ONCE         707 drivers/md/dm-stats.c 		shared->tmp.io_ticks[WRITE] += READ_ONCE(p->io_ticks[WRITE]);
READ_ONCE         708 drivers/md/dm-stats.c 		shared->tmp.io_ticks_total += READ_ONCE(p->io_ticks_total);
READ_ONCE         709 drivers/md/dm-stats.c 		shared->tmp.time_in_queue += READ_ONCE(p->time_in_queue);
READ_ONCE         713 drivers/md/dm-stats.c 				shared->tmp.histogram[i] += READ_ONCE(p->histogram[i]);
READ_ONCE         147 drivers/md/dm-switch.c 	return (READ_ONCE(sctx->region_table[region_index]) >> bit) &
READ_ONCE        2541 drivers/md/dm-thin.c 	unsigned long no_space_timeout = READ_ONCE(no_space_timeout_secs) * HZ;
READ_ONCE         590 drivers/md/dm-verity-target.c 			unsigned cluster = READ_ONCE(dm_verity_prefetch_cluster);
READ_ONCE         411 drivers/md/dm-writecache.c #define writecache_has_error(wc)	(unlikely(READ_ONCE((wc)->error)))
READ_ONCE        1491 drivers/md/dm-writecache.c 		if (READ_ONCE(wc->writeback_size) - wbl->size >= wc->max_writeback_jobs) {
READ_ONCE         169 drivers/md/dm.c 	int param = READ_ONCE(*module_param);
READ_ONCE         191 drivers/md/dm.c 	unsigned param = READ_ONCE(*module_param);
READ_ONCE        2806 drivers/md/md.c 	unsigned long flags = READ_ONCE(rdev->flags);
READ_ONCE         972 drivers/md/raid1.c 	if (!READ_ONCE(conf->array_frozen) &&
READ_ONCE        1014 drivers/md/raid1.c 	if (!READ_ONCE(conf->array_frozen))
READ_ONCE        6098 drivers/md/raid5.c 		struct md_rdev *rdev = READ_ONCE(conf->disks[i].rdev);
READ_ONCE          74 drivers/media/dvb-core/dvb_ringbuffer.c 	free = READ_ONCE(rbuf->pread) - rbuf->pwrite;
READ_ONCE         182 drivers/misc/mei/dma-ring.c 	rd_idx = READ_ONCE(ctrl->dbuf_rd_idx) & (dbuf_depth - 1);
READ_ONCE         226 drivers/misc/mei/dma-ring.c 	rd_idx = READ_ONCE(ctrl->hbuf_rd_idx);
READ_ONCE         227 drivers/misc/mei/dma-ring.c 	wr_idx = READ_ONCE(ctrl->hbuf_wr_idx);
READ_ONCE         255 drivers/misc/mei/dma-ring.c 	wr_idx = READ_ONCE(ctrl->hbuf_wr_idx) & (hbuf_depth - 1);
READ_ONCE         271 drivers/misc/mic/scif/scif_rma_list.c 		if (READ_ONCE(ep->rma_info.async_list_del))
READ_ONCE         967 drivers/misc/vmw_balloon.c 	target = READ_ONCE(b->target);
READ_ONCE         983 drivers/misc/vmw_balloon.c 	if (target > size && time_before(jiffies, READ_ONCE(b->shrink_timeout)))
READ_ONCE        1499 drivers/misc/vmw_balloon.c 			 atomic64_read(&b->size), READ_ONCE(b->target));
READ_ONCE        1678 drivers/misc/vmw_balloon.c 	seq_printf(f, "%-22s: %16lu\n", "target", READ_ONCE(b->target));
READ_ONCE         368 drivers/mmc/core/core.c 	struct mmc_request *ongoing_mrq = READ_ONCE(host->ongoing_mrq);
READ_ONCE        1364 drivers/net/bonding/bond_alb.c 				count = slaves ? READ_ONCE(slaves->count) : 0;
READ_ONCE        1498 drivers/net/bonding/bond_alb.c 			count = slaves ? READ_ONCE(slaves->count) : 0;
READ_ONCE        1192 drivers/net/bonding/bond_main.c 	recv_probe = READ_ONCE(bond->recv_probe);
READ_ONCE        3941 drivers/net/bonding/bond_main.c 	slave_cnt = READ_ONCE(bond->slave_cnt);
READ_ONCE        4106 drivers/net/bonding/bond_main.c 	count = slaves ? READ_ONCE(slaves->count) : 0;
READ_ONCE         661 drivers/net/can/usb/ucan.c 		dlc = READ_ONCE(context->dlc);
READ_ONCE         497 drivers/net/ethernet/amazon/ena/ena_com.c 	while ((READ_ONCE(cqe->acq_common_descriptor.flags) &
READ_ONCE         832 drivers/net/ethernet/amazon/ena/ena_com.c 		if (READ_ONCE(read_resp->req_id) == mmio_read->seq_num)
READ_ONCE        2027 drivers/net/ethernet/amazon/ena/ena_com.c 	while ((READ_ONCE(aenq_common->flags) &
READ_ONCE          48 drivers/net/ethernet/amazon/ena/ena_eth_com.c 	desc_phase = (READ_ONCE(cdesc->status) & ENA_ETH_IO_RX_CDESC_BASE_PHASE_MASK) >>
READ_ONCE         265 drivers/net/ethernet/amazon/ena/ena_eth_com.c 		last = (READ_ONCE(cdesc->status) & ENA_ETH_IO_RX_CDESC_BASE_LAST_MASK) >>
READ_ONCE         260 drivers/net/ethernet/amazon/ena/ena_eth_com.h 	cdesc_phase = READ_ONCE(cdesc->flags) & ENA_ETH_IO_TX_CDESC_PHASE_MASK;
READ_ONCE         266 drivers/net/ethernet/amazon/ena/ena_eth_com.h 	*req_id = READ_ONCE(cdesc->req_id);
READ_ONCE         260 drivers/net/ethernet/broadcom/bnx2.c 	diff = READ_ONCE(txr->tx_prod) - READ_ONCE(txr->tx_cons);
READ_ONCE        2840 drivers/net/ethernet/broadcom/bnx2.c 	cons = READ_ONCE(*bnapi->hw_tx_cons_ptr);
READ_ONCE        3140 drivers/net/ethernet/broadcom/bnx2.c 	cons = READ_ONCE(*bnapi->hw_rx_cons_ptr);
READ_ONCE         114 drivers/net/ethernet/broadcom/bnxt/bnxt_xdp.c 	struct bpf_prog *xdp_prog = READ_ONCE(rxr->xdp_prog);
READ_ONCE         218 drivers/net/ethernet/broadcom/bnxt/bnxt_xdp.c 	struct bpf_prog *xdp_prog = READ_ONCE(bp->xdp_prog);
READ_ONCE         299 drivers/net/ethernet/cadence/macb_ptp.c 	unsigned long tail = READ_ONCE(queue->tx_ts_tail);
READ_ONCE         370 drivers/net/ethernet/cavium/liquidio/cn23xx_pf_device.c 		while ((READ_ONCE(reg_val) & CN23XX_PKT_INPUT_CTL_RST) &&
READ_ONCE         371 drivers/net/ethernet/cavium/liquidio/cn23xx_pf_device.c 		       !(READ_ONCE(reg_val) & CN23XX_PKT_INPUT_CTL_QUIET) &&
READ_ONCE         382 drivers/net/ethernet/cavium/liquidio/cn23xx_pf_device.c 		WRITE_ONCE(reg_val, READ_ONCE(reg_val) &
READ_ONCE         385 drivers/net/ethernet/cavium/liquidio/cn23xx_pf_device.c 				   READ_ONCE(reg_val));
READ_ONCE         389 drivers/net/ethernet/cavium/liquidio/cn23xx_pf_device.c 		if (READ_ONCE(reg_val) & CN23XX_PKT_INPUT_CTL_RST) {
READ_ONCE         884 drivers/net/ethernet/cavium/liquidio/cn23xx_pf_device.c 		WRITE_ONCE(d64, READ_ONCE(d64) &
READ_ONCE         886 drivers/net/ethernet/cavium/liquidio/cn23xx_pf_device.c 		WRITE_ONCE(d64, READ_ONCE(d64) | CN23XX_PKT_INPUT_CTL_RST);
READ_ONCE         888 drivers/net/ethernet/cavium/liquidio/cn23xx_pf_device.c 				   READ_ONCE(d64));
READ_ONCE         895 drivers/net/ethernet/cavium/liquidio/cn23xx_pf_device.c 		while (!(READ_ONCE(d64) & BIT_ULL(q_no)) && loop--) {
READ_ONCE         919 drivers/net/ethernet/cavium/liquidio/cn23xx_pf_device.c 		while (!(READ_ONCE(d64) & BIT_ULL(q_no)) && loop--) {
READ_ONCE         938 drivers/net/ethernet/cavium/liquidio/cn23xx_pf_device.c 				 READ_ONCE(d32));
READ_ONCE        1052 drivers/net/ethernet/cavium/liquidio/cn23xx_pf_device.c 		lio_pci_writeq(oct, (READ_ONCE(bar1) & 0xFFFFFFFEULL),
READ_ONCE          70 drivers/net/ethernet/cavium/liquidio/cn23xx_vf_device.c 		while ((READ_ONCE(reg_val) & CN23XX_PKT_INPUT_CTL_RST) &&
READ_ONCE          71 drivers/net/ethernet/cavium/liquidio/cn23xx_vf_device.c 		       !(READ_ONCE(reg_val) & CN23XX_PKT_INPUT_CTL_QUIET) &&
READ_ONCE          83 drivers/net/ethernet/cavium/liquidio/cn23xx_vf_device.c 		WRITE_ONCE(reg_val, READ_ONCE(reg_val) &
READ_ONCE          86 drivers/net/ethernet/cavium/liquidio/cn23xx_vf_device.c 				   READ_ONCE(reg_val));
READ_ONCE          90 drivers/net/ethernet/cavium/liquidio/cn23xx_vf_device.c 		if (READ_ONCE(reg_val) & CN23XX_PKT_INPUT_CTL_RST) {
READ_ONCE         833 drivers/net/ethernet/cavium/liquidio/lio_main.c 		vfs_mask1 = READ_ONCE(oct->sriov_info.vf_drv_loaded_mask);
READ_ONCE         834 drivers/net/ethernet/cavium/liquidio/lio_main.c 		vfs_mask2 = READ_ONCE(other_oct->sriov_info.vf_drv_loaded_mask);
READ_ONCE        4019 drivers/net/ethernet/cavium/liquidio/lio_main.c 	cores_crashed = READ_ONCE(oct->cores_crashed);
READ_ONCE          93 drivers/net/ethernet/cavium/liquidio/lio_vf_rep.c 	err = READ_ONCE(rep_resp->status) ? -EBUSY : 0;
READ_ONCE         792 drivers/net/ethernet/cavium/liquidio/request_manager.c 		if (READ_ONCE(sc->caller_is_done)) {
READ_ONCE         150 drivers/net/ethernet/cavium/liquidio/response_manager.c 				if (unlikely(READ_ONCE(sc->caller_is_done))) {
READ_ONCE         399 drivers/net/ethernet/chelsio/cxgb4/sge.c 	int hw_cidx = ntohs(READ_ONCE(q->stat->cidx));
READ_ONCE        1337 drivers/net/ethernet/chelsio/cxgb4/sge.c 	hw_cidx = ntohs(READ_ONCE(q->stat->cidx));
READ_ONCE        2002 drivers/net/ethernet/chelsio/cxgb4/sge.c 	int hw_cidx = ntohs(READ_ONCE(q->stat->cidx));
READ_ONCE         291 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c 	xdp_prog = READ_ONCE(ch->xdp.prog);
READ_ONCE        1265 drivers/net/ethernet/freescale/fec_main.c 	while (bdp != READ_ONCE(txq->bd.cur)) {
READ_ONCE        1268 drivers/net/ethernet/freescale/fec_main.c 		status = fec16_to_cpu(READ_ONCE(bdp->cbd_sc));
READ_ONCE         147 drivers/net/ethernet/google/gve/gve_adminq.c 	status = be32_to_cpu(READ_ONCE(cmd->status));
READ_ONCE         571 drivers/net/ethernet/google/gve/gve_tx.c 	return READ_ONCE(priv->counter_array[counter_index]);
READ_ONCE         456 drivers/net/ethernet/hisilicon/hip04_eth.c 	count = tx_count(READ_ONCE(priv->tx_head), tx_tail);
READ_ONCE         515 drivers/net/ethernet/hisilicon/hip04_eth.c 	count = tx_count(tx_head, READ_ONCE(priv->tx_tail));
READ_ONCE         566 drivers/net/ethernet/hisilicon/hns3/hns3_enet.h 	int end = READ_ONCE(ring->next_to_use);
READ_ONCE         584 drivers/net/ethernet/hisilicon/hns3/hns3_enet.h 	u8 __iomem *reg_addr = READ_ONCE(base);
READ_ONCE        1079 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_cmd.h 	u8 __iomem *reg_addr = READ_ONCE(base);
READ_ONCE         263 drivers/net/ethernet/hisilicon/hns3/hns3vf/hclgevf_cmd.h 	u8 __iomem *reg_addr = READ_ONCE(base);
READ_ONCE         192 drivers/net/ethernet/intel/e1000/e1000.h 	unsigned int use = READ_ONCE((R)->next_to_use);			\
READ_ONCE          20 drivers/net/ethernet/intel/fm10k/fm10k_common.h 	u32 __iomem *hw_addr = READ_ONCE((hw)->hw_addr); \
READ_ONCE          28 drivers/net/ethernet/intel/fm10k/fm10k_common.h 	u32 __iomem *sw_addr = READ_ONCE((hw)->sw_addr); \
READ_ONCE         160 drivers/net/ethernet/intel/fm10k/fm10k_iov.c 	if (!READ_ONCE(interface->iov_data))
READ_ONCE         204 drivers/net/ethernet/intel/fm10k/fm10k_iov.c 	if (!READ_ONCE(interface->iov_data))
READ_ONCE        1398 drivers/net/ethernet/intel/fm10k/fm10k_main.c 	itr_round = READ_ONCE(ring_container->itr_scale) + 8;
READ_ONCE         629 drivers/net/ethernet/intel/fm10k/fm10k_netdev.c 	int num_tx_queues = READ_ONCE(interface->num_tx_queues);
READ_ONCE        1339 drivers/net/ethernet/intel/fm10k/fm10k_netdev.c 		ring = READ_ONCE(interface->rx_ring[i]);
READ_ONCE        1355 drivers/net/ethernet/intel/fm10k/fm10k_netdev.c 		ring = READ_ONCE(interface->tx_ring[i]);
READ_ONCE        1560 drivers/net/ethernet/intel/fm10k/fm10k_netdev.c 	struct fm10k_l2_accel *l2_accel = READ_ONCE(interface->l2_accel);
READ_ONCE          51 drivers/net/ethernet/intel/fm10k/fm10k_pci.c 	u32 __iomem *hw_addr = READ_ONCE(hw->hw_addr);
READ_ONCE         361 drivers/net/ethernet/intel/fm10k/fm10k_pci.c 	hw_addr = READ_ONCE(interface->uc_addr);
READ_ONCE         560 drivers/net/ethernet/intel/fm10k/fm10k_pci.c 		struct fm10k_ring *tx_ring = READ_ONCE(interface->tx_ring[i]);
READ_ONCE         582 drivers/net/ethernet/intel/fm10k/fm10k_pci.c 		struct fm10k_ring *rx_ring = READ_ONCE(interface->rx_ring[i]);
READ_ONCE        1154 drivers/net/ethernet/intel/i40e/i40e.h 	return !!READ_ONCE(vsi->xdp_prog);
READ_ONCE         242 drivers/net/ethernet/intel/i40e/i40e_debugfs.c 		struct i40e_ring *rx_ring = READ_ONCE(vsi->rx_rings[i]);
READ_ONCE         285 drivers/net/ethernet/intel/i40e/i40e_debugfs.c 		struct i40e_ring *tx_ring = READ_ONCE(vsi->tx_rings[i]);
READ_ONCE        1382 drivers/net/ethernet/intel/i40e/i40e_ethtool.c 	flags = READ_ONCE(pf->flags);
READ_ONCE        2274 drivers/net/ethernet/intel/i40e/i40e_ethtool.c 		i40e_add_queue_stats(&data, READ_ONCE(vsi->tx_rings[i]));
READ_ONCE        2275 drivers/net/ethernet/intel/i40e/i40e_ethtool.c 		i40e_add_queue_stats(&data, READ_ONCE(vsi->rx_rings[i]));
READ_ONCE        4881 drivers/net/ethernet/intel/i40e/i40e_ethtool.c 	orig_flags = READ_ONCE(pf->flags);
READ_ONCE         455 drivers/net/ethernet/intel/i40e/i40e_main.c 		ring = READ_ONCE(vsi->tx_rings[i]);
READ_ONCE         808 drivers/net/ethernet/intel/i40e/i40e_main.c 		p = READ_ONCE(vsi->tx_rings[q]);
READ_ONCE        10403 drivers/net/ethernet/intel/i40e/i40e_main.c 			port = READ_ONCE(udp_port->port);
READ_ONCE        10404 drivers/net/ethernet/intel/i40e/i40e_main.c 			type = READ_ONCE(udp_port->type);
READ_ONCE        10405 drivers/net/ethernet/intel/i40e/i40e_main.c 			filter_index = READ_ONCE(udp_port->filter_index);
READ_ONCE         130 drivers/net/ethernet/intel/i40e/i40e_ptp.c 	adj *= READ_ONCE(pf->ptp_adj_mult);
READ_ONCE        2205 drivers/net/ethernet/intel/i40e/i40e_txrx.c 	xdp_prog = READ_ONCE(rx_ring->xdp_prog);
READ_ONCE         204 drivers/net/ethernet/intel/i40e/i40e_xsk.c 	xdp_prog = READ_ONCE(rx_ring->xdp_prog);
READ_ONCE         467 drivers/net/ethernet/intel/iavf/iavf_ethtool.c 	orig_flags = READ_ONCE(adapter->flags);
READ_ONCE        1320 drivers/net/ethernet/intel/ice/ice_ethtool.c 		ring = READ_ONCE(vsi->tx_rings[j]);
READ_ONCE        1331 drivers/net/ethernet/intel/ice/ice_ethtool.c 		ring = READ_ONCE(vsi->rx_rings[j]);
READ_ONCE        3639 drivers/net/ethernet/intel/ice/ice_main.c 		ring = READ_ONCE(vsi->tx_rings[i]);
READ_ONCE        3650 drivers/net/ethernet/intel/ice/ice_main.c 		ring = READ_ONCE(vsi->rx_rings[i]);
READ_ONCE         358 drivers/net/ethernet/intel/igb/e1000_regs.h 	u8 __iomem *hw_addr = READ_ONCE((hw)->hw_addr); \
READ_ONCE         743 drivers/net/ethernet/intel/igb/igb_main.c 	u8 __iomem *hw_addr = READ_ONCE(hw->hw_addr);
READ_ONCE        4035 drivers/net/ethernet/intel/igc/igc_main.c 	u8 __iomem *hw_addr = READ_ONCE(hw->hw_addr);
READ_ONCE         225 drivers/net/ethernet/intel/igc/igc_regs.h 	u8 __iomem *hw_addr = READ_ONCE((hw)->hw_addr); \
READ_ONCE         146 drivers/net/ethernet/intel/ixgbe/ixgbe_common.h 	u8 __iomem *reg_addr = READ_ONCE(hw->hw_addr);
READ_ONCE         165 drivers/net/ethernet/intel/ixgbe/ixgbe_common.h 	u8 __iomem *reg_addr = READ_ONCE(hw->hw_addr);
READ_ONCE         297 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 	reg_addr = READ_ONCE(hw->hw_addr);
READ_ONCE         334 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 	u8 __iomem *reg_addr = READ_ONCE(hw->hw_addr);
READ_ONCE        2206 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 	xdp_prog = READ_ONCE(rx_ring->xdp_prog);
READ_ONCE        8941 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 		struct ixgbe_ring *ring = READ_ONCE(adapter->rx_ring[i]);
READ_ONCE        8957 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 		struct ixgbe_ring *ring = READ_ONCE(adapter->tx_ring[i]);
READ_ONCE        8962 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 		struct ixgbe_ring *ring = READ_ONCE(adapter->xdp_ring[i]);
READ_ONCE         459 drivers/net/ethernet/intel/ixgbe/ixgbe_ptp.c 	incval = READ_ONCE(adapter->base_incval);
READ_ONCE          14 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c 	bool xdp_on = READ_ONCE(adapter->xdp_prog);
READ_ONCE         154 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c 	xdp_prog = READ_ONCE(rx_ring->xdp_prog);
READ_ONCE         706 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c 	if (!READ_ONCE(adapter->xdp_prog))
READ_ONCE         151 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 	u8 __iomem *reg_addr = READ_ONCE(hw->hw_addr);
READ_ONCE        1070 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 	xdp_prog = READ_ONCE(rx_ring->xdp_prog);
READ_ONCE         162 drivers/net/ethernet/intel/ixgbevf/vf.h 	u8 __iomem *reg_addr = READ_ONCE(hw->hw_addr);
READ_ONCE         769 drivers/net/ethernet/mediatek/mtk_eth_soc.c 	rxd->rxd1 = READ_ONCE(dma_rxd->rxd1);
READ_ONCE         770 drivers/net/ethernet/mediatek/mtk_eth_soc.c 	rxd->rxd2 = READ_ONCE(dma_rxd->rxd2);
READ_ONCE         771 drivers/net/ethernet/mediatek/mtk_eth_soc.c 	rxd->rxd3 = READ_ONCE(dma_rxd->rxd3);
READ_ONCE         772 drivers/net/ethernet/mediatek/mtk_eth_soc.c 	rxd->rxd4 = READ_ONCE(dma_rxd->rxd4);
READ_ONCE        1466 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 	pkt_rate_low = READ_ONCE(priv->pkt_rate_low);
READ_ONCE        1467 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 	pkt_rate_high = READ_ONCE(priv->pkt_rate_high);
READ_ONCE        1470 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 		rx_packets = READ_ONCE(priv->rx_ring[ring]->packets);
READ_ONCE        1471 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 		rx_bytes = READ_ONCE(priv->rx_ring[ring]->bytes);
READ_ONCE         165 drivers/net/ethernet/mellanox/mlx4/en_port.c 		packets += READ_ONCE(ring->packets);
READ_ONCE         166 drivers/net/ethernet/mellanox/mlx4/en_port.c 		bytes   += READ_ONCE(ring->bytes);
READ_ONCE         176 drivers/net/ethernet/mellanox/mlx4/en_port.c 		packets += READ_ONCE(ring->packets);
READ_ONCE         177 drivers/net/ethernet/mellanox/mlx4/en_port.c 		bytes   += READ_ONCE(ring->bytes);
READ_ONCE         252 drivers/net/ethernet/mellanox/mlx4/en_port.c 		sw_rx_dropped			+= READ_ONCE(ring->dropped);
READ_ONCE         253 drivers/net/ethernet/mellanox/mlx4/en_port.c 		priv->port_stats.rx_chksum_good += READ_ONCE(ring->csum_ok);
READ_ONCE         254 drivers/net/ethernet/mellanox/mlx4/en_port.c 		priv->port_stats.rx_chksum_none += READ_ONCE(ring->csum_none);
READ_ONCE         255 drivers/net/ethernet/mellanox/mlx4/en_port.c 		priv->port_stats.rx_chksum_complete += READ_ONCE(ring->csum_complete);
READ_ONCE         256 drivers/net/ethernet/mellanox/mlx4/en_port.c 		priv->port_stats.rx_alloc_pages += READ_ONCE(ring->rx_alloc_pages);
READ_ONCE         257 drivers/net/ethernet/mellanox/mlx4/en_port.c 		priv->xdp_stats.rx_xdp_drop	+= READ_ONCE(ring->xdp_drop);
READ_ONCE         258 drivers/net/ethernet/mellanox/mlx4/en_port.c 		priv->xdp_stats.rx_xdp_tx	+= READ_ONCE(ring->xdp_tx);
READ_ONCE         259 drivers/net/ethernet/mellanox/mlx4/en_port.c 		priv->xdp_stats.rx_xdp_tx_full	+= READ_ONCE(ring->xdp_tx_full);
READ_ONCE         270 drivers/net/ethernet/mellanox/mlx4/en_port.c 		sw_tx_dropped			   += READ_ONCE(ring->tx_dropped);
READ_ONCE         271 drivers/net/ethernet/mellanox/mlx4/en_port.c 		priv->port_stats.tx_chksum_offload += READ_ONCE(ring->tx_csum);
READ_ONCE         272 drivers/net/ethernet/mellanox/mlx4/en_port.c 		priv->port_stats.queue_stopped     += READ_ONCE(ring->queue_stopped);
READ_ONCE         273 drivers/net/ethernet/mellanox/mlx4/en_port.c 		priv->port_stats.wake_queue        += READ_ONCE(ring->wake_queue);
READ_ONCE         274 drivers/net/ethernet/mellanox/mlx4/en_port.c 		priv->port_stats.tso_packets       += READ_ONCE(ring->tso_packets);
READ_ONCE         275 drivers/net/ethernet/mellanox/mlx4/en_port.c 		priv->port_stats.xmit_more         += READ_ONCE(ring->xmit_more);
READ_ONCE         424 drivers/net/ethernet/mellanox/mlx4/en_tx.c 	last_nr_txbb = READ_ONCE(ring->last_nr_txbb);
READ_ONCE         425 drivers/net/ethernet/mellanox/mlx4/en_tx.c 	ring_cons = READ_ONCE(ring->cons);
READ_ONCE         868 drivers/net/ethernet/mellanox/mlx4/en_tx.c 	ring_cons = READ_ONCE(ring->cons);
READ_ONCE        1078 drivers/net/ethernet/mellanox/mlx4/en_tx.c 		ring_cons = READ_ONCE(ring->cons);
READ_ONCE        1147 drivers/net/ethernet/mellanox/mlx4/en_tx.c 			 (u32)(ring->prod - READ_ONCE(ring->cons) - 1));
READ_ONCE         211 drivers/net/ethernet/mellanox/mlx5/core/cmd.c 		own = READ_ONCE(ent->lay->status_own);
READ_ONCE         124 drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c 	struct bpf_prog *prog = READ_ONCE(rq->xdp_prog);
READ_ONCE        2620 drivers/net/ethernet/neterion/vxge/vxge-main.c 		rx_frms = READ_ONCE(ring->stats.rx_frms);
READ_ONCE        1816 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 	xdp_prog = READ_ONCE(dp->xdp_prog);
READ_ONCE          45 drivers/net/ethernet/netronome/nfp/nfp_net_debugfs.c 		frag = READ_ONCE(rx_ring->rxbufs[i].frag);
READ_ONCE         112 drivers/net/ethernet/netronome/nfp/nfp_net_debugfs.c 			struct sk_buff *skb = READ_ONCE(tx_ring->txbufs[i].skb);
READ_ONCE         119 drivers/net/ethernet/netronome/nfp/nfp_net_debugfs.c 				   READ_ONCE(tx_ring->txbufs[i].frag));
READ_ONCE         540 drivers/net/ethernet/qlogic/qed/qed_iwarp.c 		while (READ_ONCE(ep->state) != QED_IWARP_EP_CLOSED &&
READ_ONCE        1205 drivers/net/ethernet/qlogic/qede/qede_fp.c 	struct bpf_prog *xdp_prog = READ_ONCE(rxq->xdp_prog);
READ_ONCE        2237 drivers/net/ethernet/sfc/ef10.c 	if (likely(READ_ONCE(efx->irq_soft_enabled))) {
READ_ONCE        2252 drivers/net/ethernet/sfc/ef10.c 	bool soft_enabled = READ_ONCE(efx->irq_soft_enabled);
READ_ONCE        3544 drivers/net/ethernet/sfc/ef10.c 	if (unlikely(READ_ONCE(efx->reset_pending)))
READ_ONCE        3693 drivers/net/ethernet/sfc/ef10.c 	if (unlikely(READ_ONCE(efx->reset_pending)))
READ_ONCE        5505 drivers/net/ethernet/sfc/ef10.c 		if (READ_ONCE(table->entry[i].spec) &
READ_ONCE        2861 drivers/net/ethernet/sfc/efx.c 	pending = READ_ONCE(efx->reset_pending);
READ_ONCE        2926 drivers/net/ethernet/sfc/efx.c 	if (READ_ONCE(efx->state) != STATE_READY)
READ_ONCE        2516 drivers/net/ethernet/sfc/falcon/efx.c 	pending = READ_ONCE(efx->reset_pending);
READ_ONCE        2576 drivers/net/ethernet/sfc/falcon/efx.c 	if (READ_ONCE(efx->state) != STATE_READY)
READ_ONCE         452 drivers/net/ethernet/sfc/falcon/falcon.c 	if (!likely(READ_ONCE(efx->irq_soft_enabled)))
READ_ONCE        1372 drivers/net/ethernet/sfc/falcon/falcon.c 	isolate = !!READ_ONCE(efx->reset_pending);
READ_ONCE         834 drivers/net/ethernet/sfc/falcon/farch.c 	if (unlikely(READ_ONCE(efx->reset_pending)))
READ_ONCE         990 drivers/net/ethernet/sfc/falcon/farch.c 	if (unlikely(READ_ONCE(efx->reset_pending)))
READ_ONCE        1504 drivers/net/ethernet/sfc/falcon/farch.c 	bool soft_enabled = READ_ONCE(efx->irq_soft_enabled);
READ_ONCE        1596 drivers/net/ethernet/sfc/falcon/farch.c 	if (!likely(READ_ONCE(efx->irq_soft_enabled)))
READ_ONCE          83 drivers/net/ethernet/sfc/falcon/nic.h 	unsigned int empty_read_count = READ_ONCE(tx_queue->empty_read_count);
READ_ONCE         466 drivers/net/ethernet/sfc/falcon/nic.h 	return READ_ONCE(channel->event_test_cpu);
READ_ONCE         470 drivers/net/ethernet/sfc/falcon/nic.h 	return READ_ONCE(efx->last_irq_cpu);
READ_ONCE         134 drivers/net/ethernet/sfc/falcon/tx.c 	txq1->old_read_count = READ_ONCE(txq1->read_count);
READ_ONCE         135 drivers/net/ethernet/sfc/falcon/tx.c 	txq2->old_read_count = READ_ONCE(txq2->read_count);
READ_ONCE         524 drivers/net/ethernet/sfc/falcon/tx.c 		tx_queue->old_write_count = READ_ONCE(tx_queue->write_count);
READ_ONCE         826 drivers/net/ethernet/sfc/farch.c 	if (unlikely(READ_ONCE(efx->reset_pending)))
READ_ONCE         978 drivers/net/ethernet/sfc/farch.c 	if (unlikely(READ_ONCE(efx->reset_pending)))
READ_ONCE        1513 drivers/net/ethernet/sfc/farch.c 	bool soft_enabled = READ_ONCE(efx->irq_soft_enabled);
READ_ONCE        1605 drivers/net/ethernet/sfc/farch.c 	if (!likely(READ_ONCE(efx->irq_soft_enabled)))
READ_ONCE          81 drivers/net/ethernet/sfc/nic.h 	unsigned int empty_read_count = READ_ONCE(tx_queue->empty_read_count);
READ_ONCE         643 drivers/net/ethernet/sfc/nic.h 	return READ_ONCE(channel->event_test_cpu);
READ_ONCE         647 drivers/net/ethernet/sfc/nic.h 	return READ_ONCE(efx->last_irq_cpu);
READ_ONCE         870 drivers/net/ethernet/sfc/ptp.c 	       READ_ONCE(*mc_running)) {
READ_ONCE         880 drivers/net/ethernet/sfc/ptp.c 			 READ_ONCE(*mc_running));
READ_ONCE        1051 drivers/net/ethernet/sfc/ptp.c 	while (!READ_ONCE(*start) && (time_before(jiffies, timeout))) {
READ_ONCE        1061 drivers/net/ethernet/sfc/ptp.c 	if (READ_ONCE(*start))
READ_ONCE         151 drivers/net/ethernet/sfc/tx.c 	txq1->old_read_count = READ_ONCE(txq1->read_count);
READ_ONCE         152 drivers/net/ethernet/sfc/tx.c 	txq2->old_read_count = READ_ONCE(txq2->read_count);
READ_ONCE         781 drivers/net/ethernet/sfc/tx.c 		tx_queue->old_write_count = READ_ONCE(tx_queue->write_count);
READ_ONCE         936 drivers/net/ethernet/socionext/netsec.c 	xdp_prog = READ_ONCE(priv->xdp_prog);
READ_ONCE        1273 drivers/net/ethernet/socionext/netsec.c 	struct bpf_prog *xdp_prog = READ_ONCE(priv->xdp_prog);
READ_ONCE        6228 drivers/net/ethernet/sun/niu.c 	rx_rings = READ_ONCE(np->rx_rings);
READ_ONCE        6259 drivers/net/ethernet/sun/niu.c 	tx_rings = READ_ONCE(np->tx_rings);
READ_ONCE         500 drivers/net/ethernet/ti/cpsw.c 	prog = READ_ONCE(priv->xdp_prog);
READ_ONCE         261 drivers/net/tap.c 	int numvtaps = READ_ONCE(tap->numvtaps);
READ_ONCE         639 drivers/net/tap.c 		vnet_hdr_len = READ_ONCE(q->vnet_hdr_sz);
READ_ONCE         780 drivers/net/tap.c 		vnet_hdr_len = READ_ONCE(q->vnet_hdr_sz);
READ_ONCE        1159 drivers/net/tap.c 		vnet_hdr_len = READ_ONCE(q->vnet_hdr_sz);
READ_ONCE         529 drivers/net/tun.c 		if (READ_ONCE(e->queue_index) != queue_index)
READ_ONCE         571 drivers/net/tun.c 	numqueues = READ_ONCE(tun->numqueues);
READ_ONCE         592 drivers/net/tun.c 	numqueues = READ_ONCE(tun->numqueues);
READ_ONCE        1297 drivers/net/tun.c 	numqueues = READ_ONCE(tun->numqueues);
READ_ONCE        1772 drivers/net/tun.c 		int vnet_hdr_sz = READ_ONCE(tun->vnet_hdr_sz);
READ_ONCE        2055 drivers/net/tun.c 		vnet_hdr_sz = READ_ONCE(tun->vnet_hdr_sz);
READ_ONCE        2093 drivers/net/tun.c 		vnet_hdr_sz = READ_ONCE(tun->vnet_hdr_sz);
READ_ONCE         503 drivers/net/wireless/ath/ath5k/desc.c 	txstat1 = READ_ONCE(tx_status->tx_status_1);
READ_ONCE         509 drivers/net/wireless/ath/ath5k/desc.c 	txstat0 = READ_ONCE(tx_status->tx_status_0);
READ_ONCE         703 drivers/net/wireless/ath/ath5k/desc.c 	rxstat1 = READ_ONCE(rx_status->rx_status_1);
READ_ONCE         710 drivers/net/wireless/ath/ath5k/desc.c 	rxstat0 = READ_ONCE(rx_status->rx_status_0);
READ_ONCE         321 drivers/net/wireless/ath/ath9k/ar9002_mac.c 	status = READ_ONCE(ads->ds_txstatus9);
READ_ONCE         335 drivers/net/wireless/ath/ath9k/ar9002_mac.c 	status = READ_ONCE(ads->ds_txstatus0);
READ_ONCE         345 drivers/net/wireless/ath/ath9k/ar9002_mac.c 	status = READ_ONCE(ads->ds_txstatus1);
READ_ONCE         374 drivers/net/wireless/ath/ath9k/ar9002_mac.c 	status = READ_ONCE(ads->ds_txstatus5);
READ_ONCE         393 drivers/net/wireless/ath/ath9k/ar9002_mac.c 		return MS(READ_ONCE(ads->ds_ctl4), AR_PacketDur0);
READ_ONCE         395 drivers/net/wireless/ath/ath9k/ar9002_mac.c 		return MS(READ_ONCE(ads->ds_ctl4), AR_PacketDur1);
READ_ONCE         397 drivers/net/wireless/ath/ath9k/ar9002_mac.c 		return MS(READ_ONCE(ads->ds_ctl5), AR_PacketDur2);
READ_ONCE         399 drivers/net/wireless/ath/ath9k/ar9002_mac.c 		return MS(READ_ONCE(ads->ds_ctl5), AR_PacketDur3);
READ_ONCE         362 drivers/net/wireless/ath/ath9k/ar9003_mac.c 	status = READ_ONCE(ads->status8);
READ_ONCE         388 drivers/net/wireless/ath/ath9k/ar9003_mac.c 	status = READ_ONCE(ads->status2);
READ_ONCE         398 drivers/net/wireless/ath/ath9k/ar9003_mac.c 	status = READ_ONCE(ads->status3);
READ_ONCE         423 drivers/net/wireless/ath/ath9k/ar9003_mac.c 	status = READ_ONCE(ads->status7);
READ_ONCE         440 drivers/net/wireless/ath/ath9k/ar9003_mac.c 		return MS(READ_ONCE(adc->ctl15), AR_PacketDur0);
READ_ONCE         442 drivers/net/wireless/ath/ath9k/ar9003_mac.c 		return MS(READ_ONCE(adc->ctl15), AR_PacketDur1);
READ_ONCE         444 drivers/net/wireless/ath/ath9k/ar9003_mac.c 		return MS(READ_ONCE(adc->ctl16), AR_PacketDur2);
READ_ONCE         446 drivers/net/wireless/ath/ath9k/ar9003_mac.c 		return MS(READ_ONCE(adc->ctl16), AR_PacketDur3);
READ_ONCE         369 drivers/net/wireless/ath/wcn36xx/dxe.c 		if (READ_ONCE(ctl->desc->ctrl) & WCN36xx_DXE_CTRL_VLD)
READ_ONCE         373 drivers/net/wireless/ath/wcn36xx/dxe.c 		    READ_ONCE(ctl->desc->ctrl) & WCN36xx_DXE_CTRL_EOP) {
READ_ONCE         565 drivers/net/wireless/ath/wcn36xx/dxe.c 	while (!(READ_ONCE(dxe->ctrl) & WCN36xx_DXE_CTRL_VLD)) {
READ_ONCE        3708 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c 	while (READ_ONCE(bus->dpc_triggered)) {
READ_ONCE         791 drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c 		u8 ap_sta_id = READ_ONCE(mvmvif->ap_sta_id);
READ_ONCE        1185 drivers/net/wireless/intel/iwlwifi/mvm/ops.c 	bool rfkill_safe_init_done = READ_ONCE(mvm->rfkill_safe_init_done);
READ_ONCE         270 drivers/net/wireless/intel/iwlwifi/pcie/internal.h 		return READ_ONCE(*rb_stts);
READ_ONCE         274 drivers/net/wireless/intel/iwlwifi/pcie/internal.h 		return READ_ONCE(rb_stts->closed_rb_num);
READ_ONCE        2366 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 	wr_ptr = READ_ONCE(txq->write_ptr);
READ_ONCE        2368 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 	while ((txq->read_ptr != READ_ONCE(txq->write_ptr) ||
READ_ONCE        2372 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 		u8 write_ptr = READ_ONCE(txq->write_ptr);
READ_ONCE        1424 drivers/net/wireless/mac80211_hwsim.c 	_portid = READ_ONCE(data->wmediumd);
READ_ONCE        1521 drivers/net/wireless/mac80211_hwsim.c 	u32 _pid = READ_ONCE(data->wmediumd);
READ_ONCE          99 drivers/net/wireless/mediatek/mt76/dma.c 	__le32 __ctrl = READ_ONCE(q->desc[idx].ctrl);
READ_ONCE         103 drivers/net/wireless/mediatek/mt76/dma.c 		__le32 addr = READ_ONCE(q->desc[idx].buf0);
READ_ONCE         111 drivers/net/wireless/mediatek/mt76/dma.c 		__le32 addr = READ_ONCE(q->desc[idx].buf1);
READ_ONCE         215 drivers/net/wireless/mediatek/mt76/dma.c 	buf_addr = le32_to_cpu(READ_ONCE(desc->buf0));
READ_ONCE         217 drivers/net/wireless/mediatek/mt76/dma.c 		u32 ctl = le32_to_cpu(READ_ONCE(desc->ctrl));
READ_ONCE        1003 drivers/net/wireless/mediatek/mt76/mt7603/mac.c 	rate_set_tsf = READ_ONCE(sta->rate_set_tsf);
READ_ONCE         886 drivers/net/wireless/mediatek/mt76/mt7615/mac.c 	rate_set_tsf = READ_ONCE(sta->rate_set_tsf);
READ_ONCE          69 drivers/net/wireless/quantenna/qtnfmac/shm_ipc.c 	if (!READ_ONCE(ipc->waiting_for_ack))
READ_ONCE         159 drivers/net/xen-netback/interface.c 	num_queues = READ_ONCE(vif->num_queues);
READ_ONCE         191 drivers/net/xen-netback/interface.c 	num_queues = READ_ONCE(vif->num_queues);
READ_ONCE         250 drivers/net/xen-netback/interface.c 	num_queues = READ_ONCE(vif->num_queues);
READ_ONCE         402 drivers/net/xen-netback/interface.c 	num_queues = READ_ONCE(vif->num_queues);
READ_ONCE         186 drivers/nvme/host/multipath.c 		if (READ_ONCE(head->subsys->iopolicy) == NVME_IOPOLICY_NUMA)
READ_ONCE         271 drivers/nvme/host/multipath.c 	if (READ_ONCE(head->subsys->iopolicy) == NVME_IOPOLICY_RR && ns)
READ_ONCE         605 drivers/nvme/host/multipath.c 			nvme_iopolicy_names[READ_ONCE(subsys->iopolicy)]);
READ_ONCE         715 drivers/nvme/target/admin-cmd.c 	nvmet_set_result(req, READ_ONCE(req->sq->ctrl->aen_enabled));
READ_ONCE         356 drivers/nvme/target/nvmet.h 	if (!(READ_ONCE(ctrl->aen_enabled) & (1 << bn)))
READ_ONCE         111 drivers/powercap/idle_inject.c 	duration_us = READ_ONCE(ii_dev->run_duration_us);
READ_ONCE         112 drivers/powercap/idle_inject.c 	duration_us += READ_ONCE(ii_dev->idle_duration_us);
READ_ONCE         141 drivers/powercap/idle_inject.c 	play_idle(READ_ONCE(ii_dev->idle_duration_us));
READ_ONCE         168 drivers/powercap/idle_inject.c 	*run_duration_us = READ_ONCE(ii_dev->run_duration_us);
READ_ONCE         169 drivers/powercap/idle_inject.c 	*idle_duration_us = READ_ONCE(ii_dev->idle_duration_us);
READ_ONCE         184 drivers/powercap/idle_inject.c 	unsigned int idle_duration_us = READ_ONCE(ii_dev->idle_duration_us);
READ_ONCE         185 drivers/powercap/idle_inject.c 	unsigned int run_duration_us = READ_ONCE(ii_dev->run_duration_us);
READ_ONCE         394 drivers/s390/net/ism_drv.c 	while ((ism->ieq_idx + 1) != READ_ONCE(ism->ieq->header.idx)) {
READ_ONCE        14352 drivers/scsi/lpfc/lpfc_sli.c 	if (READ_ONCE(eq->mode) == LPFC_EQ_POLL)
READ_ONCE        1025 drivers/scsi/qla2xxx/qla_target.c 		while (!READ_ONCE(sess->logout_completed)) {
READ_ONCE          39 drivers/scsi/scsi_debugfs.c 	const u8 *const cdb = READ_ONCE(cmd->cmnd);
READ_ONCE         324 drivers/staging/comedi/comedi_fops.c 	return READ_ONCE(cfp->read_subdev);
READ_ONCE         332 drivers/staging/comedi/comedi_fops.c 	return READ_ONCE(cfp->write_subdev);
READ_ONCE        1231 drivers/target/target_core_user.c 	while (udev->cmdr_last_cleaned != READ_ONCE(mb->cmd_tail)) {
READ_ONCE         381 drivers/thermal/intel/intel_powerclamp.c 	w_data->target_ratio = READ_ONCE(set_target_ratio);
READ_ONCE         649 drivers/thermal/tegra/soctherm.c 	temp = READ_ONCE(tz->temperature);
READ_ONCE         650 drivers/thermal/tegra/soctherm.c 	last_temp = READ_ONCE(tz->last_temperature);
READ_ONCE         208 drivers/tty/tty_audit.c 	audit_tty = READ_ONCE(current->signal->audit_tty);
READ_ONCE          30 drivers/tty/tty_port.c 	tty = READ_ONCE(port->itty);
READ_ONCE         489 drivers/usb/class/cdc-wdm.c 	cntr = READ_ONCE(desc->length);
READ_ONCE         148 drivers/usb/core/devio.c 	lim = READ_ONCE(usbfs_memory_mb);
READ_ONCE        2473 drivers/usb/core/devio.c 			READ_ONCE(ps->not_yet_resumed) == 0);
READ_ONCE        1123 drivers/usb/core/sysfs.c 	string = READ_ONCE(intf->cur_altsetting->string);
READ_ONCE        1139 drivers/usb/core/sysfs.c 	alt = READ_ONCE(intf->cur_altsetting);
READ_ONCE        1244 drivers/usb/gadget/udc/gr_udc.c 	if (READ_ONCE(req->last_desc->ctrl) & GR_DESC_IN_CTRL_EN)
READ_ONCE        1273 drivers/usb/gadget/udc/gr_udc.c 	ctrl = READ_ONCE(req->curr_desc->ctrl);
READ_ONCE         808 drivers/usb/host/ohci-hcd.c 		head = hc32_to_cpu(ohci, READ_ONCE(ed->hwHeadP)) & TD_MASK;
READ_ONCE         191 drivers/usb/host/uhci-hcd.h #define qh_element(qh)		READ_ONCE((qh)->element)
READ_ONCE         279 drivers/usb/host/uhci-hcd.h 						READ_ONCE((td)->status))
READ_ONCE         661 drivers/vfio/vfio.c 	struct device_driver *drv = READ_ONCE(dev->driver);
READ_ONCE         897 drivers/vhost/scsi.c 		tpg = READ_ONCE(vs_tpg[*vc->target]);
READ_ONCE        2098 drivers/vhost/vhost.c 	next = vhost16_to_cpu(vq, READ_ONCE(desc->next));
READ_ONCE         825 drivers/vhost/vringh.c 	*val = vringh16_to_cpu(vrh, READ_ONCE(*p));
READ_ONCE         778 drivers/xen/pvcalls-back.c 	data = READ_ONCE(queue->rskq_accept_head) != NULL;
READ_ONCE         324 drivers/xen/pvcalls-front.c 		   READ_ONCE(bedata->rsp[req_id].req_id) == req_id);
READ_ONCE         470 drivers/xen/pvcalls-front.c 		   READ_ONCE(bedata->rsp[req_id].req_id) == req_id);
READ_ONCE         701 drivers/xen/pvcalls-front.c 		   READ_ONCE(bedata->rsp[req_id].req_id) == req_id);
READ_ONCE         750 drivers/xen/pvcalls-front.c 		   READ_ONCE(bedata->rsp[req_id].req_id) == req_id);
READ_ONCE         787 drivers/xen/pvcalls-front.c 		req_id = READ_ONCE(map->passive.inflight_req_id);
READ_ONCE         789 drivers/xen/pvcalls-front.c 		    READ_ONCE(bedata->rsp[req_id].req_id) == req_id) {
READ_ONCE         866 drivers/xen/pvcalls-front.c 		READ_ONCE(bedata->rsp[req_id].req_id) == req_id)) {
READ_ONCE         908 drivers/xen/pvcalls-front.c 		uint32_t req_id = READ_ONCE(map->passive.inflight_req_id);
READ_ONCE         911 drivers/xen/pvcalls-front.c 		    READ_ONCE(bedata->rsp[req_id].req_id) == req_id)
READ_ONCE        1038 drivers/xen/pvcalls-front.c 		   READ_ONCE(bedata->rsp[req_id].req_id) == req_id);
READ_ONCE        1068 drivers/xen/pvcalls-front.c 		if (READ_ONCE(map->passive.inflight_req_id) != PVCALLS_INVALID_ID &&
READ_ONCE        1069 drivers/xen/pvcalls-front.c 			READ_ONCE(map->passive.inflight_req_id) != 0) {
READ_ONCE          42 drivers/xen/time.c 			h = READ_ONCE(p32[1]);
READ_ONCE          43 drivers/xen/time.c 			l = READ_ONCE(p32[0]);
READ_ONCE          44 drivers/xen/time.c 			h2 = READ_ONCE(p32[1]);
READ_ONCE          49 drivers/xen/time.c 		ret = READ_ONCE(*p);
READ_ONCE          67 drivers/xen/time.c 		*res = READ_ONCE(*state);
READ_ONCE         373 fs/afs/addr_list.c 	index = READ_ONCE(ac->alist->preferred);
READ_ONCE         384 fs/afs/cell.c  	unsigned int min_ttl = READ_ONCE(afs_cell_min_ttl);
READ_ONCE         385 fs/afs/cell.c  	unsigned int max_ttl = READ_ONCE(afs_cell_max_ttl);
READ_ONCE         650 fs/afs/dir.c   	afs_dataversion_t data_version = READ_ONCE(dvnode->status.data_version);
READ_ONCE         984 fs/afs/dir.c   	parent = READ_ONCE(dentry->d_parent);
READ_ONCE        1000 fs/afs/dir.c   	dir_version = (long)READ_ONCE(dvnode->status.data_version);
READ_ONCE        1001 fs/afs/dir.c   	de_version = (long)READ_ONCE(dentry->d_fsdata);
READ_ONCE        1003 fs/afs/dir.c   		dir_version = (long)READ_ONCE(dvnode->invalid_before);
READ_ONCE         754 fs/afs/flock.c 		lock_count = READ_ONCE(vnode->status.lock_count);
READ_ONCE         605 fs/afs/inode.c 		cb_v_break = READ_ONCE(volume->cb_v_break);
READ_ONCE         611 fs/afs/inode.c 			cb_s_break = READ_ONCE(server->cb_s_break);
READ_ONCE        1170 fs/afs/internal.h 	return READ_ONCE(call->state) == state;
READ_ONCE          63 fs/afs/rotate.c 	fc->index = READ_ONCE(fc->server_list->preferred);
READ_ONCE         466 fs/afs/rotate.c 		afs_prioritise_error(&e, READ_ONCE(s->probe.error),
READ_ONCE         511 fs/afs/rxrpc.c 	while (state = READ_ONCE(call->state),
READ_ONCE         542 fs/afs/rxrpc.c 		state = READ_ONCE(call->state);
READ_ONCE         937 fs/afs/rxrpc.c 	state = READ_ONCE(call->state);
READ_ONCE         583 fs/afs/server.c 	diff = READ_ONCE(server->update_at) - now;
READ_ONCE         261 fs/afs/vl_rotate.c 		afs_prioritise_error(&e, READ_ONCE(s->probe.error),
READ_ONCE        1637 fs/aio.c       	if (!READ_ONCE(req->cancelled))
READ_ONCE        1648 fs/aio.c       	if (!mask && !READ_ONCE(req->cancelled)) {
READ_ONCE         255 fs/block_dev.c 		if (!READ_ONCE(bio.bi_private))
READ_ONCE         296 fs/block_dev.c 	return blk_poll(q, READ_ONCE(kiocb->ki_cookie), wait);
READ_ONCE         448 fs/block_dev.c 		if (!READ_ONCE(dio->waiter))
READ_ONCE          73 fs/btrfs/delayed-inode.c 	node = READ_ONCE(btrfs_inode->delayed_node);
READ_ONCE        1298 fs/btrfs/delayed-inode.c 	delayed_node = READ_ONCE(inode->delayed_node);
READ_ONCE         863 fs/btrfs/transaction.c 	if (lock && READ_ONCE(cur_trans->state) == TRANS_STATE_BLOCKED) {
READ_ONCE        1974 fs/btrfs/transaction.c 	if (unlikely(READ_ONCE(cur_trans->aborted))) {
READ_ONCE        2122 fs/btrfs/transaction.c 	if (unlikely(READ_ONCE(cur_trans->aborted))) {
READ_ONCE        2240 fs/btrfs/transaction.c 	if (unlikely(READ_ONCE(cur_trans->aborted))) {
READ_ONCE          40 fs/btrfs/tree-log.h 	return READ_ONCE(trans->fs_info->last_trans_log_full_commit) ==
READ_ONCE        1646 fs/buffer.c    		create_empty_buffers(page, 1 << READ_ONCE(inode->i_blkbits),
READ_ONCE         800 fs/ceph/addr.c 	if (READ_ONCE(fsc->mount_state) == CEPH_MOUNT_SHUTDOWN) {
READ_ONCE        1229 fs/ceph/addr.c 	if (READ_ONCE(fsc->mount_state) == CEPH_MOUNT_SHUTDOWN) {
READ_ONCE        1208 fs/ceph/caps.c 	ceph_encode_32(&p, READ_ONCE(osdc->epoch_barrier));
READ_ONCE        2261 fs/ceph/caps.c 	if (errseq_check(&ci->i_meta_err, READ_ONCE(fi->meta_err))) {
READ_ONCE        2664 fs/ceph/caps.c 			if (READ_ONCE(mdsc->fsc->mount_state) ==
READ_ONCE        2757 fs/ceph/caps.c 	    fi->filp_gen != READ_ONCE(fsc->filp_gen))
READ_ONCE        2790 fs/ceph/caps.c 		    fi->filp_gen != READ_ONCE(fsc->filp_gen)) {
READ_ONCE        1559 fs/ceph/dir.c  		parent = READ_ONCE(dentry->d_parent);
READ_ONCE         238 fs/ceph/file.c 	fi->filp_gen = READ_ONCE(ceph_inode_to_client(inode)->filp_gen);
READ_ONCE        1853 fs/ceph/inode.c 	if (READ_ONCE(fsc->mount_state) == CEPH_MOUNT_SHUTDOWN) {
READ_ONCE          26 fs/ceph/io.c   	if (READ_ONCE(ci->i_ceph_flags) & CEPH_I_ODIRECT) {
READ_ONCE          57 fs/ceph/io.c   	if (!(READ_ONCE(ci->i_ceph_flags) & CEPH_I_ODIRECT))
READ_ONCE         111 fs/ceph/io.c   	if (!(READ_ONCE(ci->i_ceph_flags) & CEPH_I_ODIRECT)) {
READ_ONCE         143 fs/ceph/io.c   	if (READ_ONCE(ci->i_ceph_flags) & CEPH_I_ODIRECT)
READ_ONCE         924 fs/ceph/mds_client.c 		parent = READ_ONCE(req->r_dentry->d_parent);
READ_ONCE        1399 fs/ceph/mds_client.c 		if (READ_ONCE(fsc->mount_state) == CEPH_MOUNT_SHUTDOWN) {
READ_ONCE        2173 fs/ceph/mds_client.c 		temp = READ_ONCE(temp->d_parent);
READ_ONCE        2536 fs/ceph/mds_client.c 	if (READ_ONCE(mdsc->fsc->mount_state) == CEPH_MOUNT_SHUTDOWN) {
READ_ONCE        2541 fs/ceph/mds_client.c 	if (READ_ONCE(mdsc->fsc->mount_state) == CEPH_MOUNT_MOUNTING) {
READ_ONCE        4034 fs/ceph/mds_client.c 	if (READ_ONCE(fsc->mount_state) != CEPH_MOUNT_MOUNTED)
READ_ONCE        4037 fs/ceph/mds_client.c 	if (!READ_ONCE(fsc->blacklisted))
READ_ONCE        4301 fs/ceph/mds_client.c 	if (READ_ONCE(mdsc->fsc->mount_state) == CEPH_MOUNT_SHUTDOWN)
READ_ONCE        4332 fs/ceph/mds_client.c 	if (READ_ONCE(mdsc->fsc->mount_state) == CEPH_MOUNT_SHUTDOWN)
READ_ONCE         538 fs/ceph/super.h 	if (!(READ_ONCE(ci->i_ceph_flags) & CEPH_I_ERROR_WRITE)) {
READ_ONCE         547 fs/ceph/super.h 	if (READ_ONCE(ci->i_ceph_flags) & CEPH_I_ERROR_WRITE) {
READ_ONCE         404 fs/crypto/fscrypt_private.h 	return READ_ONCE(secret->size) != 0;
READ_ONCE         247 fs/crypto/hooks.c 	pstr.name = READ_ONCE(inode->i_link);
READ_ONCE         234 fs/crypto/keyring.c 	keyring = READ_ONCE(sb->s_master_keys);
READ_ONCE         116 fs/crypto/keysetup.c 	struct crypto_shash *tfm = READ_ONCE(essiv_hash_tfm);
READ_ONCE         221 fs/crypto/keysetup.c 	tfm = READ_ONCE(mk->mk_mode_keys[mode_num]);
READ_ONCE         568 fs/crypto/keysetup.c 	const struct fscrypt_info *ci = READ_ONCE(inode->i_crypt_info);
READ_ONCE         214 fs/crypto/policy.c 	ci = READ_ONCE(inode->i_crypt_info);
READ_ONCE         481 fs/crypto/policy.c 	ci = READ_ONCE(parent->i_crypt_info);
READ_ONCE          41 fs/d_path.c    	u32 dlen = READ_ONCE(name->len);
READ_ONCE         104 fs/d_path.c    			struct mount *parent = READ_ONCE(mnt->mnt_parent);
READ_ONCE         114 fs/d_path.c    				dentry = READ_ONCE(mnt->mnt_mountpoint);
READ_ONCE         397 fs/dax.c       		struct address_space *mapping = READ_ONCE(page->mapping);
READ_ONCE         249 fs/dcache.c    	const unsigned char *cs = READ_ONCE(dentry->d_name.name);
READ_ONCE         319 fs/dcache.c    	flags = READ_ONCE(dentry->d_flags);
READ_ONCE         327 fs/dcache.c    	unsigned flags = READ_ONCE(dentry->d_flags);
READ_ONCE         603 fs/dcache.c    	parent = READ_ONCE(dentry->d_parent);
READ_ONCE         779 fs/dcache.c    	d_flags = READ_ONCE(dentry->d_flags);
READ_ONCE         912 fs/dcache.c    	ret = READ_ONCE(dentry->d_parent);
READ_ONCE         916 fs/dcache.c    		if (likely(ret == READ_ONCE(dentry->d_parent)))
READ_ONCE        2556 fs/dcache.c    	if (unlikely(READ_ONCE(parent->d_inode->i_dir_seq) != seq)) {
READ_ONCE          86 fs/debugfs/file.c 	d_fsd = READ_ONCE(dentry->d_fsdata);
READ_ONCE         100 fs/debugfs/file.c 			fsd = READ_ONCE(dentry->d_fsdata);
READ_ONCE         133 fs/debugfs/file.c 	struct debugfs_fsdata *fsd = READ_ONCE(dentry->d_fsdata);
READ_ONCE         681 fs/debugfs/inode.c 	fsd = READ_ONCE(dentry->d_fsdata);
READ_ONCE        1169 fs/direct-io.c 	unsigned i_blkbits = READ_ONCE(inode->i_blkbits);
READ_ONCE         347 fs/ecryptfs/inode.c 	lower_inode = READ_ONCE(lower_dentry->d_inode);
READ_ONCE         183 fs/erofs/zdata.c 		if (READ_ONCE(*pages))
READ_ONCE         368 fs/erofs/zdata.c 	length = READ_ONCE(pcl->length);
READ_ONCE         384 fs/erofs/zdata.c 			length = READ_ONCE(pcl->length);
READ_ONCE         773 fs/erofs/zdata.c 	DBG_BUGON(!READ_ONCE(cl->nr_pages));
READ_ONCE         972 fs/erofs/zdata.c 		owned = READ_ONCE(pcl->next);
READ_ONCE        1009 fs/erofs/zdata.c 	page = READ_ONCE(pcl->compressed_pages[nr]);
READ_ONCE        1029 fs/erofs/zdata.c 	mapping = READ_ONCE(page->mapping);
READ_ONCE         154 fs/erofs/zdata.h 	o = READ_ONCE(*p);
READ_ONCE         172 fs/eventfd.c   	count = READ_ONCE(ctx->count);
READ_ONCE         381 fs/eventpoll.c 		READ_ONCE(ep->ovflist) != EP_UNACTIVE_PTR;
READ_ONCE         400 fs/eventpoll.c 	unsigned int napi_id = READ_ONCE(ep->napi_id);
READ_ONCE         434 fs/eventpoll.c 	napi_id = READ_ONCE(sk->sk_napi_id);
READ_ONCE         712 fs/eventpoll.c 	for (nepi = READ_ONCE(ep->ovflist); (epi = nepi) != NULL;
READ_ONCE        1248 fs/eventpoll.c 	if (READ_ONCE(ep->ovflist) != EP_UNACTIVE_PTR) {
READ_ONCE        1539 fs/exec.c      	mode = READ_ONCE(inode->i_mode);
READ_ONCE         678 fs/ext4/dir.c  	const struct dentry *parent = READ_ONCE(dentry->d_parent);
READ_ONCE         679 fs/ext4/dir.c  	const struct inode *inode = READ_ONCE(parent->d_inode);
READ_ONCE         695 fs/ext4/dir.c  	const struct inode *inode = READ_ONCE(dentry->d_inode);
READ_ONCE         782 fs/ext4/inode.c 		old_state = READ_ONCE(bh->b_state);
READ_ONCE        2576 fs/ext4/inode.c 	if (disksize > READ_ONCE(EXT4_I(inode)->i_disksize)) {
READ_ONCE        5143 fs/ext4/inode.c 	u64 i_blocks = READ_ONCE(inode->i_blocks);
READ_ONCE        2750 fs/f2fs/data.c 	unsigned i_blkbits = READ_ONCE(inode->i_blkbits);
READ_ONCE        1071 fs/f2fs/dir.c  	const struct dentry *parent = READ_ONCE(dentry->d_parent);
READ_ONCE        1072 fs/f2fs/dir.c  	const struct inode *inode = READ_ONCE(parent->d_inode);
READ_ONCE        1087 fs/f2fs/dir.c  	const struct inode *inode = READ_ONCE(dentry->d_inode);
READ_ONCE        3696 fs/f2fs/f2fs.h 	unsigned int i_blkbits = READ_ONCE(inode->i_blkbits);
READ_ONCE         732 fs/fcntl.c     	int signum = READ_ONCE(fown->signum);
READ_ONCE          76 fs/fs_pin.c    		p = READ_ONCE(m->mnt_pins.first);
READ_ONCE          90 fs/fs_pin.c    		q = READ_ONCE(p->first);
READ_ONCE          43 fs/fscache/cookie.c 	object = READ_ONCE(cookie->backing_objects.first);
READ_ONCE         177 fs/fscache/operation.c 	flags = READ_ONCE(object->flags);
READ_ONCE         264 fs/fscache/operation.c 	flags = READ_ONCE(object->flags);
READ_ONCE         113 fs/fuse/control.c 	val = READ_ONCE(fc->max_background);
READ_ONCE         155 fs/fuse/control.c 	val = READ_ONCE(fc->congestion_threshold);
READ_ONCE          40 fs/fuse/dev.c  	return READ_ONCE(file->private_data);
READ_ONCE          97 fs/fuse/dev.c  	    !READ_ONCE(fc->connected)) {
READ_ONCE         981 fs/fuse/dir.c  	else if (request_mask & READ_ONCE(fi->inval_mask))
READ_ONCE        1169 fs/fuse/dir.c  		if (perm_mask & READ_ONCE(fi->inval_mask) ||
READ_ONCE         481 fs/io_uring.c  	if (ctx->cached_cq_tail != READ_ONCE(rings->cq.tail)) {
READ_ONCE         563 fs/io_uring.c  	if (tail - READ_ONCE(rings->cq.head) == rings->cq_ring_entries)
READ_ONCE         751 fs/io_uring.c  	return READ_ONCE(rings->cq.tail) - READ_ONCE(rings->cq.head);
READ_ONCE        1102 fs/io_uring.c  	kiocb->ki_pos = READ_ONCE(sqe->off);
READ_ONCE        1106 fs/io_uring.c  	ioprio = READ_ONCE(sqe->ioprio);
READ_ONCE        1116 fs/io_uring.c  	ret = kiocb_set_rw_flags(kiocb, READ_ONCE(sqe->rw_flags));
READ_ONCE        1169 fs/io_uring.c  	size_t len = READ_ONCE(sqe->len);
READ_ONCE        1179 fs/io_uring.c  	buf_index = READ_ONCE(sqe->buf_index);
READ_ONCE        1185 fs/io_uring.c  	buf_addr = READ_ONCE(sqe->addr);
READ_ONCE        1244 fs/io_uring.c  	void __user *buf = u64_to_user_ptr(READ_ONCE(sqe->addr));
READ_ONCE        1245 fs/io_uring.c  	size_t sqe_len = READ_ONCE(sqe->len);
READ_ONCE        1256 fs/io_uring.c  	opcode = READ_ONCE(sqe->opcode);
READ_ONCE        1574 fs/io_uring.c  	loff_t sqe_off = READ_ONCE(sqe->off);
READ_ONCE        1575 fs/io_uring.c  	loff_t sqe_len = READ_ONCE(sqe->len);
READ_ONCE        1580 fs/io_uring.c  	fsync_flags = READ_ONCE(sqe->fsync_flags);
READ_ONCE        1636 fs/io_uring.c  	sqe_off = READ_ONCE(sqe->off);
READ_ONCE        1637 fs/io_uring.c  	sqe_len = READ_ONCE(sqe->len);
READ_ONCE        1638 fs/io_uring.c  	flags = READ_ONCE(sqe->sync_range_flags);
READ_ONCE        1666 fs/io_uring.c  		flags = READ_ONCE(sqe->msg_flags);
READ_ONCE        1678 fs/io_uring.c  			READ_ONCE(sqe->addr);
READ_ONCE        1768 fs/io_uring.c  		if (READ_ONCE(sqe->addr) == poll_req->user_data) {
READ_ONCE        1800 fs/io_uring.c  	if (!READ_ONCE(poll->canceled))
READ_ONCE        1811 fs/io_uring.c  	if (!mask && !READ_ONCE(poll->canceled)) {
READ_ONCE        1895 fs/io_uring.c  	events = READ_ONCE(sqe->poll_events);
READ_ONCE        2001 fs/io_uring.c  	count = READ_ONCE(sqe->off);
READ_ONCE        2097 fs/io_uring.c  	req->user_data = READ_ONCE(s->sqe->user_data);
READ_ONCE        2102 fs/io_uring.c  	opcode = READ_ONCE(s->sqe->opcode);
READ_ONCE        2184 fs/io_uring.c  	u8 opcode = READ_ONCE(sqe->opcode);
READ_ONCE        2360 fs/io_uring.c  	int op = READ_ONCE(sqe->opcode);
READ_ONCE        2378 fs/io_uring.c  	flags = READ_ONCE(s->sqe->flags);
READ_ONCE        2379 fs/io_uring.c  	fd = READ_ONCE(s->sqe->fd);
READ_ONCE        2553 fs/io_uring.c  	switch (READ_ONCE(s->sqe->opcode)) {
READ_ONCE        2627 fs/io_uring.c  	if (ctx->cached_sq_head != READ_ONCE(rings->sq.head)) {
READ_ONCE        2664 fs/io_uring.c  	head = READ_ONCE(sq_array[head & ctx->sq_mask]);
READ_ONCE        3020 fs/io_uring.c  	return READ_ONCE(rings->cq.head) == READ_ONCE(rings->cq.tail) ? ret : 0;
READ_ONCE        3662 fs/io_uring.c  	if (READ_ONCE(ctx->rings->sq.tail) - ctx->cached_sq_head !=
READ_ONCE        3665 fs/io_uring.c  	if (READ_ONCE(ctx->rings->cq.head) != ctx->cached_cq_tail)
READ_ONCE          53 fs/iomap/direct-io.c 	struct request_queue *q = READ_ONCE(kiocb->private);
READ_ONCE          57 fs/iomap/direct-io.c 	return blk_poll(q, READ_ONCE(kiocb->ki_cookie), spin);
READ_ONCE         549 fs/iomap/direct-io.c 			if (!READ_ONCE(dio->submit.waiter))
READ_ONCE        1003 fs/jbd2/commit.c 			mapping = READ_ONCE(bh->b_page->mapping);
READ_ONCE        1076 fs/jbd2/transaction.c 	jh = READ_ONCE(bh->b_private);
READ_ONCE        1082 fs/jbd2/transaction.c 	if (READ_ONCE(jh->b_transaction) != handle->h_transaction &&
READ_ONCE        1083 fs/jbd2/transaction.c 	    READ_ONCE(jh->b_next_transaction) != handle->h_transaction)
READ_ONCE        1064 fs/namei.c     	res = READ_ONCE(inode->i_link);
READ_ONCE        1223 fs/namei.c     	while (managed = READ_ONCE(path->dentry->d_flags),
READ_ONCE        1408 fs/namei.c     	while (managed = READ_ONCE(path->dentry->d_flags),
READ_ONCE        4729 fs/namei.c     	link = READ_ONCE(inode->i_link);
READ_ONCE         324 fs/namespace.c 	while (READ_ONCE(mnt->mnt.mnt_flags) & MNT_WRITE_HOLD)
READ_ONCE        1128 fs/namespace.c 	if (likely(READ_ONCE(mnt->mnt_ns))) {
READ_ONCE        1957 fs/namespace.c 	unsigned int max = READ_ONCE(sysctl_mount_max);
READ_ONCE        1264 fs/nfs/dir.c   		parent = READ_ONCE(dentry->d_parent);
READ_ONCE        1269 fs/nfs/dir.c   		if (parent != READ_ONCE(dentry->d_parent))
READ_ONCE         170 fs/nfs/inode.c 	unsigned long cache_validity = READ_ONCE(NFS_I(inode)->cache_validity);
READ_ONCE         181 fs/nfs/inode.c 	unsigned long cache_validity = READ_ONCE(NFS_I(inode)->cache_validity);
READ_ONCE         822 fs/nfs/inode.c 	cache_validity = READ_ONCE(NFS_I(inode)->cache_validity);
READ_ONCE         287 fs/nfs/nfs4proc.c 	cache_validity = READ_ONCE(NFS_I(inode)->cache_validity);
READ_ONCE         238 fs/nfsd/filecache.c 	return filemap_check_wb_err(file->f_mapping, READ_ONCE(file->f_wb_err));
READ_ONCE         368 fs/notify/fanotify/fanotify.c 		conn = READ_ONCE(iter_info->marks[type]->connector);
READ_ONCE         229 fs/notify/mark.c 	struct fsnotify_mark_connector *conn = READ_ONCE(mark->connector);
READ_ONCE         427 fs/overlayfs/file.c 	old_iflags = READ_ONCE(inode->i_flags);
READ_ONCE         120 fs/overlayfs/ovl_entry.h 	return READ_ONCE(oi->__upperdentry);
READ_ONCE         840 fs/overlayfs/readdir.c 		realfile = READ_ONCE(od->upperfile);
READ_ONCE         634 fs/pipe.c      	unsigned long soft_limit = READ_ONCE(pipe_user_pages_soft);
READ_ONCE         641 fs/pipe.c      	unsigned long hard_limit = READ_ONCE(pipe_user_pages_hard);
READ_ONCE         657 fs/pipe.c      	unsigned int max_size = READ_ONCE(pipe_max_size);
READ_ONCE         497 fs/proc/array.c 		rsslim = READ_ONCE(sig->rlim[RLIMIT_RSS].rlim_cur);
READ_ONCE          31 fs/proc_namespace.c 	event = READ_ONCE(ns->event);
READ_ONCE         255 fs/splice.c    	unsigned int buffers = READ_ONCE(pipe->buffers);
READ_ONCE         309 fs/userfaultfd.c 	_pmd = READ_ONCE(*pmd);
READ_ONCE         398 fs/userfaultfd.c 	if (unlikely(READ_ONCE(ctx->released))) {
READ_ONCE         494 fs/userfaultfd.c 	if (likely(must_wait && !READ_ONCE(ctx->released) &&
READ_ONCE         508 fs/userfaultfd.c 		while (!READ_ONCE(uwq.waken)) {
READ_ONCE         516 fs/userfaultfd.c 			if (READ_ONCE(uwq.waken) ||
READ_ONCE         517 fs/userfaultfd.c 			    READ_ONCE(ctx->released) ||
READ_ONCE         606 fs/userfaultfd.c 		if (READ_ONCE(ctx->released) ||
READ_ONCE        1712 fs/userfaultfd.c 	if (READ_ONCE(ctx->mmap_changing))
READ_ONCE        1768 fs/userfaultfd.c 	if (READ_ONCE(ctx->mmap_changing))
READ_ONCE          54 fs/verity/hash_algs.c 	tfm = READ_ONCE(alg->tfm);
READ_ONCE        4499 fs/xfs/libxfs/xfs_bmap.c 		*seq = READ_ONCE(ifp->if_seq);
READ_ONCE        4532 fs/xfs/libxfs/xfs_bmap.c 	*seq = READ_ONCE(ifp->if_seq);
READ_ONCE         621 fs/xfs/libxfs/xfs_iext_tree.c 	WRITE_ONCE(ifp->if_seq, READ_ONCE(ifp->if_seq) + 1);
READ_ONCE         446 fs/xfs/xfs_aops.c 	if (wpc->data_seq != READ_ONCE(ip->i_df.if_seq))
READ_ONCE         449 fs/xfs/xfs_aops.c 	    wpc->cow_seq != READ_ONCE(ip->i_cowfp->if_seq))
READ_ONCE         544 fs/xfs/xfs_aops.c 		wpc->cow_seq = READ_ONCE(ip->i_cowfp->if_seq);
READ_ONCE         567 fs/xfs/xfs_aops.c 	wpc->data_seq = READ_ONCE(ip->i_df.if_seq);
READ_ONCE         610 fs/xfs/xfs_extent_busy.c 		if  (busy_gen != READ_ONCE(pag->pagb_gen))
READ_ONCE         583 fs/xfs/xfs_log_priv.h 	cur_cycle = READ_ONCE(log->l_curr_cycle);
READ_ONCE         585 fs/xfs/xfs_log_priv.h 	cur_block = READ_ONCE(log->l_curr_block);
READ_ONCE         171 include/asm-generic/atomic.h #define atomic_read(v)	READ_ONCE((v)->counter)
READ_ONCE         131 include/asm-generic/barrier.h 	typeof(*p) ___p1 = READ_ONCE(*p);				\
READ_ONCE         186 include/asm-generic/barrier.h 	typeof(*p) ___p1 = READ_ONCE(*p);				\
READ_ONCE         234 include/asm-generic/barrier.h 		VAL = READ_ONCE(*__PTR);			\
READ_ONCE          38 include/asm-generic/bitops/atomic.h 	if (READ_ONCE(*p) & mask)
READ_ONCE          51 include/asm-generic/bitops/atomic.h 	if (!(READ_ONCE(*p) & mask))
READ_ONCE          25 include/asm-generic/bitops/lock.h 	if (READ_ONCE(*p) & mask)
READ_ONCE          63 include/asm-generic/bitops/lock.h 	old = READ_ONCE(*p);
READ_ONCE         119 include/asm-generic/percpu.h 	__ret = READ_ONCE(*raw_cpu_ptr(&(pcp)));			\
READ_ONCE          11 include/asm-generic/preempt.h 	return READ_ONCE(current_thread_info()->preempt_count);
READ_ONCE          60 include/clocksource/hyperv_timer.h 		sequence = READ_ONCE(tsc_pg->tsc_sequence);
READ_ONCE          69 include/clocksource/hyperv_timer.h 		scale = READ_ONCE(tsc_pg->tsc_scale);
READ_ONCE          70 include/clocksource/hyperv_timer.h 		offset = READ_ONCE(tsc_pg->tsc_offset);
READ_ONCE          79 include/clocksource/hyperv_timer.h 	} while (READ_ONCE(tsc_pg->tsc_sequence) != sequence);
READ_ONCE        1499 include/drm/drm_connector.h 	return READ_ONCE(connector->registration_state) ==
READ_ONCE          96 include/drm/spsc_queue.h 	node = READ_ONCE(queue->head);
READ_ONCE         101 include/drm/spsc_queue.h 	next = READ_ONCE(node->next);
READ_ONCE         112 include/drm/spsc_queue.h 			} while (unlikely(!(queue->head = READ_ONCE(node->next))));
READ_ONCE          56 include/linux/average.h 		unsigned long internal = READ_ONCE(e->internal);	\
READ_ONCE         252 include/linux/bitops.h 		old__ = READ_ONCE(*(ptr));			\
READ_ONCE         267 include/linux/bitops.h 		old__ = READ_ONCE(*(ptr));			\
READ_ONCE         550 include/linux/bpf.h 		while ((_prog = READ_ONCE(_item->prog))) {		\
READ_ONCE         595 include/linux/bpf.h 		while ((_prog = READ_ONCE(_item->prog))) {		\
READ_ONCE         841 include/linux/cgroup-defs.h 	struct sock_cgroup_data skcd_buf = {{ .val = READ_ONCE(skcd->val) }};
READ_ONCE         858 include/linux/cgroup-defs.h 	struct sock_cgroup_data skcd_buf = {{ .val = READ_ONCE(skcd->val) }};
READ_ONCE         836 include/linux/cgroup.h 	v = READ_ONCE(skcd->val);
READ_ONCE         527 include/linux/dcache.h 	return READ_ONCE(dentry->d_inode);
READ_ONCE         209 include/linux/dma-resv.h 	return READ_ONCE(obj->lock.ctx);
READ_ONCE          92 include/linux/dynamic_queue_limits.h 	return READ_ONCE(dql->adj_limit) - READ_ONCE(dql->num_queued);
READ_ONCE          82 include/linux/fscrypt.h 	return READ_ONCE(inode->i_crypt_info) != NULL;
READ_ONCE         114 include/linux/fsverity.h 	return READ_ONCE(inode->i_verity_info);
READ_ONCE         470 include/linux/hrtimer.h 	return !!(READ_ONCE(timer->state) & HRTIMER_STATE_ENQUEUED);
READ_ONCE         252 include/linux/huge_mm.h 	return READ_ONCE(huge_zero_page) == page;
READ_ONCE         145 include/linux/hyperv.h 	write_loc = READ_ONCE(rbi->ring_buffer->write_index);
READ_ONCE         158 include/linux/hyperv.h 	read_loc = READ_ONCE(rbi->ring_buffer->read_index);
READ_ONCE          68 include/linux/idr.h 	return READ_ONCE(idr->idr_next);
READ_ONCE         265 include/linux/if_team.h 	int en_port_count = READ_ONCE(team->en_port_count);
READ_ONCE        1212 include/linux/kvm_host.h 	return READ_ONCE(vcpu->requests);
READ_ONCE         268 include/linux/list.h 	return READ_ONCE(head->next) == head;
READ_ONCE         513 include/linux/list.h 	struct list_head *pos__ = READ_ONCE(head__->next); \
READ_ONCE         754 include/linux/list.h 	return !READ_ONCE(h->first);
READ_ONCE          74 include/linux/list_bl.h 	return !((unsigned long)READ_ONCE(h->first) & ~LIST_BL_LOCKMASK);
READ_ONCE          66 include/linux/list_nulls.h 	return is_a_nulls(READ_ONCE(h->first));
READ_ONCE         189 include/linux/llist.h 	return READ_ONCE(head->first) == NULL;
READ_ONCE         366 include/linux/memcontrol.h 		return READ_ONCE(memcg->memory.emin);
READ_ONCE         368 include/linux/memcontrol.h 	return max(READ_ONCE(memcg->memory.emin),
READ_ONCE         369 include/linux/memcontrol.h 		   READ_ONCE(memcg->memory.elow));
READ_ONCE        1307 include/linux/mm.h 	return READ_ONCE(page->mem_cgroup);
READ_ONCE         552 include/linux/netdevice.h 		val = READ_ONCE(n->state);
READ_ONCE         174 include/linux/page-flags.h 	unsigned long head = READ_ONCE(page->compound_head);
READ_ONCE         183 include/linux/page-flags.h 	return READ_ONCE(page->compound_head) & 1;
READ_ONCE         161 include/linux/percpu-refcount.h 	percpu_ptr = READ_ONCE(ref->percpu_count_ptr);
READ_ONCE          82 include/linux/percpu_counter.h 	s64 ret = READ_ONCE(fbc->count);
READ_ONCE         171 include/linux/ptr_ring.h 		return READ_ONCE(r->queue[r->consumer_head]);
READ_ONCE         196 include/linux/ptr_ring.h 		return !r->queue[READ_ONCE(r->consumer_head)];
READ_ONCE          40 include/linux/rbtree.h #define RB_EMPTY_ROOT(root)  (READ_ONCE((root)->rb_node) == NULL)
READ_ONCE          36 include/linux/rcu_sync.h 	return !READ_ONCE(rsp->gp_state); /* GP_IDLE */
READ_ONCE         296 include/linux/rculist.h 	container_of(READ_ONCE(ptr), type, member)
READ_ONCE         334 include/linux/rculist.h 	struct list_head *__next = READ_ONCE(__ptr->next); \
READ_ONCE         354 include/linux/rculist.h 	struct list_head *__next = READ_ONCE(__ptr->next); \
READ_ONCE         390 include/linux/rculist.h 	container_of((typeof(ptr))READ_ONCE(ptr), type, member)
READ_ONCE         135 include/linux/rcupdate.h 		if (READ_ONCE((t)->rcu_tasks_holdout)) \
READ_ONCE         312 include/linux/rcupdate.h 	typeof(*p) *_________p1 = (typeof(*p) *__force)READ_ONCE(p); \
READ_ONCE         319 include/linux/rcupdate.h 	typeof(*p) *________p1 = (typeof(*p) *__force)READ_ONCE(p); \
READ_ONCE         333 include/linux/rcupdate.h 	typeof(p) ________p1 = READ_ONCE(p); \
READ_ONCE         889 include/linux/rcupdate.h 	rcu_callback_t func = READ_ONCE(rhp->func);
READ_ONCE        1401 include/linux/sched.h 	unsigned int tsk_state = READ_ONCE(tsk->state);
READ_ONCE        1817 include/linux/sched.h 	return READ_ONCE(p->cpu);
READ_ONCE        1819 include/linux/sched.h 	return READ_ONCE(task_thread_info(p)->cpu);
READ_ONCE          87 include/linux/sched/cputime.h 	if (!READ_ONCE(tsk->signal->posix_cputimers.timers_active))
READ_ONCE         687 include/linux/sched/signal.h 	return READ_ONCE(task->signal->rlim[limit].rlim_cur);
READ_ONCE         693 include/linux/sched/signal.h 	return READ_ONCE(task->signal->rlim[limit].rlim_max);
READ_ONCE         113 include/linux/seqlock.h 	ret = READ_ONCE(s->sequence);
READ_ONCE         132 include/linux/seqlock.h 	unsigned ret = READ_ONCE(s->sequence);
READ_ONCE         184 include/linux/seqlock.h 	unsigned ret = READ_ONCE(s->sequence);
READ_ONCE         282 include/linux/seqlock.h 	int seq = READ_ONCE(s->sequence); /* ^^^ */
READ_ONCE        1509 include/linux/skbuff.h 	return READ_ONCE(list->next) == (const struct sk_buff *) list;
READ_ONCE        1800 include/linux/skbuff.h 	struct sk_buff *skb = READ_ONCE(list_->prev);
READ_ONCE          61 include/linux/slub_def.h #define slub_percpu_partial_read_once(c)     READ_ONCE(slub_percpu_partial(c))
READ_ONCE          62 include/linux/srcutiny.h 	idx = READ_ONCE(ssp->srcu_idx);
READ_ONCE          83 include/linux/srcutiny.h 	idx = READ_ONCE(ssp->srcu_idx) & 0x1;
READ_ONCE          86 include/linux/srcutiny.h 		 READ_ONCE(ssp->srcu_lock_nesting[!idx]),
READ_ONCE          87 include/linux/srcutiny.h 		 READ_ONCE(ssp->srcu_lock_nesting[idx]));
READ_ONCE         457 include/linux/tcp.h 	int somaxconn = READ_ONCE(sock_net(sk)->core.sysctl_somaxconn);
READ_ONCE         482 include/linux/tcp.h 	u16 user_mss = READ_ONCE(tp->rx_opt.user_mss);
READ_ONCE         198 include/linux/tpm_eventlog.h 	count = READ_ONCE(event->count);
READ_ONCE         199 include/linux/tpm_eventlog.h 	event_type = READ_ONCE(event->event_type);
READ_ONCE         308 include/linux/trace_events.h 	return !!READ_ONCE(call->prog_array);
READ_ONCE         746 include/linux/vmw_vmci_defs.h 	return READ_ONCE(*(unsigned long *)var);
READ_ONCE          60 include/net/arp.h 		if (READ_ONCE(n->confirmed) != now)
READ_ONCE          74 include/net/busy_poll.h 	unsigned long bp_usec = READ_ONCE(sysctl_net_busy_poll);
READ_ONCE          90 include/net/busy_poll.h 	unsigned long bp_usec = READ_ONCE(sk->sk_ll_usec);
READ_ONCE         105 include/net/busy_poll.h 	unsigned int napi_id = READ_ONCE(sk->sk_napi_id);
READ_ONCE         135 include/net/busy_poll.h 	if (!READ_ONCE(sk->sk_napi_id))
READ_ONCE         414 include/net/ip.h 	u8 pmtudisc = READ_ONCE(inet_sk(sk)->pmtudisc);
READ_ONCE         448 include/net/ip.h 	return min(READ_ONCE(dst->dev->mtu), IP_MAX_MTU);
READ_ONCE         460 include/net/ip.h 	return min(READ_ONCE(skb_dst(skb)->dev->mtu), IP_MAX_MTU);
READ_ONCE         988 include/net/ip_vs.h 	return READ_ONCE(ipvs->sysctl_sync_threshold[1]);
READ_ONCE         993 include/net/ip_vs.h 	return READ_ONCE(ipvs->sysctl_sync_refresh_period);
READ_ONCE        1018 include/net/ip_vs.h 	return READ_ONCE(ipvs->sysctl_sync_ports);
READ_ONCE         417 include/net/ndisc.h 		if (READ_ONCE(n->confirmed) != now)
READ_ONCE         434 include/net/ndisc.h 		if (READ_ONCE(n->confirmed) != now)
READ_ONCE         441 include/net/neighbour.h 	if (READ_ONCE(neigh->used) != now)
READ_ONCE         470 include/net/neighbour.h 		hh_len = READ_ONCE(hh->hh_len);
READ_ONCE        1265 include/net/netfilter/nf_tables.h 	return 1 << READ_ONCE(net->nft.gencursor);
READ_ONCE         188 include/net/request_sock.h 	return READ_ONCE(queue->rskq_accept_head) == NULL;
READ_ONCE         152 include/net/sch_generic.h 		return READ_ONCE(qdisc->empty);
READ_ONCE         153 include/net/sch_generic.h 	return !READ_ONCE(qdisc->q.qlen);
READ_ONCE         543 include/net/sock.h 		return READ_ONCE(sk->sk_peek_off);
READ_ONCE         551 include/net/sock.h 	s32 off = READ_ONCE(sk->sk_peek_off);
READ_ONCE         886 include/net/sock.h 	return READ_ONCE(sk->sk_wmem_queued) >> 1;
READ_ONCE         891 include/net/sock.h 	return READ_ONCE(sk->sk_sndbuf) - READ_ONCE(sk->sk_wmem_queued);
READ_ONCE         962 include/net/sock.h 	if (unlikely(READ_ONCE(sk->sk_incoming_cpu) != cpu))
READ_ONCE        1041 include/net/sock.h 	if (unlikely(READ_ONCE(sk->sk_backlog.tail))) {
READ_ONCE        1220 include/net/sock.h 	if (READ_ONCE(sk->sk_wmem_queued) >= READ_ONCE(sk->sk_sndbuf))
READ_ONCE        1948 include/net/sock.h 	if (!READ_ONCE(sk->sk_dst_pending_confirm))
READ_ONCE        1959 include/net/sock.h 		if (READ_ONCE(n->confirmed) != now)
READ_ONCE        1961 include/net/sock.h 		if (sk && READ_ONCE(sk->sk_dst_pending_confirm))
READ_ONCE        2273 include/net/sock.h 	return refcount_read(&sk->sk_wmem_alloc) < (READ_ONCE(sk->sk_sndbuf) >> 1);
READ_ONCE        2293 include/net/sock.h 	int v = waitall ? len : min_t(int, READ_ONCE(sk->sk_rcvlowat), len);
READ_ONCE        2350 include/net/sock.h 	return READ_ONCE(sk->sk_stamp);
READ_ONCE        2592 include/net/sock.h 	if (!sk || !sk_fullsock(sk) || READ_ONCE(sk->sk_pacing_shift) == val)
READ_ONCE          59 include/net/tc_act/tc_gact.h 	return READ_ONCE(a->tcfa_action) & TC_ACT_EXT_VAL_MASK;
READ_ONCE         261 include/net/tcp.h 	return READ_ONCE(tcp_memory_pressure);
READ_ONCE         496 include/net/tcp.h 			last_overflow = READ_ONCE(reuse->synq_overflow_ts);
READ_ONCE         504 include/net/tcp.h 	last_overflow = READ_ONCE(tcp_sk(sk)->rx_opt.ts_recent_stamp);
READ_ONCE         520 include/net/tcp.h 			last_overflow = READ_ONCE(reuse->synq_overflow_ts);
READ_ONCE         527 include/net/tcp.h 	last_overflow = READ_ONCE(tcp_sk(sk)->rx_opt.ts_recent_stamp);
READ_ONCE        1394 include/net/tcp.h 	return tcp_win_from_space(sk, READ_ONCE(sk->sk_rcvbuf) -
READ_ONCE        1395 include/net/tcp.h 				  READ_ONCE(sk->sk_backlog.len) -
READ_ONCE        1401 include/net/tcp.h 	return tcp_win_from_space(sk, READ_ONCE(sk->sk_rcvbuf));
READ_ONCE        1411 include/net/tcp.h 	int rcvbuf = READ_ONCE(sk->sk_rcvbuf);
READ_ONCE        1944 include/net/tcp.h 	u32 notsent_bytes = READ_ONCE(tp->write_seq) -
READ_ONCE        1945 include/net/tcp.h 			    READ_ONCE(tp->snd_nxt);
READ_ONCE         427 include/net/tls.h 	return READ_ONCE(rec->tx_ready);
READ_ONCE         248 include/net/udp.h 	return sk_rmem_alloc_get(sk) - READ_ONCE(udp_sk(sk)->forward_deficit);
READ_ONCE          85 include/trace/events/sock.h 		__entry->sk_rcvbuf  = READ_ONCE(sk->sk_rcvbuf);
READ_ONCE         118 include/trace/events/sock.h 		__entry->wmem_queued = READ_ONCE(sk->sk_wmem_queued);
READ_ONCE          13 include/vdso/helpers.h 	while ((seq = READ_ONCE(vd->seq)) & 1)
READ_ONCE          26 include/vdso/helpers.h 	seq = READ_ONCE(vd->seq);
READ_ONCE        1186 ipc/msg.c      		msg = READ_ONCE(msr_d.r_msg);
READ_ONCE        2174 ipc/sem.c      		error = READ_ONCE(queue.status);
READ_ONCE        2192 ipc/sem.c      		error = READ_ONCE(queue.status);
READ_ONCE         150 kernel/acct.c  	res = to_acct(READ_ONCE(ns->bacct));
READ_ONCE         162 kernel/acct.c  	if (res != to_acct(READ_ONCE(ns->bacct))) {
READ_ONCE        1449 kernel/audit.c 		t = READ_ONCE(current->signal->audit_tty);
READ_ONCE        1470 kernel/audit.c 			t = READ_ONCE(current->signal->audit_tty);
READ_ONCE         255 kernel/audit_tree.c 		if (READ_ONCE(p->key) == key) {
READ_ONCE         515 kernel/bpf/arraymap.c 	if (elem && (ptr = READ_ONCE(*elem)))
READ_ONCE         615 kernel/bpf/arraymap.c 		ptr = READ_ONCE(*elem);
READ_ONCE         711 kernel/bpf/arraymap.c 		ee = READ_ONCE(array->ptrs[i]);
READ_ONCE         800 kernel/bpf/arraymap.c 	return READ_ONCE(*inner_map);
READ_ONCE        1471 kernel/bpf/core.c 		prog = READ_ONCE(array->ptrs[index]);
READ_ONCE         543 kernel/bpf/cpumap.c 		rcpu = READ_ONCE(cmap->cpu_map[i]);
READ_ONCE         563 kernel/bpf/cpumap.c 	rcpu = READ_ONCE(cmap->cpu_map[key]);
READ_ONCE         429 kernel/bpf/devmap.c 	obj = READ_ONCE(dtab->netdev_map[key]);
READ_ONCE         791 kernel/bpf/devmap.c 				dev = READ_ONCE(dtab->netdev_map[i]);
READ_ONCE        1424 kernel/bpf/hashtab.c 		*value = map->ops->map_fd_sys_lookup_elem(READ_ONCE(*ptr));
READ_ONCE        1477 kernel/bpf/hashtab.c 	return READ_ONCE(*inner_map);
READ_ONCE         346 kernel/bpf/helpers.c 		ptr = &READ_ONCE(storage->buf)->data[0];
READ_ONCE         124 kernel/bpf/local_storage.c 	return &READ_ONCE(storage->buf)->data[0];
READ_ONCE         396 kernel/bpf/local_storage.c 				  &READ_ONCE(storage->buf)->data[0], m);
READ_ONCE         237 kernel/bpf/reuseport_array.c 	if (READ_ONCE(nsk->sk_user_data))
READ_ONCE         387 kernel/bpf/stackmap.c 	bucket = READ_ONCE(smap->buckets[id]);
READ_ONCE         366 kernel/bpf/syscall.c 	if (READ_ONCE(map->frozen))
READ_ONCE         401 kernel/bpf/syscall.c 		   READ_ONCE(map->frozen));
READ_ONCE        1180 kernel/bpf/syscall.c 	if (READ_ONCE(map->frozen)) {
READ_ONCE         171 kernel/bpf/xskmap.c 	xs = READ_ONCE(m->xsk_map[key]);
READ_ONCE         254 kernel/bpf/xskmap.c 	old_xs = READ_ONCE(*map_entry);
READ_ONCE         303 kernel/bpf/xskmap.c 	if (READ_ONCE(*map_entry) == xs) {
READ_ONCE        3466 kernel/cgroup/cgroup.c 	int descendants = READ_ONCE(cgrp->max_descendants);
READ_ONCE        3509 kernel/cgroup/cgroup.c 	int depth = READ_ONCE(cgrp->max_depth);
READ_ONCE         176 kernel/events/core.c 	return READ_ONCE(event->owner) == TASK_TOMBSTONE;
READ_ONCE         256 kernel/events/core.c 	struct task_struct *task = READ_ONCE(ctx->task); /* verified in event_function */
READ_ONCE         310 kernel/events/core.c 	struct task_struct *task = READ_ONCE(ctx->task);
READ_ONCE         512 kernel/events/core.c 	u64 max_len = READ_ONCE(perf_sample_allowed_ns);
READ_ONCE        1276 kernel/events/core.c 	ctx = READ_ONCE(event->ctx);
READ_ONCE        2102 kernel/events/core.c 	if (READ_ONCE(event->pending_disable) >= 0) {
READ_ONCE        2659 kernel/events/core.c 	struct task_struct *task = READ_ONCE(ctx->task);
READ_ONCE        2845 kernel/events/core.c 	if (READ_ONCE(event->state) != PERF_EVENT_STATE_ACTIVE)
READ_ONCE        2855 kernel/events/core.c 	if (READ_ONCE(event->oncpu) != smp_processor_id())
READ_ONCE        2884 kernel/events/core.c 		if (READ_ONCE(event->state) != PERF_EVENT_STATE_ACTIVE)
READ_ONCE        2895 kernel/events/core.c 		ret = cpu_function_call(READ_ONCE(event->oncpu),
READ_ONCE        4093 kernel/events/core.c 	enum perf_event_state state = READ_ONCE(event->state);
READ_ONCE        4112 kernel/events/core.c 		event_cpu = READ_ONCE(event->oncpu);
READ_ONCE        4624 kernel/events/core.c 	owner = READ_ONCE(event->owner);
READ_ONCE        4722 kernel/events/core.c 		ctx = READ_ONCE(child->ctx);
READ_ONCE        5751 kernel/events/core.c 		aux_offset = READ_ONCE(rb->user_page->aux_offset);
READ_ONCE        5752 kernel/events/core.c 		aux_size = READ_ONCE(rb->user_page->aux_size);
READ_ONCE        5975 kernel/events/core.c 	int cpu = READ_ONCE(event->pending_disable);
READ_ONCE        6998 kernel/events/core.c 			cpu = READ_ONCE(iter->oncpu);
READ_ONCE        8211 kernel/events/core.c 	READ_ONCE(event->overflow_handler)(event, data, regs);
READ_ONCE        9015 kernel/events/core.c 	event->orig_overflow_handler = READ_ONCE(event->overflow_handler);
READ_ONCE        9236 kernel/events/core.c 	struct task_struct *task = READ_ONCE(event->ctx->task);
READ_ONCE        10872 kernel/events/core.c 	gctx = READ_ONCE(group_leader->ctx);
READ_ONCE          60 kernel/events/ring_buffer.c 	nest = READ_ONCE(rb->nest);
READ_ONCE         192 kernel/events/ring_buffer.c 		tail = READ_ONCE(rb->user_page->data_tail);
READ_ONCE         396 kernel/events/ring_buffer.c 	nest = READ_ONCE(rb->aux_nest);
READ_ONCE         420 kernel/events/ring_buffer.c 		aux_tail = READ_ONCE(rb->user_page->aux_tail);
READ_ONCE        1545 kernel/events/uprobes.c 	area = READ_ONCE(mm->uprobes_state.xol_area); /* ^^^ */
READ_ONCE        1847 kernel/events/uprobes.c 	area = READ_ONCE(current->mm->uprobes_state.xol_area); /* ^^^ */
READ_ONCE        1265 kernel/exit.c  	int exit_state = READ_ONCE(p->exit_state);
READ_ONCE        1539 kernel/fork.c  	cpu_limit = READ_ONCE(sig->rlim[RLIMIT_CPU].rlim_cur);
READ_ONCE         647 kernel/futex.c 	mapping = READ_ONCE(page->mapping);
READ_ONCE         723 kernel/futex.c 		if (READ_ONCE(page->mapping) != mapping) {
READ_ONCE         730 kernel/futex.c 		inode = READ_ONCE(mapping->host);
READ_ONCE        2407 kernel/futex.c 	lock_ptr = READ_ONCE(q->lock_ptr);
READ_ONCE          70 kernel/kcov.c  	mode = READ_ONCE(t->kcov_mode);
READ_ONCE         107 kernel/kcov.c  	pos = READ_ONCE(area[0]) + 1;
READ_ONCE         135 kernel/kcov.c  	count = READ_ONCE(area[0]);
READ_ONCE         340 kernel/kcov.c  		READ_ONCE(area[offset]);
READ_ONCE         154 kernel/ksysfs.c 	return sprintf(buf, "%d\n", READ_ONCE(rcu_expedited));
READ_ONCE         171 kernel/ksysfs.c 	return sprintf(buf, "%d\n", READ_ONCE(rcu_normal));
READ_ONCE         184 kernel/livepatch/transition.c 		task->patch_state = READ_ONCE(klp_target_state);
READ_ONCE         697 kernel/locking/lockdep.c 	int i, depth = READ_ONCE(p->lockdep_depth);
READ_ONCE        2865 kernel/locking/lockdep.c 		if (READ_ONCE(chain->chain_key) == chain_key) {
READ_ONCE         104 kernel/locking/mcs_spinlock.h 	struct mcs_spinlock *next = READ_ONCE(node->next);
READ_ONCE         113 kernel/locking/mcs_spinlock.h 		while (!(next = READ_ONCE(node->next)))
READ_ONCE         520 kernel/locking/mutex.c 	if (ww_ctx->acquired > 0 && READ_ONCE(ww->ctx))
READ_ONCE         809 kernel/locking/mutex.c 	struct ww_acquire_ctx *hold_ctx = READ_ONCE(ww->ctx);
READ_ONCE         943 kernel/locking/mutex.c 		if (unlikely(ww_ctx == READ_ONCE(ww->ctx)))
READ_ONCE         137 kernel/locking/osq_lock.c 	while (!READ_ONCE(node->locked)) {
READ_ONCE         178 kernel/locking/osq_lock.c 		prev = READ_ONCE(node->prev);
READ_ONCE         481 kernel/locking/qspinlock.c 		next = READ_ONCE(node->next);
READ_ONCE         121 kernel/locking/qspinlock_paravirt.h 	return !READ_ONCE(lock->locked) &&
READ_ONCE         246 kernel/locking/qspinlock_paravirt.h 		if (READ_ONCE(he->lock) == lock) {
READ_ONCE         247 kernel/locking/qspinlock_paravirt.h 			node = READ_ONCE(he->node);
READ_ONCE         272 kernel/locking/qspinlock_paravirt.h 	return READ_ONCE(prev->state) != vcpu_running;
READ_ONCE         302 kernel/locking/qspinlock_paravirt.h 			if (READ_ONCE(node->locked))
READ_ONCE         322 kernel/locking/qspinlock_paravirt.h 		if (!READ_ONCE(node->locked)) {
READ_ONCE         343 kernel/locking/qspinlock_paravirt.h 				  !READ_ONCE(node->locked));
READ_ONCE         414 kernel/locking/qspinlock_paravirt.h 	if (READ_ONCE(pn->state) == vcpu_hashed)
READ_ONCE         134 kernel/locking/rtmutex.c 	owner = READ_ONCE(*p);
READ_ONCE         110 kernel/locking/rtmutex_common.h 	unsigned long owner = (unsigned long) READ_ONCE(lock->owner);
READ_ONCE          54 kernel/locking/spinlock_debug.c 	struct task_struct *owner = READ_ONCE(lock->owner);
READ_ONCE          63 kernel/locking/spinlock_debug.c 		lock, READ_ONCE(lock->magic),
READ_ONCE          66 kernel/locking/spinlock_debug.c 		READ_ONCE(lock->owner_cpu));
READ_ONCE          83 kernel/locking/spinlock_debug.c 	SPIN_BUG_ON(READ_ONCE(lock->magic) != SPINLOCK_MAGIC, lock, "bad magic");
READ_ONCE          84 kernel/locking/spinlock_debug.c 	SPIN_BUG_ON(READ_ONCE(lock->owner) == current, lock, "recursion");
READ_ONCE          85 kernel/locking/spinlock_debug.c 	SPIN_BUG_ON(READ_ONCE(lock->owner_cpu) == raw_smp_processor_id(),
READ_ONCE          47 kernel/pid_namespace.c 	kc = READ_ONCE(*pkc);
READ_ONCE          59 kernel/pid_namespace.c 	return READ_ONCE(*pkc);
READ_ONCE         181 kernel/power/energy_model.c 	return READ_ONCE(per_cpu(em_data, cpu));
READ_ONCE         217 kernel/power/energy_model.c 		if (READ_ONCE(per_cpu(em_data, cpu))) {
READ_ONCE        1705 kernel/printk/printk.c 	waiter = READ_ONCE(console_waiter);
READ_ONCE        1750 kernel/printk/printk.c 	owner = READ_ONCE(console_owner);
READ_ONCE        1751 kernel/printk/printk.c 	waiter = READ_ONCE(console_waiter);
READ_ONCE        1775 kernel/printk/printk.c 	while (READ_ONCE(console_waiter))
READ_ONCE          91 kernel/rcu/rcu.h 	s = (READ_ONCE(*sp) + 2 * RCU_SEQ_STATE_MASK + 1) & ~RCU_SEQ_STATE_MASK;
READ_ONCE          99 kernel/rcu/rcu.h 	return READ_ONCE(*sp);
READ_ONCE         108 kernel/rcu/rcu.h 	return ULONG_CMP_LT((s - 1) & ~RCU_SEQ_STATE_MASK, READ_ONCE(*sp));
READ_ONCE         117 kernel/rcu/rcu.h 	return ULONG_CMP_GE(READ_ONCE(*sp), s);
READ_ONCE          15 kernel/rcu/rcu_segcblist.h 	return READ_ONCE(rclp->len);
READ_ONCE          49 kernel/rcu/rcu_segcblist.h 	return !READ_ONCE(rsclp->head);
READ_ONCE          58 kernel/rcu/rcu_segcblist.h 	return READ_ONCE(rsclp->len);
READ_ONCE          96 kernel/rcu/rcu_segcblist.h 	return !READ_ONCE(*READ_ONCE(rsclp->tails[seg]));
READ_ONCE         204 kernel/rcu/rcutorture.c 	unsigned int i = READ_ONCE(rcu_torture_writer_state);
READ_ONCE        1067 kernel/rcu/rcutorture.c 		    !READ_ONCE(rcu_fwd_cb_nodelay) &&
READ_ONCE        1484 kernel/rcu/rcutorture.c 		wtp = READ_ONCE(writer_task);
READ_ONCE        1648 kernel/rcu/rcutorture.c 	if (READ_ONCE(fcsp->stop)) {
READ_ONCE        1792 kernel/rcu/rcutorture.c 	cver = READ_ONCE(rcu_torture_current_version);
READ_ONCE        1801 kernel/rcu/rcutorture.c 	       !READ_ONCE(rcu_fwd_emergency_stop) && !torture_must_stop()) {
READ_ONCE        1811 kernel/rcu/rcutorture.c 	    !READ_ONCE(rcu_fwd_emergency_stop) && !torture_must_stop()) {
READ_ONCE        1813 kernel/rcu/rcutorture.c 		cver = READ_ONCE(rcu_torture_current_version) - cver;
READ_ONCE        1825 kernel/rcu/rcutorture.c 		WARN_ON(READ_ONCE(fcs.stop) != 2);
READ_ONCE        1848 kernel/rcu/rcutorture.c 	if (READ_ONCE(rcu_fwd_emergency_stop))
READ_ONCE        1865 kernel/rcu/rcutorture.c 	cver = READ_ONCE(rcu_torture_current_version);
READ_ONCE        1870 kernel/rcu/rcutorture.c 	       !READ_ONCE(rcu_fwd_emergency_stop) && !torture_must_stop()) {
READ_ONCE        1871 kernel/rcu/rcutorture.c 		rfcp = READ_ONCE(rcu_fwd_cb_head);
READ_ONCE        1874 kernel/rcu/rcutorture.c 			rfcpn = READ_ONCE(rfcp->rfc_next);
READ_ONCE        1896 kernel/rcu/rcutorture.c 	n_launders_cb_snap = READ_ONCE(n_launders_cb);
READ_ONCE        1897 kernel/rcu/rcutorture.c 	cver = READ_ONCE(rcu_torture_current_version) - cver;
READ_ONCE        1902 kernel/rcu/rcutorture.c 	if (!torture_must_stop() && !READ_ONCE(rcu_fwd_emergency_stop) &&
READ_ONCE        1928 kernel/rcu/rcutorture.c 	rcu_fwd_progress_check(1 + (jiffies - READ_ONCE(rcu_fwd_startat)) / 2);
READ_ONCE          99 kernel/rcu/srcutiny.c 	if (!newval && READ_ONCE(ssp->srcu_gp_waiting))
READ_ONCE         117 kernel/rcu/srcutiny.c 	if (ssp->srcu_gp_running || !READ_ONCE(ssp->srcu_cb_head))
READ_ONCE         130 kernel/rcu/srcutiny.c 	swait_event_exclusive(ssp->srcu_wq, !READ_ONCE(ssp->srcu_lock_nesting[idx]));
READ_ONCE         149 kernel/rcu/srcutiny.c 	if (READ_ONCE(ssp->srcu_cb_head))
READ_ONCE         169 kernel/rcu/srcutiny.c 	if (!READ_ONCE(ssp->srcu_gp_running)) {
READ_ONCE         255 kernel/rcu/srcutree.c 		sum += READ_ONCE(cpuc->srcu_lock_count[idx]);
READ_ONCE         272 kernel/rcu/srcutree.c 		sum += READ_ONCE(cpuc->srcu_unlock_count[idx]);
READ_ONCE         341 kernel/rcu/srcutree.c 		sum += READ_ONCE(cpuc->srcu_lock_count[0]);
READ_ONCE         342 kernel/rcu/srcutree.c 		sum += READ_ONCE(cpuc->srcu_lock_count[1]);
READ_ONCE         343 kernel/rcu/srcutree.c 		sum -= READ_ONCE(cpuc->srcu_unlock_count[0]);
READ_ONCE         344 kernel/rcu/srcutree.c 		sum -= READ_ONCE(cpuc->srcu_unlock_count[1]);
READ_ONCE         357 kernel/rcu/srcutree.c 	if (ULONG_CMP_LT(READ_ONCE(ssp->srcu_gp_seq),
READ_ONCE         358 kernel/rcu/srcutree.c 			 READ_ONCE(ssp->srcu_gp_seq_needed_exp)))
READ_ONCE         387 kernel/rcu/srcutree.c 	if (WARN_ON(rcu_seq_state(READ_ONCE(ssp->srcu_gp_seq)) != SRCU_STATE_IDLE) ||
READ_ONCE         390 kernel/rcu/srcutree.c 			__func__, ssp, rcu_seq_state(READ_ONCE(ssp->srcu_gp_seq)));
READ_ONCE         407 kernel/rcu/srcutree.c 	idx = READ_ONCE(ssp->srcu_idx) & 0x1;
READ_ONCE         453 kernel/rcu/srcutree.c 	state = rcu_seq_state(READ_ONCE(ssp->srcu_gp_seq));
READ_ONCE         605 kernel/rcu/srcutree.c 		    ULONG_CMP_GE(READ_ONCE(snp->srcu_gp_seq_needed_exp), s))
READ_ONCE         784 kernel/rcu/srcutree.c 	tlast = READ_ONCE(ssp->srcu_last_gp_end);
READ_ONCE         792 kernel/rcu/srcutree.c 	if (ULONG_CMP_LT(curseq, READ_ONCE(ssp->srcu_gp_seq_needed)))
READ_ONCE        1116 kernel/rcu/srcutree.c 		idx = rcu_seq_state(READ_ONCE(ssp->srcu_gp_seq));
READ_ONCE        1126 kernel/rcu/srcutree.c 	if (rcu_seq_state(READ_ONCE(ssp->srcu_gp_seq)) == SRCU_STATE_SCAN1) {
READ_ONCE        1136 kernel/rcu/srcutree.c 	if (rcu_seq_state(READ_ONCE(ssp->srcu_gp_seq)) == SRCU_STATE_SCAN2) {
READ_ONCE          78 kernel/rcu/sync.c 	WARN_ON_ONCE(READ_ONCE(rsp->gp_state) == GP_IDLE);
READ_ONCE          79 kernel/rcu/sync.c 	WARN_ON_ONCE(READ_ONCE(rsp->gp_state) == GP_PASSED);
READ_ONCE         155 kernel/rcu/sync.c 	wait_event(rsp->gp_wait, READ_ONCE(rsp->gp_state) >= GP_PASSED);
READ_ONCE         170 kernel/rcu/sync.c 	WARN_ON_ONCE(READ_ONCE(rsp->gp_state) == GP_IDLE);
READ_ONCE         171 kernel/rcu/sync.c 	WARN_ON_ONCE(READ_ONCE(rsp->gp_count) == 0);
READ_ONCE         193 kernel/rcu/sync.c 	WARN_ON_ONCE(READ_ONCE(rsp->gp_count));
READ_ONCE         194 kernel/rcu/sync.c 	WARN_ON_ONCE(READ_ONCE(rsp->gp_state) == GP_PASSED);
READ_ONCE         193 kernel/rcu/tree.c 	return READ_ONCE(rnp->qsmaskinitnext);
READ_ONCE         452 kernel/rcu/tree.c 	j = READ_ONCE(jiffies_till_first_fqs) +
READ_ONCE         453 kernel/rcu/tree.c 		      2 * READ_ONCE(jiffies_till_next_fqs);
READ_ONCE         506 kernel/rcu/tree.c 	return READ_ONCE(rcu_state.gp_seq);
READ_ONCE         548 kernel/rcu/tree.c 		*flags = READ_ONCE(rcu_state.gp_flags);
READ_ONCE        1053 kernel/rcu/tree.c 	jtsq = READ_ONCE(jiffies_to_sched_qs);
READ_ONCE        1056 kernel/rcu/tree.c 	if (!READ_ONCE(*rnhqp) &&
READ_ONCE        1076 kernel/rcu/tree.c 			      READ_ONCE(rdp->last_fqs_resched) + jtsq * 3)) {
READ_ONCE        1090 kernel/rcu/tree.c 			       READ_ONCE(rdp->last_fqs_resched) + jtsq)) {
READ_ONCE        1189 kernel/rcu/tree.c 	trace_rcu_grace_period(rcu_state.name, READ_ONCE(rcu_state.gp_seq), TPS("newreq"));
READ_ONCE        1237 kernel/rcu/tree.c 	    !READ_ONCE(rcu_state.gp_flags) ||
READ_ONCE        1241 kernel/rcu/tree.c 	WRITE_ONCE(rcu_state.gp_wake_seq, READ_ONCE(rcu_state.gp_seq));
READ_ONCE        1382 kernel/rcu/tree.c 	    unlikely(READ_ONCE(rdp->gpwrap))) {
READ_ONCE        1393 kernel/rcu/tree.c 	    unlikely(READ_ONCE(rdp->gpwrap))) {
READ_ONCE        1422 kernel/rcu/tree.c 	     !unlikely(READ_ONCE(rdp->gpwrap))) || /* w/out lock. */
READ_ONCE        1454 kernel/rcu/tree.c 	if (!READ_ONCE(rcu_state.gp_flags)) {
READ_ONCE        1580 kernel/rcu/tree.c 	*gfp = READ_ONCE(rcu_state.gp_flags);
READ_ONCE        1585 kernel/rcu/tree.c 	if (!READ_ONCE(rnp->qsmask) && !rcu_preempt_blocked_readers_cgp(rnp))
READ_ONCE        1608 kernel/rcu/tree.c 	if (READ_ONCE(rcu_state.gp_flags) & RCU_GP_FLAG_FQS) {
READ_ONCE        1611 kernel/rcu/tree.c 			   READ_ONCE(rcu_state.gp_flags) & ~RCU_GP_FLAG_FQS);
READ_ONCE        1628 kernel/rcu/tree.c 	j = READ_ONCE(jiffies_till_first_fqs);
READ_ONCE        1637 kernel/rcu/tree.c 				       READ_ONCE(rcu_state.gp_seq),
READ_ONCE        1645 kernel/rcu/tree.c 		if (!READ_ONCE(rnp->qsmask) &&
READ_ONCE        1652 kernel/rcu/tree.c 					       READ_ONCE(rcu_state.gp_seq),
READ_ONCE        1657 kernel/rcu/tree.c 					       READ_ONCE(rcu_state.gp_seq),
READ_ONCE        1662 kernel/rcu/tree.c 			j = READ_ONCE(jiffies_till_next_fqs);
READ_ONCE        1669 kernel/rcu/tree.c 					       READ_ONCE(rcu_state.gp_seq),
READ_ONCE        1761 kernel/rcu/tree.c 				       READ_ONCE(rcu_state.gp_seq),
READ_ONCE        1781 kernel/rcu/tree.c 					       READ_ONCE(rcu_state.gp_seq),
READ_ONCE        1785 kernel/rcu/tree.c 					 READ_ONCE(rcu_state.gp_flags) &
READ_ONCE        1795 kernel/rcu/tree.c 					       READ_ONCE(rcu_state.gp_seq),
READ_ONCE        1824 kernel/rcu/tree.c 		   READ_ONCE(rcu_state.gp_flags) | RCU_GP_FLAG_FQS);
READ_ONCE        2315 kernel/rcu/tree.c 		ret = (READ_ONCE(rcu_state.gp_flags) & RCU_GP_FLAG_FQS) ||
READ_ONCE        2328 kernel/rcu/tree.c 	if (READ_ONCE(rcu_state.gp_flags) & RCU_GP_FLAG_FQS) {
READ_ONCE        2333 kernel/rcu/tree.c 		   READ_ONCE(rcu_state.gp_flags) | RCU_GP_FLAG_FQS);
READ_ONCE        2377 kernel/rcu/tree.c 	    likely(READ_ONCE(rcu_scheduler_fully_active)))
READ_ONCE        2824 kernel/rcu/tree.c 	    unlikely(READ_ONCE(rdp->gpwrap))) /* outside lock */
READ_ONCE        3149 kernel/rcu/tree.c 	rdp->rcu_onl_gp_seq = READ_ONCE(rcu_state.gp_seq);
READ_ONCE        3150 kernel/rcu/tree.c 	rdp->rcu_onl_gp_flags = READ_ONCE(rcu_state.gp_flags);
READ_ONCE        3186 kernel/rcu/tree.c 	rdp->rcu_ofl_gp_seq = READ_ONCE(rcu_state.gp_seq);
READ_ONCE        3187 kernel/rcu/tree.c 	rdp->rcu_ofl_gp_flags = READ_ONCE(rcu_state.gp_flags);
READ_ONCE         156 kernel/rcu/tree_exp.h 	       READ_ONCE(rnp->expmask) == 0;
READ_ONCE         282 kernel/rcu/tree_exp.h 	if (ULONG_CMP_LT(READ_ONCE(rnp->exp_seq_rq), s) &&
READ_ONCE         284 kernel/rcu/tree_exp.h 	     ULONG_CMP_LT(READ_ONCE(rnp_root->exp_seq_rq), s)) &&
READ_ONCE         431 kernel/rcu/tree_exp.h 		if (!READ_ONCE(rnp->expmask))
READ_ONCE         433 kernel/rcu/tree_exp.h 		if (!READ_ONCE(rcu_par_gp_wq) ||
READ_ONCE         492 kernel/rcu/tree_exp.h 				if (!(READ_ONCE(rnp->expmask) & mask))
READ_ONCE         504 kernel/rcu/tree_exp.h 			READ_ONCE(rnp_root->expmask),
READ_ONCE         515 kernel/rcu/tree_exp.h 					READ_ONCE(rnp->expmask),
READ_ONCE         523 kernel/rcu/tree_exp.h 				if (!(READ_ONCE(rnp->expmask) & mask))
READ_ONCE         552 kernel/rcu/tree_exp.h 		if (ULONG_CMP_LT(READ_ONCE(rnp->exp_seq_rq), s)) {
READ_ONCE         717 kernel/rcu/tree_exp.h 	if (!(READ_ONCE(rnp->expmask) & rdp->grpmask) ||
READ_ONCE         740 kernel/rcu/tree_exp.h 	if (!(READ_ONCE(rnp->expmask) & rdp->grpmask) ||
READ_ONCE         343 kernel/rcu/tree_plugin.h 	return READ_ONCE(rnp->gp_tasks) != NULL;
READ_ONCE         382 kernel/rcu/tree_plugin.h 		if (unlikely(READ_ONCE(t->rcu_read_unlock_special.s)))
READ_ONCE         553 kernel/rcu/tree_plugin.h 		READ_ONCE(t->rcu_read_unlock_special.s)) &&
READ_ONCE         615 kernel/rcu/tree_plugin.h 		      (rdp->grpmask & READ_ONCE(rnp->expmask)) ||
READ_ONCE         760 kernel/rcu/tree_plugin.h 		__func__, READ_ONCE(rnp->gp_tasks), rnp->boost_tasks,
READ_ONCE         970 kernel/rcu/tree_plugin.h 	if (READ_ONCE(rnp->exp_tasks) == NULL &&
READ_ONCE         971 kernel/rcu/tree_plugin.h 	    READ_ONCE(rnp->boost_tasks) == NULL)
READ_ONCE        1019 kernel/rcu/tree_plugin.h 	return READ_ONCE(rnp->exp_tasks) != NULL ||
READ_ONCE        1020 kernel/rcu/tree_plugin.h 	       READ_ONCE(rnp->boost_tasks) != NULL;
READ_ONCE        1319 kernel/rcu/tree_plugin.h 	     unlikely(READ_ONCE(rdp->gpwrap))) &&
READ_ONCE        1392 kernel/rcu/tree_plugin.h 	tne = READ_ONCE(tick_nohz_active);
READ_ONCE        1650 kernel/rcu/tree_plugin.h 	if (!READ_ONCE(rdp_gp->nocb_gp_kthread)) {
READ_ONCE        1659 kernel/rcu/tree_plugin.h 	if (force || READ_ONCE(rdp_gp->nocb_gp_sleep)) {
READ_ONCE        1815 kernel/rcu/tree_plugin.h 	if ((ncbs && j != READ_ONCE(rdp->nocb_bypass_first)) ||
READ_ONCE        1882 kernel/rcu/tree_plugin.h 	t = READ_ONCE(rdp->nocb_gp_kthread);
READ_ONCE        1969 kernel/rcu/tree_plugin.h 		    (time_after(j, READ_ONCE(rdp->nocb_bypass_first) + 1) ||
READ_ONCE        2045 kernel/rcu/tree_plugin.h 				!READ_ONCE(my_rdp->nocb_gp_sleep));
READ_ONCE        2053 kernel/rcu/tree_plugin.h 			!READ_ONCE(my_rdp->nocb_gp_sleep));
READ_ONCE        2125 kernel/rcu/tree_plugin.h 				 !READ_ONCE(rdp->nocb_cb_sleep));
READ_ONCE        2154 kernel/rcu/tree_plugin.h 	return READ_ONCE(rdp->nocb_defer_wakeup);
READ_ONCE        2168 kernel/rcu/tree_plugin.h 	ndw = READ_ONCE(rdp->nocb_defer_wakeup);
READ_ONCE        2409 kernel/rcu/tree_plugin.h 		rnp->grplo, rnp->grphi, READ_ONCE(rdp->nocb_gp_loops));
READ_ONCE        2557 kernel/rcu/tree_plugin.h 	     ULONG_CMP_LT(jiffies, READ_ONCE(rcu_state.gp_start) + HZ)))
READ_ONCE          26 kernel/rcu/tree_stall.h 	int till_stall_check = READ_ONCE(rcu_cpu_stall_timeout);
READ_ONCE         110 kernel/rcu/tree_stall.h 	rcu_state.n_force_qs_gpstart = READ_ONCE(rcu_state.n_force_qs);
READ_ONCE         131 kernel/rcu/tree_stall.h 	j = READ_ONCE(rcu_state.jiffies_kick_kthreads);
READ_ONCE         133 kernel/rcu/tree_stall.h 	    (rcu_gp_in_progress() || READ_ONCE(rcu_state.gp_flags))) {
READ_ONCE         328 kernel/rcu/tree_stall.h 	       READ_ONCE(rcu_state.n_force_qs) - rcu_state.n_force_qs_gpstart,
READ_ONCE         338 kernel/rcu/tree_stall.h 	j = jiffies - READ_ONCE(rcu_state.gp_activity);
READ_ONCE         343 kernel/rcu/tree_stall.h 		       READ_ONCE(rcu_state.gp_flags),
READ_ONCE         404 kernel/rcu/tree_stall.h 			gpa = READ_ONCE(rcu_state.gp_activity);
READ_ONCE         407 kernel/rcu/tree_stall.h 			       READ_ONCE(jiffies_till_next_fqs),
READ_ONCE         414 kernel/rcu/tree_stall.h 	if (ULONG_CMP_GE(jiffies, READ_ONCE(rcu_state.jiffies_stall)))
READ_ONCE         459 kernel/rcu/tree_stall.h 	if (ULONG_CMP_GE(jiffies, READ_ONCE(rcu_state.jiffies_stall)))
READ_ONCE         511 kernel/rcu/tree_stall.h 	gs1 = READ_ONCE(rcu_state.gp_seq);
READ_ONCE         513 kernel/rcu/tree_stall.h 	js = READ_ONCE(rcu_state.jiffies_stall);
READ_ONCE         515 kernel/rcu/tree_stall.h 	gps = READ_ONCE(rcu_state.gp_start);
READ_ONCE         517 kernel/rcu/tree_stall.h 	gs2 = READ_ONCE(rcu_state.gp_seq);
READ_ONCE         525 kernel/rcu/tree_stall.h 	    (READ_ONCE(rnp->qsmask) & rdp->grpmask) &&
READ_ONCE         563 kernel/rcu/tree_stall.h 	ja = j - READ_ONCE(rcu_state.gp_activity);
READ_ONCE         564 kernel/rcu/tree_stall.h 	jr = j - READ_ONCE(rcu_state.gp_req_activity);
READ_ONCE         565 kernel/rcu/tree_stall.h 	jw = j - READ_ONCE(rcu_state.gp_wake_time);
READ_ONCE         570 kernel/rcu/tree_stall.h 		ja, jr, jw, (long)READ_ONCE(rcu_state.gp_wake_seq),
READ_ONCE         571 kernel/rcu/tree_stall.h 		(long)READ_ONCE(rcu_state.gp_seq),
READ_ONCE         572 kernel/rcu/tree_stall.h 		(long)READ_ONCE(rcu_get_root()->gp_seq_needed),
READ_ONCE         573 kernel/rcu/tree_stall.h 		READ_ONCE(rcu_state.gp_flags));
READ_ONCE         617 kernel/rcu/tree_stall.h 	if (time_before(j, READ_ONCE(rcu_state.gp_req_activity) + gpssdelay) ||
READ_ONCE         618 kernel/rcu/tree_stall.h 	    time_before(j, READ_ONCE(rcu_state.gp_activity) + gpssdelay) ||
READ_ONCE         626 kernel/rcu/tree_stall.h 	    time_before(j, READ_ONCE(rcu_state.gp_req_activity) + gpssdelay) ||
READ_ONCE         627 kernel/rcu/tree_stall.h 	    time_before(j, READ_ONCE(rcu_state.gp_activity) + gpssdelay) ||
READ_ONCE         139 kernel/rcu/update.c 	return READ_ONCE(rcu_normal) &&
READ_ONCE         535 kernel/rcu/update.c 	if (needwake && READ_ONCE(rcu_tasks_kthread_ptr))
READ_ONCE         603 kernel/rcu/update.c 	if (!READ_ONCE(t->rcu_tasks_holdout) ||
READ_ONCE         604 kernel/rcu/update.c 	    t->rcu_tasks_nvcsw != READ_ONCE(t->nvcsw) ||
READ_ONCE         605 kernel/rcu/update.c 	    !READ_ONCE(t->on_rq) ||
READ_ONCE         694 kernel/rcu/update.c 			if (t != current && READ_ONCE(t->on_rq) &&
READ_ONCE         697 kernel/rcu/update.c 				t->rcu_tasks_nvcsw = READ_ONCE(t->nvcsw);
READ_ONCE         739 kernel/rcu/update.c 			rtst = READ_ONCE(rcu_task_stall_timeout);
READ_ONCE          30 kernel/sched/autogroup.h 	int enabled = READ_ONCE(sysctl_sched_autogroup_enabled);
READ_ONCE         291 kernel/sched/completion.c 	if (!READ_ONCE(x->done))
READ_ONCE         317 kernel/sched/completion.c 	if (!READ_ONCE(x->done))
READ_ONCE         383 kernel/sched/core.c 	typeof(ti->flags) old, val = READ_ONCE(ti->flags);
READ_ONCE         969 kernel/sched/core.c 	if (uc_se->value > READ_ONCE(uc_rq->value))
READ_ONCE        1008 kernel/sched/core.c 	rq_clamp = READ_ONCE(uc_rq->value);
READ_ONCE        1613 kernel/sched/deadline.c 	curr = READ_ONCE(rq->curr); /* unlocked access */
READ_ONCE        1095 kernel/sched/fair.c 		(lockdep_is_held(&task_rq(p)->lock) && !READ_ONCE(p->on_cpu)));
READ_ONCE        1130 kernel/sched/fair.c 	unsigned int scan_size = READ_ONCE(sysctl_numa_balancing_scan_size);
READ_ONCE        1601 kernel/sched/fair.c 	if (READ_ONCE(dst_rq->numa_migrate_on))
READ_ONCE        2138 kernel/sched/fair.c 	seq = READ_ONCE(p->mm->numa_scan_seq);
READ_ONCE        2280 kernel/sched/fair.c 	tsk = READ_ONCE(cpu_rq(cpu)->curr);
READ_ONCE        2476 kernel/sched/fair.c 	WRITE_ONCE(p->mm->numa_scan_seq, READ_ONCE(p->mm->numa_scan_seq) + 1);
READ_ONCE        2794 kernel/sched/fair.c 	typeof(*ptr) res, var = READ_ONCE(*ptr);                \
READ_ONCE        2814 kernel/sched/fair.c 	typeof(*ptr) res, var = READ_ONCE(*ptr);		\
READ_ONCE        2999 kernel/sched/fair.c 	tg_shares = READ_ONCE(tg->shares);
READ_ONCE        3091 kernel/sched/fair.c 	runnable = shares = READ_ONCE(gcfs_rq->tg->shares);
READ_ONCE        3695 kernel/sched/fair.c 	return READ_ONCE(p->se.avg.util_avg);
READ_ONCE        3700 kernel/sched/fair.c 	struct util_est ue = READ_ONCE(p->se.avg.util_est);
READ_ONCE        5192 kernel/sched/fair.c 	if (!READ_ONCE(rq->rd->overutilized) && cpu_overutilized(rq->cpu)) {
READ_ONCE        5411 kernel/sched/fair.c 	unsigned long nr_running = READ_ONCE(rq->cfs.h_nr_running);
READ_ONCE        5844 kernel/sched/fair.c 		return READ_ONCE(sds->has_idle_cores);
READ_ONCE        6107 kernel/sched/fair.c 	util = READ_ONCE(cfs_rq->avg.util_avg);
READ_ONCE        6110 kernel/sched/fair.c 		util = max(util, READ_ONCE(cfs_rq->avg.util_est.enqueued));
READ_ONCE        6134 kernel/sched/fair.c 	if (cpu != task_cpu(p) || !READ_ONCE(p->se.avg.last_update_time))
READ_ONCE        6138 kernel/sched/fair.c 	util = READ_ONCE(cfs_rq->avg.util_avg);
READ_ONCE        6171 kernel/sched/fair.c 			READ_ONCE(cfs_rq->avg.util_est.enqueued);
READ_ONCE        6238 kernel/sched/fair.c 	unsigned long util_est, util = READ_ONCE(cfs_rq->avg.util_avg);
READ_ONCE        6252 kernel/sched/fair.c 		util_est = READ_ONCE(cfs_rq->avg.util_est.enqueued);
READ_ONCE        6371 kernel/sched/fair.c 	if (!pd || READ_ONCE(rd->overutilized))
READ_ONCE        7519 kernel/sched/fair.c 	if (READ_ONCE(rq->avg_rt.util_avg))
READ_ONCE        7522 kernel/sched/fair.c 	if (READ_ONCE(rq->avg_dl.util_avg))
READ_ONCE        7526 kernel/sched/fair.c 	if (READ_ONCE(rq->avg_irq.util_avg))
READ_ONCE        7655 kernel/sched/fair.c 	while ((se = READ_ONCE(cfs_rq->h_load_next)) != NULL) {
READ_ONCE        7783 kernel/sched/fair.c 	used = READ_ONCE(rq->avg_rt.util_avg);
READ_ONCE        7784 kernel/sched/fair.c 	used += READ_ONCE(rq->avg_dl.util_avg);
READ_ONCE        8240 kernel/sched/fair.c 	if (env->idle == CPU_NEWLY_IDLE && READ_ONCE(nohz.has_blocked))
READ_ONCE        8548 kernel/sched/fair.c 		if (rcu_dereference(rd->pd) && !READ_ONCE(rd->overutilized))
READ_ONCE        9424 kernel/sched/fair.c 	if (READ_ONCE(nohz.has_blocked) &&
READ_ONCE        9425 kernel/sched/fair.c 	    time_after(now, READ_ONCE(nohz.next_blocked)))
READ_ONCE        9768 kernel/sched/fair.c 	if (!READ_ONCE(nohz.has_blocked) ||
READ_ONCE        9769 kernel/sched/fair.c 	    time_before(jiffies, READ_ONCE(nohz.next_blocked)))
READ_ONCE        9829 kernel/sched/fair.c 	    !READ_ONCE(this_rq->rd->overload)) {
READ_ONCE         339 kernel/sched/idle.c 	while (!READ_ONCE(it.done))
READ_ONCE         223 kernel/sched/loadavg.c 	if (!time_before(jiffies, READ_ONCE(calc_load_update)))
READ_ONCE         271 kernel/sched/loadavg.c 	this_rq->calc_load_update = READ_ONCE(calc_load_update);
READ_ONCE         309 kernel/sched/loadavg.c 	sample_window = READ_ONCE(calc_load_update);
READ_ONCE         355 kernel/sched/loadavg.c 	sample_window = READ_ONCE(calc_load_update);
READ_ONCE          99 kernel/sched/membarrier.c 		if (!(READ_ONCE(cpu_rq(cpu)->membarrier_state) &
READ_ONCE        1402 kernel/sched/rt.c 	curr = READ_ONCE(rq->curr); /* unlocked access */
READ_ONCE        1061 kernel/sched/sched.h 	return READ_ONCE(rq->clock);
READ_ONCE        1644 kernel/sched/sched.h 	return READ_ONCE(p->on_rq) == TASK_ON_RQ_MIGRATING;
READ_ONCE        1936 kernel/sched/sched.h 		if (!READ_ONCE(rq->rd->overload))
READ_ONCE        2324 kernel/sched/sched.h 	unsigned int min_util = READ_ONCE(rq->uclamp[UCLAMP_MIN].value);
READ_ONCE        2325 kernel/sched/sched.h 	unsigned int max_util = READ_ONCE(rq->uclamp[UCLAMP_MAX].value);
READ_ONCE        2402 kernel/sched/sched.h 	return READ_ONCE(rq->avg_dl.util_avg);
READ_ONCE        2407 kernel/sched/sched.h 	unsigned long util = READ_ONCE(rq->cfs.avg.util_avg);
READ_ONCE        2411 kernel/sched/sched.h 			     READ_ONCE(rq->cfs.avg.util_est.enqueued));
READ_ONCE        2419 kernel/sched/sched.h 	return READ_ONCE(rq->avg_rt.util_avg);
READ_ONCE        2493 kernel/sched/sched.h 	if (READ_ONCE(rq->membarrier_state) == membarrier_state)
READ_ONCE        2063 kernel/sched/topology.c 		if (rq->cpu_capacity_orig > READ_ONCE(d.rd->max_cpu_capacity))
READ_ONCE         218 kernel/sched/wait_bit.c 	unsigned long now = READ_ONCE(jiffies);
READ_ONCE         232 kernel/sched/wait_bit.c 	unsigned long now = READ_ONCE(jiffies);
READ_ONCE         260 kernel/seccomp.c 			READ_ONCE(current->seccomp.filter);
READ_ONCE        1934 kernel/signal.c 		if (tsk->parent_exec_id != READ_ONCE(tsk->parent->self_exec_id))
READ_ONCE         214 kernel/stop_machine.c 		newstate = READ_ONCE(msdata->state);
READ_ONCE          33 kernel/task_work.c 		head = READ_ONCE(task->task_works);
READ_ONCE          71 kernel/task_work.c 	while ((work = READ_ONCE(*pprev))) {
READ_ONCE         102 kernel/task_work.c 			work = READ_ONCE(task->task_works);
READ_ONCE         167 kernel/time/hrtimer.c 		base = READ_ONCE(timer->base);
READ_ONCE        1231 kernel/time/hrtimer.c 	struct hrtimer_clock_base *base = READ_ONCE(timer->base);
READ_ONCE        1445 kernel/time/hrtimer.c 		base = READ_ONCE(timer->base);
READ_ONCE        1453 kernel/time/hrtimer.c 		 base != READ_ONCE(timer->base));
READ_ONCE         305 kernel/time/posix-cpu-timers.c 	if (!READ_ONCE(pct->timers_active)) {
READ_ONCE         350 kernel/time/posix-cpu-timers.c 	if (!READ_ONCE(pct->timers_active)) {
READ_ONCE         916 kernel/time/posix-cpu-timers.c 	if (!READ_ONCE(pct->timers_active) || pct->expiry_active)
READ_ONCE        1092 kernel/time/posix-cpu-timers.c 	if (READ_ONCE(pct->timers_active) && !READ_ONCE(pct->expiry_active)) {
READ_ONCE         822 kernel/time/posix-timers.c 	const struct k_clock *kc = READ_ONCE(timer->kclock);
READ_ONCE         823 kernel/time/posix-timers.c 	timer_t timer_id = READ_ONCE(timer->it_id);
READ_ONCE          63 kernel/time/tick-sched.c 	delta = ktime_sub(now, READ_ONCE(last_jiffies_update));
READ_ONCE         171 kernel/time/timekeeping.c 	struct clocksource *clock = READ_ONCE(tkr->clock);
READ_ONCE         891 kernel/time/timer.c 	jnow = READ_ONCE(jiffies);
READ_ONCE         932 kernel/time/timer.c 		tf = READ_ONCE(timer->flags);
READ_ONCE        1279 kernel/time/timer.c 	tf = READ_ONCE(timer->flags);
READ_ONCE        1681 kernel/time/timer.c 	unsigned long now = READ_ONCE(jiffies);
READ_ONCE         466 kernel/torture.c 	while (READ_ONCE(fullstop) == FULLSTOP_SHUTDOWN) {
READ_ONCE         535 kernel/torture.c 	if (READ_ONCE(fullstop) == FULLSTOP_DONTSTOP) {
READ_ONCE         583 kernel/torture.c 	spt = READ_ONCE(stutter_pause_test);
READ_ONCE         584 kernel/torture.c 	for (; spt; spt = READ_ONCE(stutter_pause_test)) {
READ_ONCE         589 kernel/torture.c 			while (READ_ONCE(stutter_pause_test))
READ_ONCE         707 kernel/torture.c 	if (READ_ONCE(fullstop) == FULLSTOP_SHUTDOWN) {
READ_ONCE         746 kernel/torture.c 	return READ_ONCE(fullstop) != FULLSTOP_DONTSTOP;
READ_ONCE         362 kernel/trace/bpf_trace.c 	ee = READ_ONCE(array->ptrs[index]);
READ_ONCE         434 kernel/trace/bpf_trace.c 	ee = READ_ONCE(array->ptrs[index]);
READ_ONCE         571 kernel/trace/bpf_trace.c 	cgrp = READ_ONCE(array->ptrs[idx]);
READ_ONCE        1093 kernel/trace/ring_buffer.c 	if (tail_page == READ_ONCE(cpu_buffer->tail_page)) {
READ_ONCE        2099 kernel/trace/ring_buffer.c 		buffer_tail_page = READ_ONCE(cpu_buffer->tail_page);
READ_ONCE        2428 kernel/trace/ring_buffer.c 	bpage = READ_ONCE(cpu_buffer->tail_page);
READ_ONCE        2476 kernel/trace/ring_buffer.c 	while (cpu_buffer->commit_page != READ_ONCE(cpu_buffer->tail_page)) {
READ_ONCE        2511 kernel/trace/ring_buffer.c 	if (unlikely(cpu_buffer->commit_page != READ_ONCE(cpu_buffer->tail_page)))
READ_ONCE        2832 kernel/trace/ring_buffer.c 	tail_page = info->tail_page = READ_ONCE(cpu_buffer->tail_page);
READ_ONCE        2890 kernel/trace/ring_buffer.c 	if (unlikely(READ_ONCE(cpu_buffer->buffer) != buffer)) {
READ_ONCE         192 kernel/trace/ring_buffer_benchmark.c 	while (!READ_ONCE(reader_finish)) {
READ_ONCE         524 kernel/trace/trace.c 	pid_list->pid_max = READ_ONCE(pid_max);
READ_ONCE        1597 kernel/trace/trace.h 	return READ_ONCE(file_inode(filp)->i_private);
READ_ONCE         159 kernel/trace/trace_stack.c 	int frame_size = READ_ONCE(tracer_frame);
READ_ONCE         531 kernel/trace/tracing_map.c 			val = READ_ONCE(entry->val);
READ_ONCE         198 kernel/ucount.c 		max = READ_ONCE(tns->ucount_max[type]);
READ_ONCE        1125 kernel/user_namespace.c 	unsigned long userns_flags = READ_ONCE(ns->flags);
READ_ONCE         697 kernel/watchdog.c 		old = READ_ONCE(*param);
READ_ONCE         699 kernel/watchdog.c 		if (!err && old != READ_ONCE(*param))
READ_ONCE         748 kernel/watchdog.c 	old = READ_ONCE(watchdog_thresh);
READ_ONCE         751 kernel/watchdog.c 	if (!err && write && old != READ_ONCE(watchdog_thresh))
READ_ONCE        5729 kernel/workqueue.c 	unsigned long thresh = READ_ONCE(wq_watchdog_thresh) * HZ;
READ_ONCE        5746 kernel/workqueue.c 		pool_ts = READ_ONCE(pool->watchdog_ts);
READ_ONCE        5747 kernel/workqueue.c 		touched = READ_ONCE(wq_watchdog_touched);
READ_ONCE        5756 kernel/workqueue.c 				READ_ONCE(per_cpu(wq_watchdog_touched_cpu,
READ_ONCE          37 lib/assoc_array.c 		cursor = READ_ONCE(shortcut->next_node); /* Address dependency. */
READ_ONCE          52 lib/assoc_array.c 		ptr = READ_ONCE(node->slots[slot]); /* Address dependency. */
READ_ONCE          81 lib/assoc_array.c 		ptr = READ_ONCE(node->slots[slot]); /* Address dependency. */
READ_ONCE          90 lib/assoc_array.c 	parent = READ_ONCE(node->back_pointer); /* Address dependency. */
READ_ONCE          98 lib/assoc_array.c 		parent = READ_ONCE(shortcut->back_pointer); /* Address dependency. */
READ_ONCE         138 lib/assoc_array.c 	struct assoc_array_ptr *root = READ_ONCE(array->root); /* Address dependency. */
READ_ONCE         185 lib/assoc_array.c 	cursor = READ_ONCE(array->root);  /* Address dependency. */
READ_ONCE         209 lib/assoc_array.c 	ptr = READ_ONCE(node->slots[slot]); /* Address dependency. */
READ_ONCE         282 lib/assoc_array.c 	cursor = READ_ONCE(shortcut->next_node); /* Address dependency. */
READ_ONCE         324 lib/assoc_array.c 		ptr = READ_ONCE(node->slots[slot]); /* Address dependency. */
READ_ONCE         135 lib/debugobjects.c 	if (likely(READ_ONCE(obj_pool_free) >= debug_objects_pool_min_level))
READ_ONCE         146 lib/debugobjects.c 	while (READ_ONCE(obj_nr_tofree) && (READ_ONCE(obj_pool_free) < obj_pool_min_free)) {
READ_ONCE         165 lib/debugobjects.c 	while (READ_ONCE(obj_pool_free) < debug_objects_pool_min_level) {
READ_ONCE         430 lib/debugobjects.c 	if (!READ_ONCE(obj_freeing) && READ_ONCE(obj_nr_tofree)) {
READ_ONCE         989 lib/debugobjects.c 	if (!READ_ONCE(obj_freeing) && READ_ONCE(obj_nr_tofree)) {
READ_ONCE        1015 lib/debugobjects.c 	seq_printf(m, "pool_free     :%d\n", READ_ONCE(obj_pool_free) + obj_percpu_free);
READ_ONCE        1020 lib/debugobjects.c 	seq_printf(m, "on_free_list  :%d\n", READ_ONCE(obj_nr_tofree));
READ_ONCE          24 lib/dynamic_queue_limits.c 	num_queued = READ_ONCE(dql->num_queued);
READ_ONCE          71 lib/errseq.c   	old = READ_ONCE(*eseq);
READ_ONCE         124 lib/errseq.c   	errseq_t old = READ_ONCE(*eseq);
READ_ONCE         146 lib/errseq.c   	errseq_t cur = READ_ONCE(*eseq);
READ_ONCE         184 lib/errseq.c   	old = READ_ONCE(*eseq);
READ_ONCE         106 lib/fault-inject.c 		unsigned int fail_nth = READ_ONCE(current->fail_nth);
READ_ONCE          56 lib/generic-radix-tree.c 	struct genradix_root *r = READ_ONCE(radix->root);
READ_ONCE         107 lib/generic-radix-tree.c 	struct genradix_root *v = READ_ONCE(radix->root);
READ_ONCE         142 lib/generic-radix-tree.c 		n = READ_ONCE(*p);
READ_ONCE         170 lib/generic-radix-tree.c 	r = READ_ONCE(radix->root);
READ_ONCE          32 lib/llist.c    		new_last->next = first = READ_ONCE(head->first);
READ_ONCE          62 lib/llist.c    		next = READ_ONCE(entry->next);
READ_ONCE          15 lib/lockref.c  	old.lock_count = READ_ONCE(lockref->lock_count);			\
READ_ONCE         311 lib/sbitmap.c  		unsigned long word = READ_ONCE(sb->map[i].word);
READ_ONCE         312 lib/sbitmap.c  		unsigned int word_bits = READ_ONCE(sb->map[i].depth);
READ_ONCE         444 lib/sbitmap.c  	depth = READ_ONCE(sbq->sb.depth);
READ_ONCE         475 lib/sbitmap.c  	depth = READ_ONCE(sbq->sb.depth);
READ_ONCE         542 lib/sbitmap.c  		wake_batch = READ_ONCE(sbq->wake_batch);
READ_ONCE         170 lib/vdso/gettimeofday.c 	time_t t = READ_ONCE(vd[CS_HRES_COARSE].basetime[CLOCK_REALTIME].sec);
READ_ONCE         192 lib/vdso/gettimeofday.c 	hrtimer_res = READ_ONCE(vd[CS_HRES_COARSE].hrtimer_res);
READ_ONCE         854 lib/vsprintf.c 		p = READ_ONCE(d->d_parent);
READ_ONCE         855 lib/vsprintf.c 		array[i] = READ_ONCE(d->d_name.name);
READ_ONCE         735 mm/filemap.c   	errseq_t old = READ_ONCE(file->f_wb_err);
READ_ONCE         332 mm/gup.c       	pmdval = READ_ONCE(*pmd);
READ_ONCE         357 mm/gup.c       		pmdval = READ_ONCE(*pmd);
READ_ONCE        1796 mm/gup.c       	return READ_ONCE(*ptep);
READ_ONCE        2003 mm/gup.c       	pte = READ_ONCE(*ptep);
READ_ONCE        2195 mm/gup.c       		pmd_t pmd = READ_ONCE(*pmdp);
READ_ONCE        2238 mm/gup.c       		pud_t pud = READ_ONCE(*pudp);
READ_ONCE        2266 mm/gup.c       		p4d_t p4d = READ_ONCE(*p4dp);
READ_ONCE        2291 mm/gup.c       		pgd_t pgd = READ_ONCE(*pgdp);
READ_ONCE         565 mm/hmm.c       	pmd = READ_ONCE(*pmdp);
READ_ONCE         667 mm/hmm.c       	pud = READ_ONCE(*pudp);
READ_ONCE          85 mm/huge_memory.c 		return READ_ONCE(huge_zero_page);
READ_ONCE         104 mm/huge_memory.c 	return READ_ONCE(huge_zero_page);
READ_ONCE         119 mm/huge_memory.c 		return READ_ONCE(huge_zero_page);
READ_ONCE         127 mm/huge_memory.c 	return READ_ONCE(huge_zero_page);
READ_ONCE        2253 mm/huge_memory.c 			entry = mk_pte(page + i, READ_ONCE(vma->vm_page_prot));
READ_ONCE        2898 mm/huge_memory.c 	return READ_ONCE(ds_queue->split_queue_len);
READ_ONCE        5030 mm/hugetlb.c   	pud_entry = READ_ONCE(*pud);
READ_ONCE        5038 mm/hugetlb.c   	pmd_entry = READ_ONCE(*pmd);
READ_ONCE         253 mm/internal.h  #define page_order_unsafe(page)		READ_ONCE(page_private(page))
READ_ONCE         458 mm/kasan/common.c 	shadow_byte = READ_ONCE(*(s8 *)kasan_mem_to_shadow(object));
READ_ONCE         198 mm/kasan/quarantine.c 				READ_ONCE(quarantine_batch_size)) {
READ_ONCE         220 mm/kasan/quarantine.c 	if (likely(READ_ONCE(quarantine_size) <=
READ_ONCE         221 mm/kasan/quarantine.c 		   READ_ONCE(quarantine_max_size)))
READ_ONCE         704 mm/ksm.c       	kpfn = READ_ONCE(stable_node->kpfn); /* Address dependency. */
READ_ONCE         706 mm/ksm.c       	if (READ_ONCE(page->mapping) != expected_mapping)
READ_ONCE         733 mm/ksm.c       	if (READ_ONCE(page->mapping) != expected_mapping) {
READ_ONCE         747 mm/ksm.c       		if (READ_ONCE(page->mapping) != expected_mapping) {
READ_ONCE         763 mm/ksm.c       	if (READ_ONCE(stable_node->kpfn) != kpfn)
READ_ONCE        2049 mm/ksm.c       		    get_kpfn_nid(READ_ONCE(stable_node->kpfn)) !=
READ_ONCE        2424 mm/ksm.c       			sleep_ms = READ_ONCE(ksm_thread_sleep_millisecs);
READ_ONCE        2426 mm/ksm.c       				sleep_ms != READ_ONCE(ksm_thread_sleep_millisecs),
READ_ONCE        3052 mm/ksm.c       	if (READ_ONCE(ksm_max_page_sharing) == knob)
READ_ONCE         492 mm/memcontrol.c 		memcg = READ_ONCE(page->mem_cgroup);
READ_ONCE         589 mm/memcontrol.c 	unsigned long soft_limit = READ_ONCE(memcg->soft_limit);
READ_ONCE        1074 mm/memcontrol.c 			pos = READ_ONCE(iter->position);
READ_ONCE        1332 mm/memcontrol.c 	limit = READ_ONCE(memcg->memory.max);
READ_ONCE        1338 mm/memcontrol.c 		limit = READ_ONCE(memcg->memsw.max);
READ_ONCE        2442 mm/memcontrol.c 		high = READ_ONCE(memcg->high);
READ_ONCE        2953 mm/memcontrol.c 	kmemcg_id = READ_ONCE(memcg->kmemcg_id);
READ_ONCE        2964 mm/memcontrol.c 	memcg_cachep = READ_ONCE(arr->entries[kmemcg_id]);
READ_ONCE        5834 mm/memcontrol.c 	move_flags = READ_ONCE(memcg->move_charge_at_immigrate);
READ_ONCE        6085 mm/memcontrol.c 		READ_ONCE(mem_cgroup_from_seq(m)->memory.min));
READ_ONCE        6108 mm/memcontrol.c 		READ_ONCE(mem_cgroup_from_seq(m)->memory.low));
READ_ONCE        6130 mm/memcontrol.c 	return seq_puts_memcg_tunable(m, READ_ONCE(mem_cgroup_from_seq(m)->high));
READ_ONCE        6160 mm/memcontrol.c 		READ_ONCE(mem_cgroup_from_seq(m)->memory.max));
READ_ONCE        6455 mm/memcontrol.c 	parent_emin = READ_ONCE(parent->memory.emin);
READ_ONCE        6469 mm/memcontrol.c 	parent_elow = READ_ONCE(parent->memory.elow);
READ_ONCE        7178 mm/memcontrol.c 				      READ_ONCE(memcg->swap.max) -
READ_ONCE        7233 mm/memcontrol.c 		READ_ONCE(mem_cgroup_from_seq(m)->swap.max));
READ_ONCE        3442 mm/memory.c    	nr_pages = READ_ONCE(fault_around_bytes) >> PAGE_SHIFT;
READ_ONCE         235 mm/migrate.c   		pte = pte_mkold(mk_pte(new, READ_ONCE(vma->vm_page_prot)));
READ_ONCE        1275 mm/mmap.c      		struct anon_vma *anon_vma = READ_ONCE(old->anon_vma);
READ_ONCE         495 mm/oom_kill.c  		struct mm_struct *t_mm = READ_ONCE(t->mm);
READ_ONCE         541 mm/page_alloc.c 	word = READ_ONCE(bitmap[word_bitidx]);
READ_ONCE        2802 mm/page_alloc.c 	batch = READ_ONCE(pcp->batch);
READ_ONCE        3056 mm/page_alloc.c 		unsigned long batch = READ_ONCE(pcp->batch);
READ_ONCE         411 mm/page_io.c   		if (!READ_ONCE(bio->bi_private))
READ_ONCE         434 mm/page_owner.c 	handle = READ_ONCE(page_owner->handle);
READ_ONCE         442 mm/page_owner.c 	handle = READ_ONCE(page_owner->free_handle);
READ_ONCE         532 mm/page_owner.c 		handle = READ_ONCE(page_owner->handle);
READ_ONCE         182 mm/page_vma_mapped.c 	pmde = READ_ONCE(*pvmw->pmd);
READ_ONCE         471 mm/rmap.c      	anon_mapping = (unsigned long)READ_ONCE(page->mapping);
READ_ONCE         515 mm/rmap.c      	anon_mapping = (unsigned long)READ_ONCE(page->mapping);
READ_ONCE         522 mm/rmap.c      	root_anon_vma = READ_ONCE(anon_vma->root);
READ_ONCE         571 mm/shmem.c     	if (!READ_ONCE(sbinfo->shrinklist_len))
READ_ONCE         581 mm/shmem.c     	return READ_ONCE(sbinfo->shrinklist_len);
READ_ONCE         742 mm/shmem.c     	swapped = READ_ONCE(info->swapped);
READ_ONCE        2930 mm/slab.c      	shared = READ_ONCE(n->shared);
READ_ONCE        2935 mm/slab.c      	shared = READ_ONCE(n->shared);
READ_ONCE         336 mm/slab.h      	s = READ_ONCE(page->slab_cache);
READ_ONCE         338 mm/slab.h      		return READ_ONCE(s->memcg_params.memcg);
READ_ONCE         356 mm/slab.h      	memcg = READ_ONCE(s->memcg_params.memcg);
READ_ONCE         394 mm/slab.h      	memcg = READ_ONCE(s->memcg_params.memcg);
READ_ONCE        2708 mm/slub.c      		 unlikely(tid != READ_ONCE(c->tid)));
READ_ONCE        2982 mm/slub.c      		 unlikely(tid != READ_ONCE(c->tid)));
READ_ONCE        2988 mm/slub.c      		void **freelist = READ_ONCE(c->freelist);
READ_ONCE        4827 mm/slub.c      			page = READ_ONCE(c->page);
READ_ONCE         301 mm/swap_state.c 	return READ_ONCE(enable_vma_readahead) && !atomic_read(&nr_rotate_swap);
READ_ONCE         507 mm/swap_state.c 	max_pages = 1 << READ_ONCE(page_cluster);
READ_ONCE         654 mm/swap_state.c 	max_win = 1 << min_t(unsigned int, READ_ONCE(page_cluster),
READ_ONCE         104 mm/swapfile.c  	if (type >= READ_ONCE(nr_swapfiles))
READ_ONCE         108 mm/swapfile.c  	return READ_ONCE(swap_info[type]);
READ_ONCE        2105 mm/swapfile.c  		count = READ_ONCE(si->swap_map[i]);
READ_ONCE         470 mm/userfaultfd.c 	if (mmap_changing && READ_ONCE(*mmap_changing))
READ_ONCE        3479 mm/vmalloc.c   	head = READ_ONCE(vmap_purge_list.first);
READ_ONCE         292 mm/vmscan.c    	return READ_ONCE(mn->congested);
READ_ONCE         669 mm/z3fold.c    		zhdr = list_first_entry_or_null(READ_ONCE(l),
READ_ONCE         678 mm/z3fold.c    		if (unlikely(zhdr != list_first_entry(READ_ONCE(l),
READ_ONCE         724 mm/z3fold.c    			zhdr = list_first_entry_or_null(READ_ONCE(l),
READ_ONCE         101 net/core/datagram.c 	if (READ_ONCE(sk->sk_receive_queue.prev) != skb)
READ_ONCE         282 net/core/datagram.c 	} while (READ_ONCE(sk->sk_receive_queue.prev) != *last);
READ_ONCE        4068 net/core/dev.c 		cpu = READ_ONCE(rflow->cpu);
READ_ONCE        5932 net/core/dev.c 		val = READ_ONCE(n->state);
READ_ONCE        6003 net/core/dev.c 		val = READ_ONCE(n->state);
READ_ONCE        6105 net/core/dev.c 			unsigned long val = READ_ONCE(napi->state);
READ_ONCE        3604 net/core/filter.c 		if (unlikely(READ_ONCE(ri->map) == map))
READ_ONCE        3640 net/core/filter.c 	struct bpf_map *map = READ_ONCE(ri->map);
READ_ONCE        3695 net/core/filter.c 	struct bpf_map *map = READ_ONCE(ri->map);
READ_ONCE        4068 net/core/filter.c 	cgrp = READ_ONCE(array->ptrs[idx]);
READ_ONCE        1197 net/core/neighbour.c 		if (READ_ONCE(hh->hh_len)) {
READ_ONCE        1476 net/core/neighbour.c 		if (dev->header_ops->cache && !READ_ONCE(neigh->hh.hh_len))
READ_ONCE         198 net/core/page_pool.c 	u32 hold_cnt = READ_ONCE(pool->pages_state_hold_cnt);
READ_ONCE        3301 net/core/pktgen.c 	unsigned int burst = READ_ONCE(pkt_dev->burst);
READ_ONCE        4418 net/core/skbuff.c 	    (unsigned int)READ_ONCE(sk->sk_rcvbuf))
READ_ONCE         633 net/core/skmsg.c 	prog = READ_ONCE(psock->progs.msg_parser);
READ_ONCE         748 net/core/skmsg.c 	prog = READ_ONCE(psock->progs.skb_verdict);
READ_ONCE         771 net/core/skmsg.c 	prog = READ_ONCE(psock->progs.skb_parser);
READ_ONCE         525 net/core/sock.c 	} else if (sk_add_backlog(sk, skb, READ_ONCE(sk->sk_rcvbuf))) {
READ_ONCE        1479 net/core/sock.c 		v.val = READ_ONCE(sk->sk_incoming_cpu);
READ_ONCE        1497 net/core/sock.c 		v.val = READ_ONCE(sk->sk_napi_id);
READ_ONCE        2097 net/core/sock.c 	    refcount_read(&sk->sk_wmem_alloc) < READ_ONCE(sk->sk_sndbuf)) {
READ_ONCE        2200 net/core/sock.c 		if (refcount_read(&sk->sk_wmem_alloc) < READ_ONCE(sk->sk_sndbuf))
READ_ONCE        2235 net/core/sock.c 		if (sk_wmem_alloc_get(sk) < READ_ONCE(sk->sk_sndbuf))
READ_ONCE        2344 net/core/sock.c 		if (memory_pressure && READ_ONCE(*memory_pressure))
READ_ONCE        2816 net/core/sock.c 	if ((refcount_read(&sk->sk_wmem_alloc) << 1) <= READ_ONCE(sk->sk_sndbuf)) {
READ_ONCE        3214 net/core/sock.c 	mem[SK_MEMINFO_RCVBUF] = READ_ONCE(sk->sk_rcvbuf);
READ_ONCE        3216 net/core/sock.c 	mem[SK_MEMINFO_SNDBUF] = READ_ONCE(sk->sk_sndbuf);
READ_ONCE        3218 net/core/sock.c 	mem[SK_MEMINFO_WMEM_QUEUED] = READ_ONCE(sk->sk_wmem_queued);
READ_ONCE        3220 net/core/sock.c 	mem[SK_MEMINFO_BACKLOG] = READ_ONCE(sk->sk_backlog.len);
READ_ONCE         150 net/core/sock_map.c 	skb_verdict = READ_ONCE(progs->skb_verdict);
READ_ONCE         151 net/core/sock_map.c 	skb_parser = READ_ONCE(progs->skb_parser);
READ_ONCE         164 net/core/sock_map.c 	msg_parser = READ_ONCE(progs->msg_parser);
READ_ONCE         180 net/core/sock_map.c 		if ((msg_parser && READ_ONCE(psock->progs.msg_parser)) ||
READ_ONCE         181 net/core/sock_map.c 		    (skb_progs  && READ_ONCE(psock->progs.skb_parser))) {
READ_ONCE         275 net/core/sock_map.c 	return READ_ONCE(stab->sks[key]);
READ_ONCE         118 net/core/sock_reuseport.c 	more_reuse->synq_overflow_ts = READ_ONCE(reuse->synq_overflow_ts);
READ_ONCE         283 net/core/sock_reuseport.c 	socks = READ_ONCE(reuse->num_socks);
READ_ONCE         255 net/ipv4/icmp.c 	if (!READ_ONCE(icmp_global.credit)) {
READ_ONCE         256 net/ipv4/icmp.c 		delta = min_t(u32, now - READ_ONCE(icmp_global.stamp), HZ);
READ_ONCE         475 net/ipv4/igmp.c 	mtu = READ_ONCE(dev->mtu);
READ_ONCE         754 net/ipv4/inet_connection_sock.c 	defer_accept = READ_ONCE(queue->rskq_defer_accept);
READ_ONCE         210 net/ipv4/inet_diag.c 			.idiag_wmem = READ_ONCE(sk->sk_wmem_queued),
READ_ONCE         273 net/ipv4/inet_diag.c 		ca_ops = READ_ONCE(icsk->icsk_ca_ops);
READ_ONCE         293 net/ipv4/inet_diag.c 		ca_ops = READ_ONCE(icsk->icsk_ca_ops);
READ_ONCE         243 net/ipv4/inet_hashtables.c 		if (READ_ONCE(sk->sk_incoming_cpu) == raw_smp_processor_id())
READ_ONCE         167 net/ipv4/inetpeer.c 		delta = (__u32)jiffies - READ_ONCE(p->dtime);
READ_ONCE         323 net/ipv4/ip_input.c 		if (ipprot && (edemux = READ_ONCE(ipprot->early_demux))) {
READ_ONCE        1262 net/ipv4/ip_output.c 			 dst_mtu(&rt->dst) : READ_ONCE(rt->dst.dev->mtu);
READ_ONCE         255 net/ipv4/ipmr_base.c 	lastuse = READ_ONCE(c->mfc_un.res.lastuse);
READ_ONCE         206 net/ipv4/netfilter/arp_tables.c 	private = READ_ONCE(table->private); /* Address dependency. */
READ_ONCE         261 net/ipv4/netfilter/ip_tables.c 	private = READ_ONCE(table->private); /* Address dependency. */
READ_ONCE         523 net/ipv4/raw.c 	hdrincl = READ_ONCE(hdrincl);
READ_ONCE         491 net/ipv4/route.c 	u32 old = READ_ONCE(*p_tstamp);
READ_ONCE        1313 net/ipv4/route.c 	mtu = READ_ONCE(dst->dev->mtu);
READ_ONCE        1411 net/ipv4/route.c 		mtu = min(READ_ONCE(dev->mtu), IP_MAX_MTU);
READ_ONCE         329 net/ipv4/tcp.c 	if (READ_ONCE(tcp_memory_pressure))
READ_ONCE         344 net/ipv4/tcp.c 	if (!READ_ONCE(tcp_memory_pressure))
READ_ONCE         392 net/ipv4/tcp.c 	u32 rate = READ_ONCE(tp->rate_delivered);
READ_ONCE         393 net/ipv4/tcp.c 	u32 intv = READ_ONCE(tp->rate_interval_us);
READ_ONCE         480 net/ipv4/tcp.c 	int avail = READ_ONCE(tp->rcv_nxt) - READ_ONCE(tp->copied_seq);
READ_ONCE         557 net/ipv4/tcp.c 		if (READ_ONCE(tp->urg_seq) == READ_ONCE(tp->copied_seq) &&
READ_ONCE         619 net/ipv4/tcp.c 		       READ_ONCE(tp->urg_seq) == READ_ONCE(tp->copied_seq);
READ_ONCE         628 net/ipv4/tcp.c 			answ = READ_ONCE(tp->write_seq) - tp->snd_una;
READ_ONCE         637 net/ipv4/tcp.c 			answ = READ_ONCE(tp->write_seq) -
READ_ONCE         638 net/ipv4/tcp.c 			       READ_ONCE(tp->snd_nxt);
READ_ONCE        1928 net/ipv4/tcp.c 	u32 copied_seq = READ_ONCE(tp->copied_seq);
READ_ONCE        1929 net/ipv4/tcp.c 	u32 rcv_nxt = READ_ONCE(tp->rcv_nxt);
READ_ONCE        1933 net/ipv4/tcp.c 	if (unlikely(inq < 0 || copied_seq != READ_ONCE(tp->copied_seq))) {
READ_ONCE        3227 net/ipv4/tcp.c 	rate = READ_ONCE(sk->sk_pacing_rate);
READ_ONCE        3231 net/ipv4/tcp.c 	rate = READ_ONCE(sk->sk_max_pacing_rate);
READ_ONCE        3380 net/ipv4/tcp.c 	rate = READ_ONCE(sk->sk_pacing_rate);
READ_ONCE         310 net/ipv4/tcp_bbr.c 		      sk->sk_pacing_rate >> READ_ONCE(sk->sk_pacing_shift),
READ_ONCE         679 net/ipv4/tcp_bpf.c 	struct proto *ops = READ_ONCE(sk->sk_prot);
READ_ONCE          29 net/ipv4/tcp_diag.c 		r->idiag_rqueue = max_t(int, READ_ONCE(tp->rcv_nxt) -
READ_ONCE          30 net/ipv4/tcp_diag.c 					     READ_ONCE(tp->copied_seq), 0);
READ_ONCE          31 net/ipv4/tcp_diag.c 		r->idiag_wqueue = READ_ONCE(tp->write_seq) - tp->snd_una;
READ_ONCE        3478 net/ipv4/tcp_input.c 	count = READ_ONCE(challenge_count);
READ_ONCE        1636 net/ipv4/tcp_ipv4.c 			struct dst_entry *dst = READ_ONCE(sk->sk_rx_dst);
READ_ONCE        1650 net/ipv4/tcp_ipv4.c 	u32 limit = READ_ONCE(sk->sk_rcvbuf) + READ_ONCE(sk->sk_sndbuf);
READ_ONCE        2462 net/ipv4/tcp_ipv4.c 		rx_queue = max_t(int, READ_ONCE(tp->rcv_nxt) -
READ_ONCE        2463 net/ipv4/tcp_ipv4.c 				      READ_ONCE(tp->copied_seq), 0);
READ_ONCE        2468 net/ipv4/tcp_ipv4.c 		READ_ONCE(tp->write_seq) - tp->snd_una,
READ_ONCE         373 net/ipv4/tcp_minisocks.c 	window_clamp = READ_ONCE(tp->window_clamp);
READ_ONCE         942 net/ipv4/tcp_output.c 	for (oval = READ_ONCE(sk->sk_tsq_flags);; oval = nval) {
READ_ONCE        1735 net/ipv4/tcp_output.c 		      sk->sk_pacing_rate >> READ_ONCE(sk->sk_pacing_shift),
READ_ONCE        1983 net/ipv4/tcp_output.c 	win_divisor = READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_tso_win_divisor);
READ_ONCE        2270 net/ipv4/tcp_output.c 		      sk->sk_pacing_rate >> READ_ONCE(sk->sk_pacing_shift));
READ_ONCE         391 net/ipv4/udp.c 	if (READ_ONCE(sk->sk_incoming_cpu) == raw_smp_processor_id())
READ_ONCE         607 net/ipv4/udp.c 		lookup = READ_ONCE(up->encap_err_lookup);
READ_ONCE        1549 net/ipv4/udp.c 	if (unlikely(READ_ONCE(sk->sk_peek_off) >= 0)) {
READ_ONCE        2025 net/ipv4/udp.c 		encap_rcv = READ_ONCE(up->encap_rcv);
READ_ONCE        2477 net/ipv4/udp.c 	dst = READ_ONCE(sk->sk_rx_dst);
READ_ONCE        2515 net/ipv4/udp.c 			encap_destroy = READ_ONCE(up->encap_destroy);
READ_ONCE        1363 net/ipv6/addrconf.c 	cnf_temp_preferred_lft = READ_ONCE(idev->cnf.temp_prefered_lft);
READ_ONCE         108 net/ipv6/inet6_hashtables.c 		if (READ_ONCE(sk->sk_incoming_cpu) == raw_smp_processor_id())
READ_ONCE          58 net/ipv6/ip6_input.c 		if (ipprot && (edemux = READ_ONCE(ipprot->early_demux)))
READ_ONCE        1287 net/ipv6/ip6_output.c 		      READ_ONCE(rt->dst.dev->mtu) : dst_mtu(&rt->dst);
READ_ONCE        1290 net/ipv6/ip6_output.c 			READ_ONCE(rt->dst.dev->mtu) : dst_mtu(xfrm_dst_path(&rt->dst));
READ_ONCE         492 net/ipv6/ip6_tunnel.c 	tproto = READ_ONCE(t->parms.proto);
READ_ONCE         893 net/ipv6/ip6_tunnel.c 		u8 tproto = READ_ONCE(t->parms.proto);
READ_ONCE        1244 net/ipv6/ip6_tunnel.c 	tproto = READ_ONCE(t->parms.proto);
READ_ONCE        1315 net/ipv6/ip6_tunnel.c 	tproto = READ_ONCE(t->parms.proto);
READ_ONCE        1744 net/ipv6/mcast.c 	mtu = READ_ONCE(dev->mtu);
READ_ONCE         283 net/ipv6/netfilter/ip6_tables.c 	private = READ_ONCE(table->private); /* Address dependency. */
READ_ONCE         801 net/ipv6/raw.c 	hdrincl = READ_ONCE(hdrincl);
READ_ONCE         645 net/ipv6/route.c 	last_probe = READ_ONCE(fib6_nh->last_probe);
READ_ONCE        1725 net/ipv6/tcp_ipv6.c 			struct dst_entry *dst = READ_ONCE(sk->sk_rx_dst);
READ_ONCE        1899 net/ipv6/tcp_ipv6.c 		rx_queue = max_t(int, READ_ONCE(tp->rcv_nxt) -
READ_ONCE        1900 net/ipv6/tcp_ipv6.c 				      READ_ONCE(tp->copied_seq), 0);
READ_ONCE        1911 net/ipv6/tcp_ipv6.c 		   READ_ONCE(tp->write_seq) - tp->snd_una,
READ_ONCE         138 net/ipv6/udp.c 	if (READ_ONCE(sk->sk_incoming_cpu) == raw_smp_processor_id())
READ_ONCE         475 net/ipv6/udp.c 		lookup = READ_ONCE(up->encap_err_lookup);
READ_ONCE         623 net/ipv6/udp.c 		encap_rcv = READ_ONCE(up->encap_rcv);
READ_ONCE        1000 net/ipv6/udp.c 	dst = READ_ONCE(sk->sk_rx_dst);
READ_ONCE        1545 net/ipv6/udp.c 			encap_destroy = READ_ONCE(up->encap_destroy);
READ_ONCE         816 net/llc/llc_conn.c 		if (sk_add_backlog(sk, skb, READ_ONCE(sk->sk_rcvbuf)))
READ_ONCE         196 net/llc/llc_input.c 	sap_handler = dest ? READ_ONCE(llc_type_handlers[dest - 1]) : NULL;
READ_ONCE         217 net/llc/llc_input.c 	sta_handler = READ_ONCE(llc_station_handler);
READ_ONCE        2109 net/mac80211/sta_info.c 	u16 rate = READ_ONCE(sta_get_last_rx_stats(sta)->last_rate);
READ_ONCE         247 net/mpls/af_mpls.c 	alive = READ_ONCE(rt->rt_nhn_alive);
READ_ONCE         256 net/mpls/af_mpls.c 		unsigned int nh_flags = READ_ONCE(nh->nh_flags);
READ_ONCE         466 net/netfilter/ipvs/ip_vs_sync.c 	unsigned long orig = READ_ONCE(cp->sync_endtime);
READ_ONCE        1795 net/netfilter/nf_conntrack_core.c 	if (READ_ONCE(ct->timeout) != extra_jiffies)
READ_ONCE        6022 net/netfilter/nf_tables_api.c 		const struct net_device *dev = READ_ONCE(flowtable->ops[i].dev);
READ_ONCE         105 net/netfilter/nf_tables_core.c 	pstats = READ_ONCE(base_chain->stats);
READ_ONCE         159 net/netfilter/nf_tables_core.c 	bool genbit = READ_ONCE(net->nft.gencursor);
READ_ONCE         423 net/netfilter/nfnetlink_queue.c 	switch ((enum nfqnl_config_mode)READ_ONCE(queue->copy_mode)) {
READ_ONCE         434 net/netfilter/nfnetlink_queue.c 		data_len = READ_ONCE(queue->copy_range);
READ_ONCE         383 net/netlabel/netlabel_calipso.c 	return READ_ONCE(calipso_ops);
READ_ONCE        1206 net/packet/af_packet.c 	len = READ_ONCE(po->rx_ring.frame_max) + 1;
READ_ONCE        1207 net/packet/af_packet.c 	idx = READ_ONCE(po->rx_ring.head);
READ_ONCE        1219 net/packet/af_packet.c 	len = READ_ONCE(po->rx_ring.prb_bdqc.knum_blocks);
READ_ONCE        1220 net/packet/af_packet.c 	idx = READ_ONCE(po->rx_ring.prb_bdqc.kactive_blk_num);
READ_ONCE        1235 net/packet/af_packet.c 		int rcvbuf = READ_ONCE(sk->sk_rcvbuf);
READ_ONCE        1269 net/packet/af_packet.c 	if (READ_ONCE(po->pressure) != pressure)
READ_ONCE        1277 net/packet/af_packet.c 	if (READ_ONCE(po->pressure) &&
READ_ONCE        1305 net/packet/af_packet.c 		if (READ_ONCE(history[i]) == rxhash)
READ_ONCE        1311 net/packet/af_packet.c 	if (READ_ONCE(history[victim]) != rxhash)
READ_ONCE        1368 net/packet/af_packet.c 		if (po_next != po_skip && !READ_ONCE(po_next->pressure) &&
READ_ONCE        1418 net/packet/af_packet.c 	unsigned int num = READ_ONCE(f->num_members);
READ_ONCE         180 net/rds/send.c 	send_gen = READ_ONCE(cp->cp_send_gen) + 1;
READ_ONCE         452 net/rds/send.c 		raced = send_gen != READ_ONCE(cp->cp_send_gen);
READ_ONCE         227 net/rxrpc/af_rxrpc.c 		max = READ_ONCE(rxrpc_max_backlog);
READ_ONCE          56 net/rxrpc/call_accept.c 	call_tail = READ_ONCE(b->call_backlog_tail);
READ_ONCE          65 net/rxrpc/call_accept.c 	tail = READ_ONCE(b->peer_backlog_tail);
READ_ONCE          76 net/rxrpc/call_accept.c 	tail = READ_ONCE(b->conn_backlog_tail);
READ_ONCE         119 net/rxrpc/call_event.c 		ack_at += READ_ONCE(call->tx_backoff);
READ_ONCE         329 net/rxrpc/call_event.c 	t = READ_ONCE(call->expect_rx_by);
READ_ONCE         335 net/rxrpc/call_event.c 	t = READ_ONCE(call->expect_req_by);
READ_ONCE         342 net/rxrpc/call_event.c 	t = READ_ONCE(call->expect_term_by);
READ_ONCE         348 net/rxrpc/call_event.c 	t = READ_ONCE(call->ack_at);
READ_ONCE         355 net/rxrpc/call_event.c 	t = READ_ONCE(call->ack_lost_at);
READ_ONCE         362 net/rxrpc/call_event.c 	t = READ_ONCE(call->keepalive_at);
READ_ONCE         371 net/rxrpc/call_event.c 	t = READ_ONCE(call->ping_at);
READ_ONCE         378 net/rxrpc/call_event.c 	t = READ_ONCE(call->resend_at);
READ_ONCE         427 net/rxrpc/call_event.c #define set(T) { t = READ_ONCE(T); if (time_before(t, next)) next = t; }
READ_ONCE        1117 net/rxrpc/conn_client.c 		now = READ_ONCE(jiffies);
READ_ONCE          49 net/rxrpc/conn_event.c 	call_id = READ_ONCE(chan->last_call);
READ_ONCE         117 net/rxrpc/conn_event.c 	if (READ_ONCE(chan->last_call) != call_id)
READ_ONCE         416 net/rxrpc/conn_event.c 		ack_at = READ_ONCE(chan->final_ack_at);
READ_ONCE         402 net/rxrpc/conn_object.c 			idle_timestamp = READ_ONCE(conn->idle_timestamp);
READ_ONCE         308 net/rxrpc/input.c 	rxrpc_seq_t top = READ_ONCE(call->tx_top);
READ_ONCE         428 net/rxrpc/input.c 	state = READ_ONCE(call->state);
READ_ONCE         435 net/rxrpc/input.c 		unsigned long timo = READ_ONCE(call->next_req_timo);
READ_ONCE         458 net/rxrpc/input.c 	hard_ack = READ_ONCE(call->rx_hard_ack);
READ_ONCE         694 net/rxrpc/input.c 	ping_serial = READ_ONCE(call->ping_serial);
READ_ONCE         812 net/rxrpc/input.c 	rxrpc_seq_t base = READ_ONCE(call->ackr_first_seq);
READ_ONCE         929 net/rxrpc/input.c 	switch (READ_ONCE(call->state)) {
READ_ONCE        1032 net/rxrpc/input.c 	timo = READ_ONCE(call->next_rx_timo);
READ_ONCE        1087 net/rxrpc/input.c 	switch (READ_ONCE(call->state)) {
READ_ONCE          39 net/rxrpc/output.c 		u16 tx_backoff = READ_ONCE(call->tx_backoff);
READ_ONCE          84 net/rxrpc/output.c 	hard_ack = READ_ONCE(call->rx_hard_ack);
READ_ONCE          79 net/rxrpc/proc.c 		local = READ_ONCE(rx->local);
READ_ONCE          95 net/rxrpc/proc.c 		timeout = READ_ONCE(call->expect_rx_by);
READ_ONCE          99 net/rxrpc/proc.c 	tx_hard_ack = READ_ONCE(call->tx_hard_ack);
READ_ONCE         100 net/rxrpc/proc.c 	rx_hard_ack = READ_ONCE(call->rx_hard_ack);
READ_ONCE         114 net/rxrpc/proc.c 		   tx_hard_ack, READ_ONCE(call->tx_top) - tx_hard_ack,
READ_ONCE         115 net/rxrpc/proc.c 		   rx_hard_ack, READ_ONCE(call->rx_top) - rx_hard_ack,
READ_ONCE         413 net/rxrpc/recvmsg.c 			ASSERTCMP(seq, ==, READ_ONCE(call->rx_top));
READ_ONCE         553 net/rxrpc/recvmsg.c 	switch (READ_ONCE(call->state)) {
READ_ONCE         656 net/rxrpc/recvmsg.c 	switch (READ_ONCE(call->state)) {
READ_ONCE         749 net/rxrpc/recvmsg.c 	if (READ_ONCE(call->state) != RXRPC_CALL_CLIENT_RECV_REPLY)
READ_ONCE         176 net/rxrpc/rtt.c 	u8 backoff = READ_ONCE(peer->backoff);
READ_ONCE          28 net/rxrpc/sendmsg.c 	rxrpc_seq_t tx_win = READ_ONCE(call->tx_hard_ack);
READ_ONCE          71 net/rxrpc/sendmsg.c 	rtt = READ_ONCE(call->peer->srtt_us) >> 3;
READ_ONCE          77 net/rxrpc/sendmsg.c 	tx_start = READ_ONCE(call->tx_hard_ack);
READ_ONCE          82 net/rxrpc/sendmsg.c 		tx_win = READ_ONCE(call->tx_hard_ack);
READ_ONCE         687 net/rxrpc/sendmsg.c 		switch (READ_ONCE(call->state)) {
READ_ONCE         742 net/rxrpc/sendmsg.c 	state = READ_ONCE(call->state);
READ_ONCE         809 net/rxrpc/sendmsg.c 	switch (READ_ONCE(call->state)) {
READ_ONCE         585 net/sched/act_csum.c 	action = READ_ONCE(p->tcf_action);
READ_ONCE         395 net/sched/act_ct.c 	retval = READ_ONCE(c->tcf_action);
READ_ONCE          96 net/sched/act_ctinfo.c 	action = READ_ONCE(ca->tcf_action);
READ_ONCE         154 net/sched/act_gact.c 	int action = READ_ONCE(gact->tcf_action);
READ_ONCE         158 net/sched/act_gact.c 	u32 ptype = READ_ONCE(gact->tcfg_ptype);
READ_ONCE         177 net/sched/act_gact.c 	int action = READ_ONCE(gact->tcf_action);
READ_ONCE         238 net/sched/act_mirred.c 	m_mac_header_xmit = READ_ONCE(m->tcfm_mac_header_xmit);
READ_ONCE         239 net/sched/act_mirred.c 	m_eaction = READ_ONCE(m->tcfm_eaction);
READ_ONCE         240 net/sched/act_mirred.c 	retval = READ_ONCE(m->tcf_action);
READ_ONCE          74 net/sched/act_mpls.c 	ret = READ_ONCE(m->tcf_action);
READ_ONCE         229 net/sched/act_police.c 	ret = READ_ONCE(police->tcf_action);
READ_ONCE         170 net/sched/act_sample.c 	retval = READ_ONCE(s->tcf_action);
READ_ONCE          37 net/sched/act_skbedit.c 	action = READ_ONCE(d->tcf_action);
READ_ONCE          44 net/sched/act_skbmod.c 	action = READ_ONCE(d->tcf_action);
READ_ONCE          35 net/sched/act_tunnel_key.c 	action = READ_ONCE(t->tcf_action);
READ_ONCE          40 net/sched/act_vlan.c 	action = READ_ONCE(v->tcf_action);
READ_ONCE        3196 net/sched/cls_api.c 	bool take_rtnl = READ_ONCE(block->lockeddevcnt) && !rtnl_held;
READ_ONCE        3232 net/sched/cls_api.c 	bool take_rtnl = READ_ONCE(block->lockeddevcnt) && !rtnl_held;
READ_ONCE        3284 net/sched/cls_api.c 	bool take_rtnl = READ_ONCE(block->lockeddevcnt) && !rtnl_held;
READ_ONCE        3336 net/sched/cls_api.c 	bool take_rtnl = READ_ONCE(block->lockeddevcnt) && !rtnl_held;
READ_ONCE         449 net/sched/em_meta.c 	dst->value = READ_ONCE(sk->sk_wmem_queued);
READ_ONCE         557 net/sched/em_meta.c 	dst->value = READ_ONCE(sk->sk_rcvlowat);
READ_ONCE         178 net/sctp/diag.c 		mem[SK_MEMINFO_BACKLOG] = READ_ONCE(sk->sk_backlog.len);
READ_ONCE         325 net/sctp/input.c 			if (sk_add_backlog(sk, skb, READ_ONCE(sk->sk_rcvbuf)))
READ_ONCE         340 net/sctp/input.c 			if (!sk_add_backlog(sk, skb, READ_ONCE(sk->sk_rcvbuf)))
READ_ONCE         367 net/sctp/input.c 	ret = sk_add_backlog(sk, skb, READ_ONCE(sk->sk_rcvbuf));
READ_ONCE        2092 net/sunrpc/auth_gss/auth_gss.c 	if (gss_seq_is_newer(req->rq_seqno, READ_ONCE(ctx->gc_seq)))
READ_ONCE        2095 net/sunrpc/auth_gss/auth_gss.c 	seq_xmit = READ_ONCE(ctx->gc_seq_xmit);
READ_ONCE          67 net/sunrpc/sched.c 	unsigned long timeout = READ_ONCE(task->tk_timeout);
READ_ONCE         850 net/sunrpc/sched.c 	queue = READ_ONCE(task->tk_waitqueue);
READ_ONCE         383 net/sunrpc/svc_xprt.c 	xpt_flags = READ_ONCE(xprt->xpt_flags);
READ_ONCE         709 net/sunrpc/xprt.c 	return READ_ONCE(xprt->connect_cookie);
READ_ONCE        1028 net/sunrpc/xprt.c 		READ_ONCE(task->tk_rqstp->rq_reply_bytes_recvd) != 0;
READ_ONCE        1035 net/sunrpc/xprt.c 		READ_ONCE(task->tk_rqstp->rq_reply_bytes_recvd) == 0;
READ_ONCE         175 net/sunrpc/xprtmultipath.c 	if (READ_ONCE(xps->xps_iter_ops) != &rpc_xprt_iter_roundrobin)
READ_ONCE         345 net/sunrpc/xprtmultipath.c 		nactive = READ_ONCE(xps->xps_nactive);
READ_ONCE         889 net/sunrpc/xprtrdma/verbs.c 	if (next_head == READ_ONCE(buf->rb_sc_tail))
READ_ONCE        2128 net/tipc/socket.c 		return READ_ONCE(sk->sk_rcvbuf);
READ_ONCE        2131 net/tipc/socket.c 		return READ_ONCE(sk->sk_rcvbuf) << msg_importance(hdr);
READ_ONCE        2134 net/tipc/socket.c 		return READ_ONCE(sk->sk_rcvbuf);
READ_ONCE        3800 net/tipc/socket.c 	i += scnprintf(buf + i, sz - i, " | %d\n", READ_ONCE(sk->sk_backlog.len));
READ_ONCE         670 net/tls/tls_device.c 	netdev = READ_ONCE(tls_ctx->netdev);
READ_ONCE         395 net/tls/tls_sw.c 		if (READ_ONCE(rec->tx_ready)) {
READ_ONCE        15749 net/wireless/nl80211.c 	u32 nlportid = READ_ONCE(wdev->ap_unexpected_nlportid);
READ_ONCE        15917 net/wireless/nl80211.c 	u32 nlportid = READ_ONCE(wdev->conn_owner_nlportid);
READ_ONCE          58 net/x25/x25_dev.c 			queued = !sk_add_backlog(sk, skb, READ_ONCE(sk->sk_rcvbuf));
READ_ONCE          36 net/xdp/xsk.c  	return READ_ONCE(xs->rx) &&  READ_ONCE(xs->umem) &&
READ_ONCE          37 net/xdp/xsk.c  		READ_ONCE(xs->umem->fq);
READ_ONCE         192 net/xdp/xsk.c  	if (READ_ONCE(xs->state) == XSK_BOUND) {
READ_ONCE         973 net/xdp/xsk.c  	if (READ_ONCE(xs->state) != XSK_READY)
READ_ONCE         977 net/xdp/xsk.c  		q = READ_ONCE(xs->rx);
READ_ONCE         979 net/xdp/xsk.c  		q = READ_ONCE(xs->tx);
READ_ONCE         981 net/xdp/xsk.c  		umem = READ_ONCE(xs->umem);
READ_ONCE         988 net/xdp/xsk.c  			q = READ_ONCE(umem->fq);
READ_ONCE         990 net/xdp/xsk.c  			q = READ_ONCE(umem->cq);
READ_ONCE         102 net/xdp/xsk_queue.h 		q->prod_tail = READ_ONCE(q->ring->producer);
READ_ONCE         117 net/xdp/xsk_queue.h 	q->cons_tail = READ_ONCE(q->ring->consumer);
READ_ONCE         129 net/xdp/xsk_queue.h 	q->prod_tail = READ_ONCE(q->ring->producer);
READ_ONCE         181 net/xdp/xsk_queue.h 		*addr = READ_ONCE(ring->desc[idx]) & q->chunk_mask;
READ_ONCE         307 net/xdp/xsk_queue.h 		*desc = READ_ONCE(ring->desc[idx]);
READ_ONCE        2412 net/xfrm/xfrm_state.c 	const struct xfrm_type *type = READ_ONCE(x->type);
READ_ONCE         374 samples/mic/mpssd/mpssd.c 	return READ_ONCE(vr->info->avail_idx);
READ_ONCE         515 samples/mic/mpssd/mpssd.c 	while (avail_idx == le16toh(READ_ONCE(vr->vr.avail->idx))) {
READ_ONCE         549 security/apparmor/apparmorfs.c 					     READ_ONCE(rev->ns->revision)))
READ_ONCE        1721 security/apparmor/apparmorfs.c 	ns = __aa_find_or_create_ns(parent, READ_ONCE(dentry->d_name.name),
READ_ONCE         425 security/apparmor/domain.c 					long rev = READ_ONCE(ns->revision);
READ_ONCE         435 security/apparmor/domain.c 					    READ_ONCE(ns->revision))
READ_ONCE         578 security/keys/keyring.c 	unsigned long kflags = READ_ONCE(key->flags);
READ_ONCE         579 security/keys/keyring.c 	short state = READ_ONCE(key->state);
READ_ONCE         591 security/keys/keyring.c 		time64_t expiry = READ_ONCE(key->expiry);
READ_ONCE         732 security/keys/keyring.c 	ptr = READ_ONCE(keyring->keys.root);
READ_ONCE         745 security/keys/keyring.c 		ptr = READ_ONCE(shortcut->next_node);
READ_ONCE         762 security/keys/keyring.c 		ptr = READ_ONCE(shortcut->next_node);
READ_ONCE         773 security/keys/keyring.c 		ptr = READ_ONCE(node->slots[slot]);
READ_ONCE         811 security/keys/keyring.c 	ptr = READ_ONCE(node->back_pointer);
READ_ONCE         816 security/keys/keyring.c 		ptr = READ_ONCE(shortcut->back_pointer);
READ_ONCE          87 security/keys/permission.c 	unsigned long flags = READ_ONCE(key->flags);
READ_ONCE          88 security/keys/permission.c 	time64_t expiry = READ_ONCE(key->expiry);
READ_ONCE         200 security/keys/proc.c 	expiry = READ_ONCE(key->expiry);
READ_ONCE         225 security/keys/proc.c 	flags = READ_ONCE(key->flags);
READ_ONCE          40 security/keys/process_keys.c 	struct key *reg_keyring = READ_ONCE(user_ns->user_keyring_register);
READ_ONCE         187 security/keys/process_keys.c 	struct key *reg_keyring = READ_ONCE(cred->user_ns->user_keyring_register);
READ_ONCE         211 security/selinux/ss/sidtab.c 		WRITE_ONCE(s->rcache[pos], READ_ONCE(s->rcache[pos - 1]));
READ_ONCE         228 security/selinux/ss/sidtab.c 		u32 v = READ_ONCE(s->rcache[i]);
READ_ONCE          53 sound/core/pcm_lib.c 		snd_pcm_uframes_t appl_ptr = READ_ONCE(runtime->control->appl_ptr);
READ_ONCE        2205 sound/core/pcm_lib.c 		appl_ptr = READ_ONCE(runtime->control->appl_ptr);
READ_ONCE          78 sound/core/seq/seq_virmidi.c 		if (!READ_ONCE(vmidi->trigger))
READ_ONCE         143 sound/core/seq/seq_virmidi.c 	while (READ_ONCE(vmidi->trigger)) {
READ_ONCE         433 sound/firewire/amdtp-stream.c 	struct snd_pcm_substream *pcm = READ_ONCE(s->pcm);
READ_ONCE         481 sound/firewire/amdtp-stream.c 	cip_header[0] = cpu_to_be32(READ_ONCE(s->source_node_id_field) |
READ_ONCE         770 sound/firewire/amdtp-stream.c 	pcm = READ_ONCE(s->pcm);
READ_ONCE        1058 sound/firewire/amdtp-stream.c 	return READ_ONCE(s->pcm_buffer_pointer);
READ_ONCE        1132 sound/firewire/amdtp-stream.c 	pcm = READ_ONCE(s->pcm);
READ_ONCE         416 sound/firewire/fireface/ff-protocol-former.c 		substream = READ_ONCE(ff->tx_midi_substreams[0]);
READ_ONCE         577 sound/firewire/fireface/ff-protocol-former.c 			substream = READ_ONCE(ff->tx_midi_substreams[0]);
READ_ONCE         587 sound/firewire/fireface/ff-protocol-former.c 			substream = READ_ONCE(ff->tx_midi_substreams[1]);
READ_ONCE         331 sound/firewire/fireface/ff-protocol-latter.c 	substream = READ_ONCE(ff->tx_midi_substreams[index]);
READ_ONCE          14 sound/firewire/fireface/ff-transaction.c 				READ_ONCE(ff->rx_midi_substreams[port]);
READ_ONCE          56 sound/firewire/fireface/ff-transaction.c 			READ_ONCE(ff->rx_midi_substreams[port]);
READ_ONCE         114 sound/firewire/isight.c 	if (!READ_ONCE(isight->pcm_running))
READ_ONCE         134 sound/firewire/isight.c 	if (READ_ONCE(isight->pcm_active))
READ_ONCE         144 sound/firewire/isight.c 	if (!READ_ONCE(isight->pcm_running))
READ_ONCE         442 sound/firewire/isight.c 	return READ_ONCE(isight->buffer_pointer);
READ_ONCE         242 sound/firewire/motu/amdtp-motu.c 	struct snd_rawmidi_substream *midi = READ_ONCE(p->midi);
READ_ONCE         274 sound/firewire/motu/amdtp-motu.c 		midi = READ_ONCE(p->midi);
READ_ONCE         114 sound/firewire/oxfw/oxfw-scs1x.c 		stream = READ_ONCE(scs->input);
READ_ONCE         185 sound/firewire/oxfw/oxfw-scs1x.c 	stream = READ_ONCE(scs->output);
READ_ONCE         133 sound/firewire/tascam/amdtp-tascam.c 	bool used = READ_ONCE(tscm->hwdep->used);
READ_ONCE         150 sound/firewire/tascam/tascam-transaction.c 	struct snd_rawmidi_substream *substream = READ_ONCE(port->substream);
READ_ONCE         175 sound/firewire/tascam/tascam-transaction.c 	struct snd_rawmidi_substream *substream = READ_ONCE(port->substream);
READ_ONCE         284 sound/firewire/tascam/tascam-transaction.c 		substream = READ_ONCE(tscm->tx_midi_substreams[port]);
READ_ONCE         111 sound/soc/dwc/dwc-pcm.c 			ptr = READ_ONCE(dev->tx_ptr);
READ_ONCE         116 sound/soc/dwc/dwc-pcm.c 			ptr = READ_ONCE(dev->rx_ptr);
READ_ONCE         241 sound/soc/dwc/dwc-pcm.c 		pos = READ_ONCE(dev->tx_ptr);
READ_ONCE         243 sound/soc/dwc/dwc-pcm.c 		pos = READ_ONCE(dev->rx_ptr);
READ_ONCE         913 sound/soc/sh/rcar/ssi.c 	*pointer = bytes_to_frames(runtime, READ_ONCE(ssi->byte_pos));
READ_ONCE          19 sound/soc/sof/trace.c 	loff_t host_offset = READ_ONCE(sdev->host_offset);
READ_ONCE         165 sound/soc/xtensa/xtfpga-i2s.c 		unsigned tx_ptr = READ_ONCE(i2s->tx_ptr);
READ_ONCE         459 sound/soc/xtensa/xtfpga-i2s.c 	snd_pcm_uframes_t pos = READ_ONCE(i2s->tx_ptr);
READ_ONCE         102 sound/usb/bcd2000/bcd2000.c 	midi_receive_substream = READ_ONCE(bcd2k->midi_receive_substream);
READ_ONCE         133 sound/usb/bcd2000/bcd2000.c 	midi_out_substream = READ_ONCE(bcd2k->midi_out_substream);
READ_ONCE          57 tools/arch/ia64/include/asm/barrier.h 	typeof(*p) ___p1 = READ_ONCE(*p);	\
READ_ONCE          41 tools/arch/powerpc/include/asm/barrier.h 	typeof(*p) ___p1 = READ_ONCE(*p);	\
READ_ONCE          39 tools/arch/s390/include/asm/barrier.h 	typeof(*p) ___p1 = READ_ONCE(*p);	\
READ_ONCE          51 tools/arch/sparc/include/asm/barrier_64.h 	typeof(*p) ___p1 = READ_ONCE(*p);	\
READ_ONCE          28 tools/arch/x86/include/asm/atomic.h 	return READ_ONCE((v)->counter);
READ_ONCE          41 tools/arch/x86/include/asm/barrier.h 	typeof(*p) ___p1 = READ_ONCE(*p);	\
READ_ONCE          25 tools/include/asm-generic/atomic-gcc.h 	return READ_ONCE((v)->counter);
READ_ONCE          61 tools/include/asm/barrier.h 	typeof(*p) ___p1 = READ_ONCE(*p);	\
READ_ONCE          39 tools/include/linux/rbtree.h #define RB_EMPTY_ROOT(root)  (READ_ONCE((root)->rb_node) == NULL)
READ_ONCE          61 tools/include/linux/ring_buffer.h 	u64 head = READ_ONCE(base->data_head);
READ_ONCE        5992 tools/lib/bpf/libbpf.c 	tmp_cpus = READ_ONCE(cpus);
READ_ONCE         396 tools/perf/util/auxtrace.h 	u64 head = READ_ONCE(pc->aux_head);
READ_ONCE         407 tools/perf/util/auxtrace.h 	u64 head = READ_ONCE(pc->aux_head);
READ_ONCE         132 tools/perf/util/session.h #define session_done()	READ_ONCE(session_done)
READ_ONCE          44 tools/testing/selftests/bpf/map_tests/sk_storage_map.c 	return READ_ONCE(stop);
READ_ONCE          49 tools/testing/selftests/bpf/map_tests/sk_storage_map.c 	return READ_ONCE(nr_sk_threads_err);
READ_ONCE          67 tools/testing/selftests/bpf/map_tests/sk_storage_map.c 	return READ_ONCE(nr_sk_threads_done);
READ_ONCE          99 tools/testing/selftests/bpf/map_tests/sk_storage_map.c 	while (READ_ONCE(sk_storage_map) == -1 && !is_stopped())
READ_ONCE         107 tools/testing/selftests/bpf/map_tests/sk_storage_map.c 	while (READ_ONCE(sk_storage_map) != -1 && !is_stopped())
READ_ONCE         185 tools/testing/selftests/bpf/map_tests/sk_storage_map.c 		map_fd = READ_ONCE(sk_storage_map);
READ_ONCE         287 tools/testing/selftests/bpf/map_tests/sk_storage_map.c 	int map_fd = READ_ONCE(sk_storage_map);
READ_ONCE         311 tools/testing/selftests/bpf/map_tests/sk_storage_map.c 	int map_fd = READ_ONCE(sk_storage_map);
READ_ONCE         102 tools/testing/selftests/kvm/dirty_log_test.c 		*(uint64_t *)addr = READ_ONCE(iteration);
READ_ONCE         108 tools/testing/selftests/kvm/dirty_log_test.c 			addr += (READ_ONCE(random_array[i]) % guest_num_pages)
READ_ONCE         111 tools/testing/selftests/kvm/dirty_log_test.c 			*(uint64_t *)addr = READ_ONCE(iteration);
READ_ONCE         161 tools/testing/selftests/kvm/dirty_log_test.c 	while (!READ_ONCE(host_quit)) {
READ_ONCE          27 tools/testing/selftests/powerpc/dscr/dscr_default_test.c 		s1 = READ_ONCE(sequence);
READ_ONCE          21 tools/testing/selftests/rcutorture/formal/srcu-cbmc/tests/store_buffering/test.c 	__unbuffered_tpr_y = READ_ONCE(y);
READ_ONCE          41 tools/testing/selftests/rcutorture/formal/srcu-cbmc/tests/store_buffering/test.c 	__unbuffered_tpr_x = READ_ONCE(x);
READ_ONCE          26 tools/virtio/linux/uaccess.h 	x = READ_ONCE(*(__pu_ptr));				\
READ_ONCE         509 virt/kvm/arm/arm.c 	return unlikely(READ_ONCE(vmid->vmid_gen) != current_vmid_gen);
READ_ONCE         351 virt/kvm/arm/mmu.c 		if (!READ_ONCE(kvm->arch.pgd))
READ_ONCE        1005 virt/kvm/arm/mmu.c 		pgd = READ_ONCE(kvm->arch.pgd);
READ_ONCE        1519 virt/kvm/arm/mmu.c 		if (!READ_ONCE(kvm->arch.pgd))
READ_ONCE         327 virt/kvm/arm/vgic/vgic-its.c 	irq_count = READ_ONCE(dist->lpi_list_count);
READ_ONCE         141 virt/kvm/arm/vgic/vgic-mmio-v3.c 		ret = extract_bytes(READ_ONCE(irq->mpidr), addr & 7, len);
READ_ONCE         405 virt/kvm/arm/vgic/vgic-mmio-v3.c 		old_propbaser = READ_ONCE(dist->propbaser);
READ_ONCE         433 virt/kvm/arm/vgic/vgic-mmio-v3.c 		old_pendbaser = READ_ONCE(vgic_cpu->pendbaser);
READ_ONCE          77 virt/kvm/coalesced_mmio.c 	insert = READ_ONCE(ring->last);
READ_ONCE        2414 virt/kvm/kvm_main.c 	grow_start = READ_ONCE(halt_poll_ns_grow_start);
READ_ONCE        2415 virt/kvm/kvm_main.c 	grow = READ_ONCE(halt_poll_ns_grow);
READ_ONCE        2436 virt/kvm/kvm_main.c 	shrink = READ_ONCE(halt_poll_ns_shrink);
READ_ONCE        2681 virt/kvm/kvm_main.c 			if (!READ_ONCE(vcpu->ready))
READ_ONCE        2687 virt/kvm/kvm_main.c 			if (READ_ONCE(vcpu->preempted) && yield_to_kernel_mode &&