READ_ONCE 30 arch/alpha/include/asm/atomic.h #define atomic_read(v) READ_ONCE((v)->counter) READ_ONCE 31 arch/alpha/include/asm/atomic.h #define atomic64_read(v) READ_ONCE((v)->counter) READ_ONCE 21 arch/arc/include/asm/atomic.h #define atomic_read(v) READ_ONCE((v)->counter) READ_ONCE 280 arch/arc/kernel/smp.c new = old = READ_ONCE(*ipi_data_ptr); READ_ONCE 27 arch/arm/include/asm/atomic.h #define atomic_read(v) READ_ONCE((v)->counter) READ_ONCE 75 arch/arm/include/asm/spinlock.h lockval.tickets.owner = READ_ONCE(lock->tickets.owner); READ_ONCE 121 arch/arm/include/asm/spinlock.h return !arch_spin_value_unlocked(READ_ONCE(*lock)); READ_ONCE 126 arch/arm/include/asm/spinlock.h struct __raw_tickets tickets = READ_ONCE(lock->tickets); READ_ONCE 65 arch/arm/kvm/reset.c if (READ_ONCE(vcpu->arch.reset_state.reset)) { READ_ONCE 26 arch/arm/vdso/vgettimeofday.c seq = READ_ONCE(vdata->seq_count); READ_ONCE 104 arch/arm64/include/asm/atomic.h #define arch_atomic_read(v) READ_ONCE((v)->counter) READ_ONCE 135 arch/arm64/include/asm/barrier.h VAL = READ_ONCE(*__PTR); \ READ_ONCE 23 arch/arm64/include/asm/hugetlb.h return READ_ONCE(*ptep); READ_ONCE 228 arch/arm64/include/asm/kvm_mmu.h pteval = READ_ONCE(pte_val(*ptep)); READ_ONCE 239 arch/arm64/include/asm/kvm_mmu.h return (READ_ONCE(pte_val(*ptep)) & PTE_S2_RDWR) == PTE_S2_RDONLY; READ_ONCE 244 arch/arm64/include/asm/kvm_mmu.h return !(READ_ONCE(pte_val(*ptep)) & PTE_S2_XN); READ_ONCE 259 arch/arm64/include/asm/kvm_mmu.h return !(READ_ONCE(pmd_val(*pmdp)) & PMD_S2_XN); READ_ONCE 274 arch/arm64/include/asm/kvm_mmu.h return !(READ_ONCE(pud_val(*pudp)) & PUD_S2_XN); READ_ONCE 43 arch/arm64/include/asm/percpu.h return READ_ONCE(*(u##sz *)ptr); \ READ_ONCE 251 arch/arm64/include/asm/pgtable.h old_pte = READ_ONCE(*ptep); READ_ONCE 501 arch/arm64/include/asm/pgtable.h #define pte_offset_phys(dir,addr) (pmd_page_paddr(READ_ONCE(*(dir))) + pte_index(addr) * sizeof(pte_t)) READ_ONCE 560 arch/arm64/include/asm/pgtable.h #define pmd_offset_phys(dir, addr) (pud_page_paddr(READ_ONCE(*(dir))) + pmd_index(addr) * sizeof(pmd_t)) READ_ONCE 618 arch/arm64/include/asm/pgtable.h #define pud_offset_phys(dir, addr) (pgd_page_paddr(READ_ONCE(*(dir))) + pud_index(addr) * sizeof(pud_t)) READ_ONCE 707 arch/arm64/include/asm/pgtable.h pte = READ_ONCE(*ptep); READ_ONCE 781 arch/arm64/include/asm/pgtable.h pte = READ_ONCE(*ptep); READ_ONCE 12 arch/arm64/include/asm/preempt.h return READ_ONCE(current_thread_info()->preempt.count); READ_ONCE 46 arch/arm64/include/asm/preempt.h u32 pc = READ_ONCE(current_thread_info()->preempt.count); READ_ONCE 53 arch/arm64/include/asm/preempt.h u32 pc = READ_ONCE(current_thread_info()->preempt.count); READ_ONCE 61 arch/arm64/include/asm/preempt.h u64 pc = READ_ONCE(ti->preempt_count); READ_ONCE 73 arch/arm64/include/asm/preempt.h return !pc || !READ_ONCE(ti->preempt_count); READ_ONCE 78 arch/arm64/include/asm/preempt.h u64 pc = READ_ONCE(current_thread_info()->preempt_count); READ_ONCE 134 arch/arm64/include/asm/uaccess.h ttbr0 = READ_ONCE(current_thread_info()->ttbr0); READ_ONCE 220 arch/arm64/kernel/alternative.c while (!READ_ONCE(all_alternatives_applied)) READ_ONCE 88 arch/arm64/kernel/efi.c pte_t pte = READ_ONCE(*ptep); READ_ONCE 226 arch/arm64/kernel/hibernate.c if (pgd_none(READ_ONCE(*pgdp))) { READ_ONCE 236 arch/arm64/kernel/hibernate.c if (pud_none(READ_ONCE(*pudp))) { READ_ONCE 246 arch/arm64/kernel/hibernate.c if (pmd_none(READ_ONCE(*pmdp))) { READ_ONCE 344 arch/arm64/kernel/hibernate.c pte_t pte = READ_ONCE(*src_ptep); READ_ONCE 398 arch/arm64/kernel/hibernate.c if (pud_none(READ_ONCE(*dst_pudp))) { READ_ONCE 408 arch/arm64/kernel/hibernate.c pmd_t pmd = READ_ONCE(*src_pmdp); READ_ONCE 433 arch/arm64/kernel/hibernate.c if (pgd_none(READ_ONCE(*dst_pgdp))) { READ_ONCE 443 arch/arm64/kernel/hibernate.c pud_t pud = READ_ONCE(*src_pudp); READ_ONCE 470 arch/arm64/kernel/hibernate.c if (pgd_none(READ_ONCE(*src_pgdp))) READ_ONCE 1832 arch/arm64/kernel/ptrace.c unsigned long flags = READ_ONCE(current_thread_info()->flags); READ_ONCE 938 arch/arm64/kernel/signal.c thread_flags = READ_ONCE(current_thread_info()->flags); READ_ONCE 140 arch/arm64/kernel/smp.c status = READ_ONCE(secondary_data.status); READ_ONCE 144 arch/arm64/kernel/smp.c status = READ_ONCE(__early_cpu_boot_status); READ_ONCE 302 arch/arm64/mm/dump.c note_page(st, addr, 4, READ_ONCE(pte_val(*ptep))); READ_ONCE 313 arch/arm64/mm/dump.c pmd_t pmd = READ_ONCE(*pmdp); READ_ONCE 332 arch/arm64/mm/dump.c pud_t pud = READ_ONCE(*pudp); READ_ONCE 352 arch/arm64/mm/dump.c pgd_t pgd = READ_ONCE(*pgdp); READ_ONCE 155 arch/arm64/mm/fault.c pgd = READ_ONCE(*pgdp); READ_ONCE 167 arch/arm64/mm/fault.c pud = READ_ONCE(*pudp); READ_ONCE 173 arch/arm64/mm/fault.c pmd = READ_ONCE(*pmdp); READ_ONCE 179 arch/arm64/mm/fault.c pte = READ_ONCE(*ptep); READ_ONCE 202 arch/arm64/mm/fault.c pte_t pte = READ_ONCE(*ptep); READ_ONCE 247 arch/arm64/mm/hugetlbpage.c pud_none(READ_ONCE(*pudp))) READ_ONCE 268 arch/arm64/mm/hugetlbpage.c if (!pgd_present(READ_ONCE(*pgdp))) READ_ONCE 272 arch/arm64/mm/hugetlbpage.c pud = READ_ONCE(*pudp); READ_ONCE 284 arch/arm64/mm/hugetlbpage.c pmd = READ_ONCE(*pmdp); READ_ONCE 413 arch/arm64/mm/hugetlbpage.c if (!pte_cont(READ_ONCE(*ptep))) { READ_ONCE 437 arch/arm64/mm/hugetlbpage.c if (!pte_cont(READ_ONCE(*ptep))) { READ_ONCE 63 arch/arm64/mm/kasan_init.c if (pmd_none(READ_ONCE(*pmdp))) { READ_ONCE 77 arch/arm64/mm/kasan_init.c if (pud_none(READ_ONCE(*pudp))) { READ_ONCE 90 arch/arm64/mm/kasan_init.c if (pgd_none(READ_ONCE(*pgdp))) { READ_ONCE 114 arch/arm64/mm/kasan_init.c } while (ptep++, addr = next, addr != end && pte_none(READ_ONCE(*ptep))); READ_ONCE 126 arch/arm64/mm/kasan_init.c } while (pmdp++, addr = next, addr != end && pmd_none(READ_ONCE(*pmdp))); READ_ONCE 138 arch/arm64/mm/kasan_init.c } while (pudp++, addr = next, addr != end && pud_none(READ_ONCE(*pudp))); READ_ONCE 184 arch/arm64/mm/kasan_init.c set_pgd(pgdp_new, READ_ONCE(*pgdp)); READ_ONCE 147 arch/arm64/mm/mmu.c pte_t old_pte = READ_ONCE(*ptep); READ_ONCE 156 arch/arm64/mm/mmu.c READ_ONCE(pte_val(*ptep)))); READ_ONCE 171 arch/arm64/mm/mmu.c pmd_t pmd = READ_ONCE(*pmdp); READ_ONCE 179 arch/arm64/mm/mmu.c pmd = READ_ONCE(*pmdp); READ_ONCE 208 arch/arm64/mm/mmu.c pmd_t old_pmd = READ_ONCE(*pmdp); READ_ONCE 222 arch/arm64/mm/mmu.c READ_ONCE(pmd_val(*pmdp)))); READ_ONCE 228 arch/arm64/mm/mmu.c pmd_val(old_pmd) != READ_ONCE(pmd_val(*pmdp))); READ_ONCE 242 arch/arm64/mm/mmu.c pud_t pud = READ_ONCE(*pudp); READ_ONCE 253 arch/arm64/mm/mmu.c pud = READ_ONCE(*pudp); READ_ONCE 292 arch/arm64/mm/mmu.c pgd_t pgd = READ_ONCE(*pgdp); READ_ONCE 299 arch/arm64/mm/mmu.c pgd = READ_ONCE(*pgdp); READ_ONCE 305 arch/arm64/mm/mmu.c pud_t old_pud = READ_ONCE(*pudp); READ_ONCE 321 arch/arm64/mm/mmu.c READ_ONCE(pud_val(*pudp)))); READ_ONCE 327 arch/arm64/mm/mmu.c pud_val(old_pud) != READ_ONCE(pud_val(*pudp))); READ_ONCE 641 arch/arm64/mm/mmu.c if (!READ_ONCE(pgd_val(*pgd_offset_raw(pgdp, FIXADDR_START)))) { READ_ONCE 648 arch/arm64/mm/mmu.c READ_ONCE(*pgd_offset_k(FIXADDR_START))); READ_ONCE 702 arch/arm64/mm/mmu.c if (pgd_none(READ_ONCE(*pgdp))) READ_ONCE 706 arch/arm64/mm/mmu.c pud = READ_ONCE(*pudp); READ_ONCE 714 arch/arm64/mm/mmu.c pmd = READ_ONCE(*pmdp); READ_ONCE 722 arch/arm64/mm/mmu.c pte = READ_ONCE(*ptep); READ_ONCE 757 arch/arm64/mm/mmu.c if (pmd_none(READ_ONCE(*pmdp))) { READ_ONCE 781 arch/arm64/mm/mmu.c pgd_t pgd = READ_ONCE(*pgdp); READ_ONCE 791 arch/arm64/mm/mmu.c pud_t pud = READ_ONCE(*pudp); READ_ONCE 817 arch/arm64/mm/mmu.c pgd = READ_ONCE(*pgdp); READ_ONCE 832 arch/arm64/mm/mmu.c if (pud_none(READ_ONCE(*pudp))) READ_ONCE 961 arch/arm64/mm/mmu.c if (!pgattr_change_is_safe(READ_ONCE(pud_val(*pudp)), READ_ONCE 975 arch/arm64/mm/mmu.c if (!pgattr_change_is_safe(READ_ONCE(pmd_val(*pmdp)), READ_ONCE 986 arch/arm64/mm/mmu.c if (!pud_sect(READ_ONCE(*pudp))) READ_ONCE 994 arch/arm64/mm/mmu.c if (!pmd_sect(READ_ONCE(*pmdp))) READ_ONCE 1005 arch/arm64/mm/mmu.c pmd = READ_ONCE(*pmdp); READ_ONCE 1026 arch/arm64/mm/mmu.c pud = READ_ONCE(*pudp); READ_ONCE 25 arch/arm64/mm/pageattr.c pte_t pte = READ_ONCE(*ptep); READ_ONCE 210 arch/arm64/mm/pageattr.c if (pgd_none(READ_ONCE(*pgdp))) READ_ONCE 214 arch/arm64/mm/pageattr.c pud = READ_ONCE(*pudp); READ_ONCE 221 arch/arm64/mm/pageattr.c pmd = READ_ONCE(*pmdp); READ_ONCE 228 arch/arm64/mm/pageattr.c return pte_valid(READ_ONCE(*ptep)); READ_ONCE 32 arch/csky/include/asm/spinlock.h lockval.tickets.owner = READ_ONCE(lock->tickets.owner); READ_ONCE 78 arch/csky/include/asm/spinlock.h return !arch_spin_value_unlocked(READ_ONCE(*lock)); READ_ONCE 83 arch/csky/include/asm/spinlock.h struct __raw_tickets tickets = READ_ONCE(lock->tickets); READ_ONCE 145 arch/csky/include/asm/spinlock.h #define arch_spin_is_locked(x) (READ_ONCE((x)->lock) != 0) READ_ONCE 17 arch/h8300/include/asm/atomic.h #define atomic_read(v) READ_ONCE((v)->counter) READ_ONCE 39 arch/hexagon/include/asm/atomic.h #define atomic_read(v) READ_ONCE((v)->counter) READ_ONCE 25 arch/ia64/include/asm/atomic.h #define atomic_read(v) READ_ONCE((v)->counter) READ_ONCE 26 arch/ia64/include/asm/atomic.h #define atomic64_read(v) READ_ONCE((v)->counter) READ_ONCE 65 arch/ia64/include/asm/barrier.h typeof(*p) ___p1 = READ_ONCE(*p); \ READ_ONCE 65 arch/ia64/include/asm/spinlock.h int tmp = READ_ONCE(lock->lock); READ_ONCE 84 arch/ia64/include/asm/spinlock.h long tmp = READ_ONCE(lock->lock); READ_ONCE 91 arch/ia64/include/asm/spinlock.h long tmp = READ_ONCE(lock->lock); READ_ONCE 21 arch/m68k/include/asm/atomic.h #define atomic_read(v) READ_ONCE((v)->counter) READ_ONCE 308 arch/mips/cavium-octeon/executive/cvmx-l2c.c READ_ONCE(*ptr); READ_ONCE 44 arch/mips/include/asm/atomic.h #define atomic_read(v) READ_ONCE((v)->counter) READ_ONCE 248 arch/mips/include/asm/atomic.h #define atomic64_read(v) READ_ONCE((v)->counter) READ_ONCE 210 arch/mips/mm/context.c old_active_mmid = READ_ONCE(cpu_data[cpu].asid_cache); READ_ONCE 26 arch/nds32/kernel/vdso/gettimeofday.c seq = READ_ONCE(vdata->seq_count); READ_ONCE 73 arch/openrisc/include/asm/cmpxchg.h load32 = READ_ONCE(*p); READ_ONCE 106 arch/openrisc/include/asm/cmpxchg.h oldv = READ_ONCE(*p); READ_ONCE 73 arch/parisc/include/asm/atomic.h return READ_ONCE((v)->counter); READ_ONCE 218 arch/parisc/include/asm/atomic.h return READ_ONCE((v)->counter); READ_ONCE 75 arch/powerpc/include/asm/barrier.h typeof(*p) ___p1 = READ_ONCE(*p); \ READ_ONCE 428 arch/powerpc/include/asm/kvm_book3s_64.h old_pte = READ_ONCE(*ptep); READ_ONCE 22 arch/powerpc/include/asm/local.h return READ_ONCE(l->v); READ_ONCE 3282 arch/powerpc/kernel/ptrace.c flags = READ_ONCE(current_thread_info()->flags) & READ_ONCE 426 arch/powerpc/kernel/smp.c fn = READ_ONCE(nmi_ipi_function); READ_ONCE 1272 arch/powerpc/kvm/book3s_64_mmu_radix.c pgd = READ_ONCE(*pgdp); READ_ONCE 1279 arch/powerpc/kvm/book3s_64_mmu_radix.c pud = READ_ONCE(*pudp); READ_ONCE 1291 arch/powerpc/kvm/book3s_64_mmu_radix.c pmd = READ_ONCE(*pmdp); READ_ONCE 1303 arch/powerpc/kvm/book3s_64_mmu_radix.c pte = pte_val(READ_ONCE(*ptep)); READ_ONCE 172 arch/powerpc/kvm/book3s_hv.c vcpu = READ_ONCE(vc->runnable_threads[i]); READ_ONCE 240 arch/powerpc/kvm/book3s_hv.c cpu = READ_ONCE(vcpu->arch.thread_cpu); READ_ONCE 1179 arch/powerpc/kvm/book3s_hv.c pcpu = READ_ONCE(v->cpu); READ_ONCE 89 arch/powerpc/kvm/book3s_hv_rm_xics.c old = new = READ_ONCE(rm_core[core].rm_state); READ_ONCE 253 arch/powerpc/kvm/book3s_hv_rm_xics.c old_state = new_state = READ_ONCE(icp->state); READ_ONCE 452 arch/powerpc/kvm/book3s_hv_rm_xics.c old_state = new_state = READ_ONCE(icp->state); READ_ONCE 511 arch/powerpc/kvm/book3s_hv_rm_xics.c old_state = new_state = READ_ONCE(icp->state); READ_ONCE 577 arch/powerpc/kvm/book3s_hv_rm_xics.c old_state = new_state = READ_ONCE(icp->state); READ_ONCE 652 arch/powerpc/kvm/book3s_hv_rm_xics.c old_state = new_state = READ_ONCE(icp->state); READ_ONCE 346 arch/powerpc/kvm/book3s_xics.c old_state = new_state = READ_ONCE(icp->state); READ_ONCE 549 arch/powerpc/kvm/book3s_xics.c old_state = new_state = READ_ONCE(icp->state); READ_ONCE 603 arch/powerpc/kvm/book3s_xics.c old_state = new_state = READ_ONCE(icp->state); READ_ONCE 670 arch/powerpc/kvm/book3s_xics.c old_state = new_state = READ_ONCE(icp->state); READ_ONCE 715 arch/powerpc/kvm/book3s_xics.c state = READ_ONCE(icp->state); READ_ONCE 756 arch/powerpc/kvm/book3s_xics.c old_state = new_state = READ_ONCE(icp->state); READ_ONCE 971 arch/powerpc/kvm/book3s_xics.c state.raw = READ_ONCE(icp->state.raw); READ_ONCE 1168 arch/powerpc/kvm/book3s_xics.c old_state = READ_ONCE(icp->state); READ_ONCE 150 arch/powerpc/kvm/book3s_xive_template.c qpage = READ_ONCE(q->qpage); READ_ONCE 390 arch/powerpc/kvm/book3s_xive_template.c qpage = READ_ONCE(q->qpage); READ_ONCE 479 arch/powerpc/kvm/e500_mmu_host.c pte_t pte = READ_ONCE(*ptep); READ_ONCE 34 arch/powerpc/mm/book3s64/hash_4k.c pte_t pte = READ_ONCE(*ptep); READ_ONCE 52 arch/powerpc/mm/book3s64/hash_64k.c pte_t pte = READ_ONCE(*ptep); READ_ONCE 238 arch/powerpc/mm/book3s64/hash_64k.c pte_t pte = READ_ONCE(*ptep); READ_ONCE 36 arch/powerpc/mm/book3s64/hash_hugepage.c pmd_t pmd = READ_ONCE(*pmdp); READ_ONCE 1670 arch/powerpc/mm/book3s64/hash_utils.c pkey = pte_to_pkey_bits(pte_val(READ_ONCE(*ptep))); READ_ONCE 329 arch/powerpc/mm/pgtable.c pgd = READ_ONCE(*pgdp); READ_ONCE 356 arch/powerpc/mm/pgtable.c pud = READ_ONCE(*pudp); READ_ONCE 373 arch/powerpc/mm/pgtable.c pmd = READ_ONCE(*pmdp); READ_ONCE 136 arch/powerpc/perf/callchain.c pte = READ_ONCE(*ptep); READ_ONCE 973 arch/powerpc/perf/imc-pmu.c data = be64_to_cpu(READ_ONCE(*addr)); READ_ONCE 1164 arch/powerpc/perf/imc-pmu.c if (be64_to_cpu(READ_ONCE(mem->tb1)) > *prev_tb) READ_ONCE 1165 arch/powerpc/perf/imc-pmu.c *prev_tb = be64_to_cpu(READ_ONCE(mem->tb1)); READ_ONCE 1169 arch/powerpc/perf/imc-pmu.c if ((be64_to_cpu(READ_ONCE(mem->tb1)) & IMC_TRACE_RECORD_TB1_MASK) != READ_ONCE 1170 arch/powerpc/perf/imc-pmu.c be64_to_cpu(READ_ONCE(mem->tb2))) READ_ONCE 1174 arch/powerpc/perf/imc-pmu.c data->ip = be64_to_cpu(READ_ONCE(mem->ip)); READ_ONCE 261 arch/powerpc/platforms/powernv/idle.c u64 s = READ_ONCE(*state); READ_ONCE 46 arch/powerpc/platforms/powernv/opal-irqchip.c e = READ_ONCE(last_outstanding_events) & opal_event_irqchip.mask; READ_ONCE 42 arch/powerpc/platforms/powernv/opal-msglog.c out_pos = be32_to_cpu(READ_ONCE(mc->out_pos)); READ_ONCE 64 arch/powerpc/platforms/powernv/pci-ioda-tce.c unsigned long oldtce, tce = be64_to_cpu(READ_ONCE(tmp[n])); READ_ONCE 32 arch/riscv/include/asm/atomic.h return READ_ONCE(v->counter); READ_ONCE 43 arch/riscv/include/asm/atomic.h return READ_ONCE(v->counter); READ_ONCE 39 arch/riscv/include/asm/barrier.h typeof(*p) ___p1 = READ_ONCE(*p); \ READ_ONCE 20 arch/riscv/include/asm/spinlock.h #define arch_spin_is_locked(x) (READ_ONCE((x)->lock) != 0) READ_ONCE 43 arch/s390/include/asm/barrier.h typeof(*p) ___p1 = READ_ONCE(*p); \ READ_ONCE 1247 arch/s390/include/asm/pgtable.h #define pgd_offset(mm, address) pgd_offset_raw(READ_ONCE((mm)->pgd), address) READ_ONCE 17 arch/s390/include/asm/preempt.h return READ_ONCE(S390_lowcore.preempt_count) & ~PREEMPT_NEED_RESCHED; READ_ONCE 25 arch/s390/include/asm/preempt.h old = READ_ONCE(S390_lowcore.preempt_count); READ_ONCE 50 arch/s390/include/asm/preempt.h return !(READ_ONCE(S390_lowcore.preempt_count) & PREEMPT_NEED_RESCHED); READ_ONCE 73 arch/s390/include/asm/preempt.h return unlikely(READ_ONCE(S390_lowcore.preempt_count) == READ_ONCE 83 arch/s390/include/asm/preempt.h return READ_ONCE(S390_lowcore.preempt_count); READ_ONCE 55 arch/s390/include/asm/spinlock.h return READ_ONCE(lp->lock) != 0; READ_ONCE 142 arch/s390/include/asm/spinlock.h old = READ_ONCE(rw->cnts); READ_ONCE 151 arch/s390/include/asm/spinlock.h old = READ_ONCE(rw->cnts); READ_ONCE 65 arch/s390/include/asm/tlbflush.h gmap_asce = READ_ONCE(mm->context.gmap_asce); READ_ONCE 61 arch/s390/kernel/idle.c idle_count = READ_ONCE(idle->idle_count); READ_ONCE 62 arch/s390/kernel/idle.c if (READ_ONCE(idle->clock_idle_enter)) READ_ONCE 78 arch/s390/kernel/idle.c idle_time = READ_ONCE(idle->idle_time); READ_ONCE 79 arch/s390/kernel/idle.c idle_enter = READ_ONCE(idle->clock_idle_enter); READ_ONCE 80 arch/s390/kernel/idle.c idle_exit = READ_ONCE(idle->clock_idle_exit); READ_ONCE 104 arch/s390/kernel/idle.c idle_enter = READ_ONCE(idle->clock_idle_enter); READ_ONCE 105 arch/s390/kernel/idle.c idle_exit = READ_ONCE(idle->clock_idle_exit); READ_ONCE 627 arch/s390/kernel/time.c while (READ_ONCE(sync->in_sync) == 0) READ_ONCE 155 arch/s390/kernel/vtime.c READ_ONCE(S390_lowcore.user_timer)); READ_ONCE 157 arch/s390/kernel/vtime.c READ_ONCE(S390_lowcore.guest_timer)); READ_ONCE 159 arch/s390/kernel/vtime.c READ_ONCE(S390_lowcore.system_timer)); READ_ONCE 161 arch/s390/kernel/vtime.c READ_ONCE(S390_lowcore.hardirq_timer)); READ_ONCE 163 arch/s390/kernel/vtime.c READ_ONCE(S390_lowcore.softirq_timer)); READ_ONCE 289 arch/s390/kvm/gaccess.c old = READ_ONCE(*ic); READ_ONCE 314 arch/s390/kvm/gaccess.c old = READ_ONCE(*ic); READ_ONCE 332 arch/s390/kvm/gaccess.c old = READ_ONCE(*ic); READ_ONCE 352 arch/s390/kvm/gaccess.c old = READ_ONCE(*ic); READ_ONCE 245 arch/s390/kvm/interrupt.c word = READ_ONCE(gisa->u64.word[0]); READ_ONCE 268 arch/s390/kvm/interrupt.c word = READ_ONCE(gisa->u64.word[0]); READ_ONCE 289 arch/s390/kvm/interrupt.c word = READ_ONCE(gi->origin->u64.word[0]); READ_ONCE 290 arch/s390/kvm/interrupt.c alert_mask = READ_ONCE(gi->alert.mask); READ_ONCE 302 arch/s390/kvm/interrupt.c return READ_ONCE(gisa->next_alert) != (u32)(u64)gisa; READ_ONCE 312 arch/s390/kvm/interrupt.c return READ_ONCE(gisa->ipm); READ_ONCE 1850 arch/s390/kvm/interrupt.c u64 type = READ_ONCE(inti->type); READ_ONCE 287 arch/s390/kvm/vsie.c const uint32_t crycbd_o = READ_ONCE(scb_o->crycbd); READ_ONCE 364 arch/s390/kvm/vsie.c const uint32_t new_ibc = READ_ONCE(__new_ibc) & 0x0fffU; READ_ONCE 443 arch/s390/kvm/vsie.c const uint32_t new_prefix = READ_ONCE(__new_prefix); READ_ONCE 444 arch/s390/kvm/vsie.c const bool wants_tx = READ_ONCE(scb_o->ecb) & ECB_TE; READ_ONCE 492 arch/s390/kvm/vsie.c new_mso = READ_ONCE(scb_o->mso) & 0xfffffffffff00000UL; READ_ONCE 580 arch/s390/kvm/vsie.c page = READ_ONCE(kvm->arch.vsie.pages[i]); READ_ONCE 584 arch/s390/kvm/vsie.c if (READ_ONCE(cur->gmap) != gmap) READ_ONCE 727 arch/s390/kvm/vsie.c gpa = READ_ONCE(scb_o->scaol) & ~0xfUL; READ_ONCE 729 arch/s390/kvm/vsie.c gpa |= (u64) READ_ONCE(scb_o->scaoh) << 32; READ_ONCE 750 arch/s390/kvm/vsie.c gpa = READ_ONCE(scb_o->itdba) & ~0xffUL; READ_ONCE 766 arch/s390/kvm/vsie.c gpa = READ_ONCE(scb_o->gvrd) & ~0x1ffUL; READ_ONCE 785 arch/s390/kvm/vsie.c gpa = READ_ONCE(scb_o->riccbd) & ~0x3fUL; READ_ONCE 805 arch/s390/kvm/vsie.c gpa = READ_ONCE(scb_o->sdnxo) & ~0xfUL; READ_ONCE 806 arch/s390/kvm/vsie.c sdnxc = READ_ONCE(scb_o->sdnxo) & 0xfUL; READ_ONCE 972 arch/s390/kvm/vsie.c __u32 fac = READ_ONCE(vsie_page->scb_o->fac) & 0x7ffffff8U; READ_ONCE 1368 arch/s390/kvm/vsie.c struct kvm_s390_sie_block *scb = READ_ONCE(vcpu->arch.vsie_block); READ_ONCE 131 arch/s390/lib/spinlock.c old = READ_ONCE(lp->lock); READ_ONCE 168 arch/s390/lib/spinlock.c while (READ_ONCE(node->prev) != NULL) { READ_ONCE 182 arch/s390/lib/spinlock.c old = READ_ONCE(lp->lock); READ_ONCE 202 arch/s390/lib/spinlock.c while ((next = READ_ONCE(node->next)) == NULL) READ_ONCE 218 arch/s390/lib/spinlock.c owner = arch_spin_yield_target(READ_ONCE(lp->lock), NULL); READ_ONCE 259 arch/s390/lib/spinlock.c owner = READ_ONCE(lp->lock); READ_ONCE 273 arch/s390/lib/spinlock.c while (READ_ONCE(rw->cnts) & 0x10000) READ_ONCE 285 arch/s390/lib/spinlock.c while (READ_ONCE(rw->cnts) & 0x10000) READ_ONCE 302 arch/s390/lib/spinlock.c old = READ_ONCE(rw->cnts); READ_ONCE 318 arch/s390/lib/spinlock.c cpu = READ_ONCE(lp->lock) & _Q_LOCK_CPU_MASK; READ_ONCE 36 arch/s390/mm/pgtable.c asce = READ_ONCE(mm->context.gmap_asce); READ_ONCE 56 arch/s390/mm/pgtable.c asce = READ_ONCE(mm->context.gmap_asce); READ_ONCE 24 arch/sh/include/asm/atomic.h #define atomic_read(v) READ_ONCE((v)->counter) READ_ONCE 32 arch/sh/include/asm/cmpxchg-xchg.h oldv = READ_ONCE(*p); READ_ONCE 37 arch/sparc/include/asm/atomic_32.h #define atomic_read(v) READ_ONCE((v)->counter) READ_ONCE 18 arch/sparc/include/asm/atomic_64.h #define atomic_read(v) READ_ONCE((v)->counter) READ_ONCE 19 arch/sparc/include/asm/atomic_64.h #define atomic64_read(v) READ_ONCE((v)->counter) READ_ONCE 50 arch/sparc/include/asm/barrier_64.h typeof(*p) ___p1 = READ_ONCE(*p); \ READ_ONCE 45 arch/sparc/include/asm/vvar.h ret = READ_ONCE(s->seq); READ_ONCE 80 arch/x86/entry/common.c work = READ_ONCE(ti->flags); READ_ONCE 173 arch/x86/entry/common.c cached_flags = READ_ONCE(current_thread_info()->flags); READ_ONCE 191 arch/x86/entry/common.c cached_flags = READ_ONCE(ti->flags); READ_ONCE 197 arch/x86/entry/common.c cached_flags = READ_ONCE(ti->flags); READ_ONCE 256 arch/x86/entry/common.c u32 cached_flags = READ_ONCE(ti->flags); READ_ONCE 285 arch/x86/entry/common.c if (READ_ONCE(ti->flags) & _TIF_WORK_SYSCALL_ENTRY) READ_ONCE 320 arch/x86/entry/common.c if (READ_ONCE(ti->flags) & _TIF_WORK_SYSCALL_ENTRY) { READ_ONCE 957 arch/x86/events/core.c READ_ONCE(cpuc->excl_cntrs->exclusive_present)) READ_ONCE 2113 arch/x86/events/core.c if (READ_ONCE(x86_pmu.attr_rdpmc) && READ_ONCE 2407 arch/x86/events/core.c ldt = READ_ONCE(current->active_mm->context.ldt); READ_ONCE 304 arch/x86/events/intel/bts.c int state = READ_ONCE(bts->state); READ_ONCE 336 arch/x86/events/intel/bts.c int state = READ_ONCE(bts->state); READ_ONCE 361 arch/x86/events/intel/bts.c if (READ_ONCE(bts->state) != BTS_STATE_ACTIVE) READ_ONCE 464 arch/x86/events/intel/bts.c if (READ_ONCE(bts->state) == BTS_STATE_STOPPED) READ_ONCE 504 arch/x86/events/intel/pt.c if (READ_ONCE(pt->vmx_on)) READ_ONCE 513 arch/x86/events/intel/pt.c u64 ctl = READ_ONCE(event->hw.config); READ_ONCE 520 arch/x86/events/intel/pt.c if (!READ_ONCE(pt->vmx_on)) READ_ONCE 1345 arch/x86/events/intel/pt.c if (!READ_ONCE(pt->handle_nmi)) READ_ONCE 51 arch/x86/hyperv/hv_spinlock.c if (READ_ONCE(*byte) == val) READ_ONCE 31 arch/x86/include/asm/atomic.h return READ_ONCE((v)->counter); READ_ONCE 22 arch/x86/include/asm/atomic64_64.h return READ_ONCE((v)->counter); READ_ONCE 75 arch/x86/include/asm/barrier.h typeof(*p) ___p1 = READ_ONCE(*p); \ READ_ONCE 106 arch/x86/include/asm/mmu_context.h ldt = READ_ONCE(mm->context.ldt); READ_ONCE 32 arch/x86/include/asm/switch_to.h READ_ONCE(*(unsigned char *)next->thread.sp); READ_ONCE 111 arch/x86/include/asm/unwind.h val = READ_ONCE(x); \ READ_ONCE 33 arch/x86/include/asm/vdso/vsyscall.h WRITE_ONCE(vclocks_used, READ_ONCE(vclocks_used) | (1 << vclock_mode)); READ_ONCE 21 arch/x86/include/asm/vgtod.h return READ_ONCE(vclocks_used) & (1 << vclock); READ_ONCE 913 arch/x86/kernel/apic/apic.c jif_start = READ_ONCE(jiffies); READ_ONCE 936 arch/x86/kernel/apic/apic.c unsigned long jif_now = READ_ONCE(jiffies); READ_ONCE 245 arch/x86/kernel/cpu/mce/dev-mcelog.c if (READ_ONCE(mcelog.next)) READ_ONCE 879 arch/x86/kernel/cpu/resctrl/pseudo_lock.c mem_r = READ_ONCE(plr->kmem); READ_ONCE 981 arch/x86/kernel/cpu/resctrl/pseudo_lock.c line_size = READ_ONCE(plr->line_size); READ_ONCE 982 arch/x86/kernel/cpu/resctrl/pseudo_lock.c mem_r = READ_ONCE(plr->kmem); READ_ONCE 983 arch/x86/kernel/cpu/resctrl/pseudo_lock.c size = READ_ONCE(plr->size); READ_ONCE 41 arch/x86/kernel/cpu/umwait.c wrmsr(MSR_IA32_UMWAIT_CONTROL, READ_ONCE(umwait_control_cached), 0); READ_ONCE 134 arch/x86/kernel/cpu/umwait.c u32 ctrl = READ_ONCE(umwait_control_cached); READ_ONCE 153 arch/x86/kernel/cpu/umwait.c ctrl = READ_ONCE(umwait_control_cached); READ_ONCE 166 arch/x86/kernel/cpu/umwait.c u32 ctrl = READ_ONCE(umwait_control_cached); READ_ONCE 188 arch/x86/kernel/cpu/umwait.c ctrl = READ_ONCE(umwait_control_cached); READ_ONCE 150 arch/x86/kernel/espfix_64.c stack_page = READ_ONCE(espfix_pages[page]); READ_ONCE 157 arch/x86/kernel/espfix_64.c stack_page = READ_ONCE(espfix_pages[page]); READ_ONCE 1254 arch/x86/kernel/fpu/xstate.c unsigned long timestamp = READ_ONCE(task->thread.fpu.avx512_timestamp); READ_ONCE 699 arch/x86/kernel/hpet.c old.lockval = READ_ONCE(hpet.lockval); READ_ONCE 732 arch/x86/kernel/hpet.c new.lockval = READ_ONCE(hpet.lockval); READ_ONCE 598 arch/x86/kernel/kvm.c state = READ_ONCE(src->preempted); READ_ONCE 779 arch/x86/kernel/kvm.c if (READ_ONCE(*ptr) != val) READ_ONCE 109 arch/x86/kernel/nmi.c u64 whole_msecs = READ_ONCE(action->max_duration); READ_ONCE 498 arch/x86/kernel/process.c tifn = READ_ONCE(task_thread_info(next_p)->flags); READ_ONCE 499 arch/x86/kernel/process.c tifp = READ_ONCE(task_thread_info(prev_p)->flags); READ_ONCE 837 arch/x86/kernel/process.c sp = READ_ONCE(p->thread.sp); READ_ONCE 1531 arch/x86/kvm/hyperv.c return READ_ONCE(kvm->arch.hyperv.hv_hypercall) & HV_X64_MSR_HYPERCALL_ENABLE; READ_ONCE 395 arch/x86/kvm/lapic.c pir_val = READ_ONCE(pir[i]); READ_ONCE 389 arch/x86/kvm/mmu.c return READ_ONCE(nx_huge_pages); READ_ONCE 663 arch/x86/kvm/mmu.c return READ_ONCE(*sptep); READ_ONCE 5446 arch/x86/kvm/mmu.c if (!READ_ONCE(vcpu->kvm->arch.indirect_shadow_pages)) READ_ONCE 6408 arch/x86/kvm/mmu.c if (READ_ONCE(nx_huge_pages) && READ_ONCE 6434 arch/x86/kvm/mmu.c ratio = READ_ONCE(nx_huge_pages_recovery_ratio); READ_ONCE 6462 arch/x86/kvm/mmu.c return READ_ONCE(nx_huge_pages) && READ_ONCE(nx_huge_pages_recovery_ratio) READ_ONCE 159 arch/x86/kvm/page_track.c return !!READ_ONCE(slot->arch.gfn_track[mode][index]); READ_ONCE 498 arch/x86/kvm/svm.c return (READ_ONCE(*entry) & AVIC_PHYSICAL_ID_ENTRY_IS_RUNNING_MASK); READ_ONCE 2093 arch/x86/kvm/svm.c entry = READ_ONCE(*(svm->avic_physical_id_cache)); READ_ONCE 2116 arch/x86/kvm/svm.c entry = READ_ONCE(*(svm->avic_physical_id_cache)); READ_ONCE 4622 arch/x86/kvm/svm.c new_entry = READ_ONCE(*entry); READ_ONCE 7446 arch/x86/kvm/x86.c if (target && READ_ONCE(target->ready)) READ_ONCE 9993 arch/x86/kvm/x86.c if (READ_ONCE(vcpu->arch.pv.pv_unhalted)) READ_ONCE 136 arch/x86/mm/pkeys.c u32 init_pkru_value_snapshot = READ_ONCE(init_pkru_value); READ_ONCE 559 arch/x86/xen/p2m.c mid_mfn = READ_ONCE(p2m_top_mfn_p[topidx]); READ_ONCE 589 arch/x86/xen/p2m.c p2m_pfn = pte_pfn(READ_ONCE(*ptep)); READ_ONCE 52 arch/x86/xen/spinlock.c } else if (READ_ONCE(*byte) == val) { READ_ONCE 48 arch/xtensa/include/asm/atomic.h #define atomic_read(v) READ_ONCE((v)->counter) READ_ONCE 188 arch/xtensa/include/asm/cmpxchg.h oldv = READ_ONCE(*p); READ_ONCE 224 arch/xtensa/kernel/smp.c ccount = READ_ONCE(cpu_start_ccount); READ_ONCE 315 arch/xtensa/kernel/smp.c if (READ_ONCE(cpu_start_id) == -cpu) { READ_ONCE 88 arch/xtensa/platforms/iss/simdisk.c READ_ONCE(*buffer); READ_ONCE 1759 block/bio.c stamp = READ_ONCE(part->stamp); READ_ONCE 985 block/blk-iocost.c u32 active_sum = READ_ONCE(parent->child_active_sum); READ_ONCE 986 block/blk-iocost.c u32 inuse_sum = READ_ONCE(parent->child_inuse_sum); READ_ONCE 987 block/blk-iocost.c u32 active = READ_ONCE(child->active); READ_ONCE 988 block/blk-iocost.c u32 inuse = READ_ONCE(child->inuse); READ_ONCE 1289 block/blk-iocost.c u32 this_met = READ_ONCE(stat->missed[rw].nr_met); READ_ONCE 1290 block/blk-iocost.c u32 this_missed = READ_ONCE(stat->missed[rw].nr_missed); READ_ONCE 1298 block/blk-iocost.c this_rq_wait_ns = READ_ONCE(stat->rq_wait_ns); READ_ONCE 410 block/blk-iolatency.c scale_lat = READ_ONCE(lat_info->scale_lat); READ_ONCE 138 block/blk-mq-sched.c struct blk_mq_ctx *ctx = READ_ONCE(hctx->dispatch_from); READ_ONCE 878 block/blk-mq.c deadline = READ_ONCE(rq->deadline); READ_ONCE 138 block/blk-mq.h return READ_ONCE(rq->state); READ_ONCE 218 block/blk-wbt.c u64 now, issue = READ_ONCE(rwb->sync_issue); READ_ONCE 694 block/kyber-iosched.c struct blk_mq_hw_ctx *hctx = READ_ONCE(wqe->private); READ_ONCE 6022 drivers/android/binder.c int debug_id = READ_ONCE(e->debug_id_done); READ_ONCE 6041 drivers/android/binder.c seq_printf(m, debug_id && debug_id == READ_ONCE(e->debug_id_done) ? READ_ONCE 951 drivers/base/core.c drv = READ_ONCE(dev->driver); READ_ONCE 276 drivers/base/power/domain_governor.c next_hrtimer = READ_ONCE(dev->next_hrtimer); READ_ONCE 270 drivers/base/power/main.c if (READ_ONCE(link->status) != DL_STATE_DORMANT) READ_ONCE 327 drivers/base/power/main.c if (READ_ONCE(link->status) != DL_STATE_DORMANT) READ_ONCE 171 drivers/base/power/runtime.c autosuspend_delay = READ_ONCE(dev->power.autosuspend_delay); READ_ONCE 175 drivers/base/power/runtime.c expires = READ_ONCE(dev->power.last_busy); READ_ONCE 295 drivers/base/power/runtime.c READ_ONCE(link->status) == DL_STATE_SUPPLIER_UNBIND) READ_ONCE 315 drivers/base/power/runtime.c if (READ_ONCE(link->status) == DL_STATE_SUPPLIER_UNBIND) READ_ONCE 559 drivers/base/power/runtime.c u64 slack = (u64)READ_ONCE(dev->power.autosuspend_delay) * READ_ONCE 989 drivers/block/xen-blkback/blkback.c first_sect = READ_ONCE(segments[i].first_sect); READ_ONCE 990 drivers/block/xen-blkback/blkback.c last_sect = READ_ONCE(segments[i].last_sect); READ_ONCE 399 drivers/block/xen-blkback/common.h dst->operation = READ_ONCE(src->operation); READ_ONCE 447 drivers/block/xen-blkback/common.h dst->operation = READ_ONCE(src->operation); READ_ONCE 764 drivers/char/ipmi/ipmi_msghandler.c int intf_num = READ_ONCE(intf->intf_num); READ_ONCE 1957 drivers/char/ipmi/ipmi_msghandler.c chans = READ_ONCE(intf->channel_list)->c; READ_ONCE 2096 drivers/char/ipmi/ipmi_msghandler.c chans = READ_ONCE(intf->channel_list)->c; READ_ONCE 4265 drivers/char/ipmi/ipmi_msghandler.c chans = READ_ONCE(intf->channel_list)->c; READ_ONCE 4358 drivers/char/ipmi/ipmi_msghandler.c chans = READ_ONCE(intf->channel_list)->c; READ_ONCE 722 drivers/char/random.c entropy_count = orig = READ_ONCE(r->entropy_count); READ_ONCE 1300 drivers/char/random.c idx = READ_ONCE(f->reg_idx); READ_ONCE 1457 drivers/char/random.c entropy_count = orig = READ_ONCE(r->entropy_count); READ_ONCE 1683 drivers/char/random.c (previous && (caller == READ_ONCE(*previous)))) READ_ONCE 61 drivers/clocksource/bcm2835_timer.c event_handler = READ_ONCE(timer->evt.event_handler); READ_ONCE 293 drivers/cpufreq/cpufreq_governor.c lst = READ_ONCE(policy_dbs->last_sample_time); READ_ONCE 311 drivers/cpufreq/cpufreq_governor.c if (unlikely(lst != READ_ONCE(policy_dbs->last_sample_time))) { READ_ONCE 1597 drivers/cpufreq/intel_pstate.c u64 hwp_req = READ_ONCE(cpu->hwp_req_cached); READ_ONCE 1087 drivers/crypto/amcc/crypto4xx_core.c ((READ_ONCE(pd->pd_ctl.w) & READ_ONCE 205 drivers/crypto/caam/jr.c head = READ_ONCE(jrp->head); READ_ONCE 374 drivers/crypto/caam/jr.c tail = READ_ONCE(jrp->tail); READ_ONCE 510 drivers/crypto/cavium/nitrox/nitrox_reqmgr.c u64 orh = READ_ONCE(*sr->resp.orh); READ_ONCE 516 drivers/crypto/cavium/nitrox/nitrox_reqmgr.c while (READ_ONCE(*sr->resp.completion) == PENDING_SIG) { READ_ONCE 559 drivers/crypto/cavium/nitrox/nitrox_reqmgr.c READ_ONCE(*sr->resp.orh)); READ_ONCE 568 drivers/crypto/cavium/nitrox/nitrox_reqmgr.c err = READ_ONCE(*sr->resp.orh) & 0xff; READ_ONCE 186 drivers/crypto/nx/nx-842-powernv.c while (!(READ_ONCE(csb->flags) & CSB_V)) { READ_ONCE 489 drivers/devfreq/tegra30-devfreq.c cur_freq = READ_ONCE(tegra->cur_freq); READ_ONCE 485 drivers/dma-buf/st-dma-fence.c if (!READ_ONCE(cb.seen)) { READ_ONCE 568 drivers/dma/mediatek/mtk-hsdma.c desc2 = READ_ONCE(rxd->desc2); READ_ONCE 172 drivers/edac/altera_edac.c reg = READ_ONCE(ptemp[0]); READ_ONCE 173 drivers/edac/altera_edac.c read_reg = READ_ONCE(ptemp[1]); READ_ONCE 678 drivers/edac/altera_edac.c if (READ_ONCE(ptemp[i])) READ_ONCE 695 drivers/edac/altera_edac.c if (READ_ONCE(ptemp[i]) != i) READ_ONCE 724 drivers/firewire/ohci.c res_count = READ_ONCE(ctx->descriptors[i].res_count); READ_ONCE 732 drivers/firewire/ohci.c next_res_count = READ_ONCE(ctx->descriptors[next_i].res_count); READ_ONCE 748 drivers/firewire/ohci.c next_res_count = READ_ONCE(ctx->descriptors[next_i].res_count); READ_ONCE 2806 drivers/firewire/ohci.c res_count = le16_to_cpu(READ_ONCE(last->res_count)); READ_ONCE 547 drivers/firmware/efi/efi.c size = READ_ONCE(seed->size); READ_ONCE 80 drivers/firmware/qemu_fw_cfg.c u32 ctrl = be32_to_cpu(READ_ONCE(d->control)); READ_ONCE 119 drivers/firmware/qemu_fw_cfg.c if (be32_to_cpu(READ_ONCE(d->control)) & FW_CFG_DMA_CTL_ERROR) { READ_ONCE 97 drivers/firmware/tegra/ivc.c u32 tx = READ_ONCE(header->tx.count); READ_ONCE 98 drivers/firmware/tegra/ivc.c u32 rx = READ_ONCE(header->rx.count); READ_ONCE 119 drivers/firmware/tegra/ivc.c u32 tx = READ_ONCE(header->tx.count); READ_ONCE 120 drivers/firmware/tegra/ivc.c u32 rx = READ_ONCE(header->rx.count); READ_ONCE 132 drivers/firmware/tegra/ivc.c u32 tx = READ_ONCE(header->tx.count); READ_ONCE 133 drivers/firmware/tegra/ivc.c u32 rx = READ_ONCE(header->rx.count); READ_ONCE 147 drivers/firmware/tegra/ivc.c READ_ONCE(ivc->tx.channel->tx.count) + 1); READ_ONCE 158 drivers/firmware/tegra/ivc.c READ_ONCE(ivc->rx.channel->rx.count) + 1); READ_ONCE 423 drivers/firmware/tegra/ivc.c state = READ_ONCE(ivc->rx.channel->tx.state); READ_ONCE 309 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c uint64_t seq = READ_ONCE(ring->fence_drv.sync_seq); READ_ONCE 373 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c emitted += READ_ONCE(ring->fence_drv.sync_seq); READ_ONCE 821 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c pin_count = READ_ONCE(bo->pin_count); READ_ONCE 825 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c dma_buf = READ_ONCE(bo->tbo.base.dma_buf); READ_ONCE 826 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c attachment = READ_ONCE(bo->tbo.base.import_attach); READ_ONCE 566 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c wptr = READ_ONCE(*((u64 *)&adev->wb.wb[ring->wptr_offs])); READ_ONCE 635 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c wptr = READ_ONCE(*((u64 *)&adev->wb.wb[ring->wptr_offs])); READ_ONCE 864 drivers/gpu/drm/drm_dp_mst_topology.c state = READ_ONCE(txmsg->state); READ_ONCE 246 drivers/gpu/drm/drm_fb_helper.c if (READ_ONCE(fb_helper->deferred_setup)) READ_ONCE 390 drivers/gpu/drm/drm_lease.c bool universal_planes = READ_ONCE(lessor_priv->universal_planes); READ_ONCE 159 drivers/gpu/drm/drm_sysfs.c status = READ_ONCE(connector->status); READ_ONCE 172 drivers/gpu/drm/drm_sysfs.c dpms = READ_ONCE(connector->dpms); READ_ONCE 185 drivers/gpu/drm/drm_sysfs.c enabled = READ_ONCE(connector->encoder); READ_ONCE 421 drivers/gpu/drm/drm_vblank.c WARN_ON(READ_ONCE(vblank->enabled) && READ_ONCE 1249 drivers/gpu/drm/drm_vblank.c WARN_ON(!READ_ONCE(vblank->inmodeset)); READ_ONCE 1487 drivers/gpu/drm/drm_vblank.c if (!READ_ONCE(vblank->enabled)) { READ_ONCE 1631 drivers/gpu/drm/drm_vblank.c READ_ONCE(vblank->enabled)) { READ_ONCE 1678 drivers/gpu/drm/drm_vblank.c !READ_ONCE(vblank->enabled), READ_ONCE 1850 drivers/gpu/drm/drm_vblank.c vblank_enabled = dev->vblank_disable_immediate && READ_ONCE(vblank->enabled); READ_ONCE 1947 drivers/gpu/drm/drm_vblank.c if (!READ_ONCE(vblank->enabled)) { READ_ONCE 324 drivers/gpu/drm/etnaviv/etnaviv_buffer.c unsigned int new_flush_seq = READ_ONCE(gpu->mmu_context->flush_seq); READ_ONCE 1240 drivers/gpu/drm/i915/display/intel_fbc.c if (READ_ONCE(fbc->underrun_detected)) READ_ONCE 2154 drivers/gpu/drm/i915/display/intel_hdmi.c READ_ONCE(to_intel_digital_connector_state(connector->state)->force_audio) == HDMI_AUDIO_OFF_DVI; READ_ONCE 609 drivers/gpu/drm/i915/display/intel_hotplug.c enabled = READ_ONCE(dev_priv->hotplug.poll_enabled); READ_ONCE 901 drivers/gpu/drm/i915/display/intel_psr.c if (!CAN_PSR(dev_priv) || READ_ONCE(psr->dp) != intel_dp) READ_ONCE 956 drivers/gpu/drm/i915/display/intel_psr.c if (READ_ONCE(dev_priv->psr.psr2_enabled)) READ_ONCE 1100 drivers/gpu/drm/i915/display/intel_psr.c if (READ_ONCE(dev_priv->psr.irq_aux_error)) READ_ONCE 1978 drivers/gpu/drm/i915/gem/i915_gem_context.c vm = READ_ONCE(src->vm); READ_ONCE 2000 drivers/gpu/drm/i915/gem/i915_gem_context.c if (vm == READ_ONCE(src->vm)) READ_ONCE 30 drivers/gpu/drm/i915/gem/i915_gem_domain.c if (!READ_ONCE(obj->pin_global)) READ_ONCE 618 drivers/gpu/drm/i915/gem/i915_gem_domain.c if (READ_ONCE(obj->write_domain) == read_domains) { READ_ONCE 170 drivers/gpu/drm/i915/gem/i915_gem_object.h return READ_ONCE(obj->frontbuffer); READ_ONCE 256 drivers/gpu/drm/i915/gem/i915_gem_object.h return !IS_ERR_OR_NULL(READ_ONCE(obj->mm.pages)); READ_ONCE 395 drivers/gpu/drm/i915/gem/i915_gem_pages.c if (n < READ_ONCE(iter->sg_idx)) READ_ONCE 81 drivers/gpu/drm/i915/gem/i915_gem_shrinker.c if (READ_ONCE(obj->pin_global)) READ_ONCE 322 drivers/gpu/drm/i915/gem/i915_gem_shrinker.c count = READ_ONCE(i915->mm.shrink_memory) >> PAGE_SHIFT; READ_ONCE 323 drivers/gpu/drm/i915/gem/i915_gem_shrinker.c num_objects = READ_ONCE(i915->mm.shrink_count); READ_ONCE 409 drivers/gpu/drm/i915/gem/i915_gem_tiling.c READ_ONCE(obj->tiling_and_stride) & TILING_MASK; READ_ONCE 136 drivers/gpu/drm/i915/gt/intel_engine.h return READ_ONCE(*execlists->active); READ_ONCE 160 drivers/gpu/drm/i915/gt/intel_engine.h return READ_ONCE(engine->status_page.addr[reg]); READ_ONCE 449 drivers/gpu/drm/i915/gt/intel_engine.h if (READ_ONCE(engine->stats.enabled) == 0) READ_ONCE 467 drivers/gpu/drm/i915/gt/intel_engine.h if (READ_ONCE(engine->stats.enabled) == 0) READ_ONCE 1093 drivers/gpu/drm/i915/gt/intel_engine_cs.c if (!READ_ONCE(gt->awake)) READ_ONCE 1263 drivers/gpu/drm/i915/gt/intel_engine_cs.c write = READ_ONCE(*execlists->csb_write); READ_ONCE 267 drivers/gpu/drm/i915/gt/intel_hangcheck.c if (!READ_ONCE(gt->awake)) READ_ONCE 363 drivers/gpu/drm/i915/gt/intel_lrc.c next = READ_ONCE(ve->request); READ_ONCE 582 drivers/gpu/drm/i915/gt/intel_lrc.c old = READ_ONCE(ce->inflight); READ_ONCE 597 drivers/gpu/drm/i915/gt/intel_lrc.c struct i915_request *next = READ_ONCE(ve->request); READ_ONCE 636 drivers/gpu/drm/i915/gt/intel_lrc.c old = READ_ONCE(ce->inflight); READ_ONCE 928 drivers/gpu/drm/i915/gt/intel_lrc.c struct i915_request * const *last = READ_ONCE(execlists->active); READ_ONCE 1072 drivers/gpu/drm/i915/gt/intel_lrc.c struct i915_request *rq = READ_ONCE(ve->request); READ_ONCE 1220 drivers/gpu/drm/i915/gt/intel_lrc.c GEM_BUG_ON(READ_ONCE(ve->context.inflight)); READ_ONCE 1506 drivers/gpu/drm/i915/gt/intel_lrc.c tail = READ_ONCE(*execlists->csb_write); READ_ONCE 1638 drivers/gpu/drm/i915/gt/intel_lrc.c if (!READ_ONCE(engine->execlists.pending[0])) { READ_ONCE 3530 drivers/gpu/drm/i915/gt/intel_lrc.c rq = READ_ONCE(ve->request); READ_ONCE 3564 drivers/gpu/drm/i915/gt/intel_lrc.c for (n = 0; READ_ONCE(ve->request) && n < ve->num_siblings; n++) { READ_ONCE 3698 drivers/gpu/drm/i915/gt/intel_lrc.c exec = READ_ONCE(rq->execution_mask); READ_ONCE 3976 drivers/gpu/drm/i915/gt/intel_lrc.c struct i915_request *rq = READ_ONCE(ve->request); READ_ONCE 632 drivers/gpu/drm/i915/gt/intel_reset.c vma = READ_ONCE(gt->ggtt->fence_regs[i].vma); READ_ONCE 272 drivers/gpu/drm/i915/gt/selftest_hangcheck.c return READ_ONCE(h->seqno[rq->fence.context % (PAGE_SIZE/sizeof(u32))]); READ_ONCE 412 drivers/gpu/drm/i915/gt/selftest_lrc.c if (wait_for(READ_ONCE(*map), 10)) { READ_ONCE 459 drivers/gpu/drm/i915/gt/selftest_lrc.c GEM_BUG_ON(READ_ONCE(*map)); READ_ONCE 384 drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c #define done (READ_ONCE(desc->fence) == fence) READ_ONCE 430 drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c #define done INTEL_GUC_MSG_IS_RESPONSE(READ_ONCE(req->status)) READ_ONCE 416 drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c wq_off = READ_ONCE(desc->tail); READ_ONCE 417 drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c GEM_BUG_ON(CIRC_SPACE(wq_off, READ_ONCE(desc->head), READ_ONCE 456 drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c cookie = READ_ONCE(db->cookie); READ_ONCE 69 drivers/gpu/drm/i915/gt/uc/selftest_guc.c err = wait_for(READ_ONCE(desc->head) == READ_ONCE(desc->tail), 10); READ_ONCE 58 drivers/gpu/drm/i915/i915_active.c return (struct intel_engine_cs *)READ_ONCE(node->base.link.prev); READ_ONCE 195 drivers/gpu/drm/i915/i915_active.c node = READ_ONCE(ref->cache); READ_ONCE 97 drivers/gpu/drm/i915/i915_debugfs.c return READ_ONCE(obj->userfault_count) ? 'g' : ' '; READ_ONCE 1151 drivers/gpu/drm/i915/i915_debugfs.c READ_ONCE(fw_domain->wake_count)); READ_ONCE 1735 drivers/gpu/drm/i915/i915_debugfs.c seq_printf(m, "Interactive? %d\n", READ_ONCE(rps->power.interactive)); READ_ONCE 2280 drivers/gpu/drm/i915/i915_debugfs.c *val = READ_ONCE(dev_priv->psr.debug); READ_ONCE 485 drivers/gpu/drm/i915/i915_gem_fence_reg.c struct i915_vma *vma = READ_ONCE(reg->vma); READ_ONCE 798 drivers/gpu/drm/i915/i915_gpu_error.c if (READ_ONCE(error->sgl)) READ_ONCE 843 drivers/gpu/drm/i915/i915_gpu_error.c sg = READ_ONCE(error->fit); READ_ONCE 1681 drivers/gpu/drm/i915/i915_gpu_error.c error = READ_ONCE(i915->gpu_error.first_error); READ_ONCE 1745 drivers/gpu/drm/i915/i915_gpu_error.c if (READ_ONCE(i915->gpu_error.first_error)) READ_ONCE 366 drivers/gpu/drm/i915/i915_irq.c if (READ_ONCE(rps->interrupts_enabled)) READ_ONCE 393 drivers/gpu/drm/i915/i915_irq.c if (!READ_ONCE(rps->interrupts_enabled)) READ_ONCE 3090 drivers/gpu/drm/i915/i915_irq.c READ_ONCE(dev_priv->i945gm_vblank.enabled) ? READ_ONCE 262 drivers/gpu/drm/i915/i915_pmu.c if (!READ_ONCE(pmu->timer_enabled)) READ_ONCE 172 drivers/gpu/drm/i915/i915_request.c file_priv = READ_ONCE(request->file_priv); READ_ONCE 207 drivers/gpu/drm/i915/i915_request.c locked = READ_ONCE(rq->engine); READ_ONCE 209 drivers/gpu/drm/i915/i915_request.c while (unlikely(locked != (engine = READ_ONCE(rq->engine)))) { READ_ONCE 579 drivers/gpu/drm/i915/i915_request.c if (!(READ_ONCE(rq->sched.attr.priority) & I915_PRIORITY_NOSEMAPHORE)) { READ_ONCE 340 drivers/gpu/drm/i915/i915_request.h return READ_ONCE(*rq->hwsp_seqno); READ_ONCE 164 drivers/gpu/drm/i915/i915_scheduler.c while (locked != (engine = READ_ONCE(rq->engine))) { READ_ONCE 244 drivers/gpu/drm/i915/i915_scheduler.c if (prio <= READ_ONCE(node->attr.priority)) READ_ONCE 290 drivers/gpu/drm/i915/i915_scheduler.c if (prio > READ_ONCE(p->signaler->attr.priority)) READ_ONCE 380 drivers/gpu/drm/i915/i915_scheduler.c if (READ_ONCE(rq->sched.attr.priority) & bump) READ_ONCE 6692 drivers/gpu/drm/i915/intel_pm.c if (!rps->power.interactive++ && READ_ONCE(i915->gt.awake)) READ_ONCE 6898 drivers/gpu/drm/i915/intel_pm.c if (READ_ONCE(rps->cur_freq) < rps->boost_freq) READ_ONCE 150 drivers/gpu/drm/i915/intel_runtime_pm.c stack = READ_ONCE(rpm->debug.last_release); READ_ONCE 790 drivers/gpu/drm/i915/intel_uncore.c unsigned int actual = READ_ONCE(domain->wake_count); READ_ONCE 163 drivers/gpu/drm/i915/intel_wakeref.h return READ_ONCE(wf->wakeref); READ_ONCE 548 drivers/gpu/drm/i915/selftests/i915_sw_fence.c if (READ_ONCE(ipc.value)) { READ_ONCE 556 drivers/gpu/drm/i915/selftests/i915_sw_fence.c if (!READ_ONCE(ipc.value)) { READ_ONCE 175 drivers/gpu/drm/i915/selftests/igt_spinner.c return READ_ONCE(*seqno); READ_ONCE 448 drivers/gpu/drm/nouveau/nouveau_fbcon.c int state = READ_ONCE(drm->fbcon_new_state); READ_ONCE 460 drivers/gpu/drm/radeon/radeon_gem.c cur_placement = READ_ONCE(robj->tbo.mem.mem_type); READ_ONCE 490 drivers/gpu/drm/radeon/radeon_gem.c cur_placement = READ_ONCE(robj->tbo.mem.mem_type); READ_ONCE 119 drivers/gpu/drm/scheduler/sched_entity.c if (READ_ONCE(entity->dependency)) READ_ONCE 492 drivers/gpu/drm/scheduler/sched_entity.c fence = READ_ONCE(entity->last_scheduled); READ_ONCE 1235 drivers/gpu/drm/vmwgfx/vmwgfx_drv.c if (READ_ONCE(dev_priv->suspend_locked)) { READ_ONCE 1475 drivers/gpu/drm/vmwgfx/vmwgfx_drv.h return READ_ONCE(*addr); READ_ONCE 190 drivers/gpu/drm/vmwgfx/vmwgfx_fb.c if (!READ_ONCE(par->dirty.active)) READ_ONCE 86 drivers/gpu/drm/vmwgfx/vmwgfx_irq.c masked_status = status & READ_ONCE(dev_priv->irq_mask); READ_ONCE 795 drivers/hv/channel_mgmt.c message_type = READ_ONCE(msg->header.message_type); READ_ONCE 1071 drivers/hv/channel_mgmt.c while (READ_ONCE(channel->probe_done) == false) { READ_ONCE 1134 drivers/hv/channel_mgmt.c while (!READ_ONCE(channel->probe_done) || !READ_ONCE(channel->rescind)) READ_ONCE 371 drivers/hv/connection.c callback_fn = READ_ONCE(channel->onchannel_callback); READ_ONCE 55 drivers/hv/ring_buffer.c if (READ_ONCE(rbi->ring_buffer->interrupt_mask)) READ_ONCE 64 drivers/hv/ring_buffer.c if (old_write == READ_ONCE(rbi->ring_buffer->read_index)) { READ_ONCE 146 drivers/hv/ring_buffer.c read_loc = READ_ONCE(rbi->ring_buffer->read_index); READ_ONCE 147 drivers/hv/ring_buffer.c write_loc = READ_ONCE(rbi->ring_buffer->write_index); READ_ONCE 381 drivers/hv/ring_buffer.c u32 write_loc = READ_ONCE(rbi->ring_buffer->write_index); READ_ONCE 503 drivers/hv/ring_buffer.c pending_sz = READ_ONCE(rbi->ring_buffer->pending_send_sz); READ_ONCE 1162 drivers/hv/vmbus_drv.c callback_fn = READ_ONCE(channel->onchannel_callback); READ_ONCE 123 drivers/hwmon/xgene-hwmon.c val = le16_to_cpu(READ_ONCE(*addr)); READ_ONCE 151 drivers/hwmon/xgene-hwmon.c val = le16_to_cpu(READ_ONCE(generic_comm_base->status)); READ_ONCE 1139 drivers/hwtracing/coresight/coresight-tmc-etr.c sysfs_buf = READ_ONCE(drvdata->sysfs_buf); READ_ONCE 1171 drivers/hwtracing/coresight/coresight-tmc-etr.c sysfs_buf = READ_ONCE(drvdata->sysfs_buf); READ_ONCE 131 drivers/i2c/busses/i2c-xgene-slimpro.c val = le16_to_cpu(READ_ONCE(*addr)); READ_ONCE 191 drivers/i2c/busses/i2c-xgene-slimpro.c status = le16_to_cpu(READ_ONCE(generic_comm_base->status)); READ_ONCE 261 drivers/i2c/muxes/i2c-mux-pca954x.c idle_state = READ_ONCE(data->idle_state); READ_ONCE 286 drivers/i2c/muxes/i2c-mux-pca954x.c return sprintf(buf, "%d\n", READ_ONCE(data->idle_state)); READ_ONCE 504 drivers/infiniband/core/addr.c struct net_device *ndev = READ_ONCE(dst->dev); READ_ONCE 1294 drivers/infiniband/core/cache.c (ndev && ((READ_ONCE(ndev->flags) & IFF_UP) == 0))) READ_ONCE 93 drivers/infiniband/core/netlink.c cb_table = READ_ONCE(rdma_nl_types[type].cb_table); READ_ONCE 104 drivers/infiniband/core/netlink.c cb_table = READ_ONCE(rdma_nl_types[type].cb_table); READ_ONCE 115 drivers/infiniband/core/netlink.c WARN_ON(READ_ONCE(rdma_nl_types[index].cb_table))) READ_ONCE 744 drivers/infiniband/core/security.c if (!READ_ONCE(map->agent.smp_allowed)) READ_ONCE 106 drivers/infiniband/hw/efa/efa_com.c if (READ_ONCE(read_resp->req_id) == mmio_read->seq_num) READ_ONCE 466 drivers/infiniband/hw/efa/efa_com.c while ((READ_ONCE(cqe->acq_common_descriptor.flags) & READ_ONCE 863 drivers/infiniband/hw/efa/efa_com.c while ((READ_ONCE(aenq_common->flags) & READ_ONCE 1661 drivers/infiniband/hw/hfi1/file_ops.c !(READ_ONCE(dd->flags) & HFI1_FROZEN), READ_ONCE 1475 drivers/infiniband/hw/hfi1/pio.c sc->alloc_free = READ_ONCE(sc->free); READ_ONCE 1482 drivers/infiniband/hw/hfi1/pio.c sc->alloc_free = READ_ONCE(sc->free); READ_ONCE 1722 drivers/infiniband/hw/hfi1/pio.c head = READ_ONCE(sc->sr_head); /* snapshot the head */ READ_ONCE 482 drivers/infiniband/hw/hfi1/rc.c if (qp->s_last == READ_ONCE(qp->s_head)) READ_ONCE 526 drivers/infiniband/hw/hfi1/rc.c if (qp->s_tail == READ_ONCE(qp->s_head)) { READ_ONCE 2370 drivers/infiniband/hw/hfi1/rc.c if (cmp_psn(psn, READ_ONCE(qp->s_next_psn)) >= 0) READ_ONCE 1728 drivers/infiniband/hw/hfi1/sdma.c swtail = READ_ONCE(sde->descq_tail) & sde->sdma_mask; READ_ONCE 1873 drivers/infiniband/hw/hfi1/sdma.c swtail = READ_ONCE(sde->descq_tail) & sde->sdma_mask; READ_ONCE 2221 drivers/infiniband/hw/hfi1/sdma.c tail = READ_ONCE(sde->descq_tail) & sde->sdma_mask; READ_ONCE 3300 drivers/infiniband/hw/hfi1/sdma.c nr = ffz(READ_ONCE(sde->ahg_bits)); READ_ONCE 441 drivers/infiniband/hw/hfi1/sdma.h READ_ONCE(sde->descq_head)) - 1; READ_ONCE 82 drivers/infiniband/hw/hfi1/uc.c if (qp->s_last == READ_ONCE(qp->s_head)) READ_ONCE 121 drivers/infiniband/hw/hfi1/uc.c if (qp->s_cur == READ_ONCE(qp->s_head)) { READ_ONCE 494 drivers/infiniband/hw/hfi1/ud.c if (qp->s_last == READ_ONCE(qp->s_head)) READ_ONCE 507 drivers/infiniband/hw/hfi1/ud.c if (qp->s_cur == READ_ONCE(qp->s_head)) READ_ONCE 793 drivers/infiniband/hw/hfi1/user_sdma.c if (READ_ONCE(req->has_error)) READ_ONCE 817 drivers/infiniband/hw/hfi1/user_sdma.c if (READ_ONCE(req->has_error)) READ_ONCE 843 drivers/infiniband/hw/hfi1/user_sdma.c if (READ_ONCE(iovec->offset) == iovec->iov.iov_len) { READ_ONCE 176 drivers/infiniband/hw/hfi1/vnic_sdma.c if (unlikely(READ_ONCE(vnic_sdma->state) != HFI1_VNIC_SDMA_Q_ACTIVE)) READ_ONCE 274 drivers/infiniband/hw/hfi1/vnic_sdma.c return (READ_ONCE(vnic_sdma->state) == HFI1_VNIC_SDMA_Q_ACTIVE); READ_ONCE 653 drivers/infiniband/hw/mlx5/odp.c current_seq = READ_ONCE(odp->notifiers_seq); READ_ONCE 990 drivers/infiniband/hw/qedr/verbs.c while (oparams.num_cq_notif != READ_ONCE(cq->cnq_notif) && iter) { READ_ONCE 996 drivers/infiniband/hw/qedr/verbs.c while (oparams.num_cq_notif != READ_ONCE(cq->cnq_notif) && iter) { READ_ONCE 244 drivers/infiniband/hw/qib/qib_rc.c if (qp->s_last == READ_ONCE(qp->s_head)) READ_ONCE 290 drivers/infiniband/hw/qib/qib_rc.c if (qp->s_tail == READ_ONCE(qp->s_head)) READ_ONCE 1312 drivers/infiniband/hw/qib/qib_rc.c if (qib_cmp24(psn, READ_ONCE(qp->s_next_psn)) >= 0) READ_ONCE 63 drivers/infiniband/hw/qib/qib_uc.c if (qp->s_last == READ_ONCE(qp->s_head)) READ_ONCE 92 drivers/infiniband/hw/qib/qib_uc.c if (qp->s_cur == READ_ONCE(qp->s_head)) READ_ONCE 250 drivers/infiniband/hw/qib/qib_ud.c if (qp->s_last == READ_ONCE(qp->s_head)) READ_ONCE 263 drivers/infiniband/hw/qib/qib_ud.c if (qp->s_cur == READ_ONCE(qp->s_head)) READ_ONCE 955 drivers/infiniband/sw/rdmavt/mr.c if (!READ_ONCE(mr->lkey_published)) READ_ONCE 1064 drivers/infiniband/sw/rdmavt/mr.c if (!READ_ONCE(mr->lkey_published)) READ_ONCE 1354 drivers/infiniband/sw/rdmavt/qp.c if (READ_ONCE(qp->s_last) != qp->s_head) READ_ONCE 1850 drivers/infiniband/sw/rdmavt/qp.c if (next == READ_ONCE(wq->tail)) { READ_ONCE 2208 drivers/infiniband/sw/rdmavt/qp.c call_send = qp->s_head == READ_ONCE(qp->s_last) && !wr->next; READ_ONCE 2265 drivers/infiniband/sw/rdmavt/qp.c if (next == READ_ONCE(wq->tail)) { READ_ONCE 3002 drivers/infiniband/sw/rdmavt/qp.c if (sqp->s_last == READ_ONCE(sqp->s_head)) READ_ONCE 110 drivers/infiniband/sw/rdmavt/rc.c credits = READ_ONCE(qp->r_rq.kwq->count); READ_ONCE 117 drivers/infiniband/sw/rdmavt/rc.c head = READ_ONCE(qp->r_rq.kwq->head); READ_ONCE 118 drivers/infiniband/sw/rdmavt/rc.c tail = READ_ONCE(qp->r_rq.kwq->tail); READ_ONCE 643 drivers/infiniband/sw/siw/siw.h return READ_ONCE(sqe->flags) == 0; READ_ONCE 650 drivers/infiniband/sw/siw/siw.h if (READ_ONCE(sqe->flags) & SIW_WQE_VALID) READ_ONCE 670 drivers/infiniband/sw/siw/siw.h if (orq_e && READ_ONCE(orq_e->flags) == 0) READ_ONCE 685 drivers/infiniband/sw/siw/siw.h if (READ_ONCE(irq_e->flags) == 0) { READ_ONCE 56 drivers/infiniband/sw/siw/siw_cq.c if (READ_ONCE(cqe->flags) & SIW_WQE_VALID) { READ_ONCE 1029 drivers/infiniband/sw/siw/siw_qp.c cq_notify = READ_ONCE(cq->notify->flags); READ_ONCE 1064 drivers/infiniband/sw/siw/siw_qp.c if (!READ_ONCE(cqe->flags)) { READ_ONCE 1121 drivers/infiniband/sw/siw/siw_qp.c if (!READ_ONCE(cqe->flags)) { READ_ONCE 1186 drivers/infiniband/sw/siw/siw_qp.c if (!READ_ONCE(sqe->flags)) READ_ONCE 1222 drivers/infiniband/sw/siw/siw_qp.c if (!READ_ONCE(sqe->flags)) READ_ONCE 1287 drivers/infiniband/sw/siw/siw_qp.c if (!READ_ONCE(rqe->flags)) READ_ONCE 746 drivers/infiniband/sw/siw/siw_qp_rx.c if (READ_ONCE(orqe->flags) & SIW_WQE_VALID) { READ_ONCE 72 drivers/input/misc/pwm-beeper.c unsigned long period = READ_ONCE(beeper->period); READ_ONCE 233 drivers/input/misc/regulator-haptic.c magnitude = READ_ONCE(haptic->magnitude); READ_ONCE 1170 drivers/iommu/arm-smmu-v3.c llq->val = READ_ONCE(cmdq->q.llq.val); READ_ONCE 1176 drivers/iommu/arm-smmu-v3.c llq->val = READ_ONCE(smmu->cmdq.q.llq.val); READ_ONCE 1223 drivers/iommu/arm-smmu-v3.c llq->val = READ_ONCE(smmu->cmdq.q.llq.val); READ_ONCE 1322 drivers/iommu/arm-smmu-v3.c llq.val = READ_ONCE(cmdq->q.llq.val); READ_ONCE 289 drivers/iommu/intel-pasid.c old = READ_ONCE(*ptr); READ_ONCE 309 drivers/iommu/intel-pasid.c return (u16)(READ_ONCE(pe->val[1]) & GENMASK_ULL(15, 0)); READ_ONCE 59 drivers/iommu/intel-pasid.h return READ_ONCE(pde->val) & PASID_PTE_PRESENT; READ_ONCE 69 drivers/iommu/intel-pasid.h return phys_to_virt(READ_ONCE(pde->val) & PDE_PFN_MASK); READ_ONCE 75 drivers/iommu/intel-pasid.h return READ_ONCE(pte->val[0]) & PASID_PTE_PRESENT; READ_ONCE 494 drivers/iommu/io-pgtable-arm-v7s.c pte = READ_ONCE(*ptep); READ_ONCE 654 drivers/iommu/io-pgtable-arm-v7s.c pte[i] = READ_ONCE(ptep[i]); READ_ONCE 738 drivers/iommu/io-pgtable-arm-v7s.c pte = READ_ONCE(*ptep); READ_ONCE 403 drivers/iommu/io-pgtable-arm.c pte = READ_ONCE(*ptep); READ_ONCE 611 drivers/iommu/io-pgtable-arm.c pte = READ_ONCE(*ptep); READ_ONCE 678 drivers/iommu/io-pgtable-arm.c pte = READ_ONCE(*ptep); READ_ONCE 174 drivers/lightnvm/pblk-rb.c flags = READ_ONCE(w_ctx->flags); READ_ONCE 194 drivers/lightnvm/pblk-rb.c unsigned int mem = READ_ONCE(rb->mem); READ_ONCE 195 drivers/lightnvm/pblk-rb.c unsigned int sync = READ_ONCE(rb->sync); READ_ONCE 212 drivers/lightnvm/pblk-rb.c unsigned int mem = READ_ONCE(rb->mem); READ_ONCE 213 drivers/lightnvm/pblk-rb.c unsigned int subm = READ_ONCE(rb->subm); READ_ONCE 220 drivers/lightnvm/pblk-rb.c unsigned int mem = READ_ONCE(rb->mem); READ_ONCE 221 drivers/lightnvm/pblk-rb.c unsigned int sync = READ_ONCE(rb->sync); READ_ONCE 230 drivers/lightnvm/pblk-rb.c subm = READ_ONCE(rb->subm); READ_ONCE 251 drivers/lightnvm/pblk-rb.c flags = READ_ONCE(entry->w_ctx.flags); READ_ONCE 345 drivers/lightnvm/pblk-rb.c flags = READ_ONCE(entry->w_ctx.flags); READ_ONCE 369 drivers/lightnvm/pblk-rb.c flags = READ_ONCE(entry->w_ctx.flags); READ_ONCE 393 drivers/lightnvm/pblk-rb.c sync = READ_ONCE(rb->sync); READ_ONCE 425 drivers/lightnvm/pblk-rb.c sync = READ_ONCE(rb->sync); READ_ONCE 426 drivers/lightnvm/pblk-rb.c mem = READ_ONCE(rb->mem); READ_ONCE 455 drivers/lightnvm/pblk-rb.c unsigned int mem = READ_ONCE(rb->mem); READ_ONCE 583 drivers/lightnvm/pblk-rb.c flags = READ_ONCE(entry->w_ctx.flags); READ_ONCE 663 drivers/lightnvm/pblk-rb.c flags = READ_ONCE(w_ctx->flags); READ_ONCE 718 drivers/lightnvm/pblk-rb.c sync = READ_ONCE(rb->sync); READ_ONCE 719 drivers/lightnvm/pblk-rb.c flush_point = READ_ONCE(rb->flush_point); READ_ONCE 754 drivers/lightnvm/pblk-rb.c subm = READ_ONCE(rb->subm); READ_ONCE 64 drivers/lightnvm/pblk-rl.c rb_user_active = READ_ONCE(rl->rb_user_active); READ_ONCE 80 drivers/lightnvm/pblk-sysfs.c READ_ONCE(pblk->rl.rb_user_active)); READ_ONCE 36 drivers/lightnvm/pblk-write.c flags = READ_ONCE(w_ctx->flags); READ_ONCE 171 drivers/lightnvm/pblk-write.c flags = READ_ONCE(w_ctx->flags); READ_ONCE 112 drivers/md/bcache/closure.c p = READ_ONCE(s->task); READ_ONCE 342 drivers/md/dm-bufio.c dm_bufio_cache_size_latch = READ_ONCE(dm_bufio_cache_size); READ_ONCE 1544 drivers/md/dm-bufio.c unsigned long retain_bytes = READ_ONCE(dm_bufio_retain_bytes); READ_ONCE 1595 drivers/md/dm-bufio.c unsigned long count = READ_ONCE(c->n_buffers[LIST_CLEAN]) + READ_ONCE 1596 drivers/md/dm-bufio.c READ_ONCE(c->n_buffers[LIST_DIRTY]); READ_ONCE 1777 drivers/md/dm-bufio.c unsigned max_age = READ_ONCE(dm_bufio_max_age); READ_ONCE 161 drivers/md/dm-clone-target.c return READ_ONCE(clone->mode); READ_ONCE 984 drivers/md/dm-clone-target.c unsigned int max_batch_size = READ_ONCE(clone->hydration_batch_size); READ_ONCE 1103 drivers/md/dm-clone-target.c if (current_volume > READ_ONCE(clone->hydration_threshold)) READ_ONCE 1420 drivers/md/dm-clone-target.c READ_ONCE(clone->hydration_threshold), READ_ONCE 1421 drivers/md/dm-clone-target.c READ_ONCE(clone->hydration_batch_size)); READ_ONCE 381 drivers/md/dm-integrity.c return READ_ONCE(ic->failed); READ_ONCE 1826 drivers/md/dm-integrity.c if (READ_ONCE(ic->free_sectors) <= ic->free_sectors_threshold) { READ_ONCE 2120 drivers/md/dm-integrity.c if (READ_ONCE(ic->free_sectors) <= ic->free_sectors_threshold) READ_ONCE 131 drivers/md/dm-kcopyd.c throttle = READ_ONCE(t->throttle); READ_ONCE 181 drivers/md/dm-kcopyd.c if (likely(READ_ONCE(t->throttle) >= 100)) READ_ONCE 357 drivers/md/dm-mpath.c if (unlikely(READ_ONCE(m->current_pg) != pg)) { READ_ONCE 381 drivers/md/dm-mpath.c if (READ_ONCE(m->next_pg)) { READ_ONCE 397 drivers/md/dm-mpath.c pg = READ_ONCE(m->current_pg); READ_ONCE 470 drivers/md/dm-mpath.c unsigned long flags = READ_ONCE(m->flags); READ_ONCE 476 drivers/md/dm-mpath.c unsigned long flags = READ_ONCE(m->flags); READ_ONCE 496 drivers/md/dm-mpath.c pgpath = READ_ONCE(m->current_pgpath); READ_ONCE 578 drivers/md/dm-mpath.c pgpath = READ_ONCE(m->current_pgpath); READ_ONCE 1858 drivers/md/dm-mpath.c current_pgpath = READ_ONCE(m->current_pgpath); READ_ONCE 1879 drivers/md/dm-mpath.c if (!READ_ONCE(m->current_pg)) { READ_ONCE 1948 drivers/md/dm-mpath.c pg = READ_ONCE(m->current_pg); READ_ONCE 1949 drivers/md/dm-mpath.c next_pg = READ_ONCE(m->next_pg); READ_ONCE 1950 drivers/md/dm-mpath.c if (unlikely(!READ_ONCE(m->current_pgpath) && next_pg)) READ_ONCE 644 drivers/md/dm-stats.c (bi_sector == (READ_ONCE(last->last_sector) && READ_ONCE 646 drivers/md/dm-stats.c (READ_ONCE(last->last_rw) == WRITE)) READ_ONCE 698 drivers/md/dm-stats.c shared->tmp.sectors[READ] += READ_ONCE(p->sectors[READ]); READ_ONCE 699 drivers/md/dm-stats.c shared->tmp.sectors[WRITE] += READ_ONCE(p->sectors[WRITE]); READ_ONCE 700 drivers/md/dm-stats.c shared->tmp.ios[READ] += READ_ONCE(p->ios[READ]); READ_ONCE 701 drivers/md/dm-stats.c shared->tmp.ios[WRITE] += READ_ONCE(p->ios[WRITE]); READ_ONCE 702 drivers/md/dm-stats.c shared->tmp.merges[READ] += READ_ONCE(p->merges[READ]); READ_ONCE 703 drivers/md/dm-stats.c shared->tmp.merges[WRITE] += READ_ONCE(p->merges[WRITE]); READ_ONCE 704 drivers/md/dm-stats.c shared->tmp.ticks[READ] += READ_ONCE(p->ticks[READ]); READ_ONCE 705 drivers/md/dm-stats.c shared->tmp.ticks[WRITE] += READ_ONCE(p->ticks[WRITE]); READ_ONCE 706 drivers/md/dm-stats.c shared->tmp.io_ticks[READ] += READ_ONCE(p->io_ticks[READ]); READ_ONCE 707 drivers/md/dm-stats.c shared->tmp.io_ticks[WRITE] += READ_ONCE(p->io_ticks[WRITE]); READ_ONCE 708 drivers/md/dm-stats.c shared->tmp.io_ticks_total += READ_ONCE(p->io_ticks_total); READ_ONCE 709 drivers/md/dm-stats.c shared->tmp.time_in_queue += READ_ONCE(p->time_in_queue); READ_ONCE 713 drivers/md/dm-stats.c shared->tmp.histogram[i] += READ_ONCE(p->histogram[i]); READ_ONCE 147 drivers/md/dm-switch.c return (READ_ONCE(sctx->region_table[region_index]) >> bit) & READ_ONCE 2541 drivers/md/dm-thin.c unsigned long no_space_timeout = READ_ONCE(no_space_timeout_secs) * HZ; READ_ONCE 590 drivers/md/dm-verity-target.c unsigned cluster = READ_ONCE(dm_verity_prefetch_cluster); READ_ONCE 411 drivers/md/dm-writecache.c #define writecache_has_error(wc) (unlikely(READ_ONCE((wc)->error))) READ_ONCE 1491 drivers/md/dm-writecache.c if (READ_ONCE(wc->writeback_size) - wbl->size >= wc->max_writeback_jobs) { READ_ONCE 169 drivers/md/dm.c int param = READ_ONCE(*module_param); READ_ONCE 191 drivers/md/dm.c unsigned param = READ_ONCE(*module_param); READ_ONCE 2806 drivers/md/md.c unsigned long flags = READ_ONCE(rdev->flags); READ_ONCE 972 drivers/md/raid1.c if (!READ_ONCE(conf->array_frozen) && READ_ONCE 1014 drivers/md/raid1.c if (!READ_ONCE(conf->array_frozen)) READ_ONCE 6098 drivers/md/raid5.c struct md_rdev *rdev = READ_ONCE(conf->disks[i].rdev); READ_ONCE 74 drivers/media/dvb-core/dvb_ringbuffer.c free = READ_ONCE(rbuf->pread) - rbuf->pwrite; READ_ONCE 182 drivers/misc/mei/dma-ring.c rd_idx = READ_ONCE(ctrl->dbuf_rd_idx) & (dbuf_depth - 1); READ_ONCE 226 drivers/misc/mei/dma-ring.c rd_idx = READ_ONCE(ctrl->hbuf_rd_idx); READ_ONCE 227 drivers/misc/mei/dma-ring.c wr_idx = READ_ONCE(ctrl->hbuf_wr_idx); READ_ONCE 255 drivers/misc/mei/dma-ring.c wr_idx = READ_ONCE(ctrl->hbuf_wr_idx) & (hbuf_depth - 1); READ_ONCE 271 drivers/misc/mic/scif/scif_rma_list.c if (READ_ONCE(ep->rma_info.async_list_del)) READ_ONCE 967 drivers/misc/vmw_balloon.c target = READ_ONCE(b->target); READ_ONCE 983 drivers/misc/vmw_balloon.c if (target > size && time_before(jiffies, READ_ONCE(b->shrink_timeout))) READ_ONCE 1499 drivers/misc/vmw_balloon.c atomic64_read(&b->size), READ_ONCE(b->target)); READ_ONCE 1678 drivers/misc/vmw_balloon.c seq_printf(f, "%-22s: %16lu\n", "target", READ_ONCE(b->target)); READ_ONCE 368 drivers/mmc/core/core.c struct mmc_request *ongoing_mrq = READ_ONCE(host->ongoing_mrq); READ_ONCE 1364 drivers/net/bonding/bond_alb.c count = slaves ? READ_ONCE(slaves->count) : 0; READ_ONCE 1498 drivers/net/bonding/bond_alb.c count = slaves ? READ_ONCE(slaves->count) : 0; READ_ONCE 1192 drivers/net/bonding/bond_main.c recv_probe = READ_ONCE(bond->recv_probe); READ_ONCE 3941 drivers/net/bonding/bond_main.c slave_cnt = READ_ONCE(bond->slave_cnt); READ_ONCE 4106 drivers/net/bonding/bond_main.c count = slaves ? READ_ONCE(slaves->count) : 0; READ_ONCE 661 drivers/net/can/usb/ucan.c dlc = READ_ONCE(context->dlc); READ_ONCE 497 drivers/net/ethernet/amazon/ena/ena_com.c while ((READ_ONCE(cqe->acq_common_descriptor.flags) & READ_ONCE 832 drivers/net/ethernet/amazon/ena/ena_com.c if (READ_ONCE(read_resp->req_id) == mmio_read->seq_num) READ_ONCE 2027 drivers/net/ethernet/amazon/ena/ena_com.c while ((READ_ONCE(aenq_common->flags) & READ_ONCE 48 drivers/net/ethernet/amazon/ena/ena_eth_com.c desc_phase = (READ_ONCE(cdesc->status) & ENA_ETH_IO_RX_CDESC_BASE_PHASE_MASK) >> READ_ONCE 265 drivers/net/ethernet/amazon/ena/ena_eth_com.c last = (READ_ONCE(cdesc->status) & ENA_ETH_IO_RX_CDESC_BASE_LAST_MASK) >> READ_ONCE 260 drivers/net/ethernet/amazon/ena/ena_eth_com.h cdesc_phase = READ_ONCE(cdesc->flags) & ENA_ETH_IO_TX_CDESC_PHASE_MASK; READ_ONCE 266 drivers/net/ethernet/amazon/ena/ena_eth_com.h *req_id = READ_ONCE(cdesc->req_id); READ_ONCE 260 drivers/net/ethernet/broadcom/bnx2.c diff = READ_ONCE(txr->tx_prod) - READ_ONCE(txr->tx_cons); READ_ONCE 2840 drivers/net/ethernet/broadcom/bnx2.c cons = READ_ONCE(*bnapi->hw_tx_cons_ptr); READ_ONCE 3140 drivers/net/ethernet/broadcom/bnx2.c cons = READ_ONCE(*bnapi->hw_rx_cons_ptr); READ_ONCE 114 drivers/net/ethernet/broadcom/bnxt/bnxt_xdp.c struct bpf_prog *xdp_prog = READ_ONCE(rxr->xdp_prog); READ_ONCE 218 drivers/net/ethernet/broadcom/bnxt/bnxt_xdp.c struct bpf_prog *xdp_prog = READ_ONCE(bp->xdp_prog); READ_ONCE 299 drivers/net/ethernet/cadence/macb_ptp.c unsigned long tail = READ_ONCE(queue->tx_ts_tail); READ_ONCE 370 drivers/net/ethernet/cavium/liquidio/cn23xx_pf_device.c while ((READ_ONCE(reg_val) & CN23XX_PKT_INPUT_CTL_RST) && READ_ONCE 371 drivers/net/ethernet/cavium/liquidio/cn23xx_pf_device.c !(READ_ONCE(reg_val) & CN23XX_PKT_INPUT_CTL_QUIET) && READ_ONCE 382 drivers/net/ethernet/cavium/liquidio/cn23xx_pf_device.c WRITE_ONCE(reg_val, READ_ONCE(reg_val) & READ_ONCE 385 drivers/net/ethernet/cavium/liquidio/cn23xx_pf_device.c READ_ONCE(reg_val)); READ_ONCE 389 drivers/net/ethernet/cavium/liquidio/cn23xx_pf_device.c if (READ_ONCE(reg_val) & CN23XX_PKT_INPUT_CTL_RST) { READ_ONCE 884 drivers/net/ethernet/cavium/liquidio/cn23xx_pf_device.c WRITE_ONCE(d64, READ_ONCE(d64) & READ_ONCE 886 drivers/net/ethernet/cavium/liquidio/cn23xx_pf_device.c WRITE_ONCE(d64, READ_ONCE(d64) | CN23XX_PKT_INPUT_CTL_RST); READ_ONCE 888 drivers/net/ethernet/cavium/liquidio/cn23xx_pf_device.c READ_ONCE(d64)); READ_ONCE 895 drivers/net/ethernet/cavium/liquidio/cn23xx_pf_device.c while (!(READ_ONCE(d64) & BIT_ULL(q_no)) && loop--) { READ_ONCE 919 drivers/net/ethernet/cavium/liquidio/cn23xx_pf_device.c while (!(READ_ONCE(d64) & BIT_ULL(q_no)) && loop--) { READ_ONCE 938 drivers/net/ethernet/cavium/liquidio/cn23xx_pf_device.c READ_ONCE(d32)); READ_ONCE 1052 drivers/net/ethernet/cavium/liquidio/cn23xx_pf_device.c lio_pci_writeq(oct, (READ_ONCE(bar1) & 0xFFFFFFFEULL), READ_ONCE 70 drivers/net/ethernet/cavium/liquidio/cn23xx_vf_device.c while ((READ_ONCE(reg_val) & CN23XX_PKT_INPUT_CTL_RST) && READ_ONCE 71 drivers/net/ethernet/cavium/liquidio/cn23xx_vf_device.c !(READ_ONCE(reg_val) & CN23XX_PKT_INPUT_CTL_QUIET) && READ_ONCE 83 drivers/net/ethernet/cavium/liquidio/cn23xx_vf_device.c WRITE_ONCE(reg_val, READ_ONCE(reg_val) & READ_ONCE 86 drivers/net/ethernet/cavium/liquidio/cn23xx_vf_device.c READ_ONCE(reg_val)); READ_ONCE 90 drivers/net/ethernet/cavium/liquidio/cn23xx_vf_device.c if (READ_ONCE(reg_val) & CN23XX_PKT_INPUT_CTL_RST) { READ_ONCE 833 drivers/net/ethernet/cavium/liquidio/lio_main.c vfs_mask1 = READ_ONCE(oct->sriov_info.vf_drv_loaded_mask); READ_ONCE 834 drivers/net/ethernet/cavium/liquidio/lio_main.c vfs_mask2 = READ_ONCE(other_oct->sriov_info.vf_drv_loaded_mask); READ_ONCE 4019 drivers/net/ethernet/cavium/liquidio/lio_main.c cores_crashed = READ_ONCE(oct->cores_crashed); READ_ONCE 93 drivers/net/ethernet/cavium/liquidio/lio_vf_rep.c err = READ_ONCE(rep_resp->status) ? -EBUSY : 0; READ_ONCE 792 drivers/net/ethernet/cavium/liquidio/request_manager.c if (READ_ONCE(sc->caller_is_done)) { READ_ONCE 150 drivers/net/ethernet/cavium/liquidio/response_manager.c if (unlikely(READ_ONCE(sc->caller_is_done))) { READ_ONCE 399 drivers/net/ethernet/chelsio/cxgb4/sge.c int hw_cidx = ntohs(READ_ONCE(q->stat->cidx)); READ_ONCE 1337 drivers/net/ethernet/chelsio/cxgb4/sge.c hw_cidx = ntohs(READ_ONCE(q->stat->cidx)); READ_ONCE 2002 drivers/net/ethernet/chelsio/cxgb4/sge.c int hw_cidx = ntohs(READ_ONCE(q->stat->cidx)); READ_ONCE 291 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c xdp_prog = READ_ONCE(ch->xdp.prog); READ_ONCE 1265 drivers/net/ethernet/freescale/fec_main.c while (bdp != READ_ONCE(txq->bd.cur)) { READ_ONCE 1268 drivers/net/ethernet/freescale/fec_main.c status = fec16_to_cpu(READ_ONCE(bdp->cbd_sc)); READ_ONCE 147 drivers/net/ethernet/google/gve/gve_adminq.c status = be32_to_cpu(READ_ONCE(cmd->status)); READ_ONCE 571 drivers/net/ethernet/google/gve/gve_tx.c return READ_ONCE(priv->counter_array[counter_index]); READ_ONCE 456 drivers/net/ethernet/hisilicon/hip04_eth.c count = tx_count(READ_ONCE(priv->tx_head), tx_tail); READ_ONCE 515 drivers/net/ethernet/hisilicon/hip04_eth.c count = tx_count(tx_head, READ_ONCE(priv->tx_tail)); READ_ONCE 566 drivers/net/ethernet/hisilicon/hns3/hns3_enet.h int end = READ_ONCE(ring->next_to_use); READ_ONCE 584 drivers/net/ethernet/hisilicon/hns3/hns3_enet.h u8 __iomem *reg_addr = READ_ONCE(base); READ_ONCE 1079 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_cmd.h u8 __iomem *reg_addr = READ_ONCE(base); READ_ONCE 263 drivers/net/ethernet/hisilicon/hns3/hns3vf/hclgevf_cmd.h u8 __iomem *reg_addr = READ_ONCE(base); READ_ONCE 192 drivers/net/ethernet/intel/e1000/e1000.h unsigned int use = READ_ONCE((R)->next_to_use); \ READ_ONCE 20 drivers/net/ethernet/intel/fm10k/fm10k_common.h u32 __iomem *hw_addr = READ_ONCE((hw)->hw_addr); \ READ_ONCE 28 drivers/net/ethernet/intel/fm10k/fm10k_common.h u32 __iomem *sw_addr = READ_ONCE((hw)->sw_addr); \ READ_ONCE 160 drivers/net/ethernet/intel/fm10k/fm10k_iov.c if (!READ_ONCE(interface->iov_data)) READ_ONCE 204 drivers/net/ethernet/intel/fm10k/fm10k_iov.c if (!READ_ONCE(interface->iov_data)) READ_ONCE 1398 drivers/net/ethernet/intel/fm10k/fm10k_main.c itr_round = READ_ONCE(ring_container->itr_scale) + 8; READ_ONCE 629 drivers/net/ethernet/intel/fm10k/fm10k_netdev.c int num_tx_queues = READ_ONCE(interface->num_tx_queues); READ_ONCE 1339 drivers/net/ethernet/intel/fm10k/fm10k_netdev.c ring = READ_ONCE(interface->rx_ring[i]); READ_ONCE 1355 drivers/net/ethernet/intel/fm10k/fm10k_netdev.c ring = READ_ONCE(interface->tx_ring[i]); READ_ONCE 1560 drivers/net/ethernet/intel/fm10k/fm10k_netdev.c struct fm10k_l2_accel *l2_accel = READ_ONCE(interface->l2_accel); READ_ONCE 51 drivers/net/ethernet/intel/fm10k/fm10k_pci.c u32 __iomem *hw_addr = READ_ONCE(hw->hw_addr); READ_ONCE 361 drivers/net/ethernet/intel/fm10k/fm10k_pci.c hw_addr = READ_ONCE(interface->uc_addr); READ_ONCE 560 drivers/net/ethernet/intel/fm10k/fm10k_pci.c struct fm10k_ring *tx_ring = READ_ONCE(interface->tx_ring[i]); READ_ONCE 582 drivers/net/ethernet/intel/fm10k/fm10k_pci.c struct fm10k_ring *rx_ring = READ_ONCE(interface->rx_ring[i]); READ_ONCE 1154 drivers/net/ethernet/intel/i40e/i40e.h return !!READ_ONCE(vsi->xdp_prog); READ_ONCE 242 drivers/net/ethernet/intel/i40e/i40e_debugfs.c struct i40e_ring *rx_ring = READ_ONCE(vsi->rx_rings[i]); READ_ONCE 285 drivers/net/ethernet/intel/i40e/i40e_debugfs.c struct i40e_ring *tx_ring = READ_ONCE(vsi->tx_rings[i]); READ_ONCE 1382 drivers/net/ethernet/intel/i40e/i40e_ethtool.c flags = READ_ONCE(pf->flags); READ_ONCE 2274 drivers/net/ethernet/intel/i40e/i40e_ethtool.c i40e_add_queue_stats(&data, READ_ONCE(vsi->tx_rings[i])); READ_ONCE 2275 drivers/net/ethernet/intel/i40e/i40e_ethtool.c i40e_add_queue_stats(&data, READ_ONCE(vsi->rx_rings[i])); READ_ONCE 4881 drivers/net/ethernet/intel/i40e/i40e_ethtool.c orig_flags = READ_ONCE(pf->flags); READ_ONCE 455 drivers/net/ethernet/intel/i40e/i40e_main.c ring = READ_ONCE(vsi->tx_rings[i]); READ_ONCE 808 drivers/net/ethernet/intel/i40e/i40e_main.c p = READ_ONCE(vsi->tx_rings[q]); READ_ONCE 10403 drivers/net/ethernet/intel/i40e/i40e_main.c port = READ_ONCE(udp_port->port); READ_ONCE 10404 drivers/net/ethernet/intel/i40e/i40e_main.c type = READ_ONCE(udp_port->type); READ_ONCE 10405 drivers/net/ethernet/intel/i40e/i40e_main.c filter_index = READ_ONCE(udp_port->filter_index); READ_ONCE 130 drivers/net/ethernet/intel/i40e/i40e_ptp.c adj *= READ_ONCE(pf->ptp_adj_mult); READ_ONCE 2205 drivers/net/ethernet/intel/i40e/i40e_txrx.c xdp_prog = READ_ONCE(rx_ring->xdp_prog); READ_ONCE 204 drivers/net/ethernet/intel/i40e/i40e_xsk.c xdp_prog = READ_ONCE(rx_ring->xdp_prog); READ_ONCE 467 drivers/net/ethernet/intel/iavf/iavf_ethtool.c orig_flags = READ_ONCE(adapter->flags); READ_ONCE 1320 drivers/net/ethernet/intel/ice/ice_ethtool.c ring = READ_ONCE(vsi->tx_rings[j]); READ_ONCE 1331 drivers/net/ethernet/intel/ice/ice_ethtool.c ring = READ_ONCE(vsi->rx_rings[j]); READ_ONCE 3639 drivers/net/ethernet/intel/ice/ice_main.c ring = READ_ONCE(vsi->tx_rings[i]); READ_ONCE 3650 drivers/net/ethernet/intel/ice/ice_main.c ring = READ_ONCE(vsi->rx_rings[i]); READ_ONCE 358 drivers/net/ethernet/intel/igb/e1000_regs.h u8 __iomem *hw_addr = READ_ONCE((hw)->hw_addr); \ READ_ONCE 743 drivers/net/ethernet/intel/igb/igb_main.c u8 __iomem *hw_addr = READ_ONCE(hw->hw_addr); READ_ONCE 4035 drivers/net/ethernet/intel/igc/igc_main.c u8 __iomem *hw_addr = READ_ONCE(hw->hw_addr); READ_ONCE 225 drivers/net/ethernet/intel/igc/igc_regs.h u8 __iomem *hw_addr = READ_ONCE((hw)->hw_addr); \ READ_ONCE 146 drivers/net/ethernet/intel/ixgbe/ixgbe_common.h u8 __iomem *reg_addr = READ_ONCE(hw->hw_addr); READ_ONCE 165 drivers/net/ethernet/intel/ixgbe/ixgbe_common.h u8 __iomem *reg_addr = READ_ONCE(hw->hw_addr); READ_ONCE 297 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c reg_addr = READ_ONCE(hw->hw_addr); READ_ONCE 334 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c u8 __iomem *reg_addr = READ_ONCE(hw->hw_addr); READ_ONCE 2206 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c xdp_prog = READ_ONCE(rx_ring->xdp_prog); READ_ONCE 8941 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c struct ixgbe_ring *ring = READ_ONCE(adapter->rx_ring[i]); READ_ONCE 8957 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c struct ixgbe_ring *ring = READ_ONCE(adapter->tx_ring[i]); READ_ONCE 8962 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c struct ixgbe_ring *ring = READ_ONCE(adapter->xdp_ring[i]); READ_ONCE 459 drivers/net/ethernet/intel/ixgbe/ixgbe_ptp.c incval = READ_ONCE(adapter->base_incval); READ_ONCE 14 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c bool xdp_on = READ_ONCE(adapter->xdp_prog); READ_ONCE 154 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c xdp_prog = READ_ONCE(rx_ring->xdp_prog); READ_ONCE 706 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c if (!READ_ONCE(adapter->xdp_prog)) READ_ONCE 151 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c u8 __iomem *reg_addr = READ_ONCE(hw->hw_addr); READ_ONCE 1070 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c xdp_prog = READ_ONCE(rx_ring->xdp_prog); READ_ONCE 162 drivers/net/ethernet/intel/ixgbevf/vf.h u8 __iomem *reg_addr = READ_ONCE(hw->hw_addr); READ_ONCE 769 drivers/net/ethernet/mediatek/mtk_eth_soc.c rxd->rxd1 = READ_ONCE(dma_rxd->rxd1); READ_ONCE 770 drivers/net/ethernet/mediatek/mtk_eth_soc.c rxd->rxd2 = READ_ONCE(dma_rxd->rxd2); READ_ONCE 771 drivers/net/ethernet/mediatek/mtk_eth_soc.c rxd->rxd3 = READ_ONCE(dma_rxd->rxd3); READ_ONCE 772 drivers/net/ethernet/mediatek/mtk_eth_soc.c rxd->rxd4 = READ_ONCE(dma_rxd->rxd4); READ_ONCE 1466 drivers/net/ethernet/mellanox/mlx4/en_netdev.c pkt_rate_low = READ_ONCE(priv->pkt_rate_low); READ_ONCE 1467 drivers/net/ethernet/mellanox/mlx4/en_netdev.c pkt_rate_high = READ_ONCE(priv->pkt_rate_high); READ_ONCE 1470 drivers/net/ethernet/mellanox/mlx4/en_netdev.c rx_packets = READ_ONCE(priv->rx_ring[ring]->packets); READ_ONCE 1471 drivers/net/ethernet/mellanox/mlx4/en_netdev.c rx_bytes = READ_ONCE(priv->rx_ring[ring]->bytes); READ_ONCE 165 drivers/net/ethernet/mellanox/mlx4/en_port.c packets += READ_ONCE(ring->packets); READ_ONCE 166 drivers/net/ethernet/mellanox/mlx4/en_port.c bytes += READ_ONCE(ring->bytes); READ_ONCE 176 drivers/net/ethernet/mellanox/mlx4/en_port.c packets += READ_ONCE(ring->packets); READ_ONCE 177 drivers/net/ethernet/mellanox/mlx4/en_port.c bytes += READ_ONCE(ring->bytes); READ_ONCE 252 drivers/net/ethernet/mellanox/mlx4/en_port.c sw_rx_dropped += READ_ONCE(ring->dropped); READ_ONCE 253 drivers/net/ethernet/mellanox/mlx4/en_port.c priv->port_stats.rx_chksum_good += READ_ONCE(ring->csum_ok); READ_ONCE 254 drivers/net/ethernet/mellanox/mlx4/en_port.c priv->port_stats.rx_chksum_none += READ_ONCE(ring->csum_none); READ_ONCE 255 drivers/net/ethernet/mellanox/mlx4/en_port.c priv->port_stats.rx_chksum_complete += READ_ONCE(ring->csum_complete); READ_ONCE 256 drivers/net/ethernet/mellanox/mlx4/en_port.c priv->port_stats.rx_alloc_pages += READ_ONCE(ring->rx_alloc_pages); READ_ONCE 257 drivers/net/ethernet/mellanox/mlx4/en_port.c priv->xdp_stats.rx_xdp_drop += READ_ONCE(ring->xdp_drop); READ_ONCE 258 drivers/net/ethernet/mellanox/mlx4/en_port.c priv->xdp_stats.rx_xdp_tx += READ_ONCE(ring->xdp_tx); READ_ONCE 259 drivers/net/ethernet/mellanox/mlx4/en_port.c priv->xdp_stats.rx_xdp_tx_full += READ_ONCE(ring->xdp_tx_full); READ_ONCE 270 drivers/net/ethernet/mellanox/mlx4/en_port.c sw_tx_dropped += READ_ONCE(ring->tx_dropped); READ_ONCE 271 drivers/net/ethernet/mellanox/mlx4/en_port.c priv->port_stats.tx_chksum_offload += READ_ONCE(ring->tx_csum); READ_ONCE 272 drivers/net/ethernet/mellanox/mlx4/en_port.c priv->port_stats.queue_stopped += READ_ONCE(ring->queue_stopped); READ_ONCE 273 drivers/net/ethernet/mellanox/mlx4/en_port.c priv->port_stats.wake_queue += READ_ONCE(ring->wake_queue); READ_ONCE 274 drivers/net/ethernet/mellanox/mlx4/en_port.c priv->port_stats.tso_packets += READ_ONCE(ring->tso_packets); READ_ONCE 275 drivers/net/ethernet/mellanox/mlx4/en_port.c priv->port_stats.xmit_more += READ_ONCE(ring->xmit_more); READ_ONCE 424 drivers/net/ethernet/mellanox/mlx4/en_tx.c last_nr_txbb = READ_ONCE(ring->last_nr_txbb); READ_ONCE 425 drivers/net/ethernet/mellanox/mlx4/en_tx.c ring_cons = READ_ONCE(ring->cons); READ_ONCE 868 drivers/net/ethernet/mellanox/mlx4/en_tx.c ring_cons = READ_ONCE(ring->cons); READ_ONCE 1078 drivers/net/ethernet/mellanox/mlx4/en_tx.c ring_cons = READ_ONCE(ring->cons); READ_ONCE 1147 drivers/net/ethernet/mellanox/mlx4/en_tx.c (u32)(ring->prod - READ_ONCE(ring->cons) - 1)); READ_ONCE 211 drivers/net/ethernet/mellanox/mlx5/core/cmd.c own = READ_ONCE(ent->lay->status_own); READ_ONCE 124 drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c struct bpf_prog *prog = READ_ONCE(rq->xdp_prog); READ_ONCE 2620 drivers/net/ethernet/neterion/vxge/vxge-main.c rx_frms = READ_ONCE(ring->stats.rx_frms); READ_ONCE 1816 drivers/net/ethernet/netronome/nfp/nfp_net_common.c xdp_prog = READ_ONCE(dp->xdp_prog); READ_ONCE 45 drivers/net/ethernet/netronome/nfp/nfp_net_debugfs.c frag = READ_ONCE(rx_ring->rxbufs[i].frag); READ_ONCE 112 drivers/net/ethernet/netronome/nfp/nfp_net_debugfs.c struct sk_buff *skb = READ_ONCE(tx_ring->txbufs[i].skb); READ_ONCE 119 drivers/net/ethernet/netronome/nfp/nfp_net_debugfs.c READ_ONCE(tx_ring->txbufs[i].frag)); READ_ONCE 540 drivers/net/ethernet/qlogic/qed/qed_iwarp.c while (READ_ONCE(ep->state) != QED_IWARP_EP_CLOSED && READ_ONCE 1205 drivers/net/ethernet/qlogic/qede/qede_fp.c struct bpf_prog *xdp_prog = READ_ONCE(rxq->xdp_prog); READ_ONCE 2237 drivers/net/ethernet/sfc/ef10.c if (likely(READ_ONCE(efx->irq_soft_enabled))) { READ_ONCE 2252 drivers/net/ethernet/sfc/ef10.c bool soft_enabled = READ_ONCE(efx->irq_soft_enabled); READ_ONCE 3544 drivers/net/ethernet/sfc/ef10.c if (unlikely(READ_ONCE(efx->reset_pending))) READ_ONCE 3693 drivers/net/ethernet/sfc/ef10.c if (unlikely(READ_ONCE(efx->reset_pending))) READ_ONCE 5505 drivers/net/ethernet/sfc/ef10.c if (READ_ONCE(table->entry[i].spec) & READ_ONCE 2861 drivers/net/ethernet/sfc/efx.c pending = READ_ONCE(efx->reset_pending); READ_ONCE 2926 drivers/net/ethernet/sfc/efx.c if (READ_ONCE(efx->state) != STATE_READY) READ_ONCE 2516 drivers/net/ethernet/sfc/falcon/efx.c pending = READ_ONCE(efx->reset_pending); READ_ONCE 2576 drivers/net/ethernet/sfc/falcon/efx.c if (READ_ONCE(efx->state) != STATE_READY) READ_ONCE 452 drivers/net/ethernet/sfc/falcon/falcon.c if (!likely(READ_ONCE(efx->irq_soft_enabled))) READ_ONCE 1372 drivers/net/ethernet/sfc/falcon/falcon.c isolate = !!READ_ONCE(efx->reset_pending); READ_ONCE 834 drivers/net/ethernet/sfc/falcon/farch.c if (unlikely(READ_ONCE(efx->reset_pending))) READ_ONCE 990 drivers/net/ethernet/sfc/falcon/farch.c if (unlikely(READ_ONCE(efx->reset_pending))) READ_ONCE 1504 drivers/net/ethernet/sfc/falcon/farch.c bool soft_enabled = READ_ONCE(efx->irq_soft_enabled); READ_ONCE 1596 drivers/net/ethernet/sfc/falcon/farch.c if (!likely(READ_ONCE(efx->irq_soft_enabled))) READ_ONCE 83 drivers/net/ethernet/sfc/falcon/nic.h unsigned int empty_read_count = READ_ONCE(tx_queue->empty_read_count); READ_ONCE 466 drivers/net/ethernet/sfc/falcon/nic.h return READ_ONCE(channel->event_test_cpu); READ_ONCE 470 drivers/net/ethernet/sfc/falcon/nic.h return READ_ONCE(efx->last_irq_cpu); READ_ONCE 134 drivers/net/ethernet/sfc/falcon/tx.c txq1->old_read_count = READ_ONCE(txq1->read_count); READ_ONCE 135 drivers/net/ethernet/sfc/falcon/tx.c txq2->old_read_count = READ_ONCE(txq2->read_count); READ_ONCE 524 drivers/net/ethernet/sfc/falcon/tx.c tx_queue->old_write_count = READ_ONCE(tx_queue->write_count); READ_ONCE 826 drivers/net/ethernet/sfc/farch.c if (unlikely(READ_ONCE(efx->reset_pending))) READ_ONCE 978 drivers/net/ethernet/sfc/farch.c if (unlikely(READ_ONCE(efx->reset_pending))) READ_ONCE 1513 drivers/net/ethernet/sfc/farch.c bool soft_enabled = READ_ONCE(efx->irq_soft_enabled); READ_ONCE 1605 drivers/net/ethernet/sfc/farch.c if (!likely(READ_ONCE(efx->irq_soft_enabled))) READ_ONCE 81 drivers/net/ethernet/sfc/nic.h unsigned int empty_read_count = READ_ONCE(tx_queue->empty_read_count); READ_ONCE 643 drivers/net/ethernet/sfc/nic.h return READ_ONCE(channel->event_test_cpu); READ_ONCE 647 drivers/net/ethernet/sfc/nic.h return READ_ONCE(efx->last_irq_cpu); READ_ONCE 870 drivers/net/ethernet/sfc/ptp.c READ_ONCE(*mc_running)) { READ_ONCE 880 drivers/net/ethernet/sfc/ptp.c READ_ONCE(*mc_running)); READ_ONCE 1051 drivers/net/ethernet/sfc/ptp.c while (!READ_ONCE(*start) && (time_before(jiffies, timeout))) { READ_ONCE 1061 drivers/net/ethernet/sfc/ptp.c if (READ_ONCE(*start)) READ_ONCE 151 drivers/net/ethernet/sfc/tx.c txq1->old_read_count = READ_ONCE(txq1->read_count); READ_ONCE 152 drivers/net/ethernet/sfc/tx.c txq2->old_read_count = READ_ONCE(txq2->read_count); READ_ONCE 781 drivers/net/ethernet/sfc/tx.c tx_queue->old_write_count = READ_ONCE(tx_queue->write_count); READ_ONCE 936 drivers/net/ethernet/socionext/netsec.c xdp_prog = READ_ONCE(priv->xdp_prog); READ_ONCE 1273 drivers/net/ethernet/socionext/netsec.c struct bpf_prog *xdp_prog = READ_ONCE(priv->xdp_prog); READ_ONCE 6228 drivers/net/ethernet/sun/niu.c rx_rings = READ_ONCE(np->rx_rings); READ_ONCE 6259 drivers/net/ethernet/sun/niu.c tx_rings = READ_ONCE(np->tx_rings); READ_ONCE 500 drivers/net/ethernet/ti/cpsw.c prog = READ_ONCE(priv->xdp_prog); READ_ONCE 261 drivers/net/tap.c int numvtaps = READ_ONCE(tap->numvtaps); READ_ONCE 639 drivers/net/tap.c vnet_hdr_len = READ_ONCE(q->vnet_hdr_sz); READ_ONCE 780 drivers/net/tap.c vnet_hdr_len = READ_ONCE(q->vnet_hdr_sz); READ_ONCE 1159 drivers/net/tap.c vnet_hdr_len = READ_ONCE(q->vnet_hdr_sz); READ_ONCE 529 drivers/net/tun.c if (READ_ONCE(e->queue_index) != queue_index) READ_ONCE 571 drivers/net/tun.c numqueues = READ_ONCE(tun->numqueues); READ_ONCE 592 drivers/net/tun.c numqueues = READ_ONCE(tun->numqueues); READ_ONCE 1297 drivers/net/tun.c numqueues = READ_ONCE(tun->numqueues); READ_ONCE 1772 drivers/net/tun.c int vnet_hdr_sz = READ_ONCE(tun->vnet_hdr_sz); READ_ONCE 2055 drivers/net/tun.c vnet_hdr_sz = READ_ONCE(tun->vnet_hdr_sz); READ_ONCE 2093 drivers/net/tun.c vnet_hdr_sz = READ_ONCE(tun->vnet_hdr_sz); READ_ONCE 503 drivers/net/wireless/ath/ath5k/desc.c txstat1 = READ_ONCE(tx_status->tx_status_1); READ_ONCE 509 drivers/net/wireless/ath/ath5k/desc.c txstat0 = READ_ONCE(tx_status->tx_status_0); READ_ONCE 703 drivers/net/wireless/ath/ath5k/desc.c rxstat1 = READ_ONCE(rx_status->rx_status_1); READ_ONCE 710 drivers/net/wireless/ath/ath5k/desc.c rxstat0 = READ_ONCE(rx_status->rx_status_0); READ_ONCE 321 drivers/net/wireless/ath/ath9k/ar9002_mac.c status = READ_ONCE(ads->ds_txstatus9); READ_ONCE 335 drivers/net/wireless/ath/ath9k/ar9002_mac.c status = READ_ONCE(ads->ds_txstatus0); READ_ONCE 345 drivers/net/wireless/ath/ath9k/ar9002_mac.c status = READ_ONCE(ads->ds_txstatus1); READ_ONCE 374 drivers/net/wireless/ath/ath9k/ar9002_mac.c status = READ_ONCE(ads->ds_txstatus5); READ_ONCE 393 drivers/net/wireless/ath/ath9k/ar9002_mac.c return MS(READ_ONCE(ads->ds_ctl4), AR_PacketDur0); READ_ONCE 395 drivers/net/wireless/ath/ath9k/ar9002_mac.c return MS(READ_ONCE(ads->ds_ctl4), AR_PacketDur1); READ_ONCE 397 drivers/net/wireless/ath/ath9k/ar9002_mac.c return MS(READ_ONCE(ads->ds_ctl5), AR_PacketDur2); READ_ONCE 399 drivers/net/wireless/ath/ath9k/ar9002_mac.c return MS(READ_ONCE(ads->ds_ctl5), AR_PacketDur3); READ_ONCE 362 drivers/net/wireless/ath/ath9k/ar9003_mac.c status = READ_ONCE(ads->status8); READ_ONCE 388 drivers/net/wireless/ath/ath9k/ar9003_mac.c status = READ_ONCE(ads->status2); READ_ONCE 398 drivers/net/wireless/ath/ath9k/ar9003_mac.c status = READ_ONCE(ads->status3); READ_ONCE 423 drivers/net/wireless/ath/ath9k/ar9003_mac.c status = READ_ONCE(ads->status7); READ_ONCE 440 drivers/net/wireless/ath/ath9k/ar9003_mac.c return MS(READ_ONCE(adc->ctl15), AR_PacketDur0); READ_ONCE 442 drivers/net/wireless/ath/ath9k/ar9003_mac.c return MS(READ_ONCE(adc->ctl15), AR_PacketDur1); READ_ONCE 444 drivers/net/wireless/ath/ath9k/ar9003_mac.c return MS(READ_ONCE(adc->ctl16), AR_PacketDur2); READ_ONCE 446 drivers/net/wireless/ath/ath9k/ar9003_mac.c return MS(READ_ONCE(adc->ctl16), AR_PacketDur3); READ_ONCE 369 drivers/net/wireless/ath/wcn36xx/dxe.c if (READ_ONCE(ctl->desc->ctrl) & WCN36xx_DXE_CTRL_VLD) READ_ONCE 373 drivers/net/wireless/ath/wcn36xx/dxe.c READ_ONCE(ctl->desc->ctrl) & WCN36xx_DXE_CTRL_EOP) { READ_ONCE 565 drivers/net/wireless/ath/wcn36xx/dxe.c while (!(READ_ONCE(dxe->ctrl) & WCN36xx_DXE_CTRL_VLD)) { READ_ONCE 3708 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c while (READ_ONCE(bus->dpc_triggered)) { READ_ONCE 791 drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c u8 ap_sta_id = READ_ONCE(mvmvif->ap_sta_id); READ_ONCE 1185 drivers/net/wireless/intel/iwlwifi/mvm/ops.c bool rfkill_safe_init_done = READ_ONCE(mvm->rfkill_safe_init_done); READ_ONCE 270 drivers/net/wireless/intel/iwlwifi/pcie/internal.h return READ_ONCE(*rb_stts); READ_ONCE 274 drivers/net/wireless/intel/iwlwifi/pcie/internal.h return READ_ONCE(rb_stts->closed_rb_num); READ_ONCE 2366 drivers/net/wireless/intel/iwlwifi/pcie/trans.c wr_ptr = READ_ONCE(txq->write_ptr); READ_ONCE 2368 drivers/net/wireless/intel/iwlwifi/pcie/trans.c while ((txq->read_ptr != READ_ONCE(txq->write_ptr) || READ_ONCE 2372 drivers/net/wireless/intel/iwlwifi/pcie/trans.c u8 write_ptr = READ_ONCE(txq->write_ptr); READ_ONCE 1424 drivers/net/wireless/mac80211_hwsim.c _portid = READ_ONCE(data->wmediumd); READ_ONCE 1521 drivers/net/wireless/mac80211_hwsim.c u32 _pid = READ_ONCE(data->wmediumd); READ_ONCE 99 drivers/net/wireless/mediatek/mt76/dma.c __le32 __ctrl = READ_ONCE(q->desc[idx].ctrl); READ_ONCE 103 drivers/net/wireless/mediatek/mt76/dma.c __le32 addr = READ_ONCE(q->desc[idx].buf0); READ_ONCE 111 drivers/net/wireless/mediatek/mt76/dma.c __le32 addr = READ_ONCE(q->desc[idx].buf1); READ_ONCE 215 drivers/net/wireless/mediatek/mt76/dma.c buf_addr = le32_to_cpu(READ_ONCE(desc->buf0)); READ_ONCE 217 drivers/net/wireless/mediatek/mt76/dma.c u32 ctl = le32_to_cpu(READ_ONCE(desc->ctrl)); READ_ONCE 1003 drivers/net/wireless/mediatek/mt76/mt7603/mac.c rate_set_tsf = READ_ONCE(sta->rate_set_tsf); READ_ONCE 886 drivers/net/wireless/mediatek/mt76/mt7615/mac.c rate_set_tsf = READ_ONCE(sta->rate_set_tsf); READ_ONCE 69 drivers/net/wireless/quantenna/qtnfmac/shm_ipc.c if (!READ_ONCE(ipc->waiting_for_ack)) READ_ONCE 159 drivers/net/xen-netback/interface.c num_queues = READ_ONCE(vif->num_queues); READ_ONCE 191 drivers/net/xen-netback/interface.c num_queues = READ_ONCE(vif->num_queues); READ_ONCE 250 drivers/net/xen-netback/interface.c num_queues = READ_ONCE(vif->num_queues); READ_ONCE 402 drivers/net/xen-netback/interface.c num_queues = READ_ONCE(vif->num_queues); READ_ONCE 186 drivers/nvme/host/multipath.c if (READ_ONCE(head->subsys->iopolicy) == NVME_IOPOLICY_NUMA) READ_ONCE 271 drivers/nvme/host/multipath.c if (READ_ONCE(head->subsys->iopolicy) == NVME_IOPOLICY_RR && ns) READ_ONCE 605 drivers/nvme/host/multipath.c nvme_iopolicy_names[READ_ONCE(subsys->iopolicy)]); READ_ONCE 715 drivers/nvme/target/admin-cmd.c nvmet_set_result(req, READ_ONCE(req->sq->ctrl->aen_enabled)); READ_ONCE 356 drivers/nvme/target/nvmet.h if (!(READ_ONCE(ctrl->aen_enabled) & (1 << bn))) READ_ONCE 111 drivers/powercap/idle_inject.c duration_us = READ_ONCE(ii_dev->run_duration_us); READ_ONCE 112 drivers/powercap/idle_inject.c duration_us += READ_ONCE(ii_dev->idle_duration_us); READ_ONCE 141 drivers/powercap/idle_inject.c play_idle(READ_ONCE(ii_dev->idle_duration_us)); READ_ONCE 168 drivers/powercap/idle_inject.c *run_duration_us = READ_ONCE(ii_dev->run_duration_us); READ_ONCE 169 drivers/powercap/idle_inject.c *idle_duration_us = READ_ONCE(ii_dev->idle_duration_us); READ_ONCE 184 drivers/powercap/idle_inject.c unsigned int idle_duration_us = READ_ONCE(ii_dev->idle_duration_us); READ_ONCE 185 drivers/powercap/idle_inject.c unsigned int run_duration_us = READ_ONCE(ii_dev->run_duration_us); READ_ONCE 394 drivers/s390/net/ism_drv.c while ((ism->ieq_idx + 1) != READ_ONCE(ism->ieq->header.idx)) { READ_ONCE 14352 drivers/scsi/lpfc/lpfc_sli.c if (READ_ONCE(eq->mode) == LPFC_EQ_POLL) READ_ONCE 1025 drivers/scsi/qla2xxx/qla_target.c while (!READ_ONCE(sess->logout_completed)) { READ_ONCE 39 drivers/scsi/scsi_debugfs.c const u8 *const cdb = READ_ONCE(cmd->cmnd); READ_ONCE 324 drivers/staging/comedi/comedi_fops.c return READ_ONCE(cfp->read_subdev); READ_ONCE 332 drivers/staging/comedi/comedi_fops.c return READ_ONCE(cfp->write_subdev); READ_ONCE 1231 drivers/target/target_core_user.c while (udev->cmdr_last_cleaned != READ_ONCE(mb->cmd_tail)) { READ_ONCE 381 drivers/thermal/intel/intel_powerclamp.c w_data->target_ratio = READ_ONCE(set_target_ratio); READ_ONCE 649 drivers/thermal/tegra/soctherm.c temp = READ_ONCE(tz->temperature); READ_ONCE 650 drivers/thermal/tegra/soctherm.c last_temp = READ_ONCE(tz->last_temperature); READ_ONCE 208 drivers/tty/tty_audit.c audit_tty = READ_ONCE(current->signal->audit_tty); READ_ONCE 30 drivers/tty/tty_port.c tty = READ_ONCE(port->itty); READ_ONCE 489 drivers/usb/class/cdc-wdm.c cntr = READ_ONCE(desc->length); READ_ONCE 148 drivers/usb/core/devio.c lim = READ_ONCE(usbfs_memory_mb); READ_ONCE 2473 drivers/usb/core/devio.c READ_ONCE(ps->not_yet_resumed) == 0); READ_ONCE 1123 drivers/usb/core/sysfs.c string = READ_ONCE(intf->cur_altsetting->string); READ_ONCE 1139 drivers/usb/core/sysfs.c alt = READ_ONCE(intf->cur_altsetting); READ_ONCE 1244 drivers/usb/gadget/udc/gr_udc.c if (READ_ONCE(req->last_desc->ctrl) & GR_DESC_IN_CTRL_EN) READ_ONCE 1273 drivers/usb/gadget/udc/gr_udc.c ctrl = READ_ONCE(req->curr_desc->ctrl); READ_ONCE 808 drivers/usb/host/ohci-hcd.c head = hc32_to_cpu(ohci, READ_ONCE(ed->hwHeadP)) & TD_MASK; READ_ONCE 191 drivers/usb/host/uhci-hcd.h #define qh_element(qh) READ_ONCE((qh)->element) READ_ONCE 279 drivers/usb/host/uhci-hcd.h READ_ONCE((td)->status)) READ_ONCE 661 drivers/vfio/vfio.c struct device_driver *drv = READ_ONCE(dev->driver); READ_ONCE 897 drivers/vhost/scsi.c tpg = READ_ONCE(vs_tpg[*vc->target]); READ_ONCE 2098 drivers/vhost/vhost.c next = vhost16_to_cpu(vq, READ_ONCE(desc->next)); READ_ONCE 825 drivers/vhost/vringh.c *val = vringh16_to_cpu(vrh, READ_ONCE(*p)); READ_ONCE 778 drivers/xen/pvcalls-back.c data = READ_ONCE(queue->rskq_accept_head) != NULL; READ_ONCE 324 drivers/xen/pvcalls-front.c READ_ONCE(bedata->rsp[req_id].req_id) == req_id); READ_ONCE 470 drivers/xen/pvcalls-front.c READ_ONCE(bedata->rsp[req_id].req_id) == req_id); READ_ONCE 701 drivers/xen/pvcalls-front.c READ_ONCE(bedata->rsp[req_id].req_id) == req_id); READ_ONCE 750 drivers/xen/pvcalls-front.c READ_ONCE(bedata->rsp[req_id].req_id) == req_id); READ_ONCE 787 drivers/xen/pvcalls-front.c req_id = READ_ONCE(map->passive.inflight_req_id); READ_ONCE 789 drivers/xen/pvcalls-front.c READ_ONCE(bedata->rsp[req_id].req_id) == req_id) { READ_ONCE 866 drivers/xen/pvcalls-front.c READ_ONCE(bedata->rsp[req_id].req_id) == req_id)) { READ_ONCE 908 drivers/xen/pvcalls-front.c uint32_t req_id = READ_ONCE(map->passive.inflight_req_id); READ_ONCE 911 drivers/xen/pvcalls-front.c READ_ONCE(bedata->rsp[req_id].req_id) == req_id) READ_ONCE 1038 drivers/xen/pvcalls-front.c READ_ONCE(bedata->rsp[req_id].req_id) == req_id); READ_ONCE 1068 drivers/xen/pvcalls-front.c if (READ_ONCE(map->passive.inflight_req_id) != PVCALLS_INVALID_ID && READ_ONCE 1069 drivers/xen/pvcalls-front.c READ_ONCE(map->passive.inflight_req_id) != 0) { READ_ONCE 42 drivers/xen/time.c h = READ_ONCE(p32[1]); READ_ONCE 43 drivers/xen/time.c l = READ_ONCE(p32[0]); READ_ONCE 44 drivers/xen/time.c h2 = READ_ONCE(p32[1]); READ_ONCE 49 drivers/xen/time.c ret = READ_ONCE(*p); READ_ONCE 67 drivers/xen/time.c *res = READ_ONCE(*state); READ_ONCE 373 fs/afs/addr_list.c index = READ_ONCE(ac->alist->preferred); READ_ONCE 384 fs/afs/cell.c unsigned int min_ttl = READ_ONCE(afs_cell_min_ttl); READ_ONCE 385 fs/afs/cell.c unsigned int max_ttl = READ_ONCE(afs_cell_max_ttl); READ_ONCE 650 fs/afs/dir.c afs_dataversion_t data_version = READ_ONCE(dvnode->status.data_version); READ_ONCE 984 fs/afs/dir.c parent = READ_ONCE(dentry->d_parent); READ_ONCE 1000 fs/afs/dir.c dir_version = (long)READ_ONCE(dvnode->status.data_version); READ_ONCE 1001 fs/afs/dir.c de_version = (long)READ_ONCE(dentry->d_fsdata); READ_ONCE 1003 fs/afs/dir.c dir_version = (long)READ_ONCE(dvnode->invalid_before); READ_ONCE 754 fs/afs/flock.c lock_count = READ_ONCE(vnode->status.lock_count); READ_ONCE 605 fs/afs/inode.c cb_v_break = READ_ONCE(volume->cb_v_break); READ_ONCE 611 fs/afs/inode.c cb_s_break = READ_ONCE(server->cb_s_break); READ_ONCE 1170 fs/afs/internal.h return READ_ONCE(call->state) == state; READ_ONCE 63 fs/afs/rotate.c fc->index = READ_ONCE(fc->server_list->preferred); READ_ONCE 466 fs/afs/rotate.c afs_prioritise_error(&e, READ_ONCE(s->probe.error), READ_ONCE 511 fs/afs/rxrpc.c while (state = READ_ONCE(call->state), READ_ONCE 542 fs/afs/rxrpc.c state = READ_ONCE(call->state); READ_ONCE 937 fs/afs/rxrpc.c state = READ_ONCE(call->state); READ_ONCE 583 fs/afs/server.c diff = READ_ONCE(server->update_at) - now; READ_ONCE 261 fs/afs/vl_rotate.c afs_prioritise_error(&e, READ_ONCE(s->probe.error), READ_ONCE 1637 fs/aio.c if (!READ_ONCE(req->cancelled)) READ_ONCE 1648 fs/aio.c if (!mask && !READ_ONCE(req->cancelled)) { READ_ONCE 255 fs/block_dev.c if (!READ_ONCE(bio.bi_private)) READ_ONCE 296 fs/block_dev.c return blk_poll(q, READ_ONCE(kiocb->ki_cookie), wait); READ_ONCE 448 fs/block_dev.c if (!READ_ONCE(dio->waiter)) READ_ONCE 73 fs/btrfs/delayed-inode.c node = READ_ONCE(btrfs_inode->delayed_node); READ_ONCE 1298 fs/btrfs/delayed-inode.c delayed_node = READ_ONCE(inode->delayed_node); READ_ONCE 863 fs/btrfs/transaction.c if (lock && READ_ONCE(cur_trans->state) == TRANS_STATE_BLOCKED) { READ_ONCE 1974 fs/btrfs/transaction.c if (unlikely(READ_ONCE(cur_trans->aborted))) { READ_ONCE 2122 fs/btrfs/transaction.c if (unlikely(READ_ONCE(cur_trans->aborted))) { READ_ONCE 2240 fs/btrfs/transaction.c if (unlikely(READ_ONCE(cur_trans->aborted))) { READ_ONCE 40 fs/btrfs/tree-log.h return READ_ONCE(trans->fs_info->last_trans_log_full_commit) == READ_ONCE 1646 fs/buffer.c create_empty_buffers(page, 1 << READ_ONCE(inode->i_blkbits), READ_ONCE 800 fs/ceph/addr.c if (READ_ONCE(fsc->mount_state) == CEPH_MOUNT_SHUTDOWN) { READ_ONCE 1229 fs/ceph/addr.c if (READ_ONCE(fsc->mount_state) == CEPH_MOUNT_SHUTDOWN) { READ_ONCE 1208 fs/ceph/caps.c ceph_encode_32(&p, READ_ONCE(osdc->epoch_barrier)); READ_ONCE 2261 fs/ceph/caps.c if (errseq_check(&ci->i_meta_err, READ_ONCE(fi->meta_err))) { READ_ONCE 2664 fs/ceph/caps.c if (READ_ONCE(mdsc->fsc->mount_state) == READ_ONCE 2757 fs/ceph/caps.c fi->filp_gen != READ_ONCE(fsc->filp_gen)) READ_ONCE 2790 fs/ceph/caps.c fi->filp_gen != READ_ONCE(fsc->filp_gen)) { READ_ONCE 1559 fs/ceph/dir.c parent = READ_ONCE(dentry->d_parent); READ_ONCE 238 fs/ceph/file.c fi->filp_gen = READ_ONCE(ceph_inode_to_client(inode)->filp_gen); READ_ONCE 1853 fs/ceph/inode.c if (READ_ONCE(fsc->mount_state) == CEPH_MOUNT_SHUTDOWN) { READ_ONCE 26 fs/ceph/io.c if (READ_ONCE(ci->i_ceph_flags) & CEPH_I_ODIRECT) { READ_ONCE 57 fs/ceph/io.c if (!(READ_ONCE(ci->i_ceph_flags) & CEPH_I_ODIRECT)) READ_ONCE 111 fs/ceph/io.c if (!(READ_ONCE(ci->i_ceph_flags) & CEPH_I_ODIRECT)) { READ_ONCE 143 fs/ceph/io.c if (READ_ONCE(ci->i_ceph_flags) & CEPH_I_ODIRECT) READ_ONCE 924 fs/ceph/mds_client.c parent = READ_ONCE(req->r_dentry->d_parent); READ_ONCE 1399 fs/ceph/mds_client.c if (READ_ONCE(fsc->mount_state) == CEPH_MOUNT_SHUTDOWN) { READ_ONCE 2173 fs/ceph/mds_client.c temp = READ_ONCE(temp->d_parent); READ_ONCE 2536 fs/ceph/mds_client.c if (READ_ONCE(mdsc->fsc->mount_state) == CEPH_MOUNT_SHUTDOWN) { READ_ONCE 2541 fs/ceph/mds_client.c if (READ_ONCE(mdsc->fsc->mount_state) == CEPH_MOUNT_MOUNTING) { READ_ONCE 4034 fs/ceph/mds_client.c if (READ_ONCE(fsc->mount_state) != CEPH_MOUNT_MOUNTED) READ_ONCE 4037 fs/ceph/mds_client.c if (!READ_ONCE(fsc->blacklisted)) READ_ONCE 4301 fs/ceph/mds_client.c if (READ_ONCE(mdsc->fsc->mount_state) == CEPH_MOUNT_SHUTDOWN) READ_ONCE 4332 fs/ceph/mds_client.c if (READ_ONCE(mdsc->fsc->mount_state) == CEPH_MOUNT_SHUTDOWN) READ_ONCE 538 fs/ceph/super.h if (!(READ_ONCE(ci->i_ceph_flags) & CEPH_I_ERROR_WRITE)) { READ_ONCE 547 fs/ceph/super.h if (READ_ONCE(ci->i_ceph_flags) & CEPH_I_ERROR_WRITE) { READ_ONCE 404 fs/crypto/fscrypt_private.h return READ_ONCE(secret->size) != 0; READ_ONCE 247 fs/crypto/hooks.c pstr.name = READ_ONCE(inode->i_link); READ_ONCE 234 fs/crypto/keyring.c keyring = READ_ONCE(sb->s_master_keys); READ_ONCE 116 fs/crypto/keysetup.c struct crypto_shash *tfm = READ_ONCE(essiv_hash_tfm); READ_ONCE 221 fs/crypto/keysetup.c tfm = READ_ONCE(mk->mk_mode_keys[mode_num]); READ_ONCE 568 fs/crypto/keysetup.c const struct fscrypt_info *ci = READ_ONCE(inode->i_crypt_info); READ_ONCE 214 fs/crypto/policy.c ci = READ_ONCE(inode->i_crypt_info); READ_ONCE 481 fs/crypto/policy.c ci = READ_ONCE(parent->i_crypt_info); READ_ONCE 41 fs/d_path.c u32 dlen = READ_ONCE(name->len); READ_ONCE 104 fs/d_path.c struct mount *parent = READ_ONCE(mnt->mnt_parent); READ_ONCE 114 fs/d_path.c dentry = READ_ONCE(mnt->mnt_mountpoint); READ_ONCE 397 fs/dax.c struct address_space *mapping = READ_ONCE(page->mapping); READ_ONCE 249 fs/dcache.c const unsigned char *cs = READ_ONCE(dentry->d_name.name); READ_ONCE 319 fs/dcache.c flags = READ_ONCE(dentry->d_flags); READ_ONCE 327 fs/dcache.c unsigned flags = READ_ONCE(dentry->d_flags); READ_ONCE 603 fs/dcache.c parent = READ_ONCE(dentry->d_parent); READ_ONCE 779 fs/dcache.c d_flags = READ_ONCE(dentry->d_flags); READ_ONCE 912 fs/dcache.c ret = READ_ONCE(dentry->d_parent); READ_ONCE 916 fs/dcache.c if (likely(ret == READ_ONCE(dentry->d_parent))) READ_ONCE 2556 fs/dcache.c if (unlikely(READ_ONCE(parent->d_inode->i_dir_seq) != seq)) { READ_ONCE 86 fs/debugfs/file.c d_fsd = READ_ONCE(dentry->d_fsdata); READ_ONCE 100 fs/debugfs/file.c fsd = READ_ONCE(dentry->d_fsdata); READ_ONCE 133 fs/debugfs/file.c struct debugfs_fsdata *fsd = READ_ONCE(dentry->d_fsdata); READ_ONCE 681 fs/debugfs/inode.c fsd = READ_ONCE(dentry->d_fsdata); READ_ONCE 1169 fs/direct-io.c unsigned i_blkbits = READ_ONCE(inode->i_blkbits); READ_ONCE 347 fs/ecryptfs/inode.c lower_inode = READ_ONCE(lower_dentry->d_inode); READ_ONCE 183 fs/erofs/zdata.c if (READ_ONCE(*pages)) READ_ONCE 368 fs/erofs/zdata.c length = READ_ONCE(pcl->length); READ_ONCE 384 fs/erofs/zdata.c length = READ_ONCE(pcl->length); READ_ONCE 773 fs/erofs/zdata.c DBG_BUGON(!READ_ONCE(cl->nr_pages)); READ_ONCE 972 fs/erofs/zdata.c owned = READ_ONCE(pcl->next); READ_ONCE 1009 fs/erofs/zdata.c page = READ_ONCE(pcl->compressed_pages[nr]); READ_ONCE 1029 fs/erofs/zdata.c mapping = READ_ONCE(page->mapping); READ_ONCE 154 fs/erofs/zdata.h o = READ_ONCE(*p); READ_ONCE 172 fs/eventfd.c count = READ_ONCE(ctx->count); READ_ONCE 381 fs/eventpoll.c READ_ONCE(ep->ovflist) != EP_UNACTIVE_PTR; READ_ONCE 400 fs/eventpoll.c unsigned int napi_id = READ_ONCE(ep->napi_id); READ_ONCE 434 fs/eventpoll.c napi_id = READ_ONCE(sk->sk_napi_id); READ_ONCE 712 fs/eventpoll.c for (nepi = READ_ONCE(ep->ovflist); (epi = nepi) != NULL; READ_ONCE 1248 fs/eventpoll.c if (READ_ONCE(ep->ovflist) != EP_UNACTIVE_PTR) { READ_ONCE 1539 fs/exec.c mode = READ_ONCE(inode->i_mode); READ_ONCE 678 fs/ext4/dir.c const struct dentry *parent = READ_ONCE(dentry->d_parent); READ_ONCE 679 fs/ext4/dir.c const struct inode *inode = READ_ONCE(parent->d_inode); READ_ONCE 695 fs/ext4/dir.c const struct inode *inode = READ_ONCE(dentry->d_inode); READ_ONCE 782 fs/ext4/inode.c old_state = READ_ONCE(bh->b_state); READ_ONCE 2576 fs/ext4/inode.c if (disksize > READ_ONCE(EXT4_I(inode)->i_disksize)) { READ_ONCE 5143 fs/ext4/inode.c u64 i_blocks = READ_ONCE(inode->i_blocks); READ_ONCE 2750 fs/f2fs/data.c unsigned i_blkbits = READ_ONCE(inode->i_blkbits); READ_ONCE 1071 fs/f2fs/dir.c const struct dentry *parent = READ_ONCE(dentry->d_parent); READ_ONCE 1072 fs/f2fs/dir.c const struct inode *inode = READ_ONCE(parent->d_inode); READ_ONCE 1087 fs/f2fs/dir.c const struct inode *inode = READ_ONCE(dentry->d_inode); READ_ONCE 3696 fs/f2fs/f2fs.h unsigned int i_blkbits = READ_ONCE(inode->i_blkbits); READ_ONCE 732 fs/fcntl.c int signum = READ_ONCE(fown->signum); READ_ONCE 76 fs/fs_pin.c p = READ_ONCE(m->mnt_pins.first); READ_ONCE 90 fs/fs_pin.c q = READ_ONCE(p->first); READ_ONCE 43 fs/fscache/cookie.c object = READ_ONCE(cookie->backing_objects.first); READ_ONCE 177 fs/fscache/operation.c flags = READ_ONCE(object->flags); READ_ONCE 264 fs/fscache/operation.c flags = READ_ONCE(object->flags); READ_ONCE 113 fs/fuse/control.c val = READ_ONCE(fc->max_background); READ_ONCE 155 fs/fuse/control.c val = READ_ONCE(fc->congestion_threshold); READ_ONCE 40 fs/fuse/dev.c return READ_ONCE(file->private_data); READ_ONCE 97 fs/fuse/dev.c !READ_ONCE(fc->connected)) { READ_ONCE 981 fs/fuse/dir.c else if (request_mask & READ_ONCE(fi->inval_mask)) READ_ONCE 1169 fs/fuse/dir.c if (perm_mask & READ_ONCE(fi->inval_mask) || READ_ONCE 481 fs/io_uring.c if (ctx->cached_cq_tail != READ_ONCE(rings->cq.tail)) { READ_ONCE 563 fs/io_uring.c if (tail - READ_ONCE(rings->cq.head) == rings->cq_ring_entries) READ_ONCE 751 fs/io_uring.c return READ_ONCE(rings->cq.tail) - READ_ONCE(rings->cq.head); READ_ONCE 1102 fs/io_uring.c kiocb->ki_pos = READ_ONCE(sqe->off); READ_ONCE 1106 fs/io_uring.c ioprio = READ_ONCE(sqe->ioprio); READ_ONCE 1116 fs/io_uring.c ret = kiocb_set_rw_flags(kiocb, READ_ONCE(sqe->rw_flags)); READ_ONCE 1169 fs/io_uring.c size_t len = READ_ONCE(sqe->len); READ_ONCE 1179 fs/io_uring.c buf_index = READ_ONCE(sqe->buf_index); READ_ONCE 1185 fs/io_uring.c buf_addr = READ_ONCE(sqe->addr); READ_ONCE 1244 fs/io_uring.c void __user *buf = u64_to_user_ptr(READ_ONCE(sqe->addr)); READ_ONCE 1245 fs/io_uring.c size_t sqe_len = READ_ONCE(sqe->len); READ_ONCE 1256 fs/io_uring.c opcode = READ_ONCE(sqe->opcode); READ_ONCE 1574 fs/io_uring.c loff_t sqe_off = READ_ONCE(sqe->off); READ_ONCE 1575 fs/io_uring.c loff_t sqe_len = READ_ONCE(sqe->len); READ_ONCE 1580 fs/io_uring.c fsync_flags = READ_ONCE(sqe->fsync_flags); READ_ONCE 1636 fs/io_uring.c sqe_off = READ_ONCE(sqe->off); READ_ONCE 1637 fs/io_uring.c sqe_len = READ_ONCE(sqe->len); READ_ONCE 1638 fs/io_uring.c flags = READ_ONCE(sqe->sync_range_flags); READ_ONCE 1666 fs/io_uring.c flags = READ_ONCE(sqe->msg_flags); READ_ONCE 1678 fs/io_uring.c READ_ONCE(sqe->addr); READ_ONCE 1768 fs/io_uring.c if (READ_ONCE(sqe->addr) == poll_req->user_data) { READ_ONCE 1800 fs/io_uring.c if (!READ_ONCE(poll->canceled)) READ_ONCE 1811 fs/io_uring.c if (!mask && !READ_ONCE(poll->canceled)) { READ_ONCE 1895 fs/io_uring.c events = READ_ONCE(sqe->poll_events); READ_ONCE 2001 fs/io_uring.c count = READ_ONCE(sqe->off); READ_ONCE 2097 fs/io_uring.c req->user_data = READ_ONCE(s->sqe->user_data); READ_ONCE 2102 fs/io_uring.c opcode = READ_ONCE(s->sqe->opcode); READ_ONCE 2184 fs/io_uring.c u8 opcode = READ_ONCE(sqe->opcode); READ_ONCE 2360 fs/io_uring.c int op = READ_ONCE(sqe->opcode); READ_ONCE 2378 fs/io_uring.c flags = READ_ONCE(s->sqe->flags); READ_ONCE 2379 fs/io_uring.c fd = READ_ONCE(s->sqe->fd); READ_ONCE 2553 fs/io_uring.c switch (READ_ONCE(s->sqe->opcode)) { READ_ONCE 2627 fs/io_uring.c if (ctx->cached_sq_head != READ_ONCE(rings->sq.head)) { READ_ONCE 2664 fs/io_uring.c head = READ_ONCE(sq_array[head & ctx->sq_mask]); READ_ONCE 3020 fs/io_uring.c return READ_ONCE(rings->cq.head) == READ_ONCE(rings->cq.tail) ? ret : 0; READ_ONCE 3662 fs/io_uring.c if (READ_ONCE(ctx->rings->sq.tail) - ctx->cached_sq_head != READ_ONCE 3665 fs/io_uring.c if (READ_ONCE(ctx->rings->cq.head) != ctx->cached_cq_tail) READ_ONCE 53 fs/iomap/direct-io.c struct request_queue *q = READ_ONCE(kiocb->private); READ_ONCE 57 fs/iomap/direct-io.c return blk_poll(q, READ_ONCE(kiocb->ki_cookie), spin); READ_ONCE 549 fs/iomap/direct-io.c if (!READ_ONCE(dio->submit.waiter)) READ_ONCE 1003 fs/jbd2/commit.c mapping = READ_ONCE(bh->b_page->mapping); READ_ONCE 1076 fs/jbd2/transaction.c jh = READ_ONCE(bh->b_private); READ_ONCE 1082 fs/jbd2/transaction.c if (READ_ONCE(jh->b_transaction) != handle->h_transaction && READ_ONCE 1083 fs/jbd2/transaction.c READ_ONCE(jh->b_next_transaction) != handle->h_transaction) READ_ONCE 1064 fs/namei.c res = READ_ONCE(inode->i_link); READ_ONCE 1223 fs/namei.c while (managed = READ_ONCE(path->dentry->d_flags), READ_ONCE 1408 fs/namei.c while (managed = READ_ONCE(path->dentry->d_flags), READ_ONCE 4729 fs/namei.c link = READ_ONCE(inode->i_link); READ_ONCE 324 fs/namespace.c while (READ_ONCE(mnt->mnt.mnt_flags) & MNT_WRITE_HOLD) READ_ONCE 1128 fs/namespace.c if (likely(READ_ONCE(mnt->mnt_ns))) { READ_ONCE 1957 fs/namespace.c unsigned int max = READ_ONCE(sysctl_mount_max); READ_ONCE 1264 fs/nfs/dir.c parent = READ_ONCE(dentry->d_parent); READ_ONCE 1269 fs/nfs/dir.c if (parent != READ_ONCE(dentry->d_parent)) READ_ONCE 170 fs/nfs/inode.c unsigned long cache_validity = READ_ONCE(NFS_I(inode)->cache_validity); READ_ONCE 181 fs/nfs/inode.c unsigned long cache_validity = READ_ONCE(NFS_I(inode)->cache_validity); READ_ONCE 822 fs/nfs/inode.c cache_validity = READ_ONCE(NFS_I(inode)->cache_validity); READ_ONCE 287 fs/nfs/nfs4proc.c cache_validity = READ_ONCE(NFS_I(inode)->cache_validity); READ_ONCE 238 fs/nfsd/filecache.c return filemap_check_wb_err(file->f_mapping, READ_ONCE(file->f_wb_err)); READ_ONCE 368 fs/notify/fanotify/fanotify.c conn = READ_ONCE(iter_info->marks[type]->connector); READ_ONCE 229 fs/notify/mark.c struct fsnotify_mark_connector *conn = READ_ONCE(mark->connector); READ_ONCE 427 fs/overlayfs/file.c old_iflags = READ_ONCE(inode->i_flags); READ_ONCE 120 fs/overlayfs/ovl_entry.h return READ_ONCE(oi->__upperdentry); READ_ONCE 840 fs/overlayfs/readdir.c realfile = READ_ONCE(od->upperfile); READ_ONCE 634 fs/pipe.c unsigned long soft_limit = READ_ONCE(pipe_user_pages_soft); READ_ONCE 641 fs/pipe.c unsigned long hard_limit = READ_ONCE(pipe_user_pages_hard); READ_ONCE 657 fs/pipe.c unsigned int max_size = READ_ONCE(pipe_max_size); READ_ONCE 497 fs/proc/array.c rsslim = READ_ONCE(sig->rlim[RLIMIT_RSS].rlim_cur); READ_ONCE 31 fs/proc_namespace.c event = READ_ONCE(ns->event); READ_ONCE 255 fs/splice.c unsigned int buffers = READ_ONCE(pipe->buffers); READ_ONCE 309 fs/userfaultfd.c _pmd = READ_ONCE(*pmd); READ_ONCE 398 fs/userfaultfd.c if (unlikely(READ_ONCE(ctx->released))) { READ_ONCE 494 fs/userfaultfd.c if (likely(must_wait && !READ_ONCE(ctx->released) && READ_ONCE 508 fs/userfaultfd.c while (!READ_ONCE(uwq.waken)) { READ_ONCE 516 fs/userfaultfd.c if (READ_ONCE(uwq.waken) || READ_ONCE 517 fs/userfaultfd.c READ_ONCE(ctx->released) || READ_ONCE 606 fs/userfaultfd.c if (READ_ONCE(ctx->released) || READ_ONCE 1712 fs/userfaultfd.c if (READ_ONCE(ctx->mmap_changing)) READ_ONCE 1768 fs/userfaultfd.c if (READ_ONCE(ctx->mmap_changing)) READ_ONCE 54 fs/verity/hash_algs.c tfm = READ_ONCE(alg->tfm); READ_ONCE 4499 fs/xfs/libxfs/xfs_bmap.c *seq = READ_ONCE(ifp->if_seq); READ_ONCE 4532 fs/xfs/libxfs/xfs_bmap.c *seq = READ_ONCE(ifp->if_seq); READ_ONCE 621 fs/xfs/libxfs/xfs_iext_tree.c WRITE_ONCE(ifp->if_seq, READ_ONCE(ifp->if_seq) + 1); READ_ONCE 446 fs/xfs/xfs_aops.c if (wpc->data_seq != READ_ONCE(ip->i_df.if_seq)) READ_ONCE 449 fs/xfs/xfs_aops.c wpc->cow_seq != READ_ONCE(ip->i_cowfp->if_seq)) READ_ONCE 544 fs/xfs/xfs_aops.c wpc->cow_seq = READ_ONCE(ip->i_cowfp->if_seq); READ_ONCE 567 fs/xfs/xfs_aops.c wpc->data_seq = READ_ONCE(ip->i_df.if_seq); READ_ONCE 610 fs/xfs/xfs_extent_busy.c if (busy_gen != READ_ONCE(pag->pagb_gen)) READ_ONCE 583 fs/xfs/xfs_log_priv.h cur_cycle = READ_ONCE(log->l_curr_cycle); READ_ONCE 585 fs/xfs/xfs_log_priv.h cur_block = READ_ONCE(log->l_curr_block); READ_ONCE 171 include/asm-generic/atomic.h #define atomic_read(v) READ_ONCE((v)->counter) READ_ONCE 131 include/asm-generic/barrier.h typeof(*p) ___p1 = READ_ONCE(*p); \ READ_ONCE 186 include/asm-generic/barrier.h typeof(*p) ___p1 = READ_ONCE(*p); \ READ_ONCE 234 include/asm-generic/barrier.h VAL = READ_ONCE(*__PTR); \ READ_ONCE 38 include/asm-generic/bitops/atomic.h if (READ_ONCE(*p) & mask) READ_ONCE 51 include/asm-generic/bitops/atomic.h if (!(READ_ONCE(*p) & mask)) READ_ONCE 25 include/asm-generic/bitops/lock.h if (READ_ONCE(*p) & mask) READ_ONCE 63 include/asm-generic/bitops/lock.h old = READ_ONCE(*p); READ_ONCE 119 include/asm-generic/percpu.h __ret = READ_ONCE(*raw_cpu_ptr(&(pcp))); \ READ_ONCE 11 include/asm-generic/preempt.h return READ_ONCE(current_thread_info()->preempt_count); READ_ONCE 60 include/clocksource/hyperv_timer.h sequence = READ_ONCE(tsc_pg->tsc_sequence); READ_ONCE 69 include/clocksource/hyperv_timer.h scale = READ_ONCE(tsc_pg->tsc_scale); READ_ONCE 70 include/clocksource/hyperv_timer.h offset = READ_ONCE(tsc_pg->tsc_offset); READ_ONCE 79 include/clocksource/hyperv_timer.h } while (READ_ONCE(tsc_pg->tsc_sequence) != sequence); READ_ONCE 1499 include/drm/drm_connector.h return READ_ONCE(connector->registration_state) == READ_ONCE 96 include/drm/spsc_queue.h node = READ_ONCE(queue->head); READ_ONCE 101 include/drm/spsc_queue.h next = READ_ONCE(node->next); READ_ONCE 112 include/drm/spsc_queue.h } while (unlikely(!(queue->head = READ_ONCE(node->next)))); READ_ONCE 56 include/linux/average.h unsigned long internal = READ_ONCE(e->internal); \ READ_ONCE 252 include/linux/bitops.h old__ = READ_ONCE(*(ptr)); \ READ_ONCE 267 include/linux/bitops.h old__ = READ_ONCE(*(ptr)); \ READ_ONCE 550 include/linux/bpf.h while ((_prog = READ_ONCE(_item->prog))) { \ READ_ONCE 595 include/linux/bpf.h while ((_prog = READ_ONCE(_item->prog))) { \ READ_ONCE 841 include/linux/cgroup-defs.h struct sock_cgroup_data skcd_buf = {{ .val = READ_ONCE(skcd->val) }}; READ_ONCE 858 include/linux/cgroup-defs.h struct sock_cgroup_data skcd_buf = {{ .val = READ_ONCE(skcd->val) }}; READ_ONCE 836 include/linux/cgroup.h v = READ_ONCE(skcd->val); READ_ONCE 527 include/linux/dcache.h return READ_ONCE(dentry->d_inode); READ_ONCE 209 include/linux/dma-resv.h return READ_ONCE(obj->lock.ctx); READ_ONCE 92 include/linux/dynamic_queue_limits.h return READ_ONCE(dql->adj_limit) - READ_ONCE(dql->num_queued); READ_ONCE 82 include/linux/fscrypt.h return READ_ONCE(inode->i_crypt_info) != NULL; READ_ONCE 114 include/linux/fsverity.h return READ_ONCE(inode->i_verity_info); READ_ONCE 470 include/linux/hrtimer.h return !!(READ_ONCE(timer->state) & HRTIMER_STATE_ENQUEUED); READ_ONCE 252 include/linux/huge_mm.h return READ_ONCE(huge_zero_page) == page; READ_ONCE 145 include/linux/hyperv.h write_loc = READ_ONCE(rbi->ring_buffer->write_index); READ_ONCE 158 include/linux/hyperv.h read_loc = READ_ONCE(rbi->ring_buffer->read_index); READ_ONCE 68 include/linux/idr.h return READ_ONCE(idr->idr_next); READ_ONCE 265 include/linux/if_team.h int en_port_count = READ_ONCE(team->en_port_count); READ_ONCE 1212 include/linux/kvm_host.h return READ_ONCE(vcpu->requests); READ_ONCE 268 include/linux/list.h return READ_ONCE(head->next) == head; READ_ONCE 513 include/linux/list.h struct list_head *pos__ = READ_ONCE(head__->next); \ READ_ONCE 754 include/linux/list.h return !READ_ONCE(h->first); READ_ONCE 74 include/linux/list_bl.h return !((unsigned long)READ_ONCE(h->first) & ~LIST_BL_LOCKMASK); READ_ONCE 66 include/linux/list_nulls.h return is_a_nulls(READ_ONCE(h->first)); READ_ONCE 189 include/linux/llist.h return READ_ONCE(head->first) == NULL; READ_ONCE 366 include/linux/memcontrol.h return READ_ONCE(memcg->memory.emin); READ_ONCE 368 include/linux/memcontrol.h return max(READ_ONCE(memcg->memory.emin), READ_ONCE 369 include/linux/memcontrol.h READ_ONCE(memcg->memory.elow)); READ_ONCE 1307 include/linux/mm.h return READ_ONCE(page->mem_cgroup); READ_ONCE 552 include/linux/netdevice.h val = READ_ONCE(n->state); READ_ONCE 174 include/linux/page-flags.h unsigned long head = READ_ONCE(page->compound_head); READ_ONCE 183 include/linux/page-flags.h return READ_ONCE(page->compound_head) & 1; READ_ONCE 161 include/linux/percpu-refcount.h percpu_ptr = READ_ONCE(ref->percpu_count_ptr); READ_ONCE 82 include/linux/percpu_counter.h s64 ret = READ_ONCE(fbc->count); READ_ONCE 171 include/linux/ptr_ring.h return READ_ONCE(r->queue[r->consumer_head]); READ_ONCE 196 include/linux/ptr_ring.h return !r->queue[READ_ONCE(r->consumer_head)]; READ_ONCE 40 include/linux/rbtree.h #define RB_EMPTY_ROOT(root) (READ_ONCE((root)->rb_node) == NULL) READ_ONCE 36 include/linux/rcu_sync.h return !READ_ONCE(rsp->gp_state); /* GP_IDLE */ READ_ONCE 296 include/linux/rculist.h container_of(READ_ONCE(ptr), type, member) READ_ONCE 334 include/linux/rculist.h struct list_head *__next = READ_ONCE(__ptr->next); \ READ_ONCE 354 include/linux/rculist.h struct list_head *__next = READ_ONCE(__ptr->next); \ READ_ONCE 390 include/linux/rculist.h container_of((typeof(ptr))READ_ONCE(ptr), type, member) READ_ONCE 135 include/linux/rcupdate.h if (READ_ONCE((t)->rcu_tasks_holdout)) \ READ_ONCE 312 include/linux/rcupdate.h typeof(*p) *_________p1 = (typeof(*p) *__force)READ_ONCE(p); \ READ_ONCE 319 include/linux/rcupdate.h typeof(*p) *________p1 = (typeof(*p) *__force)READ_ONCE(p); \ READ_ONCE 333 include/linux/rcupdate.h typeof(p) ________p1 = READ_ONCE(p); \ READ_ONCE 889 include/linux/rcupdate.h rcu_callback_t func = READ_ONCE(rhp->func); READ_ONCE 1401 include/linux/sched.h unsigned int tsk_state = READ_ONCE(tsk->state); READ_ONCE 1817 include/linux/sched.h return READ_ONCE(p->cpu); READ_ONCE 1819 include/linux/sched.h return READ_ONCE(task_thread_info(p)->cpu); READ_ONCE 87 include/linux/sched/cputime.h if (!READ_ONCE(tsk->signal->posix_cputimers.timers_active)) READ_ONCE 687 include/linux/sched/signal.h return READ_ONCE(task->signal->rlim[limit].rlim_cur); READ_ONCE 693 include/linux/sched/signal.h return READ_ONCE(task->signal->rlim[limit].rlim_max); READ_ONCE 113 include/linux/seqlock.h ret = READ_ONCE(s->sequence); READ_ONCE 132 include/linux/seqlock.h unsigned ret = READ_ONCE(s->sequence); READ_ONCE 184 include/linux/seqlock.h unsigned ret = READ_ONCE(s->sequence); READ_ONCE 282 include/linux/seqlock.h int seq = READ_ONCE(s->sequence); /* ^^^ */ READ_ONCE 1509 include/linux/skbuff.h return READ_ONCE(list->next) == (const struct sk_buff *) list; READ_ONCE 1800 include/linux/skbuff.h struct sk_buff *skb = READ_ONCE(list_->prev); READ_ONCE 61 include/linux/slub_def.h #define slub_percpu_partial_read_once(c) READ_ONCE(slub_percpu_partial(c)) READ_ONCE 62 include/linux/srcutiny.h idx = READ_ONCE(ssp->srcu_idx); READ_ONCE 83 include/linux/srcutiny.h idx = READ_ONCE(ssp->srcu_idx) & 0x1; READ_ONCE 86 include/linux/srcutiny.h READ_ONCE(ssp->srcu_lock_nesting[!idx]), READ_ONCE 87 include/linux/srcutiny.h READ_ONCE(ssp->srcu_lock_nesting[idx])); READ_ONCE 457 include/linux/tcp.h int somaxconn = READ_ONCE(sock_net(sk)->core.sysctl_somaxconn); READ_ONCE 482 include/linux/tcp.h u16 user_mss = READ_ONCE(tp->rx_opt.user_mss); READ_ONCE 198 include/linux/tpm_eventlog.h count = READ_ONCE(event->count); READ_ONCE 199 include/linux/tpm_eventlog.h event_type = READ_ONCE(event->event_type); READ_ONCE 308 include/linux/trace_events.h return !!READ_ONCE(call->prog_array); READ_ONCE 746 include/linux/vmw_vmci_defs.h return READ_ONCE(*(unsigned long *)var); READ_ONCE 60 include/net/arp.h if (READ_ONCE(n->confirmed) != now) READ_ONCE 74 include/net/busy_poll.h unsigned long bp_usec = READ_ONCE(sysctl_net_busy_poll); READ_ONCE 90 include/net/busy_poll.h unsigned long bp_usec = READ_ONCE(sk->sk_ll_usec); READ_ONCE 105 include/net/busy_poll.h unsigned int napi_id = READ_ONCE(sk->sk_napi_id); READ_ONCE 135 include/net/busy_poll.h if (!READ_ONCE(sk->sk_napi_id)) READ_ONCE 414 include/net/ip.h u8 pmtudisc = READ_ONCE(inet_sk(sk)->pmtudisc); READ_ONCE 448 include/net/ip.h return min(READ_ONCE(dst->dev->mtu), IP_MAX_MTU); READ_ONCE 460 include/net/ip.h return min(READ_ONCE(skb_dst(skb)->dev->mtu), IP_MAX_MTU); READ_ONCE 988 include/net/ip_vs.h return READ_ONCE(ipvs->sysctl_sync_threshold[1]); READ_ONCE 993 include/net/ip_vs.h return READ_ONCE(ipvs->sysctl_sync_refresh_period); READ_ONCE 1018 include/net/ip_vs.h return READ_ONCE(ipvs->sysctl_sync_ports); READ_ONCE 417 include/net/ndisc.h if (READ_ONCE(n->confirmed) != now) READ_ONCE 434 include/net/ndisc.h if (READ_ONCE(n->confirmed) != now) READ_ONCE 441 include/net/neighbour.h if (READ_ONCE(neigh->used) != now) READ_ONCE 470 include/net/neighbour.h hh_len = READ_ONCE(hh->hh_len); READ_ONCE 1265 include/net/netfilter/nf_tables.h return 1 << READ_ONCE(net->nft.gencursor); READ_ONCE 188 include/net/request_sock.h return READ_ONCE(queue->rskq_accept_head) == NULL; READ_ONCE 152 include/net/sch_generic.h return READ_ONCE(qdisc->empty); READ_ONCE 153 include/net/sch_generic.h return !READ_ONCE(qdisc->q.qlen); READ_ONCE 543 include/net/sock.h return READ_ONCE(sk->sk_peek_off); READ_ONCE 551 include/net/sock.h s32 off = READ_ONCE(sk->sk_peek_off); READ_ONCE 886 include/net/sock.h return READ_ONCE(sk->sk_wmem_queued) >> 1; READ_ONCE 891 include/net/sock.h return READ_ONCE(sk->sk_sndbuf) - READ_ONCE(sk->sk_wmem_queued); READ_ONCE 962 include/net/sock.h if (unlikely(READ_ONCE(sk->sk_incoming_cpu) != cpu)) READ_ONCE 1041 include/net/sock.h if (unlikely(READ_ONCE(sk->sk_backlog.tail))) { READ_ONCE 1220 include/net/sock.h if (READ_ONCE(sk->sk_wmem_queued) >= READ_ONCE(sk->sk_sndbuf)) READ_ONCE 1948 include/net/sock.h if (!READ_ONCE(sk->sk_dst_pending_confirm)) READ_ONCE 1959 include/net/sock.h if (READ_ONCE(n->confirmed) != now) READ_ONCE 1961 include/net/sock.h if (sk && READ_ONCE(sk->sk_dst_pending_confirm)) READ_ONCE 2273 include/net/sock.h return refcount_read(&sk->sk_wmem_alloc) < (READ_ONCE(sk->sk_sndbuf) >> 1); READ_ONCE 2293 include/net/sock.h int v = waitall ? len : min_t(int, READ_ONCE(sk->sk_rcvlowat), len); READ_ONCE 2350 include/net/sock.h return READ_ONCE(sk->sk_stamp); READ_ONCE 2592 include/net/sock.h if (!sk || !sk_fullsock(sk) || READ_ONCE(sk->sk_pacing_shift) == val) READ_ONCE 59 include/net/tc_act/tc_gact.h return READ_ONCE(a->tcfa_action) & TC_ACT_EXT_VAL_MASK; READ_ONCE 261 include/net/tcp.h return READ_ONCE(tcp_memory_pressure); READ_ONCE 496 include/net/tcp.h last_overflow = READ_ONCE(reuse->synq_overflow_ts); READ_ONCE 504 include/net/tcp.h last_overflow = READ_ONCE(tcp_sk(sk)->rx_opt.ts_recent_stamp); READ_ONCE 520 include/net/tcp.h last_overflow = READ_ONCE(reuse->synq_overflow_ts); READ_ONCE 527 include/net/tcp.h last_overflow = READ_ONCE(tcp_sk(sk)->rx_opt.ts_recent_stamp); READ_ONCE 1394 include/net/tcp.h return tcp_win_from_space(sk, READ_ONCE(sk->sk_rcvbuf) - READ_ONCE 1395 include/net/tcp.h READ_ONCE(sk->sk_backlog.len) - READ_ONCE 1401 include/net/tcp.h return tcp_win_from_space(sk, READ_ONCE(sk->sk_rcvbuf)); READ_ONCE 1411 include/net/tcp.h int rcvbuf = READ_ONCE(sk->sk_rcvbuf); READ_ONCE 1944 include/net/tcp.h u32 notsent_bytes = READ_ONCE(tp->write_seq) - READ_ONCE 1945 include/net/tcp.h READ_ONCE(tp->snd_nxt); READ_ONCE 427 include/net/tls.h return READ_ONCE(rec->tx_ready); READ_ONCE 248 include/net/udp.h return sk_rmem_alloc_get(sk) - READ_ONCE(udp_sk(sk)->forward_deficit); READ_ONCE 85 include/trace/events/sock.h __entry->sk_rcvbuf = READ_ONCE(sk->sk_rcvbuf); READ_ONCE 118 include/trace/events/sock.h __entry->wmem_queued = READ_ONCE(sk->sk_wmem_queued); READ_ONCE 13 include/vdso/helpers.h while ((seq = READ_ONCE(vd->seq)) & 1) READ_ONCE 26 include/vdso/helpers.h seq = READ_ONCE(vd->seq); READ_ONCE 1186 ipc/msg.c msg = READ_ONCE(msr_d.r_msg); READ_ONCE 2174 ipc/sem.c error = READ_ONCE(queue.status); READ_ONCE 2192 ipc/sem.c error = READ_ONCE(queue.status); READ_ONCE 150 kernel/acct.c res = to_acct(READ_ONCE(ns->bacct)); READ_ONCE 162 kernel/acct.c if (res != to_acct(READ_ONCE(ns->bacct))) { READ_ONCE 1449 kernel/audit.c t = READ_ONCE(current->signal->audit_tty); READ_ONCE 1470 kernel/audit.c t = READ_ONCE(current->signal->audit_tty); READ_ONCE 255 kernel/audit_tree.c if (READ_ONCE(p->key) == key) { READ_ONCE 515 kernel/bpf/arraymap.c if (elem && (ptr = READ_ONCE(*elem))) READ_ONCE 615 kernel/bpf/arraymap.c ptr = READ_ONCE(*elem); READ_ONCE 711 kernel/bpf/arraymap.c ee = READ_ONCE(array->ptrs[i]); READ_ONCE 800 kernel/bpf/arraymap.c return READ_ONCE(*inner_map); READ_ONCE 1471 kernel/bpf/core.c prog = READ_ONCE(array->ptrs[index]); READ_ONCE 543 kernel/bpf/cpumap.c rcpu = READ_ONCE(cmap->cpu_map[i]); READ_ONCE 563 kernel/bpf/cpumap.c rcpu = READ_ONCE(cmap->cpu_map[key]); READ_ONCE 429 kernel/bpf/devmap.c obj = READ_ONCE(dtab->netdev_map[key]); READ_ONCE 791 kernel/bpf/devmap.c dev = READ_ONCE(dtab->netdev_map[i]); READ_ONCE 1424 kernel/bpf/hashtab.c *value = map->ops->map_fd_sys_lookup_elem(READ_ONCE(*ptr)); READ_ONCE 1477 kernel/bpf/hashtab.c return READ_ONCE(*inner_map); READ_ONCE 346 kernel/bpf/helpers.c ptr = &READ_ONCE(storage->buf)->data[0]; READ_ONCE 124 kernel/bpf/local_storage.c return &READ_ONCE(storage->buf)->data[0]; READ_ONCE 396 kernel/bpf/local_storage.c &READ_ONCE(storage->buf)->data[0], m); READ_ONCE 237 kernel/bpf/reuseport_array.c if (READ_ONCE(nsk->sk_user_data)) READ_ONCE 387 kernel/bpf/stackmap.c bucket = READ_ONCE(smap->buckets[id]); READ_ONCE 366 kernel/bpf/syscall.c if (READ_ONCE(map->frozen)) READ_ONCE 401 kernel/bpf/syscall.c READ_ONCE(map->frozen)); READ_ONCE 1180 kernel/bpf/syscall.c if (READ_ONCE(map->frozen)) { READ_ONCE 171 kernel/bpf/xskmap.c xs = READ_ONCE(m->xsk_map[key]); READ_ONCE 254 kernel/bpf/xskmap.c old_xs = READ_ONCE(*map_entry); READ_ONCE 303 kernel/bpf/xskmap.c if (READ_ONCE(*map_entry) == xs) { READ_ONCE 3466 kernel/cgroup/cgroup.c int descendants = READ_ONCE(cgrp->max_descendants); READ_ONCE 3509 kernel/cgroup/cgroup.c int depth = READ_ONCE(cgrp->max_depth); READ_ONCE 176 kernel/events/core.c return READ_ONCE(event->owner) == TASK_TOMBSTONE; READ_ONCE 256 kernel/events/core.c struct task_struct *task = READ_ONCE(ctx->task); /* verified in event_function */ READ_ONCE 310 kernel/events/core.c struct task_struct *task = READ_ONCE(ctx->task); READ_ONCE 512 kernel/events/core.c u64 max_len = READ_ONCE(perf_sample_allowed_ns); READ_ONCE 1276 kernel/events/core.c ctx = READ_ONCE(event->ctx); READ_ONCE 2102 kernel/events/core.c if (READ_ONCE(event->pending_disable) >= 0) { READ_ONCE 2659 kernel/events/core.c struct task_struct *task = READ_ONCE(ctx->task); READ_ONCE 2845 kernel/events/core.c if (READ_ONCE(event->state) != PERF_EVENT_STATE_ACTIVE) READ_ONCE 2855 kernel/events/core.c if (READ_ONCE(event->oncpu) != smp_processor_id()) READ_ONCE 2884 kernel/events/core.c if (READ_ONCE(event->state) != PERF_EVENT_STATE_ACTIVE) READ_ONCE 2895 kernel/events/core.c ret = cpu_function_call(READ_ONCE(event->oncpu), READ_ONCE 4093 kernel/events/core.c enum perf_event_state state = READ_ONCE(event->state); READ_ONCE 4112 kernel/events/core.c event_cpu = READ_ONCE(event->oncpu); READ_ONCE 4624 kernel/events/core.c owner = READ_ONCE(event->owner); READ_ONCE 4722 kernel/events/core.c ctx = READ_ONCE(child->ctx); READ_ONCE 5751 kernel/events/core.c aux_offset = READ_ONCE(rb->user_page->aux_offset); READ_ONCE 5752 kernel/events/core.c aux_size = READ_ONCE(rb->user_page->aux_size); READ_ONCE 5975 kernel/events/core.c int cpu = READ_ONCE(event->pending_disable); READ_ONCE 6998 kernel/events/core.c cpu = READ_ONCE(iter->oncpu); READ_ONCE 8211 kernel/events/core.c READ_ONCE(event->overflow_handler)(event, data, regs); READ_ONCE 9015 kernel/events/core.c event->orig_overflow_handler = READ_ONCE(event->overflow_handler); READ_ONCE 9236 kernel/events/core.c struct task_struct *task = READ_ONCE(event->ctx->task); READ_ONCE 10872 kernel/events/core.c gctx = READ_ONCE(group_leader->ctx); READ_ONCE 60 kernel/events/ring_buffer.c nest = READ_ONCE(rb->nest); READ_ONCE 192 kernel/events/ring_buffer.c tail = READ_ONCE(rb->user_page->data_tail); READ_ONCE 396 kernel/events/ring_buffer.c nest = READ_ONCE(rb->aux_nest); READ_ONCE 420 kernel/events/ring_buffer.c aux_tail = READ_ONCE(rb->user_page->aux_tail); READ_ONCE 1545 kernel/events/uprobes.c area = READ_ONCE(mm->uprobes_state.xol_area); /* ^^^ */ READ_ONCE 1847 kernel/events/uprobes.c area = READ_ONCE(current->mm->uprobes_state.xol_area); /* ^^^ */ READ_ONCE 1265 kernel/exit.c int exit_state = READ_ONCE(p->exit_state); READ_ONCE 1539 kernel/fork.c cpu_limit = READ_ONCE(sig->rlim[RLIMIT_CPU].rlim_cur); READ_ONCE 647 kernel/futex.c mapping = READ_ONCE(page->mapping); READ_ONCE 723 kernel/futex.c if (READ_ONCE(page->mapping) != mapping) { READ_ONCE 730 kernel/futex.c inode = READ_ONCE(mapping->host); READ_ONCE 2407 kernel/futex.c lock_ptr = READ_ONCE(q->lock_ptr); READ_ONCE 70 kernel/kcov.c mode = READ_ONCE(t->kcov_mode); READ_ONCE 107 kernel/kcov.c pos = READ_ONCE(area[0]) + 1; READ_ONCE 135 kernel/kcov.c count = READ_ONCE(area[0]); READ_ONCE 340 kernel/kcov.c READ_ONCE(area[offset]); READ_ONCE 154 kernel/ksysfs.c return sprintf(buf, "%d\n", READ_ONCE(rcu_expedited)); READ_ONCE 171 kernel/ksysfs.c return sprintf(buf, "%d\n", READ_ONCE(rcu_normal)); READ_ONCE 184 kernel/livepatch/transition.c task->patch_state = READ_ONCE(klp_target_state); READ_ONCE 697 kernel/locking/lockdep.c int i, depth = READ_ONCE(p->lockdep_depth); READ_ONCE 2865 kernel/locking/lockdep.c if (READ_ONCE(chain->chain_key) == chain_key) { READ_ONCE 104 kernel/locking/mcs_spinlock.h struct mcs_spinlock *next = READ_ONCE(node->next); READ_ONCE 113 kernel/locking/mcs_spinlock.h while (!(next = READ_ONCE(node->next))) READ_ONCE 520 kernel/locking/mutex.c if (ww_ctx->acquired > 0 && READ_ONCE(ww->ctx)) READ_ONCE 809 kernel/locking/mutex.c struct ww_acquire_ctx *hold_ctx = READ_ONCE(ww->ctx); READ_ONCE 943 kernel/locking/mutex.c if (unlikely(ww_ctx == READ_ONCE(ww->ctx))) READ_ONCE 137 kernel/locking/osq_lock.c while (!READ_ONCE(node->locked)) { READ_ONCE 178 kernel/locking/osq_lock.c prev = READ_ONCE(node->prev); READ_ONCE 481 kernel/locking/qspinlock.c next = READ_ONCE(node->next); READ_ONCE 121 kernel/locking/qspinlock_paravirt.h return !READ_ONCE(lock->locked) && READ_ONCE 246 kernel/locking/qspinlock_paravirt.h if (READ_ONCE(he->lock) == lock) { READ_ONCE 247 kernel/locking/qspinlock_paravirt.h node = READ_ONCE(he->node); READ_ONCE 272 kernel/locking/qspinlock_paravirt.h return READ_ONCE(prev->state) != vcpu_running; READ_ONCE 302 kernel/locking/qspinlock_paravirt.h if (READ_ONCE(node->locked)) READ_ONCE 322 kernel/locking/qspinlock_paravirt.h if (!READ_ONCE(node->locked)) { READ_ONCE 343 kernel/locking/qspinlock_paravirt.h !READ_ONCE(node->locked)); READ_ONCE 414 kernel/locking/qspinlock_paravirt.h if (READ_ONCE(pn->state) == vcpu_hashed) READ_ONCE 134 kernel/locking/rtmutex.c owner = READ_ONCE(*p); READ_ONCE 110 kernel/locking/rtmutex_common.h unsigned long owner = (unsigned long) READ_ONCE(lock->owner); READ_ONCE 54 kernel/locking/spinlock_debug.c struct task_struct *owner = READ_ONCE(lock->owner); READ_ONCE 63 kernel/locking/spinlock_debug.c lock, READ_ONCE(lock->magic), READ_ONCE 66 kernel/locking/spinlock_debug.c READ_ONCE(lock->owner_cpu)); READ_ONCE 83 kernel/locking/spinlock_debug.c SPIN_BUG_ON(READ_ONCE(lock->magic) != SPINLOCK_MAGIC, lock, "bad magic"); READ_ONCE 84 kernel/locking/spinlock_debug.c SPIN_BUG_ON(READ_ONCE(lock->owner) == current, lock, "recursion"); READ_ONCE 85 kernel/locking/spinlock_debug.c SPIN_BUG_ON(READ_ONCE(lock->owner_cpu) == raw_smp_processor_id(), READ_ONCE 47 kernel/pid_namespace.c kc = READ_ONCE(*pkc); READ_ONCE 59 kernel/pid_namespace.c return READ_ONCE(*pkc); READ_ONCE 181 kernel/power/energy_model.c return READ_ONCE(per_cpu(em_data, cpu)); READ_ONCE 217 kernel/power/energy_model.c if (READ_ONCE(per_cpu(em_data, cpu))) { READ_ONCE 1705 kernel/printk/printk.c waiter = READ_ONCE(console_waiter); READ_ONCE 1750 kernel/printk/printk.c owner = READ_ONCE(console_owner); READ_ONCE 1751 kernel/printk/printk.c waiter = READ_ONCE(console_waiter); READ_ONCE 1775 kernel/printk/printk.c while (READ_ONCE(console_waiter)) READ_ONCE 91 kernel/rcu/rcu.h s = (READ_ONCE(*sp) + 2 * RCU_SEQ_STATE_MASK + 1) & ~RCU_SEQ_STATE_MASK; READ_ONCE 99 kernel/rcu/rcu.h return READ_ONCE(*sp); READ_ONCE 108 kernel/rcu/rcu.h return ULONG_CMP_LT((s - 1) & ~RCU_SEQ_STATE_MASK, READ_ONCE(*sp)); READ_ONCE 117 kernel/rcu/rcu.h return ULONG_CMP_GE(READ_ONCE(*sp), s); READ_ONCE 15 kernel/rcu/rcu_segcblist.h return READ_ONCE(rclp->len); READ_ONCE 49 kernel/rcu/rcu_segcblist.h return !READ_ONCE(rsclp->head); READ_ONCE 58 kernel/rcu/rcu_segcblist.h return READ_ONCE(rsclp->len); READ_ONCE 96 kernel/rcu/rcu_segcblist.h return !READ_ONCE(*READ_ONCE(rsclp->tails[seg])); READ_ONCE 204 kernel/rcu/rcutorture.c unsigned int i = READ_ONCE(rcu_torture_writer_state); READ_ONCE 1067 kernel/rcu/rcutorture.c !READ_ONCE(rcu_fwd_cb_nodelay) && READ_ONCE 1484 kernel/rcu/rcutorture.c wtp = READ_ONCE(writer_task); READ_ONCE 1648 kernel/rcu/rcutorture.c if (READ_ONCE(fcsp->stop)) { READ_ONCE 1792 kernel/rcu/rcutorture.c cver = READ_ONCE(rcu_torture_current_version); READ_ONCE 1801 kernel/rcu/rcutorture.c !READ_ONCE(rcu_fwd_emergency_stop) && !torture_must_stop()) { READ_ONCE 1811 kernel/rcu/rcutorture.c !READ_ONCE(rcu_fwd_emergency_stop) && !torture_must_stop()) { READ_ONCE 1813 kernel/rcu/rcutorture.c cver = READ_ONCE(rcu_torture_current_version) - cver; READ_ONCE 1825 kernel/rcu/rcutorture.c WARN_ON(READ_ONCE(fcs.stop) != 2); READ_ONCE 1848 kernel/rcu/rcutorture.c if (READ_ONCE(rcu_fwd_emergency_stop)) READ_ONCE 1865 kernel/rcu/rcutorture.c cver = READ_ONCE(rcu_torture_current_version); READ_ONCE 1870 kernel/rcu/rcutorture.c !READ_ONCE(rcu_fwd_emergency_stop) && !torture_must_stop()) { READ_ONCE 1871 kernel/rcu/rcutorture.c rfcp = READ_ONCE(rcu_fwd_cb_head); READ_ONCE 1874 kernel/rcu/rcutorture.c rfcpn = READ_ONCE(rfcp->rfc_next); READ_ONCE 1896 kernel/rcu/rcutorture.c n_launders_cb_snap = READ_ONCE(n_launders_cb); READ_ONCE 1897 kernel/rcu/rcutorture.c cver = READ_ONCE(rcu_torture_current_version) - cver; READ_ONCE 1902 kernel/rcu/rcutorture.c if (!torture_must_stop() && !READ_ONCE(rcu_fwd_emergency_stop) && READ_ONCE 1928 kernel/rcu/rcutorture.c rcu_fwd_progress_check(1 + (jiffies - READ_ONCE(rcu_fwd_startat)) / 2); READ_ONCE 99 kernel/rcu/srcutiny.c if (!newval && READ_ONCE(ssp->srcu_gp_waiting)) READ_ONCE 117 kernel/rcu/srcutiny.c if (ssp->srcu_gp_running || !READ_ONCE(ssp->srcu_cb_head)) READ_ONCE 130 kernel/rcu/srcutiny.c swait_event_exclusive(ssp->srcu_wq, !READ_ONCE(ssp->srcu_lock_nesting[idx])); READ_ONCE 149 kernel/rcu/srcutiny.c if (READ_ONCE(ssp->srcu_cb_head)) READ_ONCE 169 kernel/rcu/srcutiny.c if (!READ_ONCE(ssp->srcu_gp_running)) { READ_ONCE 255 kernel/rcu/srcutree.c sum += READ_ONCE(cpuc->srcu_lock_count[idx]); READ_ONCE 272 kernel/rcu/srcutree.c sum += READ_ONCE(cpuc->srcu_unlock_count[idx]); READ_ONCE 341 kernel/rcu/srcutree.c sum += READ_ONCE(cpuc->srcu_lock_count[0]); READ_ONCE 342 kernel/rcu/srcutree.c sum += READ_ONCE(cpuc->srcu_lock_count[1]); READ_ONCE 343 kernel/rcu/srcutree.c sum -= READ_ONCE(cpuc->srcu_unlock_count[0]); READ_ONCE 344 kernel/rcu/srcutree.c sum -= READ_ONCE(cpuc->srcu_unlock_count[1]); READ_ONCE 357 kernel/rcu/srcutree.c if (ULONG_CMP_LT(READ_ONCE(ssp->srcu_gp_seq), READ_ONCE 358 kernel/rcu/srcutree.c READ_ONCE(ssp->srcu_gp_seq_needed_exp))) READ_ONCE 387 kernel/rcu/srcutree.c if (WARN_ON(rcu_seq_state(READ_ONCE(ssp->srcu_gp_seq)) != SRCU_STATE_IDLE) || READ_ONCE 390 kernel/rcu/srcutree.c __func__, ssp, rcu_seq_state(READ_ONCE(ssp->srcu_gp_seq))); READ_ONCE 407 kernel/rcu/srcutree.c idx = READ_ONCE(ssp->srcu_idx) & 0x1; READ_ONCE 453 kernel/rcu/srcutree.c state = rcu_seq_state(READ_ONCE(ssp->srcu_gp_seq)); READ_ONCE 605 kernel/rcu/srcutree.c ULONG_CMP_GE(READ_ONCE(snp->srcu_gp_seq_needed_exp), s)) READ_ONCE 784 kernel/rcu/srcutree.c tlast = READ_ONCE(ssp->srcu_last_gp_end); READ_ONCE 792 kernel/rcu/srcutree.c if (ULONG_CMP_LT(curseq, READ_ONCE(ssp->srcu_gp_seq_needed))) READ_ONCE 1116 kernel/rcu/srcutree.c idx = rcu_seq_state(READ_ONCE(ssp->srcu_gp_seq)); READ_ONCE 1126 kernel/rcu/srcutree.c if (rcu_seq_state(READ_ONCE(ssp->srcu_gp_seq)) == SRCU_STATE_SCAN1) { READ_ONCE 1136 kernel/rcu/srcutree.c if (rcu_seq_state(READ_ONCE(ssp->srcu_gp_seq)) == SRCU_STATE_SCAN2) { READ_ONCE 78 kernel/rcu/sync.c WARN_ON_ONCE(READ_ONCE(rsp->gp_state) == GP_IDLE); READ_ONCE 79 kernel/rcu/sync.c WARN_ON_ONCE(READ_ONCE(rsp->gp_state) == GP_PASSED); READ_ONCE 155 kernel/rcu/sync.c wait_event(rsp->gp_wait, READ_ONCE(rsp->gp_state) >= GP_PASSED); READ_ONCE 170 kernel/rcu/sync.c WARN_ON_ONCE(READ_ONCE(rsp->gp_state) == GP_IDLE); READ_ONCE 171 kernel/rcu/sync.c WARN_ON_ONCE(READ_ONCE(rsp->gp_count) == 0); READ_ONCE 193 kernel/rcu/sync.c WARN_ON_ONCE(READ_ONCE(rsp->gp_count)); READ_ONCE 194 kernel/rcu/sync.c WARN_ON_ONCE(READ_ONCE(rsp->gp_state) == GP_PASSED); READ_ONCE 193 kernel/rcu/tree.c return READ_ONCE(rnp->qsmaskinitnext); READ_ONCE 452 kernel/rcu/tree.c j = READ_ONCE(jiffies_till_first_fqs) + READ_ONCE 453 kernel/rcu/tree.c 2 * READ_ONCE(jiffies_till_next_fqs); READ_ONCE 506 kernel/rcu/tree.c return READ_ONCE(rcu_state.gp_seq); READ_ONCE 548 kernel/rcu/tree.c *flags = READ_ONCE(rcu_state.gp_flags); READ_ONCE 1053 kernel/rcu/tree.c jtsq = READ_ONCE(jiffies_to_sched_qs); READ_ONCE 1056 kernel/rcu/tree.c if (!READ_ONCE(*rnhqp) && READ_ONCE 1076 kernel/rcu/tree.c READ_ONCE(rdp->last_fqs_resched) + jtsq * 3)) { READ_ONCE 1090 kernel/rcu/tree.c READ_ONCE(rdp->last_fqs_resched) + jtsq)) { READ_ONCE 1189 kernel/rcu/tree.c trace_rcu_grace_period(rcu_state.name, READ_ONCE(rcu_state.gp_seq), TPS("newreq")); READ_ONCE 1237 kernel/rcu/tree.c !READ_ONCE(rcu_state.gp_flags) || READ_ONCE 1241 kernel/rcu/tree.c WRITE_ONCE(rcu_state.gp_wake_seq, READ_ONCE(rcu_state.gp_seq)); READ_ONCE 1382 kernel/rcu/tree.c unlikely(READ_ONCE(rdp->gpwrap))) { READ_ONCE 1393 kernel/rcu/tree.c unlikely(READ_ONCE(rdp->gpwrap))) { READ_ONCE 1422 kernel/rcu/tree.c !unlikely(READ_ONCE(rdp->gpwrap))) || /* w/out lock. */ READ_ONCE 1454 kernel/rcu/tree.c if (!READ_ONCE(rcu_state.gp_flags)) { READ_ONCE 1580 kernel/rcu/tree.c *gfp = READ_ONCE(rcu_state.gp_flags); READ_ONCE 1585 kernel/rcu/tree.c if (!READ_ONCE(rnp->qsmask) && !rcu_preempt_blocked_readers_cgp(rnp)) READ_ONCE 1608 kernel/rcu/tree.c if (READ_ONCE(rcu_state.gp_flags) & RCU_GP_FLAG_FQS) { READ_ONCE 1611 kernel/rcu/tree.c READ_ONCE(rcu_state.gp_flags) & ~RCU_GP_FLAG_FQS); READ_ONCE 1628 kernel/rcu/tree.c j = READ_ONCE(jiffies_till_first_fqs); READ_ONCE 1637 kernel/rcu/tree.c READ_ONCE(rcu_state.gp_seq), READ_ONCE 1645 kernel/rcu/tree.c if (!READ_ONCE(rnp->qsmask) && READ_ONCE 1652 kernel/rcu/tree.c READ_ONCE(rcu_state.gp_seq), READ_ONCE 1657 kernel/rcu/tree.c READ_ONCE(rcu_state.gp_seq), READ_ONCE 1662 kernel/rcu/tree.c j = READ_ONCE(jiffies_till_next_fqs); READ_ONCE 1669 kernel/rcu/tree.c READ_ONCE(rcu_state.gp_seq), READ_ONCE 1761 kernel/rcu/tree.c READ_ONCE(rcu_state.gp_seq), READ_ONCE 1781 kernel/rcu/tree.c READ_ONCE(rcu_state.gp_seq), READ_ONCE 1785 kernel/rcu/tree.c READ_ONCE(rcu_state.gp_flags) & READ_ONCE 1795 kernel/rcu/tree.c READ_ONCE(rcu_state.gp_seq), READ_ONCE 1824 kernel/rcu/tree.c READ_ONCE(rcu_state.gp_flags) | RCU_GP_FLAG_FQS); READ_ONCE 2315 kernel/rcu/tree.c ret = (READ_ONCE(rcu_state.gp_flags) & RCU_GP_FLAG_FQS) || READ_ONCE 2328 kernel/rcu/tree.c if (READ_ONCE(rcu_state.gp_flags) & RCU_GP_FLAG_FQS) { READ_ONCE 2333 kernel/rcu/tree.c READ_ONCE(rcu_state.gp_flags) | RCU_GP_FLAG_FQS); READ_ONCE 2377 kernel/rcu/tree.c likely(READ_ONCE(rcu_scheduler_fully_active))) READ_ONCE 2824 kernel/rcu/tree.c unlikely(READ_ONCE(rdp->gpwrap))) /* outside lock */ READ_ONCE 3149 kernel/rcu/tree.c rdp->rcu_onl_gp_seq = READ_ONCE(rcu_state.gp_seq); READ_ONCE 3150 kernel/rcu/tree.c rdp->rcu_onl_gp_flags = READ_ONCE(rcu_state.gp_flags); READ_ONCE 3186 kernel/rcu/tree.c rdp->rcu_ofl_gp_seq = READ_ONCE(rcu_state.gp_seq); READ_ONCE 3187 kernel/rcu/tree.c rdp->rcu_ofl_gp_flags = READ_ONCE(rcu_state.gp_flags); READ_ONCE 156 kernel/rcu/tree_exp.h READ_ONCE(rnp->expmask) == 0; READ_ONCE 282 kernel/rcu/tree_exp.h if (ULONG_CMP_LT(READ_ONCE(rnp->exp_seq_rq), s) && READ_ONCE 284 kernel/rcu/tree_exp.h ULONG_CMP_LT(READ_ONCE(rnp_root->exp_seq_rq), s)) && READ_ONCE 431 kernel/rcu/tree_exp.h if (!READ_ONCE(rnp->expmask)) READ_ONCE 433 kernel/rcu/tree_exp.h if (!READ_ONCE(rcu_par_gp_wq) || READ_ONCE 492 kernel/rcu/tree_exp.h if (!(READ_ONCE(rnp->expmask) & mask)) READ_ONCE 504 kernel/rcu/tree_exp.h READ_ONCE(rnp_root->expmask), READ_ONCE 515 kernel/rcu/tree_exp.h READ_ONCE(rnp->expmask), READ_ONCE 523 kernel/rcu/tree_exp.h if (!(READ_ONCE(rnp->expmask) & mask)) READ_ONCE 552 kernel/rcu/tree_exp.h if (ULONG_CMP_LT(READ_ONCE(rnp->exp_seq_rq), s)) { READ_ONCE 717 kernel/rcu/tree_exp.h if (!(READ_ONCE(rnp->expmask) & rdp->grpmask) || READ_ONCE 740 kernel/rcu/tree_exp.h if (!(READ_ONCE(rnp->expmask) & rdp->grpmask) || READ_ONCE 343 kernel/rcu/tree_plugin.h return READ_ONCE(rnp->gp_tasks) != NULL; READ_ONCE 382 kernel/rcu/tree_plugin.h if (unlikely(READ_ONCE(t->rcu_read_unlock_special.s))) READ_ONCE 553 kernel/rcu/tree_plugin.h READ_ONCE(t->rcu_read_unlock_special.s)) && READ_ONCE 615 kernel/rcu/tree_plugin.h (rdp->grpmask & READ_ONCE(rnp->expmask)) || READ_ONCE 760 kernel/rcu/tree_plugin.h __func__, READ_ONCE(rnp->gp_tasks), rnp->boost_tasks, READ_ONCE 970 kernel/rcu/tree_plugin.h if (READ_ONCE(rnp->exp_tasks) == NULL && READ_ONCE 971 kernel/rcu/tree_plugin.h READ_ONCE(rnp->boost_tasks) == NULL) READ_ONCE 1019 kernel/rcu/tree_plugin.h return READ_ONCE(rnp->exp_tasks) != NULL || READ_ONCE 1020 kernel/rcu/tree_plugin.h READ_ONCE(rnp->boost_tasks) != NULL; READ_ONCE 1319 kernel/rcu/tree_plugin.h unlikely(READ_ONCE(rdp->gpwrap))) && READ_ONCE 1392 kernel/rcu/tree_plugin.h tne = READ_ONCE(tick_nohz_active); READ_ONCE 1650 kernel/rcu/tree_plugin.h if (!READ_ONCE(rdp_gp->nocb_gp_kthread)) { READ_ONCE 1659 kernel/rcu/tree_plugin.h if (force || READ_ONCE(rdp_gp->nocb_gp_sleep)) { READ_ONCE 1815 kernel/rcu/tree_plugin.h if ((ncbs && j != READ_ONCE(rdp->nocb_bypass_first)) || READ_ONCE 1882 kernel/rcu/tree_plugin.h t = READ_ONCE(rdp->nocb_gp_kthread); READ_ONCE 1969 kernel/rcu/tree_plugin.h (time_after(j, READ_ONCE(rdp->nocb_bypass_first) + 1) || READ_ONCE 2045 kernel/rcu/tree_plugin.h !READ_ONCE(my_rdp->nocb_gp_sleep)); READ_ONCE 2053 kernel/rcu/tree_plugin.h !READ_ONCE(my_rdp->nocb_gp_sleep)); READ_ONCE 2125 kernel/rcu/tree_plugin.h !READ_ONCE(rdp->nocb_cb_sleep)); READ_ONCE 2154 kernel/rcu/tree_plugin.h return READ_ONCE(rdp->nocb_defer_wakeup); READ_ONCE 2168 kernel/rcu/tree_plugin.h ndw = READ_ONCE(rdp->nocb_defer_wakeup); READ_ONCE 2409 kernel/rcu/tree_plugin.h rnp->grplo, rnp->grphi, READ_ONCE(rdp->nocb_gp_loops)); READ_ONCE 2557 kernel/rcu/tree_plugin.h ULONG_CMP_LT(jiffies, READ_ONCE(rcu_state.gp_start) + HZ))) READ_ONCE 26 kernel/rcu/tree_stall.h int till_stall_check = READ_ONCE(rcu_cpu_stall_timeout); READ_ONCE 110 kernel/rcu/tree_stall.h rcu_state.n_force_qs_gpstart = READ_ONCE(rcu_state.n_force_qs); READ_ONCE 131 kernel/rcu/tree_stall.h j = READ_ONCE(rcu_state.jiffies_kick_kthreads); READ_ONCE 133 kernel/rcu/tree_stall.h (rcu_gp_in_progress() || READ_ONCE(rcu_state.gp_flags))) { READ_ONCE 328 kernel/rcu/tree_stall.h READ_ONCE(rcu_state.n_force_qs) - rcu_state.n_force_qs_gpstart, READ_ONCE 338 kernel/rcu/tree_stall.h j = jiffies - READ_ONCE(rcu_state.gp_activity); READ_ONCE 343 kernel/rcu/tree_stall.h READ_ONCE(rcu_state.gp_flags), READ_ONCE 404 kernel/rcu/tree_stall.h gpa = READ_ONCE(rcu_state.gp_activity); READ_ONCE 407 kernel/rcu/tree_stall.h READ_ONCE(jiffies_till_next_fqs), READ_ONCE 414 kernel/rcu/tree_stall.h if (ULONG_CMP_GE(jiffies, READ_ONCE(rcu_state.jiffies_stall))) READ_ONCE 459 kernel/rcu/tree_stall.h if (ULONG_CMP_GE(jiffies, READ_ONCE(rcu_state.jiffies_stall))) READ_ONCE 511 kernel/rcu/tree_stall.h gs1 = READ_ONCE(rcu_state.gp_seq); READ_ONCE 513 kernel/rcu/tree_stall.h js = READ_ONCE(rcu_state.jiffies_stall); READ_ONCE 515 kernel/rcu/tree_stall.h gps = READ_ONCE(rcu_state.gp_start); READ_ONCE 517 kernel/rcu/tree_stall.h gs2 = READ_ONCE(rcu_state.gp_seq); READ_ONCE 525 kernel/rcu/tree_stall.h (READ_ONCE(rnp->qsmask) & rdp->grpmask) && READ_ONCE 563 kernel/rcu/tree_stall.h ja = j - READ_ONCE(rcu_state.gp_activity); READ_ONCE 564 kernel/rcu/tree_stall.h jr = j - READ_ONCE(rcu_state.gp_req_activity); READ_ONCE 565 kernel/rcu/tree_stall.h jw = j - READ_ONCE(rcu_state.gp_wake_time); READ_ONCE 570 kernel/rcu/tree_stall.h ja, jr, jw, (long)READ_ONCE(rcu_state.gp_wake_seq), READ_ONCE 571 kernel/rcu/tree_stall.h (long)READ_ONCE(rcu_state.gp_seq), READ_ONCE 572 kernel/rcu/tree_stall.h (long)READ_ONCE(rcu_get_root()->gp_seq_needed), READ_ONCE 573 kernel/rcu/tree_stall.h READ_ONCE(rcu_state.gp_flags)); READ_ONCE 617 kernel/rcu/tree_stall.h if (time_before(j, READ_ONCE(rcu_state.gp_req_activity) + gpssdelay) || READ_ONCE 618 kernel/rcu/tree_stall.h time_before(j, READ_ONCE(rcu_state.gp_activity) + gpssdelay) || READ_ONCE 626 kernel/rcu/tree_stall.h time_before(j, READ_ONCE(rcu_state.gp_req_activity) + gpssdelay) || READ_ONCE 627 kernel/rcu/tree_stall.h time_before(j, READ_ONCE(rcu_state.gp_activity) + gpssdelay) || READ_ONCE 139 kernel/rcu/update.c return READ_ONCE(rcu_normal) && READ_ONCE 535 kernel/rcu/update.c if (needwake && READ_ONCE(rcu_tasks_kthread_ptr)) READ_ONCE 603 kernel/rcu/update.c if (!READ_ONCE(t->rcu_tasks_holdout) || READ_ONCE 604 kernel/rcu/update.c t->rcu_tasks_nvcsw != READ_ONCE(t->nvcsw) || READ_ONCE 605 kernel/rcu/update.c !READ_ONCE(t->on_rq) || READ_ONCE 694 kernel/rcu/update.c if (t != current && READ_ONCE(t->on_rq) && READ_ONCE 697 kernel/rcu/update.c t->rcu_tasks_nvcsw = READ_ONCE(t->nvcsw); READ_ONCE 739 kernel/rcu/update.c rtst = READ_ONCE(rcu_task_stall_timeout); READ_ONCE 30 kernel/sched/autogroup.h int enabled = READ_ONCE(sysctl_sched_autogroup_enabled); READ_ONCE 291 kernel/sched/completion.c if (!READ_ONCE(x->done)) READ_ONCE 317 kernel/sched/completion.c if (!READ_ONCE(x->done)) READ_ONCE 383 kernel/sched/core.c typeof(ti->flags) old, val = READ_ONCE(ti->flags); READ_ONCE 969 kernel/sched/core.c if (uc_se->value > READ_ONCE(uc_rq->value)) READ_ONCE 1008 kernel/sched/core.c rq_clamp = READ_ONCE(uc_rq->value); READ_ONCE 1613 kernel/sched/deadline.c curr = READ_ONCE(rq->curr); /* unlocked access */ READ_ONCE 1095 kernel/sched/fair.c (lockdep_is_held(&task_rq(p)->lock) && !READ_ONCE(p->on_cpu))); READ_ONCE 1130 kernel/sched/fair.c unsigned int scan_size = READ_ONCE(sysctl_numa_balancing_scan_size); READ_ONCE 1601 kernel/sched/fair.c if (READ_ONCE(dst_rq->numa_migrate_on)) READ_ONCE 2138 kernel/sched/fair.c seq = READ_ONCE(p->mm->numa_scan_seq); READ_ONCE 2280 kernel/sched/fair.c tsk = READ_ONCE(cpu_rq(cpu)->curr); READ_ONCE 2476 kernel/sched/fair.c WRITE_ONCE(p->mm->numa_scan_seq, READ_ONCE(p->mm->numa_scan_seq) + 1); READ_ONCE 2794 kernel/sched/fair.c typeof(*ptr) res, var = READ_ONCE(*ptr); \ READ_ONCE 2814 kernel/sched/fair.c typeof(*ptr) res, var = READ_ONCE(*ptr); \ READ_ONCE 2999 kernel/sched/fair.c tg_shares = READ_ONCE(tg->shares); READ_ONCE 3091 kernel/sched/fair.c runnable = shares = READ_ONCE(gcfs_rq->tg->shares); READ_ONCE 3695 kernel/sched/fair.c return READ_ONCE(p->se.avg.util_avg); READ_ONCE 3700 kernel/sched/fair.c struct util_est ue = READ_ONCE(p->se.avg.util_est); READ_ONCE 5192 kernel/sched/fair.c if (!READ_ONCE(rq->rd->overutilized) && cpu_overutilized(rq->cpu)) { READ_ONCE 5411 kernel/sched/fair.c unsigned long nr_running = READ_ONCE(rq->cfs.h_nr_running); READ_ONCE 5844 kernel/sched/fair.c return READ_ONCE(sds->has_idle_cores); READ_ONCE 6107 kernel/sched/fair.c util = READ_ONCE(cfs_rq->avg.util_avg); READ_ONCE 6110 kernel/sched/fair.c util = max(util, READ_ONCE(cfs_rq->avg.util_est.enqueued)); READ_ONCE 6134 kernel/sched/fair.c if (cpu != task_cpu(p) || !READ_ONCE(p->se.avg.last_update_time)) READ_ONCE 6138 kernel/sched/fair.c util = READ_ONCE(cfs_rq->avg.util_avg); READ_ONCE 6171 kernel/sched/fair.c READ_ONCE(cfs_rq->avg.util_est.enqueued); READ_ONCE 6238 kernel/sched/fair.c unsigned long util_est, util = READ_ONCE(cfs_rq->avg.util_avg); READ_ONCE 6252 kernel/sched/fair.c util_est = READ_ONCE(cfs_rq->avg.util_est.enqueued); READ_ONCE 6371 kernel/sched/fair.c if (!pd || READ_ONCE(rd->overutilized)) READ_ONCE 7519 kernel/sched/fair.c if (READ_ONCE(rq->avg_rt.util_avg)) READ_ONCE 7522 kernel/sched/fair.c if (READ_ONCE(rq->avg_dl.util_avg)) READ_ONCE 7526 kernel/sched/fair.c if (READ_ONCE(rq->avg_irq.util_avg)) READ_ONCE 7655 kernel/sched/fair.c while ((se = READ_ONCE(cfs_rq->h_load_next)) != NULL) { READ_ONCE 7783 kernel/sched/fair.c used = READ_ONCE(rq->avg_rt.util_avg); READ_ONCE 7784 kernel/sched/fair.c used += READ_ONCE(rq->avg_dl.util_avg); READ_ONCE 8240 kernel/sched/fair.c if (env->idle == CPU_NEWLY_IDLE && READ_ONCE(nohz.has_blocked)) READ_ONCE 8548 kernel/sched/fair.c if (rcu_dereference(rd->pd) && !READ_ONCE(rd->overutilized)) READ_ONCE 9424 kernel/sched/fair.c if (READ_ONCE(nohz.has_blocked) && READ_ONCE 9425 kernel/sched/fair.c time_after(now, READ_ONCE(nohz.next_blocked))) READ_ONCE 9768 kernel/sched/fair.c if (!READ_ONCE(nohz.has_blocked) || READ_ONCE 9769 kernel/sched/fair.c time_before(jiffies, READ_ONCE(nohz.next_blocked))) READ_ONCE 9829 kernel/sched/fair.c !READ_ONCE(this_rq->rd->overload)) { READ_ONCE 339 kernel/sched/idle.c while (!READ_ONCE(it.done)) READ_ONCE 223 kernel/sched/loadavg.c if (!time_before(jiffies, READ_ONCE(calc_load_update))) READ_ONCE 271 kernel/sched/loadavg.c this_rq->calc_load_update = READ_ONCE(calc_load_update); READ_ONCE 309 kernel/sched/loadavg.c sample_window = READ_ONCE(calc_load_update); READ_ONCE 355 kernel/sched/loadavg.c sample_window = READ_ONCE(calc_load_update); READ_ONCE 99 kernel/sched/membarrier.c if (!(READ_ONCE(cpu_rq(cpu)->membarrier_state) & READ_ONCE 1402 kernel/sched/rt.c curr = READ_ONCE(rq->curr); /* unlocked access */ READ_ONCE 1061 kernel/sched/sched.h return READ_ONCE(rq->clock); READ_ONCE 1644 kernel/sched/sched.h return READ_ONCE(p->on_rq) == TASK_ON_RQ_MIGRATING; READ_ONCE 1936 kernel/sched/sched.h if (!READ_ONCE(rq->rd->overload)) READ_ONCE 2324 kernel/sched/sched.h unsigned int min_util = READ_ONCE(rq->uclamp[UCLAMP_MIN].value); READ_ONCE 2325 kernel/sched/sched.h unsigned int max_util = READ_ONCE(rq->uclamp[UCLAMP_MAX].value); READ_ONCE 2402 kernel/sched/sched.h return READ_ONCE(rq->avg_dl.util_avg); READ_ONCE 2407 kernel/sched/sched.h unsigned long util = READ_ONCE(rq->cfs.avg.util_avg); READ_ONCE 2411 kernel/sched/sched.h READ_ONCE(rq->cfs.avg.util_est.enqueued)); READ_ONCE 2419 kernel/sched/sched.h return READ_ONCE(rq->avg_rt.util_avg); READ_ONCE 2493 kernel/sched/sched.h if (READ_ONCE(rq->membarrier_state) == membarrier_state) READ_ONCE 2063 kernel/sched/topology.c if (rq->cpu_capacity_orig > READ_ONCE(d.rd->max_cpu_capacity)) READ_ONCE 218 kernel/sched/wait_bit.c unsigned long now = READ_ONCE(jiffies); READ_ONCE 232 kernel/sched/wait_bit.c unsigned long now = READ_ONCE(jiffies); READ_ONCE 260 kernel/seccomp.c READ_ONCE(current->seccomp.filter); READ_ONCE 1934 kernel/signal.c if (tsk->parent_exec_id != READ_ONCE(tsk->parent->self_exec_id)) READ_ONCE 214 kernel/stop_machine.c newstate = READ_ONCE(msdata->state); READ_ONCE 33 kernel/task_work.c head = READ_ONCE(task->task_works); READ_ONCE 71 kernel/task_work.c while ((work = READ_ONCE(*pprev))) { READ_ONCE 102 kernel/task_work.c work = READ_ONCE(task->task_works); READ_ONCE 167 kernel/time/hrtimer.c base = READ_ONCE(timer->base); READ_ONCE 1231 kernel/time/hrtimer.c struct hrtimer_clock_base *base = READ_ONCE(timer->base); READ_ONCE 1445 kernel/time/hrtimer.c base = READ_ONCE(timer->base); READ_ONCE 1453 kernel/time/hrtimer.c base != READ_ONCE(timer->base)); READ_ONCE 305 kernel/time/posix-cpu-timers.c if (!READ_ONCE(pct->timers_active)) { READ_ONCE 350 kernel/time/posix-cpu-timers.c if (!READ_ONCE(pct->timers_active)) { READ_ONCE 916 kernel/time/posix-cpu-timers.c if (!READ_ONCE(pct->timers_active) || pct->expiry_active) READ_ONCE 1092 kernel/time/posix-cpu-timers.c if (READ_ONCE(pct->timers_active) && !READ_ONCE(pct->expiry_active)) { READ_ONCE 822 kernel/time/posix-timers.c const struct k_clock *kc = READ_ONCE(timer->kclock); READ_ONCE 823 kernel/time/posix-timers.c timer_t timer_id = READ_ONCE(timer->it_id); READ_ONCE 63 kernel/time/tick-sched.c delta = ktime_sub(now, READ_ONCE(last_jiffies_update)); READ_ONCE 171 kernel/time/timekeeping.c struct clocksource *clock = READ_ONCE(tkr->clock); READ_ONCE 891 kernel/time/timer.c jnow = READ_ONCE(jiffies); READ_ONCE 932 kernel/time/timer.c tf = READ_ONCE(timer->flags); READ_ONCE 1279 kernel/time/timer.c tf = READ_ONCE(timer->flags); READ_ONCE 1681 kernel/time/timer.c unsigned long now = READ_ONCE(jiffies); READ_ONCE 466 kernel/torture.c while (READ_ONCE(fullstop) == FULLSTOP_SHUTDOWN) { READ_ONCE 535 kernel/torture.c if (READ_ONCE(fullstop) == FULLSTOP_DONTSTOP) { READ_ONCE 583 kernel/torture.c spt = READ_ONCE(stutter_pause_test); READ_ONCE 584 kernel/torture.c for (; spt; spt = READ_ONCE(stutter_pause_test)) { READ_ONCE 589 kernel/torture.c while (READ_ONCE(stutter_pause_test)) READ_ONCE 707 kernel/torture.c if (READ_ONCE(fullstop) == FULLSTOP_SHUTDOWN) { READ_ONCE 746 kernel/torture.c return READ_ONCE(fullstop) != FULLSTOP_DONTSTOP; READ_ONCE 362 kernel/trace/bpf_trace.c ee = READ_ONCE(array->ptrs[index]); READ_ONCE 434 kernel/trace/bpf_trace.c ee = READ_ONCE(array->ptrs[index]); READ_ONCE 571 kernel/trace/bpf_trace.c cgrp = READ_ONCE(array->ptrs[idx]); READ_ONCE 1093 kernel/trace/ring_buffer.c if (tail_page == READ_ONCE(cpu_buffer->tail_page)) { READ_ONCE 2099 kernel/trace/ring_buffer.c buffer_tail_page = READ_ONCE(cpu_buffer->tail_page); READ_ONCE 2428 kernel/trace/ring_buffer.c bpage = READ_ONCE(cpu_buffer->tail_page); READ_ONCE 2476 kernel/trace/ring_buffer.c while (cpu_buffer->commit_page != READ_ONCE(cpu_buffer->tail_page)) { READ_ONCE 2511 kernel/trace/ring_buffer.c if (unlikely(cpu_buffer->commit_page != READ_ONCE(cpu_buffer->tail_page))) READ_ONCE 2832 kernel/trace/ring_buffer.c tail_page = info->tail_page = READ_ONCE(cpu_buffer->tail_page); READ_ONCE 2890 kernel/trace/ring_buffer.c if (unlikely(READ_ONCE(cpu_buffer->buffer) != buffer)) { READ_ONCE 192 kernel/trace/ring_buffer_benchmark.c while (!READ_ONCE(reader_finish)) { READ_ONCE 524 kernel/trace/trace.c pid_list->pid_max = READ_ONCE(pid_max); READ_ONCE 1597 kernel/trace/trace.h return READ_ONCE(file_inode(filp)->i_private); READ_ONCE 159 kernel/trace/trace_stack.c int frame_size = READ_ONCE(tracer_frame); READ_ONCE 531 kernel/trace/tracing_map.c val = READ_ONCE(entry->val); READ_ONCE 198 kernel/ucount.c max = READ_ONCE(tns->ucount_max[type]); READ_ONCE 1125 kernel/user_namespace.c unsigned long userns_flags = READ_ONCE(ns->flags); READ_ONCE 697 kernel/watchdog.c old = READ_ONCE(*param); READ_ONCE 699 kernel/watchdog.c if (!err && old != READ_ONCE(*param)) READ_ONCE 748 kernel/watchdog.c old = READ_ONCE(watchdog_thresh); READ_ONCE 751 kernel/watchdog.c if (!err && write && old != READ_ONCE(watchdog_thresh)) READ_ONCE 5729 kernel/workqueue.c unsigned long thresh = READ_ONCE(wq_watchdog_thresh) * HZ; READ_ONCE 5746 kernel/workqueue.c pool_ts = READ_ONCE(pool->watchdog_ts); READ_ONCE 5747 kernel/workqueue.c touched = READ_ONCE(wq_watchdog_touched); READ_ONCE 5756 kernel/workqueue.c READ_ONCE(per_cpu(wq_watchdog_touched_cpu, READ_ONCE 37 lib/assoc_array.c cursor = READ_ONCE(shortcut->next_node); /* Address dependency. */ READ_ONCE 52 lib/assoc_array.c ptr = READ_ONCE(node->slots[slot]); /* Address dependency. */ READ_ONCE 81 lib/assoc_array.c ptr = READ_ONCE(node->slots[slot]); /* Address dependency. */ READ_ONCE 90 lib/assoc_array.c parent = READ_ONCE(node->back_pointer); /* Address dependency. */ READ_ONCE 98 lib/assoc_array.c parent = READ_ONCE(shortcut->back_pointer); /* Address dependency. */ READ_ONCE 138 lib/assoc_array.c struct assoc_array_ptr *root = READ_ONCE(array->root); /* Address dependency. */ READ_ONCE 185 lib/assoc_array.c cursor = READ_ONCE(array->root); /* Address dependency. */ READ_ONCE 209 lib/assoc_array.c ptr = READ_ONCE(node->slots[slot]); /* Address dependency. */ READ_ONCE 282 lib/assoc_array.c cursor = READ_ONCE(shortcut->next_node); /* Address dependency. */ READ_ONCE 324 lib/assoc_array.c ptr = READ_ONCE(node->slots[slot]); /* Address dependency. */ READ_ONCE 135 lib/debugobjects.c if (likely(READ_ONCE(obj_pool_free) >= debug_objects_pool_min_level)) READ_ONCE 146 lib/debugobjects.c while (READ_ONCE(obj_nr_tofree) && (READ_ONCE(obj_pool_free) < obj_pool_min_free)) { READ_ONCE 165 lib/debugobjects.c while (READ_ONCE(obj_pool_free) < debug_objects_pool_min_level) { READ_ONCE 430 lib/debugobjects.c if (!READ_ONCE(obj_freeing) && READ_ONCE(obj_nr_tofree)) { READ_ONCE 989 lib/debugobjects.c if (!READ_ONCE(obj_freeing) && READ_ONCE(obj_nr_tofree)) { READ_ONCE 1015 lib/debugobjects.c seq_printf(m, "pool_free :%d\n", READ_ONCE(obj_pool_free) + obj_percpu_free); READ_ONCE 1020 lib/debugobjects.c seq_printf(m, "on_free_list :%d\n", READ_ONCE(obj_nr_tofree)); READ_ONCE 24 lib/dynamic_queue_limits.c num_queued = READ_ONCE(dql->num_queued); READ_ONCE 71 lib/errseq.c old = READ_ONCE(*eseq); READ_ONCE 124 lib/errseq.c errseq_t old = READ_ONCE(*eseq); READ_ONCE 146 lib/errseq.c errseq_t cur = READ_ONCE(*eseq); READ_ONCE 184 lib/errseq.c old = READ_ONCE(*eseq); READ_ONCE 106 lib/fault-inject.c unsigned int fail_nth = READ_ONCE(current->fail_nth); READ_ONCE 56 lib/generic-radix-tree.c struct genradix_root *r = READ_ONCE(radix->root); READ_ONCE 107 lib/generic-radix-tree.c struct genradix_root *v = READ_ONCE(radix->root); READ_ONCE 142 lib/generic-radix-tree.c n = READ_ONCE(*p); READ_ONCE 170 lib/generic-radix-tree.c r = READ_ONCE(radix->root); READ_ONCE 32 lib/llist.c new_last->next = first = READ_ONCE(head->first); READ_ONCE 62 lib/llist.c next = READ_ONCE(entry->next); READ_ONCE 15 lib/lockref.c old.lock_count = READ_ONCE(lockref->lock_count); \ READ_ONCE 311 lib/sbitmap.c unsigned long word = READ_ONCE(sb->map[i].word); READ_ONCE 312 lib/sbitmap.c unsigned int word_bits = READ_ONCE(sb->map[i].depth); READ_ONCE 444 lib/sbitmap.c depth = READ_ONCE(sbq->sb.depth); READ_ONCE 475 lib/sbitmap.c depth = READ_ONCE(sbq->sb.depth); READ_ONCE 542 lib/sbitmap.c wake_batch = READ_ONCE(sbq->wake_batch); READ_ONCE 170 lib/vdso/gettimeofday.c time_t t = READ_ONCE(vd[CS_HRES_COARSE].basetime[CLOCK_REALTIME].sec); READ_ONCE 192 lib/vdso/gettimeofday.c hrtimer_res = READ_ONCE(vd[CS_HRES_COARSE].hrtimer_res); READ_ONCE 854 lib/vsprintf.c p = READ_ONCE(d->d_parent); READ_ONCE 855 lib/vsprintf.c array[i] = READ_ONCE(d->d_name.name); READ_ONCE 735 mm/filemap.c errseq_t old = READ_ONCE(file->f_wb_err); READ_ONCE 332 mm/gup.c pmdval = READ_ONCE(*pmd); READ_ONCE 357 mm/gup.c pmdval = READ_ONCE(*pmd); READ_ONCE 1796 mm/gup.c return READ_ONCE(*ptep); READ_ONCE 2003 mm/gup.c pte = READ_ONCE(*ptep); READ_ONCE 2195 mm/gup.c pmd_t pmd = READ_ONCE(*pmdp); READ_ONCE 2238 mm/gup.c pud_t pud = READ_ONCE(*pudp); READ_ONCE 2266 mm/gup.c p4d_t p4d = READ_ONCE(*p4dp); READ_ONCE 2291 mm/gup.c pgd_t pgd = READ_ONCE(*pgdp); READ_ONCE 565 mm/hmm.c pmd = READ_ONCE(*pmdp); READ_ONCE 667 mm/hmm.c pud = READ_ONCE(*pudp); READ_ONCE 85 mm/huge_memory.c return READ_ONCE(huge_zero_page); READ_ONCE 104 mm/huge_memory.c return READ_ONCE(huge_zero_page); READ_ONCE 119 mm/huge_memory.c return READ_ONCE(huge_zero_page); READ_ONCE 127 mm/huge_memory.c return READ_ONCE(huge_zero_page); READ_ONCE 2253 mm/huge_memory.c entry = mk_pte(page + i, READ_ONCE(vma->vm_page_prot)); READ_ONCE 2898 mm/huge_memory.c return READ_ONCE(ds_queue->split_queue_len); READ_ONCE 5030 mm/hugetlb.c pud_entry = READ_ONCE(*pud); READ_ONCE 5038 mm/hugetlb.c pmd_entry = READ_ONCE(*pmd); READ_ONCE 253 mm/internal.h #define page_order_unsafe(page) READ_ONCE(page_private(page)) READ_ONCE 458 mm/kasan/common.c shadow_byte = READ_ONCE(*(s8 *)kasan_mem_to_shadow(object)); READ_ONCE 198 mm/kasan/quarantine.c READ_ONCE(quarantine_batch_size)) { READ_ONCE 220 mm/kasan/quarantine.c if (likely(READ_ONCE(quarantine_size) <= READ_ONCE 221 mm/kasan/quarantine.c READ_ONCE(quarantine_max_size))) READ_ONCE 704 mm/ksm.c kpfn = READ_ONCE(stable_node->kpfn); /* Address dependency. */ READ_ONCE 706 mm/ksm.c if (READ_ONCE(page->mapping) != expected_mapping) READ_ONCE 733 mm/ksm.c if (READ_ONCE(page->mapping) != expected_mapping) { READ_ONCE 747 mm/ksm.c if (READ_ONCE(page->mapping) != expected_mapping) { READ_ONCE 763 mm/ksm.c if (READ_ONCE(stable_node->kpfn) != kpfn) READ_ONCE 2049 mm/ksm.c get_kpfn_nid(READ_ONCE(stable_node->kpfn)) != READ_ONCE 2424 mm/ksm.c sleep_ms = READ_ONCE(ksm_thread_sleep_millisecs); READ_ONCE 2426 mm/ksm.c sleep_ms != READ_ONCE(ksm_thread_sleep_millisecs), READ_ONCE 3052 mm/ksm.c if (READ_ONCE(ksm_max_page_sharing) == knob) READ_ONCE 492 mm/memcontrol.c memcg = READ_ONCE(page->mem_cgroup); READ_ONCE 589 mm/memcontrol.c unsigned long soft_limit = READ_ONCE(memcg->soft_limit); READ_ONCE 1074 mm/memcontrol.c pos = READ_ONCE(iter->position); READ_ONCE 1332 mm/memcontrol.c limit = READ_ONCE(memcg->memory.max); READ_ONCE 1338 mm/memcontrol.c limit = READ_ONCE(memcg->memsw.max); READ_ONCE 2442 mm/memcontrol.c high = READ_ONCE(memcg->high); READ_ONCE 2953 mm/memcontrol.c kmemcg_id = READ_ONCE(memcg->kmemcg_id); READ_ONCE 2964 mm/memcontrol.c memcg_cachep = READ_ONCE(arr->entries[kmemcg_id]); READ_ONCE 5834 mm/memcontrol.c move_flags = READ_ONCE(memcg->move_charge_at_immigrate); READ_ONCE 6085 mm/memcontrol.c READ_ONCE(mem_cgroup_from_seq(m)->memory.min)); READ_ONCE 6108 mm/memcontrol.c READ_ONCE(mem_cgroup_from_seq(m)->memory.low)); READ_ONCE 6130 mm/memcontrol.c return seq_puts_memcg_tunable(m, READ_ONCE(mem_cgroup_from_seq(m)->high)); READ_ONCE 6160 mm/memcontrol.c READ_ONCE(mem_cgroup_from_seq(m)->memory.max)); READ_ONCE 6455 mm/memcontrol.c parent_emin = READ_ONCE(parent->memory.emin); READ_ONCE 6469 mm/memcontrol.c parent_elow = READ_ONCE(parent->memory.elow); READ_ONCE 7178 mm/memcontrol.c READ_ONCE(memcg->swap.max) - READ_ONCE 7233 mm/memcontrol.c READ_ONCE(mem_cgroup_from_seq(m)->swap.max)); READ_ONCE 3442 mm/memory.c nr_pages = READ_ONCE(fault_around_bytes) >> PAGE_SHIFT; READ_ONCE 235 mm/migrate.c pte = pte_mkold(mk_pte(new, READ_ONCE(vma->vm_page_prot))); READ_ONCE 1275 mm/mmap.c struct anon_vma *anon_vma = READ_ONCE(old->anon_vma); READ_ONCE 495 mm/oom_kill.c struct mm_struct *t_mm = READ_ONCE(t->mm); READ_ONCE 541 mm/page_alloc.c word = READ_ONCE(bitmap[word_bitidx]); READ_ONCE 2802 mm/page_alloc.c batch = READ_ONCE(pcp->batch); READ_ONCE 3056 mm/page_alloc.c unsigned long batch = READ_ONCE(pcp->batch); READ_ONCE 411 mm/page_io.c if (!READ_ONCE(bio->bi_private)) READ_ONCE 434 mm/page_owner.c handle = READ_ONCE(page_owner->handle); READ_ONCE 442 mm/page_owner.c handle = READ_ONCE(page_owner->free_handle); READ_ONCE 532 mm/page_owner.c handle = READ_ONCE(page_owner->handle); READ_ONCE 182 mm/page_vma_mapped.c pmde = READ_ONCE(*pvmw->pmd); READ_ONCE 471 mm/rmap.c anon_mapping = (unsigned long)READ_ONCE(page->mapping); READ_ONCE 515 mm/rmap.c anon_mapping = (unsigned long)READ_ONCE(page->mapping); READ_ONCE 522 mm/rmap.c root_anon_vma = READ_ONCE(anon_vma->root); READ_ONCE 571 mm/shmem.c if (!READ_ONCE(sbinfo->shrinklist_len)) READ_ONCE 581 mm/shmem.c return READ_ONCE(sbinfo->shrinklist_len); READ_ONCE 742 mm/shmem.c swapped = READ_ONCE(info->swapped); READ_ONCE 2930 mm/slab.c shared = READ_ONCE(n->shared); READ_ONCE 2935 mm/slab.c shared = READ_ONCE(n->shared); READ_ONCE 336 mm/slab.h s = READ_ONCE(page->slab_cache); READ_ONCE 338 mm/slab.h return READ_ONCE(s->memcg_params.memcg); READ_ONCE 356 mm/slab.h memcg = READ_ONCE(s->memcg_params.memcg); READ_ONCE 394 mm/slab.h memcg = READ_ONCE(s->memcg_params.memcg); READ_ONCE 2708 mm/slub.c unlikely(tid != READ_ONCE(c->tid))); READ_ONCE 2982 mm/slub.c unlikely(tid != READ_ONCE(c->tid))); READ_ONCE 2988 mm/slub.c void **freelist = READ_ONCE(c->freelist); READ_ONCE 4827 mm/slub.c page = READ_ONCE(c->page); READ_ONCE 301 mm/swap_state.c return READ_ONCE(enable_vma_readahead) && !atomic_read(&nr_rotate_swap); READ_ONCE 507 mm/swap_state.c max_pages = 1 << READ_ONCE(page_cluster); READ_ONCE 654 mm/swap_state.c max_win = 1 << min_t(unsigned int, READ_ONCE(page_cluster), READ_ONCE 104 mm/swapfile.c if (type >= READ_ONCE(nr_swapfiles)) READ_ONCE 108 mm/swapfile.c return READ_ONCE(swap_info[type]); READ_ONCE 2105 mm/swapfile.c count = READ_ONCE(si->swap_map[i]); READ_ONCE 470 mm/userfaultfd.c if (mmap_changing && READ_ONCE(*mmap_changing)) READ_ONCE 3479 mm/vmalloc.c head = READ_ONCE(vmap_purge_list.first); READ_ONCE 292 mm/vmscan.c return READ_ONCE(mn->congested); READ_ONCE 669 mm/z3fold.c zhdr = list_first_entry_or_null(READ_ONCE(l), READ_ONCE 678 mm/z3fold.c if (unlikely(zhdr != list_first_entry(READ_ONCE(l), READ_ONCE 724 mm/z3fold.c zhdr = list_first_entry_or_null(READ_ONCE(l), READ_ONCE 101 net/core/datagram.c if (READ_ONCE(sk->sk_receive_queue.prev) != skb) READ_ONCE 282 net/core/datagram.c } while (READ_ONCE(sk->sk_receive_queue.prev) != *last); READ_ONCE 4068 net/core/dev.c cpu = READ_ONCE(rflow->cpu); READ_ONCE 5932 net/core/dev.c val = READ_ONCE(n->state); READ_ONCE 6003 net/core/dev.c val = READ_ONCE(n->state); READ_ONCE 6105 net/core/dev.c unsigned long val = READ_ONCE(napi->state); READ_ONCE 3604 net/core/filter.c if (unlikely(READ_ONCE(ri->map) == map)) READ_ONCE 3640 net/core/filter.c struct bpf_map *map = READ_ONCE(ri->map); READ_ONCE 3695 net/core/filter.c struct bpf_map *map = READ_ONCE(ri->map); READ_ONCE 4068 net/core/filter.c cgrp = READ_ONCE(array->ptrs[idx]); READ_ONCE 1197 net/core/neighbour.c if (READ_ONCE(hh->hh_len)) { READ_ONCE 1476 net/core/neighbour.c if (dev->header_ops->cache && !READ_ONCE(neigh->hh.hh_len)) READ_ONCE 198 net/core/page_pool.c u32 hold_cnt = READ_ONCE(pool->pages_state_hold_cnt); READ_ONCE 3301 net/core/pktgen.c unsigned int burst = READ_ONCE(pkt_dev->burst); READ_ONCE 4418 net/core/skbuff.c (unsigned int)READ_ONCE(sk->sk_rcvbuf)) READ_ONCE 633 net/core/skmsg.c prog = READ_ONCE(psock->progs.msg_parser); READ_ONCE 748 net/core/skmsg.c prog = READ_ONCE(psock->progs.skb_verdict); READ_ONCE 771 net/core/skmsg.c prog = READ_ONCE(psock->progs.skb_parser); READ_ONCE 525 net/core/sock.c } else if (sk_add_backlog(sk, skb, READ_ONCE(sk->sk_rcvbuf))) { READ_ONCE 1479 net/core/sock.c v.val = READ_ONCE(sk->sk_incoming_cpu); READ_ONCE 1497 net/core/sock.c v.val = READ_ONCE(sk->sk_napi_id); READ_ONCE 2097 net/core/sock.c refcount_read(&sk->sk_wmem_alloc) < READ_ONCE(sk->sk_sndbuf)) { READ_ONCE 2200 net/core/sock.c if (refcount_read(&sk->sk_wmem_alloc) < READ_ONCE(sk->sk_sndbuf)) READ_ONCE 2235 net/core/sock.c if (sk_wmem_alloc_get(sk) < READ_ONCE(sk->sk_sndbuf)) READ_ONCE 2344 net/core/sock.c if (memory_pressure && READ_ONCE(*memory_pressure)) READ_ONCE 2816 net/core/sock.c if ((refcount_read(&sk->sk_wmem_alloc) << 1) <= READ_ONCE(sk->sk_sndbuf)) { READ_ONCE 3214 net/core/sock.c mem[SK_MEMINFO_RCVBUF] = READ_ONCE(sk->sk_rcvbuf); READ_ONCE 3216 net/core/sock.c mem[SK_MEMINFO_SNDBUF] = READ_ONCE(sk->sk_sndbuf); READ_ONCE 3218 net/core/sock.c mem[SK_MEMINFO_WMEM_QUEUED] = READ_ONCE(sk->sk_wmem_queued); READ_ONCE 3220 net/core/sock.c mem[SK_MEMINFO_BACKLOG] = READ_ONCE(sk->sk_backlog.len); READ_ONCE 150 net/core/sock_map.c skb_verdict = READ_ONCE(progs->skb_verdict); READ_ONCE 151 net/core/sock_map.c skb_parser = READ_ONCE(progs->skb_parser); READ_ONCE 164 net/core/sock_map.c msg_parser = READ_ONCE(progs->msg_parser); READ_ONCE 180 net/core/sock_map.c if ((msg_parser && READ_ONCE(psock->progs.msg_parser)) || READ_ONCE 181 net/core/sock_map.c (skb_progs && READ_ONCE(psock->progs.skb_parser))) { READ_ONCE 275 net/core/sock_map.c return READ_ONCE(stab->sks[key]); READ_ONCE 118 net/core/sock_reuseport.c more_reuse->synq_overflow_ts = READ_ONCE(reuse->synq_overflow_ts); READ_ONCE 283 net/core/sock_reuseport.c socks = READ_ONCE(reuse->num_socks); READ_ONCE 255 net/ipv4/icmp.c if (!READ_ONCE(icmp_global.credit)) { READ_ONCE 256 net/ipv4/icmp.c delta = min_t(u32, now - READ_ONCE(icmp_global.stamp), HZ); READ_ONCE 475 net/ipv4/igmp.c mtu = READ_ONCE(dev->mtu); READ_ONCE 754 net/ipv4/inet_connection_sock.c defer_accept = READ_ONCE(queue->rskq_defer_accept); READ_ONCE 210 net/ipv4/inet_diag.c .idiag_wmem = READ_ONCE(sk->sk_wmem_queued), READ_ONCE 273 net/ipv4/inet_diag.c ca_ops = READ_ONCE(icsk->icsk_ca_ops); READ_ONCE 293 net/ipv4/inet_diag.c ca_ops = READ_ONCE(icsk->icsk_ca_ops); READ_ONCE 243 net/ipv4/inet_hashtables.c if (READ_ONCE(sk->sk_incoming_cpu) == raw_smp_processor_id()) READ_ONCE 167 net/ipv4/inetpeer.c delta = (__u32)jiffies - READ_ONCE(p->dtime); READ_ONCE 323 net/ipv4/ip_input.c if (ipprot && (edemux = READ_ONCE(ipprot->early_demux))) { READ_ONCE 1262 net/ipv4/ip_output.c dst_mtu(&rt->dst) : READ_ONCE(rt->dst.dev->mtu); READ_ONCE 255 net/ipv4/ipmr_base.c lastuse = READ_ONCE(c->mfc_un.res.lastuse); READ_ONCE 206 net/ipv4/netfilter/arp_tables.c private = READ_ONCE(table->private); /* Address dependency. */ READ_ONCE 261 net/ipv4/netfilter/ip_tables.c private = READ_ONCE(table->private); /* Address dependency. */ READ_ONCE 523 net/ipv4/raw.c hdrincl = READ_ONCE(hdrincl); READ_ONCE 491 net/ipv4/route.c u32 old = READ_ONCE(*p_tstamp); READ_ONCE 1313 net/ipv4/route.c mtu = READ_ONCE(dst->dev->mtu); READ_ONCE 1411 net/ipv4/route.c mtu = min(READ_ONCE(dev->mtu), IP_MAX_MTU); READ_ONCE 329 net/ipv4/tcp.c if (READ_ONCE(tcp_memory_pressure)) READ_ONCE 344 net/ipv4/tcp.c if (!READ_ONCE(tcp_memory_pressure)) READ_ONCE 392 net/ipv4/tcp.c u32 rate = READ_ONCE(tp->rate_delivered); READ_ONCE 393 net/ipv4/tcp.c u32 intv = READ_ONCE(tp->rate_interval_us); READ_ONCE 480 net/ipv4/tcp.c int avail = READ_ONCE(tp->rcv_nxt) - READ_ONCE(tp->copied_seq); READ_ONCE 557 net/ipv4/tcp.c if (READ_ONCE(tp->urg_seq) == READ_ONCE(tp->copied_seq) && READ_ONCE 619 net/ipv4/tcp.c READ_ONCE(tp->urg_seq) == READ_ONCE(tp->copied_seq); READ_ONCE 628 net/ipv4/tcp.c answ = READ_ONCE(tp->write_seq) - tp->snd_una; READ_ONCE 637 net/ipv4/tcp.c answ = READ_ONCE(tp->write_seq) - READ_ONCE 638 net/ipv4/tcp.c READ_ONCE(tp->snd_nxt); READ_ONCE 1928 net/ipv4/tcp.c u32 copied_seq = READ_ONCE(tp->copied_seq); READ_ONCE 1929 net/ipv4/tcp.c u32 rcv_nxt = READ_ONCE(tp->rcv_nxt); READ_ONCE 1933 net/ipv4/tcp.c if (unlikely(inq < 0 || copied_seq != READ_ONCE(tp->copied_seq))) { READ_ONCE 3227 net/ipv4/tcp.c rate = READ_ONCE(sk->sk_pacing_rate); READ_ONCE 3231 net/ipv4/tcp.c rate = READ_ONCE(sk->sk_max_pacing_rate); READ_ONCE 3380 net/ipv4/tcp.c rate = READ_ONCE(sk->sk_pacing_rate); READ_ONCE 310 net/ipv4/tcp_bbr.c sk->sk_pacing_rate >> READ_ONCE(sk->sk_pacing_shift), READ_ONCE 679 net/ipv4/tcp_bpf.c struct proto *ops = READ_ONCE(sk->sk_prot); READ_ONCE 29 net/ipv4/tcp_diag.c r->idiag_rqueue = max_t(int, READ_ONCE(tp->rcv_nxt) - READ_ONCE 30 net/ipv4/tcp_diag.c READ_ONCE(tp->copied_seq), 0); READ_ONCE 31 net/ipv4/tcp_diag.c r->idiag_wqueue = READ_ONCE(tp->write_seq) - tp->snd_una; READ_ONCE 3478 net/ipv4/tcp_input.c count = READ_ONCE(challenge_count); READ_ONCE 1636 net/ipv4/tcp_ipv4.c struct dst_entry *dst = READ_ONCE(sk->sk_rx_dst); READ_ONCE 1650 net/ipv4/tcp_ipv4.c u32 limit = READ_ONCE(sk->sk_rcvbuf) + READ_ONCE(sk->sk_sndbuf); READ_ONCE 2462 net/ipv4/tcp_ipv4.c rx_queue = max_t(int, READ_ONCE(tp->rcv_nxt) - READ_ONCE 2463 net/ipv4/tcp_ipv4.c READ_ONCE(tp->copied_seq), 0); READ_ONCE 2468 net/ipv4/tcp_ipv4.c READ_ONCE(tp->write_seq) - tp->snd_una, READ_ONCE 373 net/ipv4/tcp_minisocks.c window_clamp = READ_ONCE(tp->window_clamp); READ_ONCE 942 net/ipv4/tcp_output.c for (oval = READ_ONCE(sk->sk_tsq_flags);; oval = nval) { READ_ONCE 1735 net/ipv4/tcp_output.c sk->sk_pacing_rate >> READ_ONCE(sk->sk_pacing_shift), READ_ONCE 1983 net/ipv4/tcp_output.c win_divisor = READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_tso_win_divisor); READ_ONCE 2270 net/ipv4/tcp_output.c sk->sk_pacing_rate >> READ_ONCE(sk->sk_pacing_shift)); READ_ONCE 391 net/ipv4/udp.c if (READ_ONCE(sk->sk_incoming_cpu) == raw_smp_processor_id()) READ_ONCE 607 net/ipv4/udp.c lookup = READ_ONCE(up->encap_err_lookup); READ_ONCE 1549 net/ipv4/udp.c if (unlikely(READ_ONCE(sk->sk_peek_off) >= 0)) { READ_ONCE 2025 net/ipv4/udp.c encap_rcv = READ_ONCE(up->encap_rcv); READ_ONCE 2477 net/ipv4/udp.c dst = READ_ONCE(sk->sk_rx_dst); READ_ONCE 2515 net/ipv4/udp.c encap_destroy = READ_ONCE(up->encap_destroy); READ_ONCE 1363 net/ipv6/addrconf.c cnf_temp_preferred_lft = READ_ONCE(idev->cnf.temp_prefered_lft); READ_ONCE 108 net/ipv6/inet6_hashtables.c if (READ_ONCE(sk->sk_incoming_cpu) == raw_smp_processor_id()) READ_ONCE 58 net/ipv6/ip6_input.c if (ipprot && (edemux = READ_ONCE(ipprot->early_demux))) READ_ONCE 1287 net/ipv6/ip6_output.c READ_ONCE(rt->dst.dev->mtu) : dst_mtu(&rt->dst); READ_ONCE 1290 net/ipv6/ip6_output.c READ_ONCE(rt->dst.dev->mtu) : dst_mtu(xfrm_dst_path(&rt->dst)); READ_ONCE 492 net/ipv6/ip6_tunnel.c tproto = READ_ONCE(t->parms.proto); READ_ONCE 893 net/ipv6/ip6_tunnel.c u8 tproto = READ_ONCE(t->parms.proto); READ_ONCE 1244 net/ipv6/ip6_tunnel.c tproto = READ_ONCE(t->parms.proto); READ_ONCE 1315 net/ipv6/ip6_tunnel.c tproto = READ_ONCE(t->parms.proto); READ_ONCE 1744 net/ipv6/mcast.c mtu = READ_ONCE(dev->mtu); READ_ONCE 283 net/ipv6/netfilter/ip6_tables.c private = READ_ONCE(table->private); /* Address dependency. */ READ_ONCE 801 net/ipv6/raw.c hdrincl = READ_ONCE(hdrincl); READ_ONCE 645 net/ipv6/route.c last_probe = READ_ONCE(fib6_nh->last_probe); READ_ONCE 1725 net/ipv6/tcp_ipv6.c struct dst_entry *dst = READ_ONCE(sk->sk_rx_dst); READ_ONCE 1899 net/ipv6/tcp_ipv6.c rx_queue = max_t(int, READ_ONCE(tp->rcv_nxt) - READ_ONCE 1900 net/ipv6/tcp_ipv6.c READ_ONCE(tp->copied_seq), 0); READ_ONCE 1911 net/ipv6/tcp_ipv6.c READ_ONCE(tp->write_seq) - tp->snd_una, READ_ONCE 138 net/ipv6/udp.c if (READ_ONCE(sk->sk_incoming_cpu) == raw_smp_processor_id()) READ_ONCE 475 net/ipv6/udp.c lookup = READ_ONCE(up->encap_err_lookup); READ_ONCE 623 net/ipv6/udp.c encap_rcv = READ_ONCE(up->encap_rcv); READ_ONCE 1000 net/ipv6/udp.c dst = READ_ONCE(sk->sk_rx_dst); READ_ONCE 1545 net/ipv6/udp.c encap_destroy = READ_ONCE(up->encap_destroy); READ_ONCE 816 net/llc/llc_conn.c if (sk_add_backlog(sk, skb, READ_ONCE(sk->sk_rcvbuf))) READ_ONCE 196 net/llc/llc_input.c sap_handler = dest ? READ_ONCE(llc_type_handlers[dest - 1]) : NULL; READ_ONCE 217 net/llc/llc_input.c sta_handler = READ_ONCE(llc_station_handler); READ_ONCE 2109 net/mac80211/sta_info.c u16 rate = READ_ONCE(sta_get_last_rx_stats(sta)->last_rate); READ_ONCE 247 net/mpls/af_mpls.c alive = READ_ONCE(rt->rt_nhn_alive); READ_ONCE 256 net/mpls/af_mpls.c unsigned int nh_flags = READ_ONCE(nh->nh_flags); READ_ONCE 466 net/netfilter/ipvs/ip_vs_sync.c unsigned long orig = READ_ONCE(cp->sync_endtime); READ_ONCE 1795 net/netfilter/nf_conntrack_core.c if (READ_ONCE(ct->timeout) != extra_jiffies) READ_ONCE 6022 net/netfilter/nf_tables_api.c const struct net_device *dev = READ_ONCE(flowtable->ops[i].dev); READ_ONCE 105 net/netfilter/nf_tables_core.c pstats = READ_ONCE(base_chain->stats); READ_ONCE 159 net/netfilter/nf_tables_core.c bool genbit = READ_ONCE(net->nft.gencursor); READ_ONCE 423 net/netfilter/nfnetlink_queue.c switch ((enum nfqnl_config_mode)READ_ONCE(queue->copy_mode)) { READ_ONCE 434 net/netfilter/nfnetlink_queue.c data_len = READ_ONCE(queue->copy_range); READ_ONCE 383 net/netlabel/netlabel_calipso.c return READ_ONCE(calipso_ops); READ_ONCE 1206 net/packet/af_packet.c len = READ_ONCE(po->rx_ring.frame_max) + 1; READ_ONCE 1207 net/packet/af_packet.c idx = READ_ONCE(po->rx_ring.head); READ_ONCE 1219 net/packet/af_packet.c len = READ_ONCE(po->rx_ring.prb_bdqc.knum_blocks); READ_ONCE 1220 net/packet/af_packet.c idx = READ_ONCE(po->rx_ring.prb_bdqc.kactive_blk_num); READ_ONCE 1235 net/packet/af_packet.c int rcvbuf = READ_ONCE(sk->sk_rcvbuf); READ_ONCE 1269 net/packet/af_packet.c if (READ_ONCE(po->pressure) != pressure) READ_ONCE 1277 net/packet/af_packet.c if (READ_ONCE(po->pressure) && READ_ONCE 1305 net/packet/af_packet.c if (READ_ONCE(history[i]) == rxhash) READ_ONCE 1311 net/packet/af_packet.c if (READ_ONCE(history[victim]) != rxhash) READ_ONCE 1368 net/packet/af_packet.c if (po_next != po_skip && !READ_ONCE(po_next->pressure) && READ_ONCE 1418 net/packet/af_packet.c unsigned int num = READ_ONCE(f->num_members); READ_ONCE 180 net/rds/send.c send_gen = READ_ONCE(cp->cp_send_gen) + 1; READ_ONCE 452 net/rds/send.c raced = send_gen != READ_ONCE(cp->cp_send_gen); READ_ONCE 227 net/rxrpc/af_rxrpc.c max = READ_ONCE(rxrpc_max_backlog); READ_ONCE 56 net/rxrpc/call_accept.c call_tail = READ_ONCE(b->call_backlog_tail); READ_ONCE 65 net/rxrpc/call_accept.c tail = READ_ONCE(b->peer_backlog_tail); READ_ONCE 76 net/rxrpc/call_accept.c tail = READ_ONCE(b->conn_backlog_tail); READ_ONCE 119 net/rxrpc/call_event.c ack_at += READ_ONCE(call->tx_backoff); READ_ONCE 329 net/rxrpc/call_event.c t = READ_ONCE(call->expect_rx_by); READ_ONCE 335 net/rxrpc/call_event.c t = READ_ONCE(call->expect_req_by); READ_ONCE 342 net/rxrpc/call_event.c t = READ_ONCE(call->expect_term_by); READ_ONCE 348 net/rxrpc/call_event.c t = READ_ONCE(call->ack_at); READ_ONCE 355 net/rxrpc/call_event.c t = READ_ONCE(call->ack_lost_at); READ_ONCE 362 net/rxrpc/call_event.c t = READ_ONCE(call->keepalive_at); READ_ONCE 371 net/rxrpc/call_event.c t = READ_ONCE(call->ping_at); READ_ONCE 378 net/rxrpc/call_event.c t = READ_ONCE(call->resend_at); READ_ONCE 427 net/rxrpc/call_event.c #define set(T) { t = READ_ONCE(T); if (time_before(t, next)) next = t; } READ_ONCE 1117 net/rxrpc/conn_client.c now = READ_ONCE(jiffies); READ_ONCE 49 net/rxrpc/conn_event.c call_id = READ_ONCE(chan->last_call); READ_ONCE 117 net/rxrpc/conn_event.c if (READ_ONCE(chan->last_call) != call_id) READ_ONCE 416 net/rxrpc/conn_event.c ack_at = READ_ONCE(chan->final_ack_at); READ_ONCE 402 net/rxrpc/conn_object.c idle_timestamp = READ_ONCE(conn->idle_timestamp); READ_ONCE 308 net/rxrpc/input.c rxrpc_seq_t top = READ_ONCE(call->tx_top); READ_ONCE 428 net/rxrpc/input.c state = READ_ONCE(call->state); READ_ONCE 435 net/rxrpc/input.c unsigned long timo = READ_ONCE(call->next_req_timo); READ_ONCE 458 net/rxrpc/input.c hard_ack = READ_ONCE(call->rx_hard_ack); READ_ONCE 694 net/rxrpc/input.c ping_serial = READ_ONCE(call->ping_serial); READ_ONCE 812 net/rxrpc/input.c rxrpc_seq_t base = READ_ONCE(call->ackr_first_seq); READ_ONCE 929 net/rxrpc/input.c switch (READ_ONCE(call->state)) { READ_ONCE 1032 net/rxrpc/input.c timo = READ_ONCE(call->next_rx_timo); READ_ONCE 1087 net/rxrpc/input.c switch (READ_ONCE(call->state)) { READ_ONCE 39 net/rxrpc/output.c u16 tx_backoff = READ_ONCE(call->tx_backoff); READ_ONCE 84 net/rxrpc/output.c hard_ack = READ_ONCE(call->rx_hard_ack); READ_ONCE 79 net/rxrpc/proc.c local = READ_ONCE(rx->local); READ_ONCE 95 net/rxrpc/proc.c timeout = READ_ONCE(call->expect_rx_by); READ_ONCE 99 net/rxrpc/proc.c tx_hard_ack = READ_ONCE(call->tx_hard_ack); READ_ONCE 100 net/rxrpc/proc.c rx_hard_ack = READ_ONCE(call->rx_hard_ack); READ_ONCE 114 net/rxrpc/proc.c tx_hard_ack, READ_ONCE(call->tx_top) - tx_hard_ack, READ_ONCE 115 net/rxrpc/proc.c rx_hard_ack, READ_ONCE(call->rx_top) - rx_hard_ack, READ_ONCE 413 net/rxrpc/recvmsg.c ASSERTCMP(seq, ==, READ_ONCE(call->rx_top)); READ_ONCE 553 net/rxrpc/recvmsg.c switch (READ_ONCE(call->state)) { READ_ONCE 656 net/rxrpc/recvmsg.c switch (READ_ONCE(call->state)) { READ_ONCE 749 net/rxrpc/recvmsg.c if (READ_ONCE(call->state) != RXRPC_CALL_CLIENT_RECV_REPLY) READ_ONCE 176 net/rxrpc/rtt.c u8 backoff = READ_ONCE(peer->backoff); READ_ONCE 28 net/rxrpc/sendmsg.c rxrpc_seq_t tx_win = READ_ONCE(call->tx_hard_ack); READ_ONCE 71 net/rxrpc/sendmsg.c rtt = READ_ONCE(call->peer->srtt_us) >> 3; READ_ONCE 77 net/rxrpc/sendmsg.c tx_start = READ_ONCE(call->tx_hard_ack); READ_ONCE 82 net/rxrpc/sendmsg.c tx_win = READ_ONCE(call->tx_hard_ack); READ_ONCE 687 net/rxrpc/sendmsg.c switch (READ_ONCE(call->state)) { READ_ONCE 742 net/rxrpc/sendmsg.c state = READ_ONCE(call->state); READ_ONCE 809 net/rxrpc/sendmsg.c switch (READ_ONCE(call->state)) { READ_ONCE 585 net/sched/act_csum.c action = READ_ONCE(p->tcf_action); READ_ONCE 395 net/sched/act_ct.c retval = READ_ONCE(c->tcf_action); READ_ONCE 96 net/sched/act_ctinfo.c action = READ_ONCE(ca->tcf_action); READ_ONCE 154 net/sched/act_gact.c int action = READ_ONCE(gact->tcf_action); READ_ONCE 158 net/sched/act_gact.c u32 ptype = READ_ONCE(gact->tcfg_ptype); READ_ONCE 177 net/sched/act_gact.c int action = READ_ONCE(gact->tcf_action); READ_ONCE 238 net/sched/act_mirred.c m_mac_header_xmit = READ_ONCE(m->tcfm_mac_header_xmit); READ_ONCE 239 net/sched/act_mirred.c m_eaction = READ_ONCE(m->tcfm_eaction); READ_ONCE 240 net/sched/act_mirred.c retval = READ_ONCE(m->tcf_action); READ_ONCE 74 net/sched/act_mpls.c ret = READ_ONCE(m->tcf_action); READ_ONCE 229 net/sched/act_police.c ret = READ_ONCE(police->tcf_action); READ_ONCE 170 net/sched/act_sample.c retval = READ_ONCE(s->tcf_action); READ_ONCE 37 net/sched/act_skbedit.c action = READ_ONCE(d->tcf_action); READ_ONCE 44 net/sched/act_skbmod.c action = READ_ONCE(d->tcf_action); READ_ONCE 35 net/sched/act_tunnel_key.c action = READ_ONCE(t->tcf_action); READ_ONCE 40 net/sched/act_vlan.c action = READ_ONCE(v->tcf_action); READ_ONCE 3196 net/sched/cls_api.c bool take_rtnl = READ_ONCE(block->lockeddevcnt) && !rtnl_held; READ_ONCE 3232 net/sched/cls_api.c bool take_rtnl = READ_ONCE(block->lockeddevcnt) && !rtnl_held; READ_ONCE 3284 net/sched/cls_api.c bool take_rtnl = READ_ONCE(block->lockeddevcnt) && !rtnl_held; READ_ONCE 3336 net/sched/cls_api.c bool take_rtnl = READ_ONCE(block->lockeddevcnt) && !rtnl_held; READ_ONCE 449 net/sched/em_meta.c dst->value = READ_ONCE(sk->sk_wmem_queued); READ_ONCE 557 net/sched/em_meta.c dst->value = READ_ONCE(sk->sk_rcvlowat); READ_ONCE 178 net/sctp/diag.c mem[SK_MEMINFO_BACKLOG] = READ_ONCE(sk->sk_backlog.len); READ_ONCE 325 net/sctp/input.c if (sk_add_backlog(sk, skb, READ_ONCE(sk->sk_rcvbuf))) READ_ONCE 340 net/sctp/input.c if (!sk_add_backlog(sk, skb, READ_ONCE(sk->sk_rcvbuf))) READ_ONCE 367 net/sctp/input.c ret = sk_add_backlog(sk, skb, READ_ONCE(sk->sk_rcvbuf)); READ_ONCE 2092 net/sunrpc/auth_gss/auth_gss.c if (gss_seq_is_newer(req->rq_seqno, READ_ONCE(ctx->gc_seq))) READ_ONCE 2095 net/sunrpc/auth_gss/auth_gss.c seq_xmit = READ_ONCE(ctx->gc_seq_xmit); READ_ONCE 67 net/sunrpc/sched.c unsigned long timeout = READ_ONCE(task->tk_timeout); READ_ONCE 850 net/sunrpc/sched.c queue = READ_ONCE(task->tk_waitqueue); READ_ONCE 383 net/sunrpc/svc_xprt.c xpt_flags = READ_ONCE(xprt->xpt_flags); READ_ONCE 709 net/sunrpc/xprt.c return READ_ONCE(xprt->connect_cookie); READ_ONCE 1028 net/sunrpc/xprt.c READ_ONCE(task->tk_rqstp->rq_reply_bytes_recvd) != 0; READ_ONCE 1035 net/sunrpc/xprt.c READ_ONCE(task->tk_rqstp->rq_reply_bytes_recvd) == 0; READ_ONCE 175 net/sunrpc/xprtmultipath.c if (READ_ONCE(xps->xps_iter_ops) != &rpc_xprt_iter_roundrobin) READ_ONCE 345 net/sunrpc/xprtmultipath.c nactive = READ_ONCE(xps->xps_nactive); READ_ONCE 889 net/sunrpc/xprtrdma/verbs.c if (next_head == READ_ONCE(buf->rb_sc_tail)) READ_ONCE 2128 net/tipc/socket.c return READ_ONCE(sk->sk_rcvbuf); READ_ONCE 2131 net/tipc/socket.c return READ_ONCE(sk->sk_rcvbuf) << msg_importance(hdr); READ_ONCE 2134 net/tipc/socket.c return READ_ONCE(sk->sk_rcvbuf); READ_ONCE 3800 net/tipc/socket.c i += scnprintf(buf + i, sz - i, " | %d\n", READ_ONCE(sk->sk_backlog.len)); READ_ONCE 670 net/tls/tls_device.c netdev = READ_ONCE(tls_ctx->netdev); READ_ONCE 395 net/tls/tls_sw.c if (READ_ONCE(rec->tx_ready)) { READ_ONCE 15749 net/wireless/nl80211.c u32 nlportid = READ_ONCE(wdev->ap_unexpected_nlportid); READ_ONCE 15917 net/wireless/nl80211.c u32 nlportid = READ_ONCE(wdev->conn_owner_nlportid); READ_ONCE 58 net/x25/x25_dev.c queued = !sk_add_backlog(sk, skb, READ_ONCE(sk->sk_rcvbuf)); READ_ONCE 36 net/xdp/xsk.c return READ_ONCE(xs->rx) && READ_ONCE(xs->umem) && READ_ONCE 37 net/xdp/xsk.c READ_ONCE(xs->umem->fq); READ_ONCE 192 net/xdp/xsk.c if (READ_ONCE(xs->state) == XSK_BOUND) { READ_ONCE 973 net/xdp/xsk.c if (READ_ONCE(xs->state) != XSK_READY) READ_ONCE 977 net/xdp/xsk.c q = READ_ONCE(xs->rx); READ_ONCE 979 net/xdp/xsk.c q = READ_ONCE(xs->tx); READ_ONCE 981 net/xdp/xsk.c umem = READ_ONCE(xs->umem); READ_ONCE 988 net/xdp/xsk.c q = READ_ONCE(umem->fq); READ_ONCE 990 net/xdp/xsk.c q = READ_ONCE(umem->cq); READ_ONCE 102 net/xdp/xsk_queue.h q->prod_tail = READ_ONCE(q->ring->producer); READ_ONCE 117 net/xdp/xsk_queue.h q->cons_tail = READ_ONCE(q->ring->consumer); READ_ONCE 129 net/xdp/xsk_queue.h q->prod_tail = READ_ONCE(q->ring->producer); READ_ONCE 181 net/xdp/xsk_queue.h *addr = READ_ONCE(ring->desc[idx]) & q->chunk_mask; READ_ONCE 307 net/xdp/xsk_queue.h *desc = READ_ONCE(ring->desc[idx]); READ_ONCE 2412 net/xfrm/xfrm_state.c const struct xfrm_type *type = READ_ONCE(x->type); READ_ONCE 374 samples/mic/mpssd/mpssd.c return READ_ONCE(vr->info->avail_idx); READ_ONCE 515 samples/mic/mpssd/mpssd.c while (avail_idx == le16toh(READ_ONCE(vr->vr.avail->idx))) { READ_ONCE 549 security/apparmor/apparmorfs.c READ_ONCE(rev->ns->revision))) READ_ONCE 1721 security/apparmor/apparmorfs.c ns = __aa_find_or_create_ns(parent, READ_ONCE(dentry->d_name.name), READ_ONCE 425 security/apparmor/domain.c long rev = READ_ONCE(ns->revision); READ_ONCE 435 security/apparmor/domain.c READ_ONCE(ns->revision)) READ_ONCE 578 security/keys/keyring.c unsigned long kflags = READ_ONCE(key->flags); READ_ONCE 579 security/keys/keyring.c short state = READ_ONCE(key->state); READ_ONCE 591 security/keys/keyring.c time64_t expiry = READ_ONCE(key->expiry); READ_ONCE 732 security/keys/keyring.c ptr = READ_ONCE(keyring->keys.root); READ_ONCE 745 security/keys/keyring.c ptr = READ_ONCE(shortcut->next_node); READ_ONCE 762 security/keys/keyring.c ptr = READ_ONCE(shortcut->next_node); READ_ONCE 773 security/keys/keyring.c ptr = READ_ONCE(node->slots[slot]); READ_ONCE 811 security/keys/keyring.c ptr = READ_ONCE(node->back_pointer); READ_ONCE 816 security/keys/keyring.c ptr = READ_ONCE(shortcut->back_pointer); READ_ONCE 87 security/keys/permission.c unsigned long flags = READ_ONCE(key->flags); READ_ONCE 88 security/keys/permission.c time64_t expiry = READ_ONCE(key->expiry); READ_ONCE 200 security/keys/proc.c expiry = READ_ONCE(key->expiry); READ_ONCE 225 security/keys/proc.c flags = READ_ONCE(key->flags); READ_ONCE 40 security/keys/process_keys.c struct key *reg_keyring = READ_ONCE(user_ns->user_keyring_register); READ_ONCE 187 security/keys/process_keys.c struct key *reg_keyring = READ_ONCE(cred->user_ns->user_keyring_register); READ_ONCE 211 security/selinux/ss/sidtab.c WRITE_ONCE(s->rcache[pos], READ_ONCE(s->rcache[pos - 1])); READ_ONCE 228 security/selinux/ss/sidtab.c u32 v = READ_ONCE(s->rcache[i]); READ_ONCE 53 sound/core/pcm_lib.c snd_pcm_uframes_t appl_ptr = READ_ONCE(runtime->control->appl_ptr); READ_ONCE 2205 sound/core/pcm_lib.c appl_ptr = READ_ONCE(runtime->control->appl_ptr); READ_ONCE 78 sound/core/seq/seq_virmidi.c if (!READ_ONCE(vmidi->trigger)) READ_ONCE 143 sound/core/seq/seq_virmidi.c while (READ_ONCE(vmidi->trigger)) { READ_ONCE 433 sound/firewire/amdtp-stream.c struct snd_pcm_substream *pcm = READ_ONCE(s->pcm); READ_ONCE 481 sound/firewire/amdtp-stream.c cip_header[0] = cpu_to_be32(READ_ONCE(s->source_node_id_field) | READ_ONCE 770 sound/firewire/amdtp-stream.c pcm = READ_ONCE(s->pcm); READ_ONCE 1058 sound/firewire/amdtp-stream.c return READ_ONCE(s->pcm_buffer_pointer); READ_ONCE 1132 sound/firewire/amdtp-stream.c pcm = READ_ONCE(s->pcm); READ_ONCE 416 sound/firewire/fireface/ff-protocol-former.c substream = READ_ONCE(ff->tx_midi_substreams[0]); READ_ONCE 577 sound/firewire/fireface/ff-protocol-former.c substream = READ_ONCE(ff->tx_midi_substreams[0]); READ_ONCE 587 sound/firewire/fireface/ff-protocol-former.c substream = READ_ONCE(ff->tx_midi_substreams[1]); READ_ONCE 331 sound/firewire/fireface/ff-protocol-latter.c substream = READ_ONCE(ff->tx_midi_substreams[index]); READ_ONCE 14 sound/firewire/fireface/ff-transaction.c READ_ONCE(ff->rx_midi_substreams[port]); READ_ONCE 56 sound/firewire/fireface/ff-transaction.c READ_ONCE(ff->rx_midi_substreams[port]); READ_ONCE 114 sound/firewire/isight.c if (!READ_ONCE(isight->pcm_running)) READ_ONCE 134 sound/firewire/isight.c if (READ_ONCE(isight->pcm_active)) READ_ONCE 144 sound/firewire/isight.c if (!READ_ONCE(isight->pcm_running)) READ_ONCE 442 sound/firewire/isight.c return READ_ONCE(isight->buffer_pointer); READ_ONCE 242 sound/firewire/motu/amdtp-motu.c struct snd_rawmidi_substream *midi = READ_ONCE(p->midi); READ_ONCE 274 sound/firewire/motu/amdtp-motu.c midi = READ_ONCE(p->midi); READ_ONCE 114 sound/firewire/oxfw/oxfw-scs1x.c stream = READ_ONCE(scs->input); READ_ONCE 185 sound/firewire/oxfw/oxfw-scs1x.c stream = READ_ONCE(scs->output); READ_ONCE 133 sound/firewire/tascam/amdtp-tascam.c bool used = READ_ONCE(tscm->hwdep->used); READ_ONCE 150 sound/firewire/tascam/tascam-transaction.c struct snd_rawmidi_substream *substream = READ_ONCE(port->substream); READ_ONCE 175 sound/firewire/tascam/tascam-transaction.c struct snd_rawmidi_substream *substream = READ_ONCE(port->substream); READ_ONCE 284 sound/firewire/tascam/tascam-transaction.c substream = READ_ONCE(tscm->tx_midi_substreams[port]); READ_ONCE 111 sound/soc/dwc/dwc-pcm.c ptr = READ_ONCE(dev->tx_ptr); READ_ONCE 116 sound/soc/dwc/dwc-pcm.c ptr = READ_ONCE(dev->rx_ptr); READ_ONCE 241 sound/soc/dwc/dwc-pcm.c pos = READ_ONCE(dev->tx_ptr); READ_ONCE 243 sound/soc/dwc/dwc-pcm.c pos = READ_ONCE(dev->rx_ptr); READ_ONCE 913 sound/soc/sh/rcar/ssi.c *pointer = bytes_to_frames(runtime, READ_ONCE(ssi->byte_pos)); READ_ONCE 19 sound/soc/sof/trace.c loff_t host_offset = READ_ONCE(sdev->host_offset); READ_ONCE 165 sound/soc/xtensa/xtfpga-i2s.c unsigned tx_ptr = READ_ONCE(i2s->tx_ptr); READ_ONCE 459 sound/soc/xtensa/xtfpga-i2s.c snd_pcm_uframes_t pos = READ_ONCE(i2s->tx_ptr); READ_ONCE 102 sound/usb/bcd2000/bcd2000.c midi_receive_substream = READ_ONCE(bcd2k->midi_receive_substream); READ_ONCE 133 sound/usb/bcd2000/bcd2000.c midi_out_substream = READ_ONCE(bcd2k->midi_out_substream); READ_ONCE 57 tools/arch/ia64/include/asm/barrier.h typeof(*p) ___p1 = READ_ONCE(*p); \ READ_ONCE 41 tools/arch/powerpc/include/asm/barrier.h typeof(*p) ___p1 = READ_ONCE(*p); \ READ_ONCE 39 tools/arch/s390/include/asm/barrier.h typeof(*p) ___p1 = READ_ONCE(*p); \ READ_ONCE 51 tools/arch/sparc/include/asm/barrier_64.h typeof(*p) ___p1 = READ_ONCE(*p); \ READ_ONCE 28 tools/arch/x86/include/asm/atomic.h return READ_ONCE((v)->counter); READ_ONCE 41 tools/arch/x86/include/asm/barrier.h typeof(*p) ___p1 = READ_ONCE(*p); \ READ_ONCE 25 tools/include/asm-generic/atomic-gcc.h return READ_ONCE((v)->counter); READ_ONCE 61 tools/include/asm/barrier.h typeof(*p) ___p1 = READ_ONCE(*p); \ READ_ONCE 39 tools/include/linux/rbtree.h #define RB_EMPTY_ROOT(root) (READ_ONCE((root)->rb_node) == NULL) READ_ONCE 61 tools/include/linux/ring_buffer.h u64 head = READ_ONCE(base->data_head); READ_ONCE 5992 tools/lib/bpf/libbpf.c tmp_cpus = READ_ONCE(cpus); READ_ONCE 396 tools/perf/util/auxtrace.h u64 head = READ_ONCE(pc->aux_head); READ_ONCE 407 tools/perf/util/auxtrace.h u64 head = READ_ONCE(pc->aux_head); READ_ONCE 132 tools/perf/util/session.h #define session_done() READ_ONCE(session_done) READ_ONCE 44 tools/testing/selftests/bpf/map_tests/sk_storage_map.c return READ_ONCE(stop); READ_ONCE 49 tools/testing/selftests/bpf/map_tests/sk_storage_map.c return READ_ONCE(nr_sk_threads_err); READ_ONCE 67 tools/testing/selftests/bpf/map_tests/sk_storage_map.c return READ_ONCE(nr_sk_threads_done); READ_ONCE 99 tools/testing/selftests/bpf/map_tests/sk_storage_map.c while (READ_ONCE(sk_storage_map) == -1 && !is_stopped()) READ_ONCE 107 tools/testing/selftests/bpf/map_tests/sk_storage_map.c while (READ_ONCE(sk_storage_map) != -1 && !is_stopped()) READ_ONCE 185 tools/testing/selftests/bpf/map_tests/sk_storage_map.c map_fd = READ_ONCE(sk_storage_map); READ_ONCE 287 tools/testing/selftests/bpf/map_tests/sk_storage_map.c int map_fd = READ_ONCE(sk_storage_map); READ_ONCE 311 tools/testing/selftests/bpf/map_tests/sk_storage_map.c int map_fd = READ_ONCE(sk_storage_map); READ_ONCE 102 tools/testing/selftests/kvm/dirty_log_test.c *(uint64_t *)addr = READ_ONCE(iteration); READ_ONCE 108 tools/testing/selftests/kvm/dirty_log_test.c addr += (READ_ONCE(random_array[i]) % guest_num_pages) READ_ONCE 111 tools/testing/selftests/kvm/dirty_log_test.c *(uint64_t *)addr = READ_ONCE(iteration); READ_ONCE 161 tools/testing/selftests/kvm/dirty_log_test.c while (!READ_ONCE(host_quit)) { READ_ONCE 27 tools/testing/selftests/powerpc/dscr/dscr_default_test.c s1 = READ_ONCE(sequence); READ_ONCE 21 tools/testing/selftests/rcutorture/formal/srcu-cbmc/tests/store_buffering/test.c __unbuffered_tpr_y = READ_ONCE(y); READ_ONCE 41 tools/testing/selftests/rcutorture/formal/srcu-cbmc/tests/store_buffering/test.c __unbuffered_tpr_x = READ_ONCE(x); READ_ONCE 26 tools/virtio/linux/uaccess.h x = READ_ONCE(*(__pu_ptr)); \ READ_ONCE 509 virt/kvm/arm/arm.c return unlikely(READ_ONCE(vmid->vmid_gen) != current_vmid_gen); READ_ONCE 351 virt/kvm/arm/mmu.c if (!READ_ONCE(kvm->arch.pgd)) READ_ONCE 1005 virt/kvm/arm/mmu.c pgd = READ_ONCE(kvm->arch.pgd); READ_ONCE 1519 virt/kvm/arm/mmu.c if (!READ_ONCE(kvm->arch.pgd)) READ_ONCE 327 virt/kvm/arm/vgic/vgic-its.c irq_count = READ_ONCE(dist->lpi_list_count); READ_ONCE 141 virt/kvm/arm/vgic/vgic-mmio-v3.c ret = extract_bytes(READ_ONCE(irq->mpidr), addr & 7, len); READ_ONCE 405 virt/kvm/arm/vgic/vgic-mmio-v3.c old_propbaser = READ_ONCE(dist->propbaser); READ_ONCE 433 virt/kvm/arm/vgic/vgic-mmio-v3.c old_pendbaser = READ_ONCE(vgic_cpu->pendbaser); READ_ONCE 77 virt/kvm/coalesced_mmio.c insert = READ_ONCE(ring->last); READ_ONCE 2414 virt/kvm/kvm_main.c grow_start = READ_ONCE(halt_poll_ns_grow_start); READ_ONCE 2415 virt/kvm/kvm_main.c grow = READ_ONCE(halt_poll_ns_grow); READ_ONCE 2436 virt/kvm/kvm_main.c shrink = READ_ONCE(halt_poll_ns_shrink); READ_ONCE 2681 virt/kvm/kvm_main.c if (!READ_ONCE(vcpu->ready)) READ_ONCE 2687 virt/kvm/kvm_main.c if (READ_ONCE(vcpu->preempted) && yield_to_kernel_mode &&